Catboost

Latest version: v1.2.5

Safety actively analyzes 629788 Python packages for vulnerabilities to keep your Python projects secure.

Scan your dependencies

Page 10 of 14

0.12.2

Not secure
Changes:
* Fixed loading of `epsilon` dataset into memory
* Fixed multiclass learning on GPU for >255 classes
* Improved error handling
* Some other minor fixes

0.12.1.1

Not secure
Changes:
* Fixed Python compatibility issue in dataset downloading
* Added `sampling_type` parameter for `YetiRankPairwise` loss

0.12.1

Not secure
Changes:
* Support saving models in ONNX format (only for models without categorical features).
* Added new dataset to our `catboost.datasets()` -- dataset [epsilon](catboost/benchmarks/model_evaluation_speed), a large dense dataset for binary classification.
* Speedup of Python `cv` on GPU.
* Fixed creation of `Pool` from `pandas.DataFrame` with `pandas.Categorical` columns.

0.12.0

Not secure
Breaking changes:
* Class weights are now taken into account by `eval_metrics()`,
`get_feature_importance()`, and `get_object_importance()`.
In previous versions the weights were ignored.
* Parameter `random-strength` for pairwise training (`PairLogitPairwise`,
`QueryCrossEntropy`, `YetiRankPairwise`) is not supported anymore.
* Simultaneous use of `MultiClass` and `MultiClassOneVsAll` metrics is now
deprecated.

New functionality:
* `cv` method is now supported on GPU.
* String labels for classes are supported in Python.
In multiclassification the string class names are inferred from the data.
In binary classification for using string labels you should employ `class_names`
parameter and specify which class is negative (0) and which is positive (1).
You can also use `class_names` in multiclassification mode to pass all
possible class names to the fit function.
* Borders can now be saved and reused.
To save the feature quantization information obtained during training data
preprocessing into a text file use cli option `--output-borders-file`.
To use the borders for training use cli option `--input-borders-file`.
This functionanlity is now supported on CPU and GPU (it was GPU-only in previous versions).
File format for the borders is described [here](https://tech.yandex.com/catboost/doc/dg/concepts/input-data_custom-borders-docpage).
* CLI option `--eval-file` is now supported on GPU.

Quality improvement:
* Some cases in binary classification are fixed where training could diverge

Optimizations:
* A great speedup of the Python applier (10x)
* Reduced memory consumption in Python `cv` function (times fold count)

Benchmarks and tutorials:
* Added [speed benchmarks](catboost/benchmarks/gpu_vs_cpu_training_speed) for CPU and GPU on a variety of different datasets.
* Added [benchmarks](catboost/benchmarks/ranking) of different ranking modes. In [this tutorial](catboost/tutorials/ranking/ranking_tutorial.ipynb) we compare
different ranking modes in CatBoost, XGBoost and LightGBM.
* Added [tutorial](catboost/tutorials/apply_model/catboost4j_prediction_tutorial.ipynb) for applying model in Java.
* Added [benchmarks](catboost/benchmarks/shap_speed) of SHAP values calculation for CatBoost, XGBoost and LightGBM.
The benchmarks also contain explanation of complexity of this calculation
in all the libraries.

We also made a list of stability improvements
and stricter checks of input data and parameters.

And we are so grateful to our community members canorbal and neer201
for their contribution in this release. Thank you.

0.11.2

Not secure
Changes:
* Pure GPU implementation of NDCG metric
* Enabled LQ loss function
* Fixed NDCG metric on CPU
* Added `model_sum` mode to command line interface
* Added SHAP values benchmark (566)
* Fixed `random_strength` for `Plain` boosting (448)
* Enabled passing a test pool to caret training (544)
* Fixed a bug in exporting the model as python code (556)
* Fixed label mapper for multiclassification custom labels (523)
* Fixed hash type of categorical features (558)
* Fixed handling of cross-validation fold count options in python package (568)

0.11.1

Not secure
Changes:
* Accelerated formula evaluation by ~15%
* Improved model application interface
* Improved compilation time for building GPU version
* Better handling of stray commas in list arguments
* Added a benchmark that employs Rossman Store Sales dataset to compare quality of GBDT packages
* Added references to Catboost papers in R-package CITATION file
* Fixed a build issue in compilation for GPU
* Fixed a bug in model applicator
* Fixed model conversion, 533
* Returned pre 0.11 behaviour for `best_score_` and `evals_result_` (issue 539)
* Make valid RECORD in wheel (issue 534)

Page 10 of 14

© 2024 Safety CLI Cybersecurity Inc. All Rights Reserved.