categorical_features | Indicates the categorical features | default: null |
early_stopping | If 'auto', early stopping is enabled if the sample size is larger than
10000. If True, early stopping is enabled, otherwise early stopping is
disabled
.. versionadded:: 0.23 | default: "auto" |
l2_regularization | The L2 regularization parameter. Use 0 for no regularization | default: 0.2244 |
learning_rate | The learning rate, also known as *shrinkage*. This is used as a
multiplicative factor for the leaves values. Use ``1`` for no
shrinkage | default: 0.4915 |
loss | | default: "auto" |
max_bins | The maximum number of bins to use for non-missing values. Before
training, each feature of the input array `X` is binned into
integer-valued bins, which allows for a much faster training stage
Features with a small number of unique values may use less than
``max_bins`` bins. In addition to the ``max_bins`` bins, one more bin
is always reserved for missing values. Must be no larger than 255 | default: 255 |
max_depth | The maximum depth of each tree. The depth of a tree is the number of
edges to go from the root to the deepest leaf
Depth isn't constrained by default | default: 32 |
max_iter | The maximum number of iterations of the boosting process, i.e. the
maximum number of trees for binary classification. For multiclass
classification, `n_classes` trees per iteration are built | default: 100 |
max_leaf_nodes | The maximum number of leaves for each tree. Must be strictly greater
than 1. If None, there is no maximum limit | default: 31 |
min_samples_leaf | The minimum number of samples per leaf. For small datasets with less
than a few hundred samples, it is recommended to lower this value
since only very shallow trees would be built | default: 20 |
monotonic_cst | Indicates the monotonic constraint to enforce on each feature. -1, 1
and 0 respectively correspond to a negative constraint, positive
constraint and no constraint. Read more in the :ref:`User Guide
`
.. versionadded:: 0.23 | default: null |
n_iter_no_change | Used to determine when to "early stop". The fitting process is
stopped when none of the last ``n_iter_no_change`` scores are better
than the ``n_iter_no_change - 1`` -th-to-last one, up to some
tolerance. Only used if early stopping is performed | default: 10 |
random_state | Pseudo-random number generator to control the subsampling in the
binning process, and the train/validation data split if early stopping
is enabled
Pass an int for reproducible output across multiple function calls
See :term:`Glossary `. | default: null |
scoring | Scoring parameter to use for early stopping. It can be a single
string (see :ref:`scoring_parameter`) or a callable (see
:ref:`scoring`). If None, the estimator's default scorer
is used. If ``scoring='loss'``, early stopping is checked
w.r.t the loss value. Only used if early stopping is performed | default: "loss" |
tol | The absolute tolerance to use when comparing scores. The higher the
tolerance, the more likely we are to early stop: higher tolerance
means that it will be harder for subsequent iterations to be
considered an improvement upon the reference score | default: 1e-07 |
validation_fraction | Proportion (or absolute size) of training data to set aside as
validation data for early stopping. If None, early stopping is done on
the training data. Only used if early stopping is performed | default: 0.1 |
verbose | The verbosity level. If not zero, print some information about the
fitting process | default: 0 |
warm_start | When set to ``True``, reuse the solution of the previous call to fit
and add more estimators to the ensemble. For results to be valid, the
estimator should be re-trained on the same data only
See :term:`the Glossary ` | default: false |