Classification LightGBM Learner
Source:R/learner_lightgbm_classif_lightgbm.R
mlr_learners_classif.lightgbm.RdGradient boosting algorithm.
Calls lightgbm::lightgbm() from lightgbm.
The list of parameters can be found here
and in the documentation of lightgbm::lgb.train().
Meta Information
Task type: “classif”
Predict Types: “response”, “prob”
Feature Types: “logical”, “integer”, “numeric”, “factor”
Required Packages: mlr3, mlr3extralearners, lightgbm
Parameters
| Id | Type | Default | Levels | Range |
| objective | character | - | binary, multiclass, multiclassova | - |
| eval | untyped | - | - | |
| verbose | integer | 1 | \((-\infty, \infty)\) | |
| record | logical | TRUE | TRUE, FALSE | - |
| eval_freq | integer | 1 | \([1, \infty)\) | |
| callbacks | untyped | - | - | |
| reset_data | logical | FALSE | TRUE, FALSE | - |
| boosting | character | gbdt | gbdt, rf, dart, goss | - |
| linear_tree | logical | FALSE | TRUE, FALSE | - |
| learning_rate | numeric | 0.1 | \([0, \infty)\) | |
| num_leaves | integer | 31 | \([1, 131072]\) | |
| tree_learner | character | serial | serial, feature, data, voting | - |
| num_threads | integer | 0 | \([0, \infty)\) | |
| device_type | character | cpu | cpu, gpu | - |
| seed | integer | - | \((-\infty, \infty)\) | |
| deterministic | logical | FALSE | TRUE, FALSE | - |
| data_sample_strategy | character | bagging | bagging, goss | - |
| force_col_wise | logical | FALSE | TRUE, FALSE | - |
| force_row_wise | logical | FALSE | TRUE, FALSE | - |
| histogram_pool_size | numeric | -1 | \((-\infty, \infty)\) | |
| max_depth | integer | -1 | \((-\infty, \infty)\) | |
| min_data_in_leaf | integer | 20 | \([0, \infty)\) | |
| min_sum_hessian_in_leaf | numeric | 0.001 | \([0, \infty)\) | |
| bagging_fraction | numeric | 1 | \([0, 1]\) | |
| pos_bagging_fraction | numeric | 1 | \([0, 1]\) | |
| neg_bagging_fraction | numeric | 1 | \([0, 1]\) | |
| bagging_freq | integer | 0 | \([0, \infty)\) | |
| bagging_seed | integer | 3 | \((-\infty, \infty)\) | |
| bagging_by_query | logical | FALSE | TRUE, FALSE | - |
| feature_fraction | numeric | 1 | \([0, 1]\) | |
| feature_fraction_bynode | numeric | 1 | \([0, 1]\) | |
| feature_fraction_seed | integer | 2 | \((-\infty, \infty)\) | |
| extra_trees | logical | FALSE | TRUE, FALSE | - |
| extra_seed | integer | 6 | \((-\infty, \infty)\) | |
| max_delta_step | numeric | 0 | \((-\infty, \infty)\) | |
| lambda_l1 | numeric | 0 | \([0, \infty)\) | |
| lambda_l2 | numeric | 0 | \([0, \infty)\) | |
| linear_lambda | numeric | 0 | \([0, \infty)\) | |
| min_gain_to_split | numeric | 0 | \([0, \infty)\) | |
| drop_rate | numeric | 0.1 | \([0, 1]\) | |
| max_drop | integer | 50 | \((-\infty, \infty)\) | |
| skip_drop | numeric | 0.5 | \([0, 1]\) | |
| xgboost_dart_mode | logical | FALSE | TRUE, FALSE | - |
| uniform_drop | logical | FALSE | TRUE, FALSE | - |
| drop_seed | integer | 4 | \((-\infty, \infty)\) | |
| top_rate | numeric | 0.2 | \([0, 1]\) | |
| other_rate | numeric | 0.1 | \([0, 1]\) | |
| min_data_per_group | integer | 100 | \([1, \infty)\) | |
| max_cat_threshold | integer | 32 | \([1, \infty)\) | |
| cat_l2 | numeric | 10 | \([0, \infty)\) | |
| cat_smooth | numeric | 10 | \([0, \infty)\) | |
| max_cat_to_onehot | integer | 4 | \([1, \infty)\) | |
| top_k | integer | 20 | \([1, \infty)\) | |
| monotone_constraints | untyped | NULL | - | |
| monotone_constraints_method | character | basic | basic, intermediate, advanced | - |
| monotone_penalty | numeric | 0 | \([0, \infty)\) | |
| feature_contri | untyped | NULL | - | |
| forcedsplits_filename | untyped | "" | - | |
| refit_decay_rate | numeric | 0.9 | \([0, 1]\) | |
| cegb_tradeoff | numeric | 1 | \([0, \infty)\) | |
| cegb_penalty_split | numeric | 0 | \([0, \infty)\) | |
| cegb_penalty_feature_lazy | untyped | - | - | |
| cegb_penalty_feature_coupled | untyped | - | - | |
| path_smooth | numeric | 0 | \([0, \infty)\) | |
| interaction_constraints | untyped | - | - | |
| use_quantized_grad | logical | TRUE | TRUE, FALSE | - |
| num_grad_quant_bins | integer | 4 | \((-\infty, \infty)\) | |
| quant_train_renew_leaf | logical | FALSE | TRUE, FALSE | - |
| stochastic_rounding | logical | TRUE | TRUE, FALSE | - |
| serializable | logical | TRUE | TRUE, FALSE | - |
| max_bin | integer | 255 | \([2, \infty)\) | |
| max_bin_by_feature | untyped | NULL | - | |
| min_data_in_bin | integer | 3 | \([1, \infty)\) | |
| bin_construct_sample_cnt | integer | 200000 | \([1, \infty)\) | |
| data_random_seed | integer | 1 | \((-\infty, \infty)\) | |
| is_enable_sparse | logical | TRUE | TRUE, FALSE | - |
| enable_bundle | logical | TRUE | TRUE, FALSE | - |
| use_missing | logical | TRUE | TRUE, FALSE | - |
| zero_as_missing | logical | FALSE | TRUE, FALSE | - |
| feature_pre_filter | logical | TRUE | TRUE, FALSE | - |
| pre_partition | logical | FALSE | TRUE, FALSE | - |
| two_round | logical | FALSE | TRUE, FALSE | - |
| forcedbins_filename | untyped | "" | - | |
| is_unbalance | logical | FALSE | TRUE, FALSE | - |
| scale_pos_weight | numeric | 1 | \([0, \infty)\) | |
| sigmoid | numeric | 1 | \([0, \infty)\) | |
| boost_from_average | logical | TRUE | TRUE, FALSE | - |
| eval_at | untyped | 1:5 | - | |
| multi_error_top_k | integer | 1 | \([1, \infty)\) | |
| auc_mu_weights | untyped | NULL | - | |
| num_machines | integer | 1 | \([1, \infty)\) | |
| local_listen_port | integer | 12400 | \([1, \infty)\) | |
| time_out | integer | 120 | \([1, \infty)\) | |
| machines | untyped | "" | - | |
| gpu_platform_id | integer | -1 | \((-\infty, \infty)\) | |
| gpu_device_id | integer | -1 | \((-\infty, \infty)\) | |
| gpu_use_dp | logical | FALSE | TRUE, FALSE | - |
| num_gpu | integer | 1 | \([1, \infty)\) | |
| start_iteration_predict | integer | 0 | \((-\infty, \infty)\) | |
| num_iteration_predict | integer | -1 | \((-\infty, \infty)\) | |
| pred_early_stop | logical | FALSE | TRUE, FALSE | - |
| pred_early_stop_freq | integer | 10 | \((-\infty, \infty)\) | |
| pred_early_stop_margin | numeric | 10 | \((-\infty, \infty)\) | |
| num_iterations | integer | 100 | \([1, \infty)\) | |
| early_stopping_rounds | integer | - | \([1, \infty)\) | |
| early_stopping_min_delta | numeric | - | \([0, \infty)\) | |
| first_metric_only | logical | FALSE | TRUE, FALSE | - |
Initial parameter values
num_threads:Actual default: 0L
Initial value: 1L
Reason for change: Prevents accidental conflicts with
future.
verbose:Actual default: 1L
Initial value: -1L
Reason for change: Prevents accidental conflicts with mlr messaging system.
objective:Depends on the task: if binary classification, then this parameter is set to
"binary", otherwise"multiclasss"and cannot be changed.
Custom mlr3 parameters
num_class: This parameter is automatically inferred for multiclass tasks and does not have to be set.
Early Stopping and Validation
Early stopping can be used to find the optimal number of boosting rounds.
Set early_stopping_rounds to an integer value to monitor the performance of the model on the validation set while training.
For information on how to configure the validation set, see the Validation section of mlr3::Learner.
The internal validation measure can be set the eval parameter which should be a list of mlr3::Measures, functions, or strings for the internal lightgbm measures.
If first_metric_only = FALSE (default), the learner stops when any metric fails to improve.
References
Ke, Guolin, Meng, Qi, Finley, Thomas, Wang, Taifeng, Chen, Wei, Ma, Weidong, Ye, Qiwei, Liu, Tie-Yan (2017). “Lightgbm: A highly efficient gradient boosting decision tree.” Advances in neural information processing systems, 30.
See also
as.data.table(mlr_learners)for a table of available Learners in the running session (depending on the loaded packages).Chapter in the mlr3book: https://mlr3book.mlr-org.com/basics.html#learners
mlr3learners for a selection of recommended learners.
mlr3cluster for unsupervised clustering learners.
mlr3pipelines to combine learners with pre- and postprocessing steps.
mlr3tuning for tuning of hyperparameters, mlr3tuningspaces for established default tuning spaces.
Super classes
mlr3::Learner -> mlr3::LearnerClassif -> LearnerClassifLightGBM
Active bindings
internal_valid_scoresThe last observation of the validation scores for all metrics. Extracted from
model$evaluation_loginternal_tuned_valuesReturns the early stopped iterations if
early_stopping_roundswas set during training.validateHow to construct the internal validation data. This parameter can be either
NULL, a ratio,"test", or"predefined".
Methods
Inherited methods
mlr3::Learner$base_learner()mlr3::Learner$configure()mlr3::Learner$encapsulate()mlr3::Learner$format()mlr3::Learner$help()mlr3::Learner$predict()mlr3::Learner$predict_newdata()mlr3::Learner$print()mlr3::Learner$reset()mlr3::Learner$selected_features()mlr3::Learner$train()mlr3::LearnerClassif$predict_newdata_fast()
Method importance()
The importance scores are extracted from lbg.importance.
Returns
Named numeric().
Examples
# Define the Learner
learner = lrn("classif.lightgbm")
print(learner)
#>
#> ── <LearnerClassifLightGBM> (classif.lightgbm): Gradient Boosting ──────────────
#> • Model: -
#> • Parameters: verbose=-1, num_threads=1
#> • Validate: NULL
#> • Packages: mlr3, mlr3extralearners, and lightgbm
#> • Predict Types: response and [prob]
#> • Feature Types: logical, integer, numeric, and factor
#> • Encapsulation: none (fallback: -)
#> • Properties: hotstart_forward, importance, internal_tuning, missings,
#> multiclass, twoclass, validation, and weights
#> • Other settings: use_weights = 'use'
# Define a Task
task = tsk("sonar")
# Create train and test set
ids = partition(task)
# Train the learner on the training ids
learner$train(task, row_ids = ids$train)
print(learner$model)
#> LightGBM Model (100 trees)
#> Objective: binary
#> Fitted to dataset with 60 columns
print(learner$importance())
#> V11 V51 V10 V48 V16 V45
#> 0.1175685048 0.0996352161 0.0799883050 0.0589855664 0.0422894292 0.0413389552
#> V12 V55 V37 V20 V23 V9
#> 0.0393963235 0.0391865408 0.0374155120 0.0350573161 0.0295352050 0.0286870264
#> V21 V5 V36 V49 V14 V52
#> 0.0282606206 0.0275384064 0.0255988295 0.0233831689 0.0167181334 0.0165077947
#> V44 V31 V39 V46 V15 V47
#> 0.0160095529 0.0154331675 0.0147853001 0.0145050657 0.0143469625 0.0132516912
#> V27 V28 V4 V42 V54 V6
#> 0.0119526752 0.0118831131 0.0116653318 0.0102885344 0.0099136682 0.0082618361
#> V58 V56 V43 V19 V59 V41
#> 0.0079736413 0.0061120165 0.0056508372 0.0055857007 0.0052007311 0.0043727889
#> V30 V40 V57 V3 V18 V1
#> 0.0034172675 0.0033745812 0.0032462351 0.0024161349 0.0024078178 0.0018574481
#> V35 V38 V17 V50 V22 V29
#> 0.0016676594 0.0014216397 0.0009405054 0.0008438011 0.0008360256 0.0006197544
#> V53 V2 V24 V13 V26 V8
#> 0.0005746870 0.0005519064 0.0005283045 0.0004046483 0.0003248526 0.0002832627
# Make predictions for the test rows
predictions = learner$predict(task, row_ids = ids$test)
# Score the predictions
predictions$score()
#> classif.ce
#> 0.1884058