Skip to contents

Random forest for classification. Calls randomForest::randomForest() from randomForest.

Dictionary

This Learner can be instantiated via lrn():

lrn("classif.randomForest")

Meta Information

  • Task type: “classif”

  • Predict Types: “response”, “prob”

  • Feature Types: “logical”, “integer”, “numeric”, “factor”, “ordered”

  • Required Packages: mlr3, mlr3extralearners, randomForest

Parameters

IdTypeDefaultLevelsRange
ntreeinteger500\([1, \infty)\)
mtryinteger-\([1, \infty)\)
replacelogicalTRUETRUE, FALSE-
classwtuntypedNULL-
cutoffuntyped--
stratauntyped--
sampsizeuntyped--
nodesizeinteger1\([1, \infty)\)
maxnodesinteger-\([1, \infty)\)
importancecharacterFALSEaccuracy, gini, none-
localImplogicalFALSETRUE, FALSE-
proximitylogicalFALSETRUE, FALSE-
oob.proxlogical-TRUE, FALSE-
norm.voteslogicalTRUETRUE, FALSE-
do.tracelogicalFALSETRUE, FALSE-
keep.forestlogicalTRUETRUE, FALSE-
keep.inbaglogicalFALSETRUE, FALSE-
predict.alllogicalFALSETRUE, FALSE-
nodeslogicalFALSETRUE, FALSE-

References

Breiman, Leo (2001). “Random Forests.” Machine Learning, 45(1), 5–32. ISSN 1573-0565, doi:10.1023/A:1010933404324 .

See also

Author

pat-s

Super classes

mlr3::Learner -> mlr3::LearnerClassif -> LearnerClassifRandomForest

Methods

Inherited methods


Method new()

Creates a new instance of this R6 class.


Method importance()

The importance scores are extracted from the slot importance. Parameter 'importance' must be set to either "accuracy" or "gini".

Usage

LearnerClassifRandomForest$importance()

Returns

Named numeric().


Method oob_error()

OOB errors are extracted from the model slot err.rate.

Usage

LearnerClassifRandomForest$oob_error()

Returns

numeric(1).


Method clone()

The objects of this class are cloneable with this method.

Usage

LearnerClassifRandomForest$clone(deep = FALSE)

Arguments

deep

Whether to make a deep clone.

Examples

# Define the Learner
learner = lrn("classif.randomForest", importance = "accuracy")
print(learner)
#> 
#> ── <LearnerClassifRandomForest> (classif.randomForest): Random Forest ──────────
#> • Model: -
#> • Parameters: importance=accuracy
#> • Packages: mlr3, mlr3extralearners, and randomForest
#> • Predict Types: [response] and prob
#> • Feature Types: logical, integer, numeric, factor, and ordered
#> • Encapsulation: none (fallback: -)
#> • Properties: importance, multiclass, oob_error, twoclass, and weights
#> • Other settings: use_weights = 'use', predict_raw = 'FALSE'

# Define a Task
task = tsk("sonar")
# Create train and test set
ids = partition(task)

# Train the learner on the training ids
learner$train(task, row_ids = ids$train)

print(learner$model)
#> 
#> Call:
#>  randomForest(formula = formula, data = data, classwt = classwt,      cutoff = cutoff, importance = TRUE) 
#>                Type of random forest: classification
#>                      Number of trees: 500
#> No. of variables tried at each split: 7
#> 
#>         OOB estimate of  error rate: 20.86%
#> Confusion matrix:
#>    M  R class.error
#> M 69  6    0.080000
#> R 23 41    0.359375
print(learner$importance())
#>           V11            V4           V12            V9           V48 
#>  0.0316129139  0.0138589383  0.0133359000  0.0126389133  0.0124607107 
#>           V10           V36           V51           V31           V49 
#>  0.0083676655  0.0076419594  0.0065931679  0.0064340265  0.0054390245 
#>           V28           V45           V16           V13            V5 
#>  0.0049338732  0.0048667085  0.0048192976  0.0047306763  0.0041316585 
#>           V46           V59           V47           V43           V52 
#>  0.0041222678  0.0041142089  0.0033366197  0.0032947241  0.0032565882 
#>           V15           V35           V21           V44           V14 
#>  0.0032442556  0.0032353950  0.0031816291  0.0029761916  0.0029201734 
#>            V2           V30           V18           V27           V32 
#>  0.0023341066  0.0022783049  0.0021528476  0.0021382423  0.0020739860 
#>           V37            V8           V42           V39            V1 
#>  0.0020433651  0.0020374758  0.0019601152  0.0017587228  0.0017437847 
#>           V17           V20           V19           V40            V3 
#>  0.0016247670  0.0014696451  0.0013199943  0.0013130621  0.0012997490 
#>           V55            V6           V29           V22           V26 
#>  0.0012692226  0.0011832342  0.0011045891  0.0010692009  0.0010115940 
#>           V58           V53           V33           V23           V25 
#>  0.0008854017  0.0008502598  0.0008065125  0.0007588061  0.0007192420 
#>           V34           V54           V38           V41           V50 
#>  0.0006316297  0.0006045380  0.0005649073  0.0005336866  0.0004084154 
#>           V56           V60            V7           V57           V24 
#>  0.0002929881  0.0001088185 -0.0001429174 -0.0003437627 -0.0009607324 

# Make predictions for the test rows
predictions = learner$predict(task, row_ids = ids$test)

# Score the predictions
predictions$score()
#> classif.ce 
#>  0.1594203