Classification Stochastic Gradient Descent Learner
Source:R/learner_RWeka_classif_sgd.R
mlr_learners_classif.sgd.Rd
Stochastic Gradient Descent for learning various linear models.
Calls RWeka::make_Weka_classifier()
from RWeka.
Initial parameter values
F
:Has only 2 out of 5 original loss functions: 0 = hinge loss (SVM) and 1 = log loss (logistic regression) with 0 (hinge loss) still being the default
Reason for change: this learner should only contain loss functions appropriate for classification tasks
Custom mlr3 parameters
output_debug_info
:original id: output-debug-info
do_not_check_capabilities
:original id: do-not-check-capabilities
num_decimal_places
:original id: num-decimal-places
batch_size
:original id: batch-size
Reason for change: This learner contains changed ids of the following control arguments since their ids contain irregular pattern
Parameters
Id | Type | Default | Levels | Range |
subset | untyped | - | - | |
na.action | untyped | - | - | |
F | character | 0 | 0, 1 | - |
L | numeric | 0.01 | \((-\infty, \infty)\) | |
R | numeric | 1e-04 | \((-\infty, \infty)\) | |
E | integer | 500 | \((-\infty, \infty)\) | |
C | numeric | 0.001 | \((-\infty, \infty)\) | |
N | logical | - | TRUE, FALSE | - |
M | logical | - | TRUE, FALSE | - |
S | integer | 1 | \((-\infty, \infty)\) | |
output_debug_info | logical | FALSE | TRUE, FALSE | - |
do_not_check_capabilities | logical | FALSE | TRUE, FALSE | - |
num_decimal_places | integer | 2 | \([1, \infty)\) | |
batch_size | integer | 100 | \([1, \infty)\) | |
options | untyped | NULL | - |
See also
as.data.table(mlr_learners)
for a table of available Learners in the running session (depending on the loaded packages).Chapter in the mlr3book: https://mlr3book.mlr-org.com/basics.html#learners
mlr3learners for a selection of recommended learners.
mlr3cluster for unsupervised clustering learners.
mlr3pipelines to combine learners with pre- and postprocessing steps.
mlr3tuning for tuning of hyperparameters, mlr3tuningspaces for established default tuning spaces.
Super classes
mlr3::Learner
-> mlr3::LearnerClassif
-> LearnerClassifSGD
Methods
Inherited methods
mlr3::Learner$base_learner()
mlr3::Learner$configure()
mlr3::Learner$encapsulate()
mlr3::Learner$format()
mlr3::Learner$help()
mlr3::Learner$predict()
mlr3::Learner$predict_newdata()
mlr3::Learner$print()
mlr3::Learner$reset()
mlr3::Learner$selected_features()
mlr3::Learner$train()
mlr3::LearnerClassif$predict_newdata_fast()
Method marshal()
Marshal the learner's model.
Arguments
...
(any)
Additional arguments passed tomlr3::marshal_model()
.
Method unmarshal()
Unmarshal the learner's model.
Arguments
...
(any)
Additional arguments passed tomlr3::unmarshal_model()
.
Examples
# Define the Learner
learner = lrn("classif.sgd")
print(learner)
#>
#> ── <LearnerClassifSGD> (classif.sgd): Stochastic Gradient Descent ──────────────
#> • Model: -
#> • Parameters: F=0
#> • Packages: mlr3 and RWeka
#> • Predict Types: [response] and prob
#> • Feature Types: logical, integer, numeric, factor, and ordered
#> • Encapsulation: none (fallback: -)
#> • Properties: marshal, missings, and twoclass
#> • Other settings: use_weights = 'error'
# Define a Task
task = tsk("sonar")
# Create train and test set
ids = partition(task)
# Train the learner on the training ids
learner$train(task, row_ids = ids$train)
print(learner$model)
#> Loss function: Hinge loss (SVM)
#>
#> Class =
#>
#> -1.1672 (normalized) V1
#> + -3.0147 (normalized) V10
#> + -3.2917 (normalized) V11
#> + -0.8023 (normalized) V12
#> + -0.8786 (normalized) V13
#> + 0.1862 (normalized) V14
#> + 0.0602 (normalized) V15
#> + 1.1144 (normalized) V16
#> + 2.5364 (normalized) V17
#> + -0.1094 (normalized) V18
#> + -0.6725 (normalized) V19
#> + -2.0367 (normalized) V2
#> + -0.9125 (normalized) V20
#> + -0.2266 (normalized) V21
#> + 0.5839 (normalized) V22
#> + -1.9306 (normalized) V23
#> + -1.8917 (normalized) V24
#> + 1.7985 (normalized) V25
#> + 0.0104 (normalized) V26
#> + -0.6607 (normalized) V27
#> + -1.5057 (normalized) V28
#> + 0.0212 (normalized) V29
#> + 0.6665 (normalized) V3
#> + -0.6203 (normalized) V30
#> + 3.9802 (normalized) V31
#> + 0.1471 (normalized) V32
#> + -1.5077 (normalized) V33
#> + 0.9172 (normalized) V34
#> + -0.2664 (normalized) V35
#> + 2.2081 (normalized) V36
#> + 2.1386 (normalized) V37
#> + -0.76 (normalized) V38
#> + 0.8816 (normalized) V39
#> + -1.558 (normalized) V4
#> + 1.1732 (normalized) V40
#> + -0.1865 (normalized) V41
#> + 1.2767 (normalized) V42
#> + -1.5716 (normalized) V43
#> + -2.6812 (normalized) V44
#> + -1.1456 (normalized) V45
#> + -0.4154 (normalized) V46
#> + 0.0491 (normalized) V47
#> + -2.9454 (normalized) V48
#> + -3.1848 (normalized) V49
#> + -1.5855 (normalized) V5
#> + 1.1506 (normalized) V50
#> + -1.7079 (normalized) V51
#> + -2.5509 (normalized) V52
#> + -0.1821 (normalized) V53
#> + -0.3289 (normalized) V54
#> + 1.213 (normalized) V55
#> + -0.5846 (normalized) V56
#> + 2.1777 (normalized) V57
#> + 0.3063 (normalized) V58
#> + -0.11 (normalized) V59
#> + 1.5445 (normalized) V6
#> + -0.4346 (normalized) V60
#> + 2.6715 (normalized) V7
#> + 0.7876 (normalized) V8
#> + -1.91 (normalized) V9
#> + 3.6
# Make predictions for the test rows
predictions = learner$predict(task, row_ids = ids$test)
# Score the predictions
predictions$score()
#> classif.ce
#> 0.2173913