Skip to contents

Adaptive best-subset selection for regression. Calls abess::abess() from abess.

Dictionary

This Learner can be instantiated via the dictionary mlr_learners or with the associated sugar function lrn():

mlr_learners$get("regr.abess")
lrn("regr.abess")

Meta Information

  • Task type: “regr”

  • Predict Types: “response”

  • Feature Types: “integer”, “numeric”

  • Required Packages: mlr3, abess

Parameters

IdTypeDefaultLevelsRange
familycharactergaussiangaussian, mgaussian, poisson, gamma-
tune.pathcharactersequencesequence, gsection-
tune.typecharactergicgic, aic, bic, ebic, cv-
normalizeintegerNULL\((-\infty, \infty)\)
support.sizeuntypedNULL-
c.maxinteger2\([1, \infty)\)
gs.rangeuntypedNULL-
lambdanumeric0\([0, \infty)\)
always.includeuntypedNULL-
group.indexuntypedNULL-
init.active.setuntypedNULL-
splicing.typeinteger2\([1, 2]\)
max.splicing.iterinteger20\([1, \infty)\)
screening.numintegerNULL\([0, \infty)\)
important.searchintegerNULL\([0, \infty)\)
warm.startlogicalTRUETRUE, FALSE-
nfoldsinteger5\((-\infty, \infty)\)
foldiduntypedNULL-
cov.updatelogicalFALSETRUE, FALSE-
newtoncharacterexactexact, approx-
newton.threshnumeric1e-06\([0, \infty)\)
max.newton.iterintegerNULL\([1, \infty)\)
early.stoplogicalFALSETRUE, FALSE-
ic.scalenumeric1\([0, \infty)\)
num.threadsinteger0\([0, \infty)\)
seedinteger1\((-\infty, \infty)\)

Initial parameter values

  • num.threads: This parameter is initialized to 1 (default is 0) to avoid conflicts with the mlr3 parallelization.

See also

Author

abess-team

Super classes

mlr3::Learner -> mlr3::LearnerRegr -> LearnerRegrAbess

Methods

Inherited methods


Method new()

Creates a new instance of this R6 class.

Usage


Method selected_features()

Extract the name of selected features from the model by abess::extract().

Usage

LearnerRegrAbess$selected_features()

Returns

The names of selected features


Method clone()

The objects of this class are cloneable with this method.

Usage

LearnerRegrAbess$clone(deep = FALSE)

Arguments

deep

Whether to make a deep clone.

Examples

learner = mlr3::lrn("regr.abess")
print(learner)
#> <LearnerRegrAbess:regr.abess>: Fast Best Subset Selection for Regression
#> * Model: -
#> * Parameters: num.threads=1
#> * Packages: mlr3, abess
#> * Predict Types:  [response]
#> * Feature Types: integer, numeric
#> * Properties: selected_features, weights

# available parameters:
learner$param_set$ids()
#>  [1] "family"            "tune.path"         "tune.type"        
#>  [4] "normalize"         "support.size"      "c.max"            
#>  [7] "gs.range"          "lambda"            "always.include"   
#> [10] "group.index"       "init.active.set"   "splicing.type"    
#> [13] "max.splicing.iter" "screening.num"     "important.search" 
#> [16] "warm.start"        "nfolds"            "foldid"           
#> [19] "cov.update"        "newton"            "newton.thresh"    
#> [22] "max.newton.iter"   "early.stop"        "ic.scale"         
#> [25] "num.threads"       "seed"