Skip to contents

Function to optimize the features of a mlr3::Learner. The function internally creates a FSelectInstanceBatchSingleCrit or FSelectInstanceBatchMultiCrit which describes the feature selection problem. It executes the feature selection with the FSelector (method) and returns the result with the fselect instance ($result). The ArchiveBatchFSelect ($archive) stores all evaluated hyperparameter configurations and performance scores.

Usage

fselect(
  fselector,
  task,
  learner,
  resampling,
  measures = NULL,
  term_evals = NULL,
  term_time = NULL,
  terminator = NULL,
  store_benchmark_result = TRUE,
  store_models = FALSE,
  check_values = FALSE,
  callbacks = NULL,
  ties_method = "least_features"
)

Arguments

fselector

(FSelector)
Optimization algorithm.

task

(mlr3::Task)
Task to operate on.

learner

(mlr3::Learner)
Learner to optimize the feature subset for.

resampling

(mlr3::Resampling)
Resampling that is used to evaluated the performance of the feature subsets. Uninstantiated resamplings are instantiated during construction so that all feature subsets are evaluated on the same data splits. Already instantiated resamplings are kept unchanged.

measures

(mlr3::Measure or list of mlr3::Measure)
A single measure creates a FSelectInstanceBatchSingleCrit and multiple measures a FSelectInstanceBatchMultiCrit. If NULL, default measure is used.

term_evals

(integer(1))
Number of allowed evaluations. Ignored if terminator is passed.

term_time

(integer(1))
Maximum allowed time in seconds. Ignored if terminator is passed.

terminator

(bbotk::Terminator)
Stop criterion of the feature selection.

store_benchmark_result

(logical(1))
Store benchmark result in archive?

store_models

(logical(1)). Store models in benchmark result?

check_values

(logical(1))
Check the parameters before the evaluation and the results for validity?

callbacks

(list of CallbackBatchFSelect)
List of callbacks.

ties_method

(character(1))
The method to break ties when selecting sets while optimizing and when selecting the best set. Can be "least_features" or "random". The option "least_features" (default) selects the feature set with the least features. If there are multiple best feature sets with the same number of features, one is selected randomly. The random method returns a random feature set from the best feature sets. Ignored if multiple measures are used.

Details

The mlr3::Task, mlr3::Learner, mlr3::Resampling, mlr3::Measure and bbotk::Terminator are used to construct a FSelectInstanceBatchSingleCrit. If multiple performance Measures are supplied, a FSelectInstanceBatchMultiCrit is created. The parameter term_evals and term_time are shortcuts to create a bbotk::Terminator. If both parameters are passed, a bbotk::TerminatorCombo is constructed. For other Terminators, pass one with terminator. If no termination criterion is needed, set term_evals, term_time and terminator to NULL.

Resources

There are several sections about feature selection in the mlr3book.

The gallery features a collection of case studies and demos about optimization.

Analysis

For analyzing the feature selection results, it is recommended to pass the archive to as.data.table(). The returned data table is joined with the benchmark result which adds the mlr3::ResampleResult for each feature set.

The archive provides various getters (e.g. $learners()) to ease the access. All getters extract by position (i) or unique hash (uhash). For a complete list of all getters see the methods section.

The benchmark result ($benchmark_result) allows to score the feature sets again on a different measure. Alternatively, measures can be supplied to as.data.table().

Examples

# Feature selection on the Palmer Penguins data set
task = tsk("pima")
learner = lrn("classif.rpart")

# Run feature selection
instance = fselect(
  fselector = fs("random_search"),
  task = task,
  learner = learner,
  resampling = rsmp ("holdout"),
  measures = msr("classif.ce"),
  term_evals = 4)

# Subset task to optimized feature set
task$select(instance$result_feature_set)

# Train the learner with optimal feature set on the full data set
learner$train(task)

# Inspect all evaluated configurations
as.data.table(instance$archive)
#>        age glucose insulin   mass pedigree pregnant pressure triceps classif.ce
#>     <lgcl>  <lgcl>  <lgcl> <lgcl>   <lgcl>   <lgcl>   <lgcl>  <lgcl>      <num>
#>  1:   TRUE    TRUE    TRUE   TRUE     TRUE     TRUE     TRUE    TRUE  0.2500000
#>  2:  FALSE   FALSE    TRUE  FALSE     TRUE     TRUE    FALSE   FALSE  0.2929688
#>  3:  FALSE   FALSE   FALSE  FALSE     TRUE    FALSE    FALSE   FALSE  0.3203125
#>  4:  FALSE   FALSE   FALSE  FALSE    FALSE     TRUE    FALSE   FALSE  0.3281250
#>  5:   TRUE    TRUE    TRUE  FALSE     TRUE    FALSE     TRUE    TRUE  0.2851562
#>  6:   TRUE    TRUE   FALSE  FALSE     TRUE    FALSE     TRUE    TRUE  0.2851562
#>  7:   TRUE    TRUE    TRUE   TRUE     TRUE     TRUE     TRUE    TRUE  0.2500000
#>  8:   TRUE    TRUE    TRUE   TRUE     TRUE     TRUE     TRUE    TRUE  0.2500000
#>  9:   TRUE    TRUE    TRUE  FALSE     TRUE     TRUE    FALSE   FALSE  0.2734375
#> 10:   TRUE    TRUE    TRUE   TRUE     TRUE    FALSE     TRUE   FALSE  0.2734375
#>     runtime_learners           timestamp batch_nr warnings errors
#>                <num>              <POSc>    <int>    <int>  <int>
#>  1:            0.009 2024-09-11 08:05:03        1        0      0
#>  2:            0.008 2024-09-11 08:05:03        1        0      0
#>  3:            0.008 2024-09-11 08:05:03        1        0      0
#>  4:            0.007 2024-09-11 08:05:03        1        0      0
#>  5:            0.008 2024-09-11 08:05:03        1        0      0
#>  6:            0.009 2024-09-11 08:05:03        1        0      0
#>  7:            0.008 2024-09-11 08:05:03        1        0      0
#>  8:            0.008 2024-09-11 08:05:03        1        0      0
#>  9:            0.008 2024-09-11 08:05:03        1        0      0
#> 10:            0.009 2024-09-11 08:05:03        1        0      0
#>                                           features n_features  resample_result
#>                                             <list>     <list>           <list>
#>  1: age,glucose,insulin,mass,pedigree,pregnant,...          8 <ResampleResult>
#>  2:                      insulin,pedigree,pregnant          3 <ResampleResult>
#>  3:                                       pedigree          1 <ResampleResult>
#>  4:                                       pregnant          1 <ResampleResult>
#>  5:  age,glucose,insulin,pedigree,pressure,triceps          6 <ResampleResult>
#>  6:          age,glucose,pedigree,pressure,triceps          5 <ResampleResult>
#>  7: age,glucose,insulin,mass,pedigree,pregnant,...          8 <ResampleResult>
#>  8: age,glucose,insulin,mass,pedigree,pregnant,...          8 <ResampleResult>
#>  9:          age,glucose,insulin,pedigree,pregnant          5 <ResampleResult>
#> 10:     age,glucose,insulin,mass,pedigree,pressure          6 <ResampleResult>