The hardware and bandwidth for this mirror is donated by METANET, the Webhosting and Full Service-Cloud Provider.
If you wish to report a bug, or if you are interested in having us mirror your free-software or open-source project, please feel free to contact us at mirror[@]metanet.ch.

1. First Steps

Martin Binder

2024-02-20

CPO Vignette Navigation

## Loading required package: ParamHelpers
## Loading required package: mlr
## Warning message: 'mlr' is in 'maintenance-only' mode since July 2019.
## Future development will only happen in 'mlr3'
## (<https://mlr3.mlr-org.com>). Due to the focus on 'mlr3' there might be
## uncaught bugs meanwhile in {mlr} - please consider switching.
    1. First Steps (compact version)
  1. mlrCPO Core (compact version)
  2. CPOs Built Into mlrCPO (compact version)
  3. Building Custom CPOs (compact version)

Table of Contents

About the Vignettes

Since mlrCPO is a package with some depth to it, it comes with a few vignettes that each explain different aspects of its operation. These are the current document (“First Steps”), offering a short introduction and information on where to get started, “mlrCPO Core”, describing all the functions and tools offered by mlrCPO that are independent from specific CPOs, “CPOs Built Into mlrCPO”, listing all CPOs included in the mlrCPO package, and “Building Custom CPOs”, describing the process of creating new CPOs that offer new functionality.

All vignettes also have a “compact version” with the R output suppressed for readability. They are linked in the navigation section at the top.

All vignettes assume that mlrCPO (and therefore its requirement mlr) is installed successfully and loaded using library("mlrCPO"). Help with installation is provided on the project’s GitHub page.

What is mlrCPO?

“Composable Preprocessing Operators”, “CPO”, are an extension for the mlr (“Machine Learning in R”) project which present preprocessing operations in the form of R objects. These CPO objects can be composed to form complex operations, they can be applied to data sets, and can be attached to mlr Learner objects to generate machine learning pipelines that combine preprocessing and model fitting.

What is Preprocessing

“Preprocessing”, as understood by mlrCPO, is any manipulation of data used in a machine learning process to get it from its form as found in the wild into a form more fitting for the machine learning algorithm (“Learner”) used for model fitting. It is important that the exact method of preprocessing is kept track of, to be able to perform this method when the resulting model is used to make predictions on new data. It is also important, when evaluating preprocessing methods e.g. using resampling, that the parameters of these methods are independent of the validation dataset and only depend on the training data set.

mlrCPO tries to support the user in all these aspects of preprocessing:

  1. By providing a large set of atomic preprocessing CPOs that can perform many different operations. Operations that go beyond the provided toolset can be implemented in custom CPOs.
  2. By using “CPOTrained” objects that represent the preprocessing done on training data that should, in that way, be re-applied to new prediction data.
  3. By making it possible to combine preprocessing objects with mlrLearner” objects that represent the entinre machine learning pipeline to be tuned and evaluated.

Preprocessing Operations

At the centre of mlrCPO are “CPO” objects. To get a CPO object, it is necessary to call a CPO Constructor. A CPO Constructor sets up the parameters of a CPO and provides further options for its behaviour. Internally, CPO Constructors are functions that have a common interface and a friendly printer method.

cpoScale  # a cpo constructor
#> <<CPO scale(center = TRUE, scale = TRUE)>>
cpoAddCols
#> <<CPO new.cols(..., .make.factors = TRUE)>>
cpoScale(center = FALSE)  # create a CPO object that scales, but does not center, data
#> scale(center = FALSE, scale = TRUE)
cpoAddCols(Sepal.Area = Sepal.Length * Sepal.Width)  #  this would add a column
#> new.cols()[not exp'd: expr = Sepal.Area=<call>, superceding.env = <environment>, make.factors = TRUE, add.cols = TRUE]

CPOs exist first to be applied to data. Every CPO represents a certain data transformation, and this transformation is performed when the CPO is applied. This can be done using the applyCPO function, or the %>>% operator. CPOs can be applied to data.frame objects, and to mlrTask” objects.

iris.demo = iris[c(1, 2, 3, 51, 52, 102, 103), ]
tail(iris.demo %>>% cpoQuantileBinNumerics())  # bin the data in below & above median
#>     Sepal.Length Sepal.Width Petal.Length Petal.Width    Species
#> 2     (-Inf,5.8]  (-Inf,3.2]   (-Inf,4.5]  (-Inf,1.4]     setosa
#> 3     (-Inf,5.8]  (-Inf,3.2]   (-Inf,4.5]  (-Inf,1.4]     setosa
#> 51    (5.8, Inf]  (-Inf,3.2]   (4.5, Inf]  (-Inf,1.4] versicolor
#> 52    (5.8, Inf]  (-Inf,3.2]   (-Inf,4.5]  (1.4, Inf] versicolor
#> 102   (-Inf,5.8]  (-Inf,3.2]   (4.5, Inf]  (1.4, Inf]  virginica
#> 103   (5.8, Inf]  (-Inf,3.2]   (4.5, Inf]  (1.4, Inf]  virginica

A useful feature of CPOs is that they can be concatenated to form new operations. Two CPOs can be combined using the composeCPO function or, as before, the %>>% operator. When two CPOs are combined, the product is a new CPO that can itself be composed or applied. The result of a composition represents the operation of first applying the first CPO and then the second CPO. Therefore, data %>>% (cpo1 %>>% cpo2) is the same as (data %>>% cpo1) %>>% cpo2.

# first create three quantile bins, then as.numeric() all columns to
# get 1, 2 or 3 as the bin number
quantilenum = cpoQuantileBinNumerics(numsplits = 3) %>>% cpoAsNumeric()
iris.demo %>>% quantilenum
#>     Sepal.Length Sepal.Width Petal.Length Petal.Width Species
#> 1              1           3            1           1       1
#> 2              1           1            1           1       1
#> 3              1           2            1           1       1
#> 51             3           2            2           2       2
#> 52             2           2            2           2       2
#> 102            2           1            3           3       3
#> 103            3           1            3           3       3

The last example shows that it is sometimes not a good idea to have a CPO affect the whole dataset. Therefore, when a CPO is created, it is possible to choose what columns the CPO should affect. The CPO Constructor has a variety of parameters, starting with affect., that can be used to choose what columns the CPO operates on. To prevent cpoAsNumeric from influencing the Species column, we can thus do

quantilenum.restricted = cpoQuantileBinNumerics(numsplits = 3) %>>%
  cpoAsNumeric(affect.names = "Species", affect.invert = TRUE)
iris.demo %>>% quantilenum.restricted
#>     Sepal.Length Sepal.Width Petal.Length Petal.Width    Species
#> 1              1           3            1           1     setosa
#> 2              1           1            1           1     setosa
#> 3              1           2            1           1     setosa
#> 51             3           2            2           2 versicolor
#> 52             2           2            2           2 versicolor
#> 102            2           1            3           3  virginica
#> 103            3           1            3           3  virginica

A more convenient method in this case, however, is to use an mlrTask”, which keeps track of the target column. “Feature Operation” CPOs (as all the ones shown) do not influence the target column.

demo.task = makeClassifTask(data = iris.demo, target = "Species")
result = demo.task %>>% quantilenum
getTaskData(result)
#>     Sepal.Length Sepal.Width Petal.Length Petal.Width    Species
#> 1              1           3            1           1     setosa
#> 2              1           1            1           1     setosa
#> 3              1           2            1           1     setosa
#> 51             3           2            2           2 versicolor
#> 52             2           2            2           2 versicolor
#> 102            2           1            3           3  virginica
#> 103            3           1            3           3  virginica

Hyperparameters

When performing preprocessing, it is sometimes necessary to change a small aspect of a long preprocessing pipeline. Instead of having to re-construct the whole pipeline, mlrCPO offers the possibility to change hyperparameters of a CPO. This makes it very easy e.g. for tuning of preprocessing in combination with a machine learning algorithm.

Hyperparameters of CPOs can be manipulated in the same way as they are manipulated for Learners in mlr, using getParamSet (to list the parameters), getHyperPars (to list the parameter values), and setHyperPars (to change these values). To get the parameter set of a CPO, it is also possible to use verbose printing using the ! (exclamation mark) operator.

cpo = cpoScale()
cpo
#> scale(center = TRUE, scale = TRUE)
getHyperPars(cpo)  # list of parameter names and values
#> $scale.center
#> [1] TRUE
#> 
#> $scale.scale
#> [1] TRUE
getParamSet(cpo)  # more detailed view of parameters and their type / range
#>                 Type len  Def Constr Req Tunable Trafo
#> scale.center logical   - TRUE      -   -    TRUE     -
#> scale.scale  logical   - TRUE      -   -    TRUE     -
!cpo  # equivalent to print(cpo, verbose = TRUE)
#> Trafo chain of 1 cpos:
#> scale(center = TRUE, scale = TRUE)
#> Operating: feature
#> ParamSet:
#>                 Type len  Def Constr Req Tunable Trafo
#> scale.center logical   - TRUE      -   -    TRUE     -
#> scale.scale  logical   - TRUE      -   -    TRUE     -

CPOs use copy semantics, therefore setHyperPars creates a copy of a CPO that has the changed hyperparameters.

cpo2 = setHyperPars(cpo, scale.scale = FALSE)
cpo2
#> scale(center = TRUE, scale = FALSE)
iris.demo %>>% cpo  # scales and centers
#>     Sepal.Length Sepal.Width Petal.Length Petal.Width    Species
#> 1    -0.75858917   1.5562541   -1.0268892  -1.0276683     setosa
#> 2    -0.95897121  -0.4611123   -1.0268892  -1.0276683     setosa
#> 3    -1.15935326   0.3458343   -1.0764631  -1.0276683     setosa
#> 51    1.14504025   0.3458343    0.6090515   0.3874815 versicolor
#> 52    0.54389412   0.3458343    0.5099036   0.5054106 versicolor
#> 102  -0.05725201  -1.6715322    0.8073474   0.9771273  virginica
#> 103   1.24523128  -0.4611123    1.2039390   1.2129856  virginica
iris.demo %>>% cpo2 # only centers
#>     Sepal.Length Sepal.Width Petal.Length Petal.Width    Species
#> 1    -0.75714286  0.38571429    -2.071429  -0.8714286     setosa
#> 2    -0.95714286 -0.11428571    -2.071429  -0.8714286     setosa
#> 3    -1.15714286  0.08571429    -2.171429  -0.8714286     setosa
#> 51    1.14285714  0.08571429     1.228571   0.3285714 versicolor
#> 52    0.54285714  0.08571429     1.028571   0.4285714 versicolor
#> 102  -0.05714286 -0.41428571     1.628571   0.8285714  virginica
#> 103   1.24285714 -0.11428571     2.428571   1.0285714  virginica

When chaining many CPOs, it is possible for the many hyperparameters to lead to very cluttered ParamSets, or even for hyperparameter names to clash. mlrCPO has two remedies for that.

First, any CPO also has an id that is always prepended to the hyperparameter names. It can be set during construction, using the id parameter, or changed later using setCPOId. The latter one only works on primitive, i.e. not compound, CPOs. Set the id to NULL to use the CPO’s hyperparameters without a prefix.

cpo = cpoScale(id = "a") %>>% cpoScale(id = "b")  # not very useful example
getHyperPars(cpo)
#> $a.center
#> [1] TRUE
#> 
#> $a.scale
#> [1] TRUE
#> 
#> $b.center
#> [1] TRUE
#> 
#> $b.scale
#> [1] TRUE

The second remedy against hyperparameter clashes is different “exports” of hyperparameters: The hyperparameters that can be changed using setHyperPars, i.e. that are exported by a CPO, are a subset of the parameters of the CPOConstructor. For each kind of CPO, there is a standard set of parameters that are exported, but during construction, it is possible to influence the parameters that actually get exported via the export parameter. export can be one of a set of standard export settings (among them “export.all” and “export.none”) or a character vector of the parameters to export.

cpo = cpoPca(export = c("center", "rank"))
getParamSet(cpo)
#>               Type len    Def   Constr Req Tunable Trafo
#> pca.center logical   -   TRUE        -   -    TRUE     -
#> pca.rank   integer   - <NULL> 1 to Inf   -    TRUE     -

Retrafo

Manipulating data for preprocessing itself is relatively easy. A challenge comes when one wants to integrate preprocessing into a machine-learning pipeline: The same preprocessing steps that are performed on the training data need to be performed on the new prediction data. However, the transformation performed for prediction often needs information from the training step. For example, if training entail performing PCA, then for prediction, the data must not undergo another PCA, instead it needs to be rotated by the rotation matrix found by the training PCA. The process of obtaining the rotation matrix will be called “training” the CPO, and the object that contains the trained information is called CPOTrained. For preprocessing operations that operate only on features of a task (as opposed to the target column), the CPOTrained will always be applied to new incoming data, and hence be of class CPORetrafo and called a “retrafo” object. To obtain this retrafo object, one can use retrafo(). Retrafo objects can be applied to data just as CPOs can, by using the %>>% operator.

transformed = iris.demo %>>% cpoPca(rank = 3)
transformed
#>        Species       PC1         PC2         PC3
#> 1       setosa -2.380452 -0.27548909  0.17256456
#> 2       setosa -2.434101  0.09825278 -0.21123031
#> 3       setosa -2.606806  0.15705569  0.02579403
#> 51  versicolor  1.600832 -0.59734881 -0.09942477
#> 52  versicolor  1.232156 -0.11798420  0.10744449
#> 102  virginica  1.672966  0.84503760  0.01891907
#> 103  virginica  2.915405 -0.10952398 -0.01406707
ret = retrafo(transformed)
ret
#> CPO Retrafo chain
#> [RETRAFO pca(center = TRUE, scale = FALSE)]

To show that ret actually represents the exact same preprocessing operation, we can feed the first line of iris.demo back to it, to verify that the transformation is the same.

iris.demo[1, ] %>>% ret
#>   Species       PC1        PC2       PC3
#> 1  setosa -2.380452 -0.2754891 0.1725646

We obviously would not have gotten there by feeding the first line to cpoPca directly:

iris.demo[1, ] %>>% cpoPca(rank = 3)
#>   Species PC1
#> 1  setosa   0

CPOTrained objects associated with an object are automatically chained when another CPO is applied. To prevent this from happening, it is necessary to “clear” the retrafos and inverters associated with the object using clearRI().

t2 = transformed %>>% cpoScale()
retrafo(t2)
#> CPO Retrafo chain
#> [RETRAFO pca(center = TRUE, scale = FALSE)] =>
#> [RETRAFO scale(center = TRUE, scale = TRUE)]
t3 = clearRI(transformed) %>>% cpoScale()
retrafo(t3)
#> CPO Retrafo chain
#> [RETRAFO scale(center = TRUE, scale = TRUE)]

Note that clearRI has no influence on the CPO operations themselves, and the resulting data is the same:

all.equal(t2, t3, check.attributes = FALSE)
#> [1] TRUE

It is also possible to chain CPOTrained object using composeCPO() or %>>%. This can be useful if the trafo chain loses access to the retrafo attribute for some reason. In general, it is only recommended to compose CPOTrained objects that were created in the same process and in correct order, since they are usually closely associated with the training data in a particular place within the preprocessing chain.

retrafo(transformed) %>>% retrafo(t3)  # is the same as retrafo(t2) above.
#> CPO Retrafo chain
#> [RETRAFO pca(center = TRUE, scale = FALSE)] =>
#> [RETRAFO scale(center = TRUE, scale = TRUE)]

Inverter

So far only CPOs were introduced that change the feature columns of a Task. (“Feature Operation CPOs”–FOCPOs). There is another class of CPOs, “Target Operation CPOs” or TOCPOs, that can change a Task’s target columns.

This comes at the cost of some complexity when performing prediction: Since the training data that was ultimately fed into a Learner had a transformed target column, the predictions made by the resulting model will not be directly comparable to the original target values. Consider cpoLogTrafoRegr, a CPO that log-transforms the target variable of a regression Task. The predictions made with a Learner on a log-transformed target variable will be in log-space and need to be exponentiated (or otherwise re-transformed). This inversion operation is represented by an “inverter” object that is attached to a transformation result similarly to a retrafo object, and can be obtained using the inverter() function. It is of class CPOInverter, a subclass of CPOTrained.

iris.regr = makeRegrTask(data = iris.demo, target = "Petal.Width")
iris.logd = iris.regr %>>% cpoLogTrafoRegr()

getTaskData(iris.logd)  # log-transformed target 'Petal.Width'
#>     Sepal.Length Sepal.Width Petal.Length Petal.Width    Species
#> 1            5.1         3.5          1.4  -1.6094379     setosa
#> 2            4.9         3.0          1.4  -1.6094379     setosa
#> 3            4.7         3.2          1.3  -1.6094379     setosa
#> 51           7.0         3.2          4.7   0.3364722 versicolor
#> 52           6.4         3.2          4.5   0.4054651 versicolor
#> 102          5.8         2.7          5.1   0.6418539  virginica
#> 103          7.1         3.0          5.9   0.7419373  virginica
inv = inverter(iris.logd)  # inverter object
inv
#> CPO Inverter chain {type:regr} (able to predict 'response', 'se')
#> [INVERTER fun.apply.regr.target(){type:regr}]

The inverter object is used by the invert() function that inverts the prediction made by a model trained on the transformed task, and re-transforms this prediction to fit the space of the original target data. The inverter object caches the “truth” of the data being inverted (iris.logd, in the example), so invert can give information on the truth of the inverted data.

logmodel = train("regr.lm", iris.logd)
pred = predict(logmodel, iris.logd)  # prediction on the task itself
pred
#> Prediction: 7 observations
#> predict.type: response
#> threshold: 
#> time: 0.00
#>     id      truth   response
#> 1    1 -1.6094379 -1.6155677
#> 2    2 -1.6094379 -1.6182648
#> 3    3 -1.6094379 -1.5944813
#> 51   4  0.3364722  0.3429698
#> 52   5  0.4054651  0.3989676
#> 102  6  0.6418539  0.6416087
#> ... (#rows: 7, #cols: 3)
invert(inv, pred)
#> Prediction: 7 observations
#> predict.type: response
#> threshold: 
#> time: 0.00
#>     id truth  response
#> 1    1   0.2 0.1987778
#> 2    2   0.2 0.1982424
#> 3    3   0.2 0.2030138
#> 51   4   1.4 1.4091262
#> 52   5   1.5 1.4902853
#> 102  6   1.9 1.8995342
#> ... (#rows: 7, #cols: 3)

This procedure can also be done with new incoming data. In general, more than just the cpoLogTrafoRegr operation could be done on the iris.regr task in the example, so to perform the complete preprocessing and inversion, one needs to use the retrafo object as well. When applying the retrafo object, a new inverter object is generated, which is specific to the exact new data that was being retransformed:

newdata = makeRegrTask("newiris", iris[7:9, ], target = "Petal.Width",
  fixup.data = "no", check.data = FALSE)
# the retrafo does the same transformation(s) on newdata that were
# done on the training data of the model, iris.logd. In general, this
# could be more than just the target log transformation.
newdata.transformed = newdata %>>% retrafo(iris.logd)
getTaskData(newdata.transformed)
#>   Sepal.Length Sepal.Width Petal.Length Petal.Width Species
#> 7          4.6         3.4          1.4   -1.203973  setosa
#> 8          5.0         3.4          1.5   -1.609438  setosa
#> 9          4.4         2.9          1.4   -1.609438  setosa
pred = predict(logmodel, newdata.transformed)
pred
#> Prediction: 3 observations
#> predict.type: response
#> threshold: 
#> time: 0.00
#>   id     truth  response
#> 7  1 -1.203973 -1.494340
#> 8  2 -1.609438 -1.548810
#> 9  3 -1.609438 -1.497037
# the inverter of the newly transformed data contains information specific
# to the newly transformed data. In the current case, that is just the
# new "truth" column for the new data.
inv.newdata = inverter(newdata.transformed)
invert(inv.newdata, pred)
#> Prediction: 3 observations
#> predict.type: response
#> threshold: 
#> time: 0.00
#>   id truth  response
#> 7  1   0.3 0.2243966
#> 8  2   0.2 0.2125006
#> 9  3   0.2 0.2237922

Constant Inverters

The cpoLogTrafoRegr is a special case of TOCPO in that its inversion operation is constant: It does not depend on the new incoming data, so in theory it is not necessary to get a new inverter object for every piece of data that is being transformed. Therefore, it is possible to use the retrafo object for inversion in this case. However, the “truth” column will not be available in this case:

invert(retrafo(iris.logd), pred)
#> Prediction: 3 observations
#> predict.type: response
#> threshold: 
#> time: 0.00
#>   id  response
#> 7  1 0.2243966
#> 8  2 0.2125006
#> 9  3 0.2237922

Whether a retrafo object is capable of performing inversion can be checked with the getCPOTrainedCapability() function. It returns a vector with named elements "retrafo" and "invert", indicating whether a CPOTrained is capable of performing retrafo or inversion. A 1 indicates that the object can perform the action and has an effect, a 0 indicates that the action would have no effect (but also throws no error), and a -1 means that the object is not capable of performing the action.

getCPOTrainedCapability(retrafo(iris.logd))  # can do both retrafo and inversion
#> retrafo  invert 
#>       1       1
getCPOTrainedCapability(inv)  # a pure inverter, can not be used for retrafo
#> retrafo  invert 
#>      -1       1

General Inverters

As an example of a CPO that does not have a constant inverter, consider cpoRegrResiduals, wich fits a regression model on training data and returns the residuals of this fit. When performing prediction, the invert action is to add predictions by the CPO’s model to the incoming predictions made by a model trained on the residuals.

set.seed(123)  # for reproducibility
iris.resid = iris.regr %>>% cpoRegrResiduals("regr.lm")
getTaskData(iris.resid)
#>     Sepal.Length Sepal.Width Petal.Length   Petal.Width    Species
#> 1            5.1         3.5          1.4  0.0086897922     setosa
#> 2            4.9         3.0          1.4  0.0125133007     setosa
#> 3            4.7         3.2          1.3 -0.0212030929     setosa
#> 51           7.0         3.2          4.7 -0.0092111797 versicolor
#> 52           6.4         3.2          4.5  0.0092111797 versicolor
#> 102          5.8         2.7          5.1  0.0003475917  virginica
#> 103          7.1         3.0          5.9 -0.0003475917  virginica
model.resid = train("regr.randomForest", iris.resid)

newdata.resid = newdata %>>% retrafo(iris.resid)
getTaskData(newdata.resid)  # Petal.Width are now the residuals of lm model predictions
#>   Sepal.Length Sepal.Width Petal.Length Petal.Width Species
#> 7          4.6         3.4          1.4  -0.0883876  setosa
#> 8          5.0         3.4          1.5  -0.1045896  setosa
#> 9          4.4         2.9          1.4  -0.1845641  setosa
pred = predict(model.resid, newdata.resid)
pred
#> Prediction: 3 observations
#> predict.type: response
#> threshold: 
#> time: 0.00
#>   id      truth     response
#> 7  1 -0.0883876 -0.001403442
#> 8  2 -0.1045896  0.002960334
#> 9  3 -0.1845641 -0.001075935
# transforming this prediction back to compare
# it to the original 'Petal.Width'
inv.newdata = inverter(newdata.resid)
invert(inv.newdata, pred)
#> Prediction: 3 observations
#> predict.type: response
#> threshold: 
#> time: 0.00
#>   id truth  response
#> 7  1   0.3 0.3869842
#> 8  2   0.2 0.3075500
#> 9  3   0.2 0.3834882

Retrafoless CPOs

Besides FOCPOs and TOCPOs, there are also “RetrafolessCPOs (ROCPOs). These only perform operation in the training part of a machine learning pipeline, but in turn are the only CPOs that may change the number of rows in a dataset. The goal of ROCPOs is to change the number of data samples, but not to transform the data or target values themselves. Examples of ROCPOs are cpoUndersample, cpoSmote, and cpoSample.

sampled = iris %>>% cpoSample(size = 3)
sampled
#>    Sepal.Length Sepal.Width Petal.Length Petal.Width    Species
#> 82          5.5         2.4          3.7         1.0 versicolor
#> 99          5.1         2.5          3.0         1.1 versicolor
#> 84          6.0         2.7          5.1         1.6 versicolor

There is no retrafo or inverter associated with the result. Instead, both of them are NULLCPO

retrafo(sampled)
#> NULLCPO
inverter(sampled)
#> NULLCPO

CPO Learners

Until now, the CPOs have been invoked explicitly to manipulate data and get retrafo and inverter objects. It is good to be aware of the data flows in a machine learning process involving preprocessing, but mlrCPO makes it very easy to automatize this. It is possible to attach a CPO to a Learner using attachCPO or the %>>%-operator. When a CPO is attached to a Learner, a CPOLearner is created. The CPOLearner performs the preprocessing operation dictated by the CPO before training the underlying model, and stores and uses the retrafo and inverter objects necessary during prediction. It is possible to attach compound CPOs, and it is possible to attach further CPOs to a CPOLearner to extend the preprocessing pipeline. Exported hyperparamters of a CPO are also present in a CPOLearner and can be changed using setHyperPars, as usual with other Learner objects.

Recreating the pipeline from General Inverters with a CPOLearner looks like the following. Note the prediction pred made in the end is identical with the one made above.

set.seed(123)  # for reproducibility
lrn = cpoRegrResiduals("regr.lm") %>>% makeLearner("regr.randomForest")
lrn
#> Learner regr.randomForest.regr.residuals from package randomForest
#> Type: regr
#> Name: ; Short name: 
#> Class: CPOLearner
#> Properties: numerics,factors,oobpreds,featimp
#> Predict-Type: response
#> Hyperparameters: regr.residuals.crr.train.residuals=plain,regr.residuals.crr.resampling=<CVDesc>
model = train(lrn, iris.regr)

pred = predict(model, newdata)
pred
#> Prediction: 3 observations
#> predict.type: response
#> threshold: 
#> time: 0.00
#>   id truth  response
#> 7  1   0.3 0.3869842
#> 8  2   0.2 0.3075500
#> 9  3   0.2 0.3834882

It is possible to get the retrafo object from a model trained with a CPOLearner using the retrafo() function. In this example, it is identical with the retrafo(iris.resid) gotten in the example in General Inverters.

retrafo(model)
#> CPO Retrafo chain {type:regr}
#> [RETRAFO regr.residuals(crr.train.residuals = plain, crr.resampling = <CVDesc>)]

CPO Tuning

Since the hyperparameters of a CPO are present in a CPOLearner, is possible to tune hyperparameters of preprocessing operations. It can be done using mlr’s tuneParams() function and works identically to tuning common Learner-parameters.

icalrn = cpoIca() %>>% makeLearner("classif.logreg")

getParamSet(icalrn)
#>                 Type len      Def             Constr Req Tunable Trafo
#> ica.n.comp   integer   -   <NULL>           1 to Inf   -    TRUE     -
#> ica.alg.typ discrete   - parallel parallel,deflation   -    TRUE     -
#> ica.fun     discrete   -  logcosh        logcosh,exp   -    TRUE     -
#> ica.alpha    numeric   -        1             1 to 2   Y    TRUE     -
#> model        logical   -     TRUE                  -   -   FALSE     -
ps = makeParamSet(
    makeIntegerParam("ica.n.comp", lower = 1, upper = 8),
    makeDiscreteParam("ica.alg.typ", values = c("parallel", "deflation")))
# shorter version using pSS:
# ps = pSS(ica.n.comp: integer[1, 8], ica.alg.typ: discrete[parallel, deflation])
tuneParams(icalrn, pid.task, cv5, par.set = ps,
  control = makeTuneControlGrid(),
  show.info = FALSE)
#> Tune result:
#> Op. pars: ica.n.comp=7; ica.alg.typ=parallel
#> mmce.test.mean=0.2265427

Syntactic Sugar

Besides the %>>% operator, there are a few related operators which are short forms of operations that otherwise take more typing.

Inspecting CPOs

As described before, it is possible to compose CPOs to create relatively complex preprocessing pipelines. It is therefore necessary to have tools to inspect a CPO pipeline or related objects.

The first line of attack when inspecting a CPO is always the print function. print(x, verbose = TRUE) will often print more information about a CPO than the ordinary print function. A shorthand alias for this is the exclamation point “!”. When verbosely printing a CPOConstructor, the transformation functions are shown. When verbosely printing a CPO, the constituent elements are separately printed, each showing their parameter sets.

cpoAsNumeric  # plain print
#> <<CPO as.numeric()>>
!cpoAsNumeric  # verbose print
#> <<CPO as.numeric()>>
#> 
#> cpo.retrafo:
#> function (data) 
#> {
#>     as.data.frame(lapply(data, as.numeric), row.names = rownames(data))
#> }
#> <environment: namespace:mlrCPO>
cpoScale() %>>% cpoIca()  # plain print
#> (scale >> ica)(scale.center = TRUE, scale.scale = TRUE, ica.n.comp = <NULL>, ica.alg.typ = parallel, ica.fun = logcosh, ica.alpha = 1)
!cpoScale() %>>% cpoIca()  # verbose print
#> Trafo chain of 2 cpos:
#> scale(center = TRUE, scale = TRUE)
#> Operating: feature
#> ParamSet:
#>                 Type len  Def Constr Req Tunable Trafo
#> scale.center logical   - TRUE      -   -    TRUE     -
#> scale.scale  logical   - TRUE      -   -    TRUE     -
#>   ====>
#> ica(n.comp = <NULL>, alg.typ = parallel, fun = logcosh, alpha = 1)[not exp'd: method = C, maxit = 200, tol = 0.0001, verbose = FALSE]
#> Operating: feature
#> ParamSet:
#>                 Type len      Def             Constr Req Tunable Trafo
#> ica.n.comp   integer   -   <NULL>           1 to Inf   -    TRUE     -
#> ica.alg.typ discrete   - parallel parallel,deflation   -    TRUE     -
#> ica.fun     discrete   -  logcosh        logcosh,exp   -    TRUE     -
#> ica.alpha    numeric   -        1             1 to 2   Y    TRUE     -

When working with compound CPOs, it is sometimes necessary to manipulate a CPO inside a compound CPO pipeline. For this purpose, the as.list() generic is implemented for both CPO and CPOTrained for splitting a pipeline into a list of the primitive elements. The inverse is pipeCPO(), which takes a list of CPO or CPOTrained and concatenates them using composeCPO().

as.list(cpoScale() %>>% cpoIca())
#> [[1]]
#> scale(center = TRUE, scale = TRUE)
#> 
#> [[2]]
#> ica(n.comp = <NULL>, alg.typ = parallel, fun = logcosh, alpha = 1)[not exp'd: method = C, maxit = 200, tol = 0.0001, verbose = FALSE]
pipeCPO(list(cpoScale(), cpoIca()))
#> (scale >> ica)(scale.center = TRUE, scale.scale = TRUE, ica.n.comp = <NULL>, ica.alg.typ = parallel, ica.fun = logcosh, ica.alpha = 1)

CPOTrained objects contain information about the retrafo or inversion to be performed for a CPO. It is possible to access this information using getCPOTrainedState(). The “state” of a CPOTrained object often contains a $data slot with information about the expected input and output format (“ShapeInfo”) of incoming data, a slot for each of its hyperparameters, and a $control slot that is specific to the CPO in question. The cpoPca state, for example, contains the PCA rotation matrix and a vector for scaling and centering. The contents of a state’s $control object are described in a CPO’s help page.

repca = retrafo(iris.demo %>>% cpoPca())
state = getCPOTrainedState(repca)
state
#> $center
#> [1] TRUE
#> 
#> $scale
#> [1] FALSE
#> 
#> $tol
#> NULL
#> 
#> $rank
#> NULL
#> 
#> $control
#> $control$rotation
#>                      PC1        PC2         PC3        PC4
#> Sepal.Length  0.39017855 -0.8270920 -0.33170933  0.2316213
#> Sepal.Width  -0.04877314 -0.4166469  0.90027349 -0.1163364
#> Petal.Length  0.84914741  0.2328611  0.09372219 -0.4646940
#> Petal.Width   0.35260539  0.2968162  0.26588101  0.8466858
#> 
#> $control$scale
#> [1] FALSE
#> 
#> $control$center
#> Sepal.Length  Sepal.Width Petal.Length  Petal.Width 
#>     5.857143     3.114286     3.471429     1.071429 
#> 
#> 
#> $data
#> $data$shapeinfo.input
#> <ShapeInfo (input) Sepal.Length: num, Sepal.Width: num, Petal.Length: num, Petal.Width: num, Species: fac>
#> 
#> $data$shapeinfo.output
#> <ShapeInfo (output)>:
#> numeric:
#> <ShapeInfo PC1: num, PC2: num, PC3: num, PC4: num>
#> factor:
#> <ShapeInfo Species: fac>
#> other:
#> <ShapeInfo (empty)>

It is even possible to change the “state” of a CPOTrained and construct a new CPOTrained using makeCPOTrainedFromState(). This is fairly advanced usage and only recommended for users familiar with the inner workings of the particular CPO. If we get familiar with the cpoPca CPO using the !-print (i.e. !cpoPca) to look at the retrafo function, we notice that the control$center and control$scale values are given to a call of scale(). If we want to create a new CPOTrained that does not perform centering or scaling during before applying the rotation matrix, we can change these values.

state$control$center = FALSE
state$control$scale = FALSE
nosc.repca = makeCPOTrainedFromState(cpoPca, state)

Comparing this to the original “repca” retrafo shows that the result of applying repca has generally smaller values because of the centering.

iris.demo %>>% repca
#>        Species       PC1         PC2         PC3           PC4
#> 1       setosa -2.380452 -0.27548909  0.17256456  0.0045112647
#> 2       setosa -2.434101  0.09825278 -0.21123031  0.0163551932
#> 3       setosa -2.606806  0.15705569  0.02579403 -0.0067669498
#> 51  versicolor  1.600832 -0.59734881 -0.09942477 -0.0379746458
#> 52  versicolor  1.232156 -0.11798420  0.10744449  0.0006599373
#> 102  virginica  1.672966  0.84503760  0.01891907 -0.0202867763
#> 103  virginica  2.915405 -0.10952398 -0.01406707  0.0435019768
iris.demo %>>% nosc.repca
#>        Species      PC1       PC2      PC3       PC4
#> 1       setosa 3.078532 -5.291065 1.643627 0.2928570
#> 2       setosa 3.024883 -4.917323 1.259832 0.3047009
#> 3       setosa 2.852178 -4.858520 1.496856 0.2815788
#> 51  versicolor 7.059816 -5.612924 1.371638 0.2503711
#> 52  versicolor 6.691140 -5.133560 1.578507 0.2890056
#> 102  virginica 7.131950 -4.170538 1.489981 0.2680589
#> 103  virginica 8.374389 -5.125100 1.456995 0.3318477

Special CPOs

There is a large and growing variety of CPOs that perform many different operations. It is advisable to browse through CPOs Built Into mlrCPO for an overview. To get a list of all built-in CPOs, use listCPO(). A few important or “meta” CPOs that can be used to influence the behaviour of other CPOs are described here.

NULLCPO

The value associated with “no operation” is the NULLCPO value. It is the neutral element of the %>>% operations, and the value of retrafo() and inverter() when there are otherwise no associated retrafo or inverter values.

NULLCPO
#> NULLCPO
all.equal(iris %>>% NULLCPO, iris)
#> [1] TRUE
cpoPca() %>>% NULLCPO
#> pca(center = TRUE, scale = FALSE)[not exp'd: tol = <NULL>, rank = <NULL>]

CPO Multiplexer

The multiplexer makes it possible to combine many CPOs into one, with an extra selected.cpo parameter that chooses between them. This makes it possible to tune over many different tuner configurations at once.

cpm = cpoMultiplex(list(cpoIca, cpoPca(export = "export.all")))
!cpm
#> Trafo chain of 1 cpos:
#> multiplex(selected.cpo = ica, ica.n.comp = <NULL>, ica.alg.typ = parallel, ica.fun = logcosh, ica.alpha = 1, pca.center = TRUE, pca.scale = FALSE, pca.tol = <NULL>, pca.rank = <NULL>)
#> Operating: feature
#> ParamSet:
#>                  Type len      Def             Constr Req Tunable Trafo
#> selected.cpo discrete   -      ica            ica,pca   -    TRUE     -
#> ica.n.comp    integer   -   <NULL>           1 to Inf   Y    TRUE     -
#> ica.alg.typ  discrete   - parallel parallel,deflation   Y    TRUE     -
#> ica.fun      discrete   -  logcosh        logcosh,exp   Y    TRUE     -
#> ica.alpha     numeric   -        1             1 to 2   Y    TRUE     -
#> pca.center    logical   -     TRUE                  -   Y    TRUE     -
#> pca.scale     logical   -    FALSE                  -   Y    TRUE     -
#> pca.tol       numeric   -   <NULL>           0 to Inf   Y    TRUE     -
#> pca.rank      integer   -   <NULL>           1 to Inf   Y    TRUE     -
iris.demo %>>% setHyperPars(cpm, selected.cpo = "ica", ica.n.comp = 3)
#>        Species          V1         V2         V3
#> 1       setosa  1.23124905  1.4008953 -0.5331872
#> 2       setosa -0.15798393 -1.5434691 -1.4475137
#> 3       setosa  0.08546248  0.4435976 -1.1814006
#> 51  versicolor  0.94186702 -1.2301687  0.9556098
#> 52  versicolor  0.23294666  0.7485481  0.7833032
#> 102  virginica -2.11220122  0.5195243  0.1262646
#> 103  virginica -0.22134005 -0.3389275  1.2969239
iris.demo %>>% setHyperPars(cpm, selected.cpo = "pca", pca.rank = 3)
#>        Species       PC1         PC2         PC3
#> 1       setosa -2.380452 -0.27548909  0.17256456
#> 2       setosa -2.434101  0.09825278 -0.21123031
#> 3       setosa -2.606806  0.15705569  0.02579403
#> 51  versicolor  1.600832 -0.59734881 -0.09942477
#> 52  versicolor  1.232156 -0.11798420  0.10744449
#> 102  virginica  1.672966  0.84503760  0.01891907
#> 103  virginica  2.915405 -0.10952398 -0.01406707

CPO Wrapper

A simple CPO with one parameter which gets applied to the data as CPO. This is different from a multiplexer in that its parameter is free and can take any value that behaves like a CPO. On the downside, this does not expose the argument’s parameters to the outside.

cpa = cpoWrap()
!cpa
#> Trafo chain of 1 cpos:
#> wrap()
#> Operating: feature
#> ParamSet:
#>             Type len Def Constr Req Tunable Trafo
#> wrap.cpo untyped   -   -      -   -    TRUE     -
iris.demo %>>% setHyperPars(cpa, wrap.cpo = cpoScale())
#>     Sepal.Length Sepal.Width Petal.Length Petal.Width    Species
#> 1    -0.75858917   1.5562541   -1.0268892  -1.0276683     setosa
#> 2    -0.95897121  -0.4611123   -1.0268892  -1.0276683     setosa
#> 3    -1.15935326   0.3458343   -1.0764631  -1.0276683     setosa
#> 51    1.14504025   0.3458343    0.6090515   0.3874815 versicolor
#> 52    0.54389412   0.3458343    0.5099036   0.5054106 versicolor
#> 102  -0.05725201  -1.6715322    0.8073474   0.9771273  virginica
#> 103   1.24523128  -0.4611123    1.2039390   1.2129856  virginica
iris.demo %>>% setHyperPars(cpa, wrap.cpo = cpoPca())
#>        Species       PC1         PC2         PC3           PC4
#> 1       setosa -2.380452 -0.27548909  0.17256456  0.0045112647
#> 2       setosa -2.434101  0.09825278 -0.21123031  0.0163551932
#> 3       setosa -2.606806  0.15705569  0.02579403 -0.0067669498
#> 51  versicolor  1.600832 -0.59734881 -0.09942477 -0.0379746458
#> 52  versicolor  1.232156 -0.11798420  0.10744449  0.0006599373
#> 102  virginica  1.672966  0.84503760  0.01891907 -0.0202867763
#> 103  virginica  2.915405 -0.10952398 -0.01406707  0.0435019768

Attaching the cpo applicator to a learner gives this learner a “cpo” hyperparameter that can be set to any CPO.

getParamSet(cpoWrap() %>>% makeLearner("classif.logreg"))
#>             Type len  Def Constr Req Tunable Trafo
#> wrap.cpo untyped   -    -      -   -    TRUE     -
#> model    logical   - TRUE      -   -   FALSE     -

CBind CPO

cbind other CPOs as operation. The cbinder makes it possible to build DAGs of CPOs that perform different operations on data and paste the results next to each other. It is often useful to combine cpoCbind with cpoSelect to filter out columns that would otherwise be duplciated.

scale = cpoSelect(pattern = "Sepal", id = "first") %>>% cpoScale(id = "scale")
scale.pca = scale %>>% cpoPca()
cbinder = cpoCbind(scale, scale.pca, cpoSelect(pattern = "Petal", id = "second"))

cpoCbind recognises that "scale" happens before "pca", but is also fed to the result directly. The verbose print draws a (crude) ascii-art graph.

!cbinder
#> Trafo chain of 1 cpos:
#> cbind(scale.center = TRUE, scale.scale = TRUE, pca.center = TRUE, pca.scale = FALSE)
#> Operating: feature
#> ParamSet:
#>                 Type len   Def Constr Req Tunable Trafo
#> scale.center logical   -  TRUE      -   -    TRUE     -
#> scale.scale  logical   -  TRUE      -   -    TRUE     -
#> pca.center   logical   -  TRUE      -   -    TRUE     -
#> pca.scale    logical   - FALSE      -   -    TRUE     -
#> O       first<select>()[not exp'd: type = character(0), index = integer(0),
#> |      names = character(0), pattern = Sepal, pattern.ignore.case = FALSE,
#> |      pattern.perl = FALSE, pattern.fixed = FALSE, invert = FALSE]
#> |      
#> O>+     scale(center = TRUE, scale = TRUE)
#> | |    
#> +--<O   second<select>()[not exp'd: type = character(0), index = integer(0),
#> | |  names = character(0), pattern = Petal, pattern.ignore.case = FALSE,
#> | |  pattern.perl = FALSE, pattern.fixed = FALSE, invert = FALSE]
#> | |  
#> +<O   pca(center = TRUE, scale = FALSE)[not exp'd: tol = <NULL>, rank =
#> |  <NULL>]
#> |  
#> O   CBIND[,,]
#> 
iris.demo %>>% cbinder
#>     Sepal.Length Sepal.Width        PC1        PC2 Petal.Length Petal.Width
#> 1    -0.75858917   1.5562541 -1.6368414  0.5640343          1.4         0.2
#> 2    -0.95897121  -0.4611123 -0.3520394 -1.0041507          1.4         0.2
#> 3    -1.15935326   0.3458343 -1.0643283 -0.5752448          1.3         0.2
#> 51    1.14504025   0.3458343  0.5651240  1.0542075          4.7         1.4
#> 52    0.54389412   0.3458343  0.1400495  0.6291330          4.5         1.5
#> 102  -0.05725201  -1.6715322  1.1414685 -1.2224350          5.1         1.9
#> 103   1.24523128  -0.4611123  1.2065671  0.5544558          5.9         2.1

Custom CPOs

Even though CPOs are very flexible and can be combined in many ways, it may be necessary to create completely custom CPOs. Custom CPOs can be created using the makeCPO() and related functions. “Building Custom CPOs” is a wide topic which has its own vignette.

Summary

These binaries (installable software) and packages are in development.
They may not be fully stable and should be used with caution. We make no claims about them.