The hardware and bandwidth for this mirror is donated by METANET, the Webhosting and Full Service-Cloud Provider.
If you wish to report a bug, or if you are interested in having us mirror your free-software or open-source project, please feel free to contact us at mirror[@]metanet.ch.
PPforest
packageN. da Silva, D. Cook & E. Lee
The PPforest
package (projection pursuit random forest)
contains functions to run a projection pursuit random forest for
classification problems. This method utilize combinations of variables
in each tree construction. In a random forest each split is based on a
single variable, chosen from a subset of predictors. In the
PPforest
, each split is based on a linear combination of
randomly chosen variables. The linear combination is computed by
optimizing a projection pursuit index, to get a projection of the
variables that best separates the classes. The PPforest
uses the PPtree
algorithm, which fits a single tree to the
data. Utilizing linear combinations of variables to separate classes
takes the correlation between variables into account, and can outperform
the basic forest when separations between groups occurs on combinations
of variables. Two projection pursuit indexes, LDA and PDA, are used for
PPforest
.
To improve the speed performance PPforest
package,
PPtree
algorithm was translated to Rcpp.
PPforest
package utilizes a number of R packages some of
them included in “suggests” not to load them all at package start-up.
The development version ofPPforest
can be installed from
github using:
library(devtools)
install_github("natydasilva/PPforest")
library(PPforest)
PPforest
package implements a classification random
forest using projection pursuit classification trees. The following
table present all the functions in PPforest
package.
Function | Description |
---|---|
baggtree | For each bootstrap sample grow a projection pursuit tree (PPtree object). |
node_data | Data structure with the projected and boundary by node and class |
permute_importance | Obtain the permuted importance variable measure |
ppf_avg_imp | Computes a global importance measure for a PPforest object, average importance measure for a pptree over all the trees. |
PPclassify | Predict class for the test set and calculate prediction error after finding the PPtree structure |
ppf_global_imp | Computes a global importance measure for a PPforest object |
PPforest | Runs a Projection pursuit random forest |
PPtree_split | Projection pursuit classification tree with random variable selection in each split |
print.PPforest | Print PPforest object |
predict.PPforest | Predict class for the test set and calculate prediction error |
ternary_str | Data structure with the projected and boundary by node and class |
tree_pred | Obtain predicted class for new data using PPforest t. |
Also PPforest
package includes some data set that were
used to test the predictive performance of our method. The data sets
included are: crab, fishcatch, glass, image, leukemia, lymphoma NCI60,
parkinson and wine.
Australian crab data set will be used as example. This data contains measurements on rock crabs of the genus Leptograpsus. There are 200 observations from two species (blue and orange) and for each specie (50 in each one) there are 50 males and 50 females. Class variable has 4 classes with the combinations of specie and sex (BlueMale, BlueFemale, OrangeMale and OrangeFemale). The data were collected on site at Fremantle, Western Australia. For each specimen, five measurements were made, using vernier calipers.
PPforest
function runs a projection pursuit random
forest. The arguments are a data frame with the data information, class
with the name of the class variable argument. size.tr to specify the
proportion of observations using in the training. Using this function we
have the option to split the data in training and test using size.tr
directly. size.tr
is the proportion of data used in the
training and the test proportion will be 1- size.tr
. The
number of trees in the forest is specified using the argument
m
. The argument size.p is the sample proportion of the
variables used in each node split, PPmethod
is the
projection pursuit index to be optimized, two options LDA and PDA are
available.
<- PPforest::PPforest(data = crab, class = "Type", size.tr = 1, m = 200,
pprf.crab size.p = .5, PPmethod = 'LDA', parallel =TRUE, cores = 2)
pprf.crab
:
Call::PPforest(data = crab, class = "Type", size.tr = 1, m = 200, PPmethod = "LDA", size.p = 0.5, parallel = TRUE, cores = 2)
PPforest: Classification
Type of random forest: 200
Number of trees: 2
No. of variables tried at each split
: 5%
OOB estimate of error rate:
Confusion matrix
BlueFemale BlueMale OrangeFemale OrangeMale class.error49 1 0 0 0.02
BlueFemale 6 44 0 0 0.12
BlueMale 0 0 47 3 0.06
OrangeFemale 0 0 0 50 0.00 OrangeMale
PPforest
print a summary result from the model with the
confusion matrix information and the oob-error rate in a similar way
randomForest packages does.
This function returns the predicted values of the training data, training error, test error and predicted test values. Also there is the information about out of bag error for the forest and also for each tree in the forest. Bootstrap samples, output of all the trees in the forest from , proximity matrix and vote matrix, number of trees grown in the forest, number of predictor variables selected to use for splitting at each node. Confusion matrix of the prediction (based on OOb data), the training data and test data and vote matrix are also returned.
The printed version of a PPforest
object follows the
randomForest
printed version to make them comparable. Based
on confusion matrix, we can observe that the biggest error is for
BlueMale class. Most of the wrong classified values are between
BlueFemale and BlueMale.
str(pprf.crab,max.level=1 )
21
List of $ predicting.training: Factor w/ 4 levels "BlueFemale","BlueMale",..: 2 1 2 2 2 2 1 2 2 1 ...
$ training.error : num 0.045
$ prediction.test : NULL
$ error.test : NULL
$ oob.error.forest : num 0.05
$ oob.error.tree : num [1:200, 1] 0.1918 0.0597 0.12 0.1519 0.303 ...
$ boot.samp :List of 200
$ output.trees :List of 200
$ proximity : num [1:200, 1:200] 0 0.72 0.835 0.85 0.78 0.865 0.32 0.875 0.765 0.29 ...
$ votes : num [1:200, 1:4] 0.375 0.605 0.372 0.417 0.253 ...
$ prediction.oob : Factor w/ 4 levels "BlueFemale","BlueMale",..: 2 1 2 2 2 2 1 2 2 1 ...
$ n.tree : num 200
$ n.var : num 2
$ type : chr "Classification"
$ confusion : num [1:4, 1:5] 49 6 0 0 1 44 0 0 0 0 ...
- attr(*, "dimnames")=List of 2
..$ call : language PPforest::PPforest(data = crab, class = "Type", size.tr = 1, m = 200, PPmethod = "LDA", size.p = 0.5, parall| __truncated__
$ train :Classes ‘tbl_df’, ‘tbl’ and 'data.frame': 200 obs. of 6 variables:
$ test : NULL
$ vote.mat : num [1:200, 1:200] 1 2 4 2 1 2 1 2 4 2 ...
- attr(*, "dimnames")=List of 2
..$ class.var : chr "Type"
$ oob.obs : num [1:200, 1:200] 1 0 1 0 0 1 0 1 0 0 ...
- attr(*, "class")= chr "PPforest"
These binaries (installable software) and packages are in development.
They may not be fully stable and should be used with caution. We make no claims about them.