The hardware and bandwidth for this mirror is donated by METANET, the Webhosting and Full Service-Cloud Provider.
If you wish to report a bug, or if you are interested in having us mirror your free-software or open-source project, please feel free to contact us at mirror[@]metanet.ch.
Improved predictive models by indirect classification and bagging for classification, regression and survival problems as well as resampling based estimators of prediction error.
Version: | 0.9-15 |
Depends: | R (≥ 2.10) |
Imports: | rpart (≥ 3.1-8), MASS, survival, nnet, class, prodlim |
Suggests: | mvtnorm, mlbench, TH.data, randomForest, party |
Published: | 2024-07-18 |
DOI: | 10.32614/CRAN.package.ipred |
Author: | Andrea Peters [aut], Torsten Hothorn [aut, cre], Brian D. Ripley [ctb], Terry Therneau [ctb], Beth Atkinson [ctb] |
Maintainer: | Torsten Hothorn <Torsten.Hothorn at R-project.org> |
License: | GPL-2 | GPL-3 [expanded from: GPL (≥ 2)] |
Copyright: | see file COPYRIGHTS |
NeedsCompilation: | yes |
Materials: | NEWS |
In views: | Environmetrics, MachineLearning, Survival |
CRAN checks: | ipred results |
Reference manual: | ipred.pdf |
Vignettes: |
Some more or less useful examples for illustration. |
Package source: | ipred_0.9-15.tar.gz |
Windows binaries: | r-devel: ipred_0.9-15.zip, r-release: ipred_0.9-15.zip, r-oldrel: ipred_0.9-15.zip |
macOS binaries: | r-release (arm64): ipred_0.9-15.tgz, r-oldrel (arm64): ipred_0.9-15.tgz, r-release (x86_64): ipred_0.9-15.tgz, r-oldrel (x86_64): ipred_0.9-15.tgz |
Old sources: | ipred archive |
Reverse imports: | ICcforest, LTRCforests, mlearning, nlcv, optBiomarker, permimp, pheble, pomodoro, recipes, RecordLinkage, RTextTools, survcomp, utsf |
Reverse suggests: | butcher, caret, censored, fastml, flowml, fscaret, MLInterfaces, party, pdp, subsemble, SuperLearner, superMICE |
Please use the canonical form https://CRAN.R-project.org/package=ipred to link to this page.
These binaries (installable software) and packages are in development.
They may not be fully stable and should be used with caution. We make no claims about them.