The hardware and bandwidth for this mirror is donated by METANET, the Webhosting and Full Service-Cloud Provider.
If you wish to report a bug, or if you are interested in having us mirror your free-software or open-source project, please feel free to contact us at mirror[@]metanet.ch.
This file contains some examples of the transfo
function to robustly transform
variables toward central normality
library(cellWise)
library(robustHD) # for the TopGear data
We start with a small toy example, in which standard normal data with one NA is used.
set.seed(1)
X = rnorm(100)
X[50] = NA
qqnorm(X)
We specify the transformation type as "bestObj"
, which chooses between Box-Cox and
Yeo-Johnson based on the value of the objective function. When there are negative
values, Yeo-Johnson is chosen automatically. First, the classical MLE transformation
is fit, indicated by the robust = FALSE
argument:
ML.out <- transfo(X, type = "bestObj", robust=FALSE)
##
## The input data has 100 rows and 1 columns.
ML.out$lambdahat
## [1] 1.033255
The value of the transformation parameter is close to 1, indicating no transformation is necesarry. Among the other outputs are the value of the objective function, the transformed data, estimates of the location and scale of the transformed data:
ML.out$objective
## [1] -139.9404
ML.out$Y[45:55]
## [1] -0.8798848 -0.9002897 0.2827949 0.7382678 -0.2480593 NA
## [7] 0.3204162 -0.7962567 0.2564864 -1.3578235 1.4957508
ML.out$muhat[1:20]
## [1] 0.0115562 NA NA NA NA NA NA
## [8] NA NA NA NA NA NA NA
## [15] NA NA NA NA NA NA
ML.out$sigmahat[1:20]
## [1] 0.9997822 NA NA NA NA NA NA
## [8] NA NA NA NA NA NA NA
## [15] NA NA NA NA NA NA
qqnorm(ML.out$Y); abline(0,1)
plot(ML.out$weights)
ML.out$ttypes[1:20]
## [1] "YJ" NA NA NA NA NA NA NA NA NA NA NA NA NA NA
## [16] NA NA NA NA NA
We can now repeat the same procedure with the robust estimator.
This is specified using the robust = TRUE
argument. The results do not differ
too much from the classical MLE estimates on this dataset, since there are no
outliers.
RewML.out <- transfo(X, type = "bestObj", robust=TRUE)
##
## The input data has 100 rows and 1 columns.
RewML.out$lambdahat
## [1] 1.099335
RewML.out$objective
## [1] 0.03190617
RewML.out$Y[45:55]
## [1] -0.8844857 -0.9047186 0.3005872 0.7842931 -0.2495009 NA
## [7] 0.3401820 -0.8014036 0.2729412 -1.3547824 1.6057686
RewML.out$muhat
## [1] 0.03539293
RewML.out$sigmahat
## [1] 0.9554364
qqnorm(RewML.out$Y); abline(0,1)
plot(RewML.out$weights)
X[61]
## [1] 2.401618
RewML.out$ttypes
## [1] "YJ"
In order to illustrate the same concepts for the Box-Cox transformation, we take the exponential of the generated data. This yields a sample from a log-normal distribution.
X = exp(X)
We first consider the classical maximum likelihood estimator. We obtain a transformation parameter close to 0, which corresponds with the logarithmic transformation.
ML.out <- transfo(X, type = "BC", robust=FALSE)
##
## The input data has 100 rows and 1 columns.
ML.out$lambdahat
## [1] 0.01965332
ML.out$objective
## [1] 8.401073
ML.out$Y[45:55]
## [1] -0.8803983 -0.9009148 0.2850604 0.7384490 -0.2456236 NA
## [7] 0.3225503 -0.7963140 0.2588371 -1.3607939 1.4921386
qqnorm(ML.out$Y); abline(0,1)
The weights are again all equal to 1, except for the NA (number 50) which gets weight zero:
plot(ML.out$weights)
ML.out$ttypes
## [1] "BC"
We now repeat the example with the robust transformation. The estimated transformation parameter is again fairly close to 0, as expected.
RewML.out <- transfo(X, type = "bestObj", robust=TRUE)
##
## The input data has 100 rows and 1 columns.
RewML.out$lambdahat
## [1] 0.1143505
RewML.out$objective
## [1] 0.03654314
RewML.out$Y[45:55]
## [1] -0.8926229 -0.9129574 0.3233614 0.8299447 -0.2453834 NA
## [7] 0.3645181 -0.8089065 0.2946509 -1.3593892 1.7158910
qqnorm(RewML.out$Y); abline(0,1)
RewML.out$ttypes
## [1] "BC"
We briefly illustrate the functions transfo_newdata() and tranfo_transformback. We start with transfo_newdata(). Given the output of transfo(), this function applies the fitted transformation to new data.
set.seed(123); tempraw <- matrix(rnorm(2000), ncol = 2)
tempx <- cbind(tempraw[, 1],exp(tempraw[, 2]))
tempy <- 0.5 * tempraw[, 1] + 0.5 * tempraw[, 2] + 1
x <- tempx[1:900, ]
y <- tempy[1:900]
tx.out <- transfo(x, type = "bestObj")
##
## The input data has 900 rows and 2 columns.
tx.out$ttypes
## [1] "YJ" "BC"
tx.out$lambdahats
## [1] 0.98744959 0.05316413
tx <- tx.out$Y
lm.out <- lm(y ~ tx)
summary(lm.out)
##
## Call:
## lm(formula = y ~ tx)
##
## Residuals:
## Min 1Q Median 3Q Max
## -0.141203 -0.004470 0.006658 0.011011 0.028085
##
## Coefficients:
## Estimate Std. Error t value Pr(>|t|)
## (Intercept) 1.0204157 0.0006331 1611.8 <2e-16 ***
## txX1 0.4763250 0.0006115 779.0 <2e-16 ***
## txX2 0.4821366 0.0006122 787.6 <2e-16 ***
## ---
## Signif. codes: 0 '***' 0.001 '**' 0.01 '*' 0.05 '.' 0.1 ' ' 1
##
## Residual standard error: 0.01899 on 897 degrees of freedom
## Multiple R-squared: 0.9993, Adjusted R-squared: 0.9993
## F-statistic: 6.62e+05 on 2 and 897 DF, p-value: < 2.2e-16
xnew <- tempx[901:1000, ]
xtnew <- transfo_newdata(xnew, tx.out)
yhatnew <- cbind(1, xtnew) %*% lm.out$coefficients
plot(tempy[901:1000], yhatnew); abline(0, 1)
We now illustrat the transfo_transformback() function. Given a transformation fitted by transfo(), this function applies the inverse of this transformation to the input.
set.seed(123); x <- matrix(rnorm(2000), ncol = 2)
y <- sqrt(abs(0.3 * x[, 1] + 0.5 * x[, 2] + 4))
ty.out <- transfo(y, type = "BC")
##
## The input data has 1000 rows and 1 columns.
ty.out$lambdahats
## [1] 2.028286
ty <- ty.out$Y
lm.out <- lm(ty ~ x)
yhat <- transfo_transformback(lm.out$fitted.values, ty.out)
plot(y, yhat); abline(0, 1)
This dataset is part of the robustHD
package. We use two of its variables
to illustrate the effect outliers can have on estimating a transformation
using maximum likelihood. The robust estimators yield more reasonable results.
data(TopGear)
First, we do some preprocessing to avoid zeroes.
CD.out <- checkDataSet(TopGear)
##
## The input data has 297 rows and 32 columns.
##
## The input data contained 19 non-numeric columns (variables).
## Their column names are:
##
## [1] Maker Model Type Fuel
## [5] DriveWheel AdaptiveHeadlights AdjustableSteering AlarmSystem
## [9] Automatic Bluetooth ClimateControl CruiseControl
## [13] ElectricSeats Leather ParkingSensors PowerSteering
## [17] SatNav ESP Origin
##
## These columns will be ignored in the analysis.
## We continue with the remaining 13 numeric columns:
##
## [1] Price Cylinders Displacement BHP Torque
## [6] Acceleration TopSpeed MPG Weight Length
## [11] Width Height Verdict
##
## The data contained 1 rows with over 50% of NAs.
## Their row names are:
##
## [1] 70
##
## These rows will be ignored in the analysis.
## We continue with the remaining 296 rows:
##
## [1] 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18
## [19] 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36
## [37] 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54
## [55] 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 71 72 73
## [73] 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91
## [91] 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109
## [109] 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127
## [127] 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145
## [145] 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163
## [163] 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181
## [181] 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199
## [199] 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217
## [217] 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235
## [235] 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253
## [253] 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271
## [271] 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289
## [289] 290 291 292 293 294 295 296 297
##
## The data contained 1 columns with zero or tiny median absolute deviation.
## Their column names are:
##
## [1] Cylinders
##
## These columns will be ignored in the analysis.
## We continue with the remaining 12 columns:
##
## [1] Price Displacement BHP Torque Acceleration
## [6] TopSpeed MPG Weight Length Width
## [11] Height Verdict
##
## The final data set we will analyze has 296 rows and 12 columns.
##
colnames(CD.out$remX)
## [1] "Price" "Displacement" "BHP" "Torque" "Acceleration"
## [6] "TopSpeed" "MPG" "Weight" "Length" "Width"
## [11] "Height" "Verdict"
# remove the subjective variable `Verdict':
X <- CD.out$remX[,-12]
carnames <- TopGear[CD.out$rowInAnalysis, 1:2]
X <- pmax(X, 1e-8) # avoids zeroes
Now we estimate transformation parameters (takes under a second) using classical MLE as well as the reweighted MLE which is much more robust.
ML.out <- transfo(X, type = "BC", robust=F)
##
## The input data has 296 rows and 11 columns.
RewML.out <- transfo(X, type = "BC", robust=T)
##
## The input data has 296 rows and 11 columns.
Now consider the transformation of the variable MPG
.
ML.out$lambdahat[7]
## [1] -0.1077643
RewML.out$lambdahat[7]
## [1] 0.8360462
The classical method estimates a transformation parameter of -0.11, which is close to the log transform. The robust method barely transforms the original variable with a transformation parameter of 0.84, relatively close to 1. The QQplots below suggest that the robust transformation is much more reasonable:
qqnorm(X[, 7], main = "", cex.lab = 1.5, cex.axis = 1.5)
qqnorm(ML.out$Y[, 7], main = "Classical transform",
cex.lab = 1.5, cex.axis = 1.5)
abline(0, 1)
qqnorm(RewML.out$Y[, 7], main = "Robustly transformed",
cex.lab = 1.5, cex.axis = 1.5)
abline(0, 1)
Now consider the transformation of the variable Weight
.
ML.out$lambdahat[8]
## [1] 0.8260084
RewML.out$lambdahat[8]
## [1] 0.09032881
Here, the situation is reversed: the classical estimator finds a transformation parameter (0.83) relatively close to one, whereas the robust transformation finds a parameter (0.09) close to the log transform. The QQplots below again suggest that the robust transformation is much more reasonable:
qqnorm(X[, 8], main = "Original variable",
cex.lab = 1.5, cex.axis = 1.5)
qqnorm(ML.out$Y[, 8], main = "Classical transform",
cex.lab = 1.5, cex.axis = 1.5)
abline(0, 1)
qqnorm(RewML.out$Y[, 8], main = "Robustly transformed",
cex.lab = 1.5, cex.axis = 1.5)
abline(0, 1)
In this example we study the glass data. It consists of spectra with 750 wavelengths of 180 archaeological glass samples.
data("data_glass")
First, we do some preprocessing to avoid variables with very small scales (in fact, the first 13 variables have mad equal to zero). Afterwards, we focus on the first 500 wavelengths since this is where most of the activity occurs.
X <- as.matrix(data_glass[, -c(1:13)])
X <- X[, 1:500]
Z <- scale(X, center=FALSE, robustbase::colMedians(X))
dim(Z)
## [1] 180 500
Now we estimate transformation parameters using classical MLE as well as the reweighted MLE which is much more robust. This only takes a few seconds for the 500 variables.
ML.out <- transfo(Z, type = "YJ", robust=F)
##
## The input data has 180 rows and 500 columns.
RewML.out <- transfo(Z, type = "YJ", robust=T)
##
## The input data has 180 rows and 500 columns.
We now construct the cellmaps:
indcells_clas = which(abs(ML.out$Y) > sqrt(qchisq(0.99, 1)))
indcells_rob = which(abs(RewML.out$Y) > sqrt(qchisq(0.99, 1)))
n = dim(ML.out$Y)[1]
d = dim(ML.out$Y)[2]; d
## [1] 500
nrowsinblock = 5
rowlabels = rep("", floor(n/nrowsinblock));
ncolumnsinblock = 5
columnlabels = rep("",floor(d/ncolumnsinblock));
columnlabels[3] = "1";
columnlabels[floor(d/ncolumnsinblock)] = "500"
CM_clas = cellMap(ML.out$Y,
nrowsinblock = nrowsinblock,
ncolumnsinblock = ncolumnsinblock,
rowblocklabels = rowlabels,
columnblocklabels = columnlabels,
mTitle = "YJ transformed variables by ML",
rowtitle = "", columntitle = "wavelengths",
columnangle = 0)
plot(CM_clas)
CM_rob = cellMap(RewML.out$Y,
nrowsinblock = nrowsinblock,
ncolumnsinblock = ncolumnsinblock,
rowblocklabels = rowlabels,
columnblocklabels = columnlabels,
mTitle = "YJ transformed variables by RewML",
rowtitle = "", columntitle = "wavelengths",
columnangle = 0)
plot(CM_rob)
# pdf("Glass_YJ_ML_RewML.pdf",width=8,height=8)
# gridExtra::grid.arrange(CM_clas, CM_rob,ncol=1)
# dev.off()
As a last example, we analyze the DPOSS data. This is a random subset of 20’000 stars from the Digitized Palomar Sky Survey described by Odewahn et al (1998).
data("data_dposs") # in package cellWise
n = nrow(data_dposs); n
## [1] 20000
ncol(data_dposs)
## [1] 21
There are lots of missing values in this dataset. Therefore, we first do some preprocessing to select the band of wavelengths which contain the fewest missing values.
missmat = is.na(data_dposs)
sizemat = nrow(missmat)*ncol(missmat); sizemat
## [1] 420000
100*sum(as.vector(missmat))/sizemat
## [1] 50.20952
missrow = length(which(rowSums(missmat) > 0))
100*missrow/nrow(missmat)
## [1] 84.61
# Missingness by band:
# F band:
300*sum(as.vector(missmat[,1:7]))/sizemat
## [1] 44.75
100*length(which(rowSums(missmat[,1:7]) > 0))/20000
## [1] 44.75
# J band:
300*sum(as.vector(missmat[,8:14]))/sizemat
## [1] 42.43857
100*length(which(rowSums(missmat[,8:14]) > 0))/20000
## [1] 42.61
# N band:
300*sum(as.vector(missmat[,15:21]))/sizemat
## [1] 63.44
100*length(which(rowSums(missmat[,15:21]) > 0))/20000
## [1] 63.44
# So typically the whole band is missing or not.
# We focus on the J band which has the most available rows.
indx = which(rowSums(missmat[,8:14]) ==0)
dpossJ = data_dposs[indx,8:14]
dim(dpossJ)
## [1] 11478 7
A quick exploration of the skewness in the data shows that there is skewness in both directions:
par(mfrow = c(1, 1))
boxplot(scale(dpossJ))
Now we fit the Yeo-Johnson transformation robustly and transform the data with the estimated parameters. The analysis shows that there are both lambdas larger than 1 and smaller than 1, corresponding with the skewness in both directions.
transfoJ_YJ <- transfo(dpossJ, type = "YJ", robust = T)
##
## The input data has 11478 rows and 7 columns.
plot(transfoJ_YJ$lambdahat)
dpossJ_YJ = transfoJ_YJ$Y
We now apply cellwise robust PCA to further analyze the data. We perform the PCA using both the original and the transformed data.
DDCPars = list(fastDDC=F,fracNA=0.5)
MacroPCAPars = list(DDCpars=DDCPars,scale=TRUE,silent=T)
MacroPCAdpossJ = MacroPCA(dpossJ,k=4,MacroPCApars=MacroPCAPars)
MacroPCAdpossJ_YJ = MacroPCA(dpossJ_YJ,k=4,MacroPCApars=MacroPCAPars)
Let’s make a plot of the scores of YJ-transformed data:
MacroPCAdpossJ_YJ$scores[, 1] <- -MacroPCAdpossJ_YJ$scores[, 1]
cols <- rep("black", dim(MacroPCAdpossJ$scores)[1])
cols[c(98, 10894)] <- "darkorange"
cols[which(MacroPCAdpossJ_YJ$SD > 14)] <- "skyblue3"
cols[which(MacroPCAdpossJ_YJ$SD > 25)] <- "firebrick"
# pdf("dpossJ_scores_YJ.pdf")
pairs(MacroPCAdpossJ_YJ$scores,gap=0,main="",pch=19,col=cols)
# dev.off()
We finally show the outlier maps on the transformed and the untransformed data:
# pdf("dpossJ_outliermap_YJ.pdf",height=5,width=5)
outlierMap(MacroPCAdpossJ_YJ, title="Outlier map of transformed data", col=cols, labelOut=FALSE)
# dev.off()
# pdf("dpossJ_outliermap_rawdata.pdf",height=5,width=5)
outlierMap(MacroPCAdpossJ, title="Outlier map of raw data", col=cols, labelOut=FALSE)
# dev.off()
These binaries (installable software) and packages are in development.
They may not be fully stable and should be used with caution. We make no claims about them.