The hardware and bandwidth for this mirror is donated by METANET, the Webhosting and Full Service-Cloud Provider.
If you wish to report a bug, or if you are interested in having us mirror your free-software or open-source project, please feel free to contact us at mirror[@]metanet.ch.
DiscreteGapStatistic estimates the number of clusters from (multiple choice) categorical response format data given any clustering algorithm extending the well-known gap statistic using a discrete distance based approach.
You can install the development version of DiscreteGapStatistic from GitHub with:
# install.packages("devtools")
::install_github("ecortesgomez/DiscreteGapStatistic") devtools
Basic usage of DiscreteGapStatistic
is shown using
likert
’s Math Anxiety data.
library(DiscreteGapStatistic)
library(dplyr)
#>
#> Attaching package: 'dplyr'
#> The following objects are masked from 'package:stats':
#>
#> filter, lag
#> The following objects are masked from 'package:base':
#>
#> intersect, setdiff, setequal, union
Libraries are uploaded. Questions are lightly reformatted and the categories are shortened.
## mass dataset is loaded automatically with DiscreteGapStatistic
<- colnames(mass)
massQ <- c('Gender',
massQsh 'Q1: Math Interesting', 'Q2: Uptight Math Test', 'Q3: Use Math In Future',
'Q4: Mind Goes Blank', 'Q5: Math Relates to Life', 'Q6: Ability Solve Math Probls',
'Q7: Sinking Feeling', 'Q8: Math Challenging', 'Q9: Math Makes Me Nervous',
'Q10: Take More Math Classes', 'Q11: Math Makes Me Uneasy','Q12: Math Favorite Subj.',
'Q13: Enjoy Learning Math', 'Q14: Math Makes Feel Confused')
<- mass
massData colnames(massData) <- massQsh
<- setNames(object = c('SD', 'D', 'N', 'A', 'SA'),
Cats nm = c('Strongly Disagree', 'Disagree',
'Neutral',
'Agree', 'Strongly Agree') )
<- data.frame(Gender = massData$Gender,
massData apply(massData[, -1], 2,
function(x) Cats[as.character(x)] %>%
factor(levels = Cats)),
check.names=FALSE)
rownames(massData) <- c(paste0('F', 1:2), paste0('M', 1),
paste0('F', 3:11), paste0('M', 2),
paste0('F', 12), paste0('M', 3),
paste0('F', 13), paste0('M', 4:5),
paste0('F', 14), paste0('M', 6))
The dataset is visualized using a heatmap similar to the one produced
by likert::likert.heatmap.plot
.
library(ggplot2)
likert.heat.plot2(massData[, -1],
allLevels = Cats,
text.size = 1.5)+
labs(title = 'Math Anxiety Data Likert Heatmap Summary')+
theme(axis.text = element_text(size = 6),
title = element_text(size = 6))
Five categorical distance functions are introduced to quantify
dissimilarities/discrepancies between two categorical vectors (see
function distancematrix
). The following table describes the
available distances and the names used within the package.
#>
#> Attaching package: 'kableExtra'
#> The following object is masked from 'package:dplyr':
#>
#> group_rows
Distance | Name |
---|---|
Hamming | hamming |
chi-square | chisquare |
Cramer’s V | cramerV |
Hellinger | hellinger |
Bhattacharyya | bhattacharyya |
The resulting distance matrix from a rectangular dataset can easily
be displayed and organized using heatmaps. The following plots visualize
the massData
using the defined distances with the function
distanceHeat
. This function only requires the dataset
object and the name of the distance. It can also take advantage of the
options and functionalities available in the pheatmap
function (see the code below) from the pheatmap
R package
[@kolde2019]. For
instance, by default, the columns are clustered using
clustering_method = 'complete'
but this parameter can be
specified according to pheatmap
’s options. Different
aesthetic options are exemplified in the plots below using the
introduced distances.
distanceHeat(x = massData[, -1], distName = 'hamming',
main = 'Hamming Distance\nMath Anxiety Data', fontsize = 6.5)
distanceHeat(x = massData[, -1], distName = 'cramerV',
main = "Cramer's V Distance\nMath Anxiety Data", fontsize = 6.5,
show_rownames = FALSE, cluster_rows = FALSE, cluster_cols=FALSE)
distanceHeat(x = massData[, -1], distName = 'hellinger',
main = 'Hellinger Distance\nMath Anxiety Data', fontsize = 6.5,
show_rownames = TRUE, border_color = 'black',
cluster_rows = FALSE, cluster_cols=FALSE)
,
distanceHeat(x = massData[, -1], distName = 'bhattacharyya',
main = 'Bhattacharyya Distance\nMath Anxiety Data', fontsize = 6.5,
show_rownames = TRUE, border_color = 'lightgrey',
cluster_rows = FALSE, cluster_cols=FALSE) ,
clusGapDiscr
:
The Gap Statistic for discrete dataThe Gap Statistic (GS) in its original formulation by tibshirani et.
al aims to determine an optimal number of clusters given a rectangular
dataset with continuous columns and a pre-specified clustering method
given \(k\) number of clusters
(i.e. \(k\)-means clustering or
partitioning around medoids (pam)). This is done by comparing the
within-cluster dispersion of the data to a randomly generated reference
null distribution of the data obtaining estimations via bootstrapping.
The idea is to quantify whether the clustering structure observed in the
data is considerably different compared to what would be expected by
random chance. clusGap
is a widely used implementation of
the GS from the cluster
package [@maechler2023cluster] with the
following basic arguments:
clusGap(x, FUNcluster, K.max, B = 100,
verbose = interactive(), ...)
x
: argument can be data.frame
or
matrix
object where rows represent observations and columns
correspond to data features or variables. This matrix is expected to be
contain exclusively numerical input.FUNcluster
: clustering function accepting on its first
argument a rectangular data object (like x
) and a second
one specifying a desired k
number of clusters.K.max
: maximum number of clusters to consider.B
: integer indicating the number of times to perform
the Montecarlo repetitions.The discrete GS (dGS) proposed applies and adapts the principle
behind the original formulation, but for categorical data using a
distance-based approach. The GS assumes that the distance between
observations can be correctly described with the Euclidean distance,
since it assumes the data is continuous. In the case of categorical
data, this distance assumption is not appropriate and different class of
distances needs to be defined. clusGapDiscr
is the main
function that performs and implements the dGS. The first four parameters
found in clusGap
are identical and have the same usage to
the ones found in clusGapDiscr
. Additionally, the function
requires a categorical distance from the list mentioned above. Another
feature of the dGS is the reference null distribution used, which can be
a discrete uniform distribution of the unique column-wise categories
found in the data; this setting is defined as the Data Support (DS). In
other settings, the categories to be used must be user-specified via a
character vector or list. This setting will be referred as the Known
Support (KS).
clusGapDiscr(x, FUNcluster, K.max, B = nrow(x),
value.range = 'DS',
distName, verbose = interactive(), useLog = TRUE ...)
distaName
: name of discrete distance. The available
options are 'hamming'
, 'chisquare'
,
'cramersV'
, 'bhattacharyya'
and
'hellinger'
.
value.range
: specifies the values of the reference
null distribution. Possible values: 'DS'
or a character
vector with unique values. 'DS'
option extracts the unique
values found in x
.
useLog
: Binary truth variable specifying whether to
evaluate the Gap Statistic with or without \(log\) function.
Similar to clusGap
, clusGapDiscr
returns a
matrix object with K.max
rows corresponding to the
user-specified number of clusters with an estimate of the GS and its
corresponding standard error. This information would then determine the
number of clusters according to the cluster-selection criterion. The one
used in DiscreteGapStatistic
is the 1-SE criterion
implemented by the function findK
, which accepts
clusGap
objects. The following example applies the dGS on
the Math Anxiety data using the Hamming, \(\chi^2\), and Cramer’s V distances applying
the cluster::pam
algorithm. A plot is then generated
displaying the dGS against the number of clusters. Each estimate is
accompanied by corresponding standard-error bars and a blue dashed line
indicating the chosen number of clusters.
## Recall Cats:
# Cats <- setNames(object = c('SD', 'D', 'N', 'A', 'SA'),
# nm = c('Strongly Disagree', 'Disagree',
# 'Neutral',
# 'Agree', 'Strongly Agree') )
<- clusGapDiscr(x = massData[, -1],
HammRun FUNcluster = cluster::pam,
B = 100,
K.max = 9,
value.range = 'DS',
distName = 'hamming')
#> Found levels: A, D, N, SA, SD
<- clusGapDiscr(x = massData[, -1],
chisqRun FUNcluster = cluster::pam,
B = 100,
K.max = 9,
value.range = Cats,
distName = 'chisquare')
#> Found levels: A, D, N, SA, SD
<- clusGapDiscr(x = massData[, -1],
crVRun FUNcluster = cluster::pam,
B = 100,
K.max = 9,
value.range = 'DS',
distName = 'cramerV')
#> Found levels: A, D, N, SA, SD
plot(HammRun,
main = "Discrete Gap statistic: Hamming Distance\nMath Anxiety Data",
cex = 2, cex.lab=1.2, cex.axis=1.5, cex.main=1.5)
abline(v = findK(HammRun), lty=3, lwd=2, col="Blue")
plot(chisqRun,
main = "Discrete Gap statistic: chi-square Distance\nMath Anxiety Data",
cex = 2, cex.lab=1.2, cex.axis=1.5, cex.main=1.5)
abline(v = findK(chisqRun), lty=3, lwd=2, col="Blue")
plot(crVRun,
main = "Discrete Gap statistic: Cramer's V Distance\nMath Anxiety Data",
cex = 2, cex.lab=1.2, cex.axis=1.5, cex.main=1.5)
abline(v = findK(crVRun), lty=3, lwd=2, col="Blue")
Notice that since all possible categorical values are available in
the data, using the option value.range = Cats
would yield
exact same results.
Heatmaps can be created to visualize the commonalities within each cluster and the differences between them.
ResHeatmap(massData[, -1],
distName = 'hamming',
catVals = Cats,
nCl = findK(HammRun),
out = 'heatmap',
prefObs = NULL, height = 6)
#> Warning: The input is a data frame-like object, convert it to a matrix.
#> Warning: Note: not all columns in the data frame are numeric. The data frame
#> will be converted into a character matrix.
ResHeatmap(massData[, -1],
distName = 'chisquare',
catVals = Cats,
nCl = findK(chisqRun),
out = 'heatmap',
prefObs = NULL, height = 6)
#> Warning: The input is a data frame-like object, convert it to a matrix.
#> Warning: Note: not all columns in the data frame are numeric. The data frame
#> will be converted into a character matrix.
ResHeatmap(massData[, -1],
distName = 'cramerV',
catVals = Cats,
nCl = findK(crVRun),
out = 'heatmap',
prefObs = NULL, height = 6)
#> Warning: The input is a data frame-like object, convert it to a matrix.
#> Warning: Note: not all columns in the data frame are numeric. The data frame
#> will be converted into a character matrix.
Notice that Hamming distance detects subclusters present in Cluster 2
from the \(\chi^2\) distance run. To
compare the similar clusters side-to-side, the parameter
nCl
can be used to reorder the clusters relative to the
given ordering. This case nCl = 2:1
will alter (invert) the
order of the two clusters on the \(\chi^2\) based cluster. A third heatmap is
displayed relabeling the clusters from the \(\chi^2\) run using the
clusterNames = 'renumber'
argument.
ResHeatmap(massData[, -1],
distName = 'hamming',
catVals = Cats,
nCl = findK(HammRun),
out = 'heatmap',
height = 6)
ResHeatmap(massData[, -1],
distName = 'chisquare',
catVals = Cats,
nCl = 2:1,
out = 'heatmap',
height = 6)
ResHeatmap(massData[, -1],
distName = 'chisquare',
catVals = Cats,
nCl = 2:1,
clusterNames = 'renumber',
out = 'heatmap',
height = 6)
These binaries (installable software) and packages are in development.
They may not be fully stable and should be used with caution. We make no claims about them.