The hardware and bandwidth for this mirror is donated by METANET, the Webhosting and Full Service-Cloud Provider.
If you wish to report a bug, or if you are interested in having us mirror your free-software or open-source project, please feel free to contact us at mirror[@]metanet.ch.

Introduction to the QuadratiK Package

Overview

The QuadratiK package provides the first implementation, in R and Python, of a comprehensive set of goodness-of-fit tests and a clustering technique for spherical data using kernel-based quadratic distances. The primary goal of QuadratiK is to offer flexible tools for testing multivariate and high-dimensional data for uniformity, normality, comparing two or more samples.

This package includes several novel algorithms that are designed to handle spherical data, which is often encountered in fields like directional statistics, geospatial data analysis, and signal processing. In particular, it offers functions for clustering spherical data efficiently, for computing the density value and for generating random samples from a Poisson kernel-based density.

Installation

You can install the version published on CRAN of QuadratiK:

install.packages("QuadratiK")

Or the development version on GitHub:

library(devtools)
devtools::install_github('giovsaraceno/QuadratiK-package')

The QuadratiK package is also available in Python on PyPI and as a Dashboard application. Usage instruction for the Dashboard can be found at https://quadratik.readthedocs.io/en/latest/user_guide/dashboard_application_usage.html.

Citation

If you use this package in your research or work, please cite it as follows:

Saraceno G, Markatou M, Mukhopadhyay R, Golzy M (2024). QuadratiK: A Collection of Methods Constructed using Kernel-Based Quadratic Distances. https://cran.r-project.org/package=QuadratiK

@Manual{saraceno2024QuadratiK, 
   title = {QuadratiK: Collection of Methods Constructed using Kernel-Based 
            Quadratic Distances}, 
   author = {Giovanni Saraceno and Marianthi Markatou and Raktim Mukhopadhyay 
            and Mojgan Golzy}, 
   year = {2024}, 
   note = {<https://cran.r-project.org/package=QuadratiK>,
   <https://github.com/giovsaraceno/QuadratiK-package>, 
   <https://giovsaraceno.github.io/QuadratiK-package/>} 
}

and the associated paper:

Saraceno Giovanni, Markatou Marianthi, Mukhopadhyay Raktim, Golzy Mojgan (2024). Goodness-of-Fit and Clustering of Spherical Data: the QuadratiK package in R and Python. arXiv preprint arXiv:2402.02290.

@misc{saraceno2024package, 
   title={Goodness-of-Fit and Clustering of Spherical Data: the QuadratiK package in
          R and Python}, 
   author={Giovanni Saraceno and Marianthi Markatou and Raktim Mukhopadhyay and
           Mojgan Golzy}, 
   year={2024}, 
   eprint={2402.02290}, 
   archivePrefix={arXiv}, 
   primaryClass={stat.CO},    
   url={<https://arxiv.org/abs/2402.02290>}
}

Key features and basic usage

Goodness-of-Fit Tests

library(QuadratiK)

The software implements one, two, and k-sample tests for goodness of fit, offering an efficient and mathematically sound way to assess the fit of probability distributions. Our tests are particularly useful for large, high dimensional data sets where the assessment of fit of probability models is of interest.

The provided goodness-of-fit tests can be performed using the kb.test() function. The kernel-based quadratic distance tests are constructed using the normal kernel which depends on the tuning parameter \(h\). If a value for \(h\) is not provided, the function perform the select_h() algorithm searching for an optimal value. For more details please visit the relative help documentations.

?kb.test
?select_h

The proposed tests perform well in terms of level and power for contiguous alternatives, heavy tailed distributions and in higher dimensions.

Test for normality

To test the null hypothesis of normality \(H_0:F=\mathcal{N}_d(\mu, \Sigma)\)

x <- matrix(rnorm(100), ncol = 2)
# Does x come from a multivariate standard normal distribution?
kb.test(x, h=0.4)
## 
##  Kernel-based quadratic distance Normality test 
##      U-statistic V-statistic
## ------------------------------------------------
## Test Statistic:   -0.9832709      0.668447 
## Critical Value:   1.717843    8.901682 
## H0 is rejected:   FALSE       FALSE 
## Selected tuning parameter h:  0.4

If needed, we can specify \(\mu\) and \(\Sigma\), otherwise the standard normal distribution is considered.

x <- matrix(rnorm(100,4), ncol = 2)
# Does x come from the specified multivariate normal distribution?
kb.test(x, mu_hat = c(4,4), Sigma_hat = diag(2), h = 0.4)
## 
##  Kernel-based quadratic distance Normality test 
##      U-statistic V-statistic
## ------------------------------------------------
## Test Statistic:   -0.7180988      0.7387403 
## Critical Value:   1.768274    8.901682 
## H0 is rejected:   FALSE       FALSE 
## Selected tuning parameter h:  0.4

Two-sample test

In case we want to compare two samples \(X \sim F\) and \(Y \sim G\) with the null hypothesis \(H_0:F=G\) vs \(H_1:F\not =G\).

x <- matrix(rnorm(100), ncol = 2)
y <- matrix(rnorm(100,mean = 5), ncol = 2)
# Do x and y come from the same distribution?
kb.test(x, y, h = 0.4)
## 
##  Kernel-based quadratic distance two-sample test 
## U-statistic   Dn          Trace 
## ------------------------------------------------
## Test Statistic:   5.771118    11.68892 
## Critical Value:   0.6535748   1.325249 
## H0 is rejected:   TRUE        TRUE 
## CV method:  subsampling 
## Selected tuning parameter h:  0.4

k-sample test

In case we want to compare \(k\) samples, with \(k>2\), that is \(H_0:F_1=F_2=\ldots=F_k\) vs \(H_1:F_i\not =F_j\) for some \(i\not = j\).

x1 <- matrix(rnorm(100), ncol = 2)
x2 <- matrix(rnorm(100), ncol = 2)
x3 <- matrix(rnorm(100, mean = 5), ncol = 2)
y <- rep(c(1, 2, 3), each = 50)
# Do x1, x2 and x3 come from the same distribution?
x <- rbind(x1, x2, x3)
kb.test(x, y, h = 0.4)
## 
##  Kernel-based quadratic distance k-sample test 
## U-statistic   Dn          Trace 
## ------------------------------------------------
## Test Statistic:   7.568824    10.9979 
## Critical Value:   0.8262628   1.201499 
## H0 is rejected:   TRUE        TRUE 
## CV method:  subsampling 
## Selected tuning parameter h:  0.4

Test for uniformity on the sphere

Expanded capabilities include supporting tests for uniformity on the d-dimensional Sphere based on Poisson kernel. The Poisson kernel depends on the concentration parameter \(\rho\) and a location vector \(\mu\). For more details please visit the help documentation of the pk.test() function.

?pk.test

To test the null hypothesis of uniformity on the \(d\)-dimensional sphere \(\mathcal{S}^{d-1} = \{x \in \mathbb{R}^d : ||x||=1 \}\)

# Generate points on the sphere from the uniform ditribution 
x <- sample_hypersphere(d = 3, n_points = 100)
# Does x come from the uniform distribution on the sphere?
pk.test(x, rho = 0.7)
## 
##  Poisson Kernel-based quadratic distance test of 
##                         Uniformity on the Sphere 
## Selected consentration parameter rho:  0.7 
## 
## U-statistic:
## 
## H0 is rejected:  FALSE 
## Statistic Un:  1.620118 
## Critical value:  1.862317 
## 
## V-statistic:
## 
## H0 is rejected:  FALSE 
## Statistic Vn:  22.84617 
## Critical value:  23.22949

Poisson kernel-based distribution (PKBD)

The package offers functions for computing the density value and for generating random samples from a PKBD. The Poisson kernel-based densities are based on the normalized Poisson kernel and are defined on the \(d\)-dimensional unit sphere. For more details please visit the help documentation of the dpkb() and rpkb() functions.

?dpkb
?rpkb

Example

mu <- c(1,0,0)
rho <- 0.9
x <- rpkb(n = 100, mu = mu, rho = rho)
head(x$x)
##           [,1]         [,2]        [,3]
## [1,] 0.9939084 -0.081245323  0.07446703
## [2,] 0.9981673 -0.031731342  0.05152923
## [3,] 0.9968513 -0.006737282  0.07900642
## [4,] 0.9968730  0.060027451  0.05138932
## [5,] 0.9945320  0.011481164 -0.10379956
## [6,] 0.9682762 -0.091252086 -0.23262456
dens_x <- dpkb(x$x, mu = mu, rho = rho)
head(dens_x)
##           [,1]
## [1,] 4.9808540
## [2,] 9.8586575
## [3,] 7.7097546
## [4,] 7.7386652
## [5,] 5.4094177
## [6,] 0.8698273

Clustering Algorithm for Spherical Data

The package incorporates a unique clustering algorithm specifically tailored for spherical data and it is especially useful in the presence of noise in the data and the presence of non-negligible overlap between clusters. This algorithm leverages a mixture of Poisson kernel-based densities on the Sphere, enabling effective clustering of spherical data or data that has been spherically transformed. For more details please visit the help documentation of the pkbc() function.

?pkbc

Example

# Generate 3 samples from the PKBD with different location directions
x1 <- rpkb(n = 100, mu = c(1,0,0), rho = rho)
x2 <- rpkb(n = 100, mu = c(-1,0,0), rho = rho)
x3 <- rpkb(n = 100, mu = c(0,0,1), rho = rho)
x <- rbind(x1$x, x2$x, x3$x)
# Perform the clustering algorithm
# Serch for 2, 3 or 4 clusters
cluster_res <- pkbc(dat = x, nClust = c(2, 3, 4))
summary(cluster_res)
## Poisson Kernel-Based Clustering on the Sphere (pkbc) Results
## ------------------------------------------------------------
## 
## Summary:
##         LogLik     WCSS
## [1,] -600.0535 399.7122
## [2,] -315.6083 320.8486
## [3,] -304.9507 320.1200
## 
## Results for 2 clusters:
## Estimated Mixing Proportions (alpha):
## [1] 0.7074514 0.2925486
## 
## Clustering table:
## 
##   1   2 
## 209  91 
## 
## 
## Results for 3 clusters:
## Estimated Mixing Proportions (alpha):
## [1] 0.3238438 0.3325603 0.3435959
## 
## Clustering table:
## 
##   1   2   3 
##  95 101 104 
## 
## 
## Results for 4 clusters:
## Estimated Mixing Proportions (alpha):
## [1] 0.332708607 0.005782234 0.339847442 0.321661717
## 
## Clustering table:
## 
##   1   2   3   4 
## 101   2 102  95

The software includes additional graphical functions, aiding users in validating and representing the cluster results as well as enhancing the interpretability and usability of the analysis.

# Predict the membership of new data with respect to the clustering results
x_new <- rpkb(n = 10, mu = c(1,0,0), rho = rho)
memb_mew <- predict(cluster_res, k = 3, newdata = x_new$x)
memb_mew$Memb
##  [1] 2 2 2 2 2 2 2 2 2 2
# Compute measures for evaluating the clustering results
val_res <- pkbc_validation(cluster_res)
val_res
## $metrics
##            2         3         4
## ASW 0.518038 0.6928325 0.6049581
## 
## $IGP
## $IGP[[1]]
## NULL
## 
## $IGP[[2]]
## [1] 0.9948980 0.9903846
## 
## $IGP[[3]]
## [1] 0.988764 0.990099 1.000000
## 
## $IGP[[4]]
## [1] 0.990099 1.000000 1.000000 0.988764
# Plot method for the pkbc object:
# - scatter plot of data points on the sphere
# - elbow plot for helping the choice of the number of clusters
plot(cluster_res)

Additional Resources

For more detailed information about the QuadratiK package, you can explore the following resources:

If you’re new to the package, we recommend starting with the available vignettes:

References

For more information on the methods implemented in this package, refer to the associated research papers:

These binaries (installable software) and packages are in development.
They may not be fully stable and should be used with caution. We make no claims about them.