---
title: "MetabolSSMF"
output: rmarkdown::html_vignette
vignette: >
%\VignetteIndexEntry{MetabolSSMF}
%\VignetteEngine{knitr::rmarkdown}
%\VignetteEncoding{UTF-8}
%\usepackage[utf8]{inputenc}
%\setmainfont{Times New Roman}\n
bibliography: references.bib
---
```{r setup, include = FALSE}
knitr::opts_chunk$set(
collapse = TRUE,
comment = "#>"
)
knitr::opts_chunk$set(fig.width=5, fig.height=4)
```
```{r LoadPackage, include = FALSE}
library(MetabolSSMF)
set.seed(12345)
```
# Introduction
This vignette presents the $\textbf{MetabolSSMF}$ package, which implements a framework for the simplex-structured matrix factorisation (SSMF) algorithm in $\mathrm{R}$. This package provides a standard algorithm, which allows users to apply the SSMF to perform soft clustering of the data set that they are interested in. The usage of this package will be illustrated through an application to a simulated metabolomic data set. This vignette also includes the usage of:
- **gap statistic** to select the number of clusters;
- **bootstrap resampling** for estimating confidence intervals for the cluster prototypes;
- **soft adjusted Rand index** for comparing known and estimated partitions;
- **Shannon diversity index** for evaluating the average uncertainty in a soft clustering partition.
This document gives a quick tour of $\textbf{MetabolSSMF}$ functionalities. It was written in R Markdown, using the knitr package for production. See help(package="MetabolSSMF") for further details.
$\textbf{Installation:}$ The latest version of the package can be installed from $\textbf{Github}$:
```{r, eval=FALSE}
# Install the package
devtools::install_github("WenxuanLiu1996/MetabolSSMF")
# Load the package
library(MetabolSSMF)
```
# Data
The simulated data set is created by the product of a soft membership matrix $H$ and a prototype matrix $W$. A user can use a Dirichlet distribution to generate the values for the soft membership matrix $H$ with simplex structure. $W$ can be defined using a designed pattern. The package provides a simulated data set $X$, simulated prototype matrix $W$ and simulated membership matrix $H$. These can be loaded through the code `data(SimulatedDataset)`, `data(SimulatedPrototypes)` and `data(SimulatedMemberships)`, respectively.
```{r Data}
# Simulated data set X
X <- as.matrix(SimulatedDataset)
# Simulated W matrix (prototype matrix)
W <- as.matrix(SimulatedPrototypes)
# Simulated H matrix (membership matrix)
H <- as.matrix(SimulatedMemberships)
```
# Visualise the simulated data
**Visualise the values of simulated prototype matrix** $W$
```{r Visulisation1, warning=FALSE}
# Define 4 different colors
color <- c(ggsci::pal_jco("default", alpha=0.8)(7)[2], ggsci::pal_jama("default", alpha=0.8)(7)[c(3:5)])
op1 <- par(mar=c(5.1, 4.1, 4.1, 4.1), xpd=TRUE)
matplot(t(W), type='l', lty = 1, ylab='Prototypes', col = color, lwd=2, ylim = c(0,1))
legend('topright', inset=c(-0.15,0), legend = 1:4, fill=color[1:4], cex = 0.8)
par(op1)
```
**Visualise the values of simulated soft membership matrix** $H$ **in each observation**
The scatter pies plot of the 2-dimensional projected clusters. It represents the simulated data with 4 soft clusters, the area of colour in the pies depicts the values of the soft membership of the observations.
```{r Visulisation2, warning=FALSE}
pca_X <- prcomp(X, center = F, scale. = F)
op1 <- par(mar=c(5.1, 4.1, 4.1, 4.1), xpd=TRUE)
caroline::pies(lapply(apply(H, 1, list), unlist), color.table = caroline::nv(color[1:4], paste0('Pty', 1:4)), x0=pca_X$x[,1], y0=pca_X$x[,2], radii = 1.5, ylab='PC2', xlab='PC1')
legend('topright', inset=c(-0.15,0), legend = 1:4, fill=color[1:4], cex = 0.8)
par(op1)
```
# Choose the number of clusters
The SSMF algorithm can be run with $k=1, 2, \ldots, K$ using a `for` loop and the results can be saved to a list. The 'elbow criterion' or the gap statistic can be done as a subsequent step.
```{r SSMF, eval=FALSE}
# Set the maximum k that is used in loop.
K <- 10
# Run the SSMF algorithm with various k and save the results as a list
fit_SSMF <- list()
for(k in 1:K){
fit <- ssmf(X, k = k, lr = 0.001, meth='kmeans')
fit_SSMF[[k]] <- fit
}
```
**Plot the residual sum of squares ("elbow criterion")**
```{r, include=FALSE}
K <- 10
```
```{r RSS}
# Extract the RSS and save as a vector
rss_SSMF <- unlist(lapply(fit_SSMF, function(x) x$SSE))
# Plot RSS
plot(1:K, rss_SSMF, type="b", xlab = "K", ylab = "RSS", main='Elbow Criterion')
```
If the 'elbow' is not clear, the gap statistic is suggested.
**Gap statistic**
```{r Gap, eval=FALSE}
# Apply the gap statistic to the simulated data
fit_gap <- gap(X, rss = rss_SSMF)
```
```{r Visulisation3}
# Visualise the results
op2 <- par(mar=c(5.1, 4.1, 4.1, 8.1), xpd=TRUE)
plot(x = c(1:K), y = fit_gap$gap, type='b', ylim=range((fit_gap$gap), (fit_gap$gap - fit_gap$standard.error)[-1]), xlab='K', ylab='GAP', main='Gap statistic')
lines(x = c(1:(K-1)), y = (fit_gap$gap - fit_gap$standard.error)[-1], col='blue', lty=2, type='b')
lines(x = rep(fit_gap$optimal.k, 2), y = range(fit_gap$gap), lty = 2, col='red')
legend('topright', inset=c(-0.45,0), col=c('black', 'blue', 'red'), lty=c(1, 2, 2), legend = c('Gap', '-1 SE', 'Optimal K'), cex=0.8)
par(op2)
```
The estimated number of prototypes is 4 from both the 'elbow criterion' and gap statistic.
# Bootstrap
```{r Bootstrap, eval=FALSE}
# Set the number of times to bootstrap
M <- 1000
# Bootstrap
# Initial H (membership) matrix to start the algorithm
# Based on the results above, k=4 here
initialisation <- init(data = X, k = 4, method = 'kmeans')
fit_boot <- bootstrap(data = X, k = 4, H = initialisation$H, mtimes=M)
```
Note: Usually the $H$ matrix is the output of the function `ssmf( )`. If $H$ is not supplied, the bootstrapped $W$ matrix might have different prototype orders from the outputs of the function `ssmf( )`.
**Visualise the values of the bootstrap estimation and confidence intervals for the prototypes**
```{r Visulisation4}
op1 <- par(mar=c(5.1, 4.1, 4.1, 4.1), xpd=TRUE)
matplot(t(fit_boot$W.est[c(2,4,1,3),]), type='n', xaxt='n', ylab='Features', ylim = range(fit_boot$lower,fit_boot$upper))
for(i in 1:4){
alpha <- 0.2
plat <- c(ggsci::pal_jco("default",alpha=alpha)(7)[2], ggsci::pal_jama("default", alpha=alpha)(7)[c(3:5)])
col <- switch(i, plat[1], plat[2], plat[3], plat[4])
polygon(c(1:138, 138:1), c(fit_boot$upper[c(2,4,1,3),][i,], rev(fit_boot$lower[c(2,4,1,3),][i,])), col = col, border= col, lty = 1)
}
matplot(t(fit_boot$W.est[c(2,4,1,3),]), type='l', lty = 1, ylab='Features', add=T, col=color, lwd=2)
legend('topright', inset=c(-0.2,0), legend = 1:4, fill=color[1:4])
axis(1, at=c(1, seq(10, 130, 10), 138), labels = c(1, seq(10, 130, 10), 138), cex=0.8)
par(op1)
```
Estimated prototypes are represented by the solid lines and the 4 colours represent 4 clusters. Shadows around the lines are confidence intervals for the prototypes.
**Visualise the values of the estimated soft membership matrix**
```{r Visulisation5, warning=FALSE}
op1 <- par(mar=c(5.1, 4.1, 4.1, 4.1), xpd=TRUE)
caroline::pies(lapply(apply(fit_SSMF[[4]]$H, 1, list), unlist), color.table = caroline::nv(color[1:4], c(4,3,2,1)), x0=pca_X$x[,1], y0=pca_X$x[,2], radii = 1.5, ylab='PC2', xlab='PC1')
legend('topright', inset=c(-0.15,0), legend = 1:4, fill=color[1:4], cex=0.8)
par(op1)
```
# Soft Adjusted Rand Index (sARI)
```{r sARI}
# Compare the true and estimated soft membership matrix
sARI(fit_SSMF[[4]]$H, H)
# Self comparison of the true soft membership matrix
sARI(H, H)
```
The sARI between estimated soft membership matrix and the true soft membership matrix is `r round(sARI(fit_SSMF[[4]]$H, H), 3)`. The self comparison of the true soft membership matrix shows the upper bound of this sARI is `r round(sARI(H, H), 3)`. We can use the percentage of these two values `r round(sARI(fit_SSMF[[4]]$H, H), 3)`/`r round(sARI(H, H), 3)` $=$ `r round(sARI(fit_SSMF[[4]]$H, H)/sARI(H, H), 3)` to states that the estimated soft membership matrix explained `r round(sARI(fit_SSMF[[4]]$H, H)/sARI(H, H), 3)*100`$\%$ information of the true soft membership matrix.
# Shannon diversity index
```{r Shannon}
E <- rep(NA, nrow(fit_SSMF[[4]]$H))
for(i in 1:nrow(fit_SSMF[[4]]$H)){
E[i] <- diversity(fit_SSMF[[4]]$H[i,], two.power=T)
}
round(mean(E), 1)
```
The estimated soft memberships of the simulated observations have an average values of $2^{\mathrm{E}(h_{i \cdot})}=$ `r round(mean(E), 1)`, indicating that most observations softly clustered by SSMF are fuzzy between `r as.integer(mean(E))` and `r as.integer(mean(E))+1`clusters.
# Additional NOTE
The SSMF algorithm has randomness, this document presents one of the situation using `set.seed(12345)` at the beginning. The results will vary with different seeds that are used in $\mathrm{R}$. It is suggested to not set the seed and run the algorithm multiple times using the same values of $k$ (the number of prototypes). Then, select a result with the lowest RSS, save it to the list and continue processing next steps.