Introduction

The MXM R Package, short for the latin 'Mens ex Machina' ( Mind from the Machine ), is a collection of utility functions for feature selection, cross validation and Bayesian Networks. MXM offers many feature selection algorithms focused on providing one or more minimal feature subsets, refered also as variable signatures, that can be used to improve the performance of downstream analysis tasks such as regression and classification, by excluding irrelevant and redundant variables. In this tutorial we will learn how to use the SES algorithm. For simplicity, we will use a dataset referred as “The Wine Dataset”.

Loading Data

The Wine Dataset contains the results of a chemical analysis of wines grown in a specific area of Italy. Three types of wine are represented in the 178 samples, with the results of 13 chemical analyses recorded for each sample. Note that the “Type” variable was transformed into a categorical variable.

So, first of all, for this tutorial analysis, we are loading the 'MXM' library and 'dplyr' library for handling easier the dataset.

### ~ ~ ~ Load Packages ~ ~ ~ ###
library(MXM) 
library(dplyr)

And on a next step we are downloading and opening the dataset, defining also the column names.

### ~ ~ ~ Load The Dataset ~ ~ ~ ###
wine.url <- "https://archive.ics.uci.edu/ml/machine-learning-databases/wine/wine.data"
wine <- read.csv(wine.url,
                 check.names = FALSE,
                 header = FALSE) 
head(wine)
##   V1    V2   V3   V4   V5  V6   V7   V8   V9  V10  V11  V12  V13  V14
## 1  1 14.23 1.71 2.43 15.6 127 2.80 3.06 0.28 2.29 5.64 1.04 3.92 1065
## 2  1 13.20 1.78 2.14 11.2 100 2.65 2.76 0.26 1.28 4.38 1.05 3.40 1050
## 3  1 13.16 2.36 2.67 18.6 101 2.80 3.24 0.30 2.81 5.68 1.03 3.17 1185
## 4  1 14.37 1.95 2.50 16.8 113 3.85 3.49 0.24 2.18 7.80 0.86 3.45 1480
## 5  1 13.24 2.59 2.87 21.0 118 2.80 2.69 0.39 1.82 4.32 1.04 2.93  735
## 6  1 14.20 1.76 2.45 15.2 112 3.27 3.39 0.34 1.97 6.75 1.05 2.85 1450
str(wine)
## 'data.frame':    178 obs. of  14 variables:
##  $ V1 : int  1 1 1 1 1 1 1 1 1 1 ...
##  $ V2 : num  14.2 13.2 13.2 14.4 13.2 ...
##  $ V3 : num  1.71 1.78 2.36 1.95 2.59 1.76 1.87 2.15 1.64 1.35 ...
##  $ V4 : num  2.43 2.14 2.67 2.5 2.87 2.45 2.45 2.61 2.17 2.27 ...
##  $ V5 : num  15.6 11.2 18.6 16.8 21 15.2 14.6 17.6 14 16 ...
##  $ V6 : int  127 100 101 113 118 112 96 121 97 98 ...
##  $ V7 : num  2.8 2.65 2.8 3.85 2.8 3.27 2.5 2.6 2.8 2.98 ...
##  $ V8 : num  3.06 2.76 3.24 3.49 2.69 3.39 2.52 2.51 2.98 3.15 ...
##  $ V9 : num  0.28 0.26 0.3 0.24 0.39 0.34 0.3 0.31 0.29 0.22 ...
##  $ V10: num  2.29 1.28 2.81 2.18 1.82 1.97 1.98 1.25 1.98 1.85 ...
##  $ V11: num  5.64 4.38 5.68 7.8 4.32 6.75 5.25 5.05 5.2 7.22 ...
##  $ V12: num  1.04 1.05 1.03 0.86 1.04 1.05 1.02 1.06 1.08 1.01 ...
##  $ V13: num  3.92 3.4 3.17 3.45 2.93 2.85 3.58 3.58 2.85 3.55 ...
##  $ V14: int  1065 1050 1185 1480 735 1450 1290 1295 1045 1045 ...
colnames(wine) <- c('Type', 'Alcohol', 'Malic', 'Ash', 
                    'Alcalinity', 'Magnesium', 'Phenols', 
                    'Flavanoids', 'Nonflavanoids',
                    'Proanthocyanins', 'Color', 'Hue', 
                    'Dilution', 'Proline')

SES for Continuous

For this tutorial example, we are going to apply the SES algorithm on the above dataset, using as data and as target variables only continuous variables.

Running SES for 1st time

Selecting Appropriate Conditional Independence Test

The selection of the appropriate conditional independence test is a crucial decision for the validity and success of downstream statistical analysis and machine learning tasks. Currently the __ MXM R package__ supports numerous tests for different combinations of target ( dependent ) and predictor ( independent ) variables. A detailed summary table to guide you through the selection of the most suitable test can be found in MXM's reference manual (p.21 “CondInditional independence tests” ) here: https://CRAN.R-project.org/package=MXM. In our example we will use the MXMX::SES(), which is the implementation of the SES algorithm and since we are going to examine only continuous variables, we will use the Fisher's Independence Test.

Creating Data & Target Matrices

dataset - A numeric matrix ( or a data.frame in case of categorical predictors), containing the variables for performing the test. The rows should refer to the different samples and columns to the features. For the purposes of this example analysis, we are going to use only the continuous variables, therefore we are removing the “Type” variable from the dataset. Furthermore, we are removing the “Nonflavanoids” variable, because we will use it as target.

### ~ ~ ~ Removing The Categorical ('Type') and The Target ('Nonflavanoids') Variables ~ ~ ~ ###

wine_dataset <- dplyr::select(wine,
                              -contains("Type"),
                              -contains("Nonflavanoids")) 
head(wine_dataset)
##   Alcohol Malic  Ash Alcalinity Magnesium Phenols Flavanoids Proanthocyanins
## 1   14.23  1.71 2.43       15.6       127    2.80       3.06            2.29
## 2   13.20  1.78 2.14       11.2       100    2.65       2.76            1.28
## 3   13.16  2.36 2.67       18.6       101    2.80       3.24            2.81
## 4   14.37  1.95 2.50       16.8       113    3.85       3.49            2.18
## 5   13.24  2.59 2.87       21.0       118    2.80       2.69            1.82
## 6   14.20  1.76 2.45       15.2       112    3.27       3.39            1.97
##   Color  Hue Dilution Proline
## 1  5.64 1.04     3.92    1065
## 2  4.38 1.05     3.40    1050
## 3  5.68 1.03     3.17    1185
## 4  7.80 0.86     3.45    1480
## 5  4.32 1.04     2.93     735
## 6  6.75 1.05     2.85    1450

target - The class variable including the values of the target variable. We should provide either a string, an integer, a numeric value, a vector, a factor, an ordered factor or a Surv object. For the purposes of this example analysis, we are going to use as the dependent variable Nonflavanoids.

wine_target <- wine$Nonflavanoids
head(wine_target)
## [1] 0.28 0.26 0.30 0.24 0.39 0.34

Function's Arguments

This is the first time that we are running the algorithm, so we are going to explain what each Argument refers to:

target : The class variable. Provide either a string, an integer, a numeric value, a vector, a factor, an ordered factor or a Surv object. As explained above, this will be the dependent variable. If the target is a single integer value or a string, it has to corresponds to the column number or to the name of the target feature in the dataset. Here we choose wine$Nonflavanoids

dataset : The dataset. Provide either a data frame or a matrix. If the dataset (predictor variables) contains missing (NA) values, they will automatically be replaced by the current variable (column) mean value with an appropriate warning to the user after the execution. Here we choose the whole wine dataset, except from the Type (categorical) and Nonflavanoids (target) variables.

max_k : The maximum conditioning set to use in the conditional independence test (see The Rule of Thumb in TIPS session). Integer, default value is 3.Here we choose the default value 3.

threshold : Threshold (suitable values in [0,1]) for assessing p-values significance. Default value is 0.05. Here we choose the default value (0.05)

test : The conditional independence test to use. Default value is NULL. Here since our dataset includes only continuous features (remember: Categorical variable 'Type' was removed) and our dependent variable is also continuous, we choose 'testIndFisher'.

ini : After running SES with some hyper-parameters you might want to run SES again with different hyper-parameters. To avoid calculating the univariate associations (first step of SES) again, you can extract them from the first run of SES and plug them here. This can speed up the second run (and subsequent runs of course) by 50%. Here, since we are going to apply the analysis for the first time, we place the argument being equal to NULL.

wei : A vector of weights to be used for weighted regression. The default value is NULL. It is not taken into account when robust is set to TRUE. Here we are not going to use weights.

user_test : A user-defined conditional independence test (provide a closure type object). Default value is NULL. If this is defined, the “test” argument is ignored. Here we have not created a different conditional independence test. We are going to use the 'testIndFisher', as explained in the test argument.

hash : A boolean variable which indicates whether (TRUE) or not (FALSE) to store the statistics calculated during SES execution in a hash-type object. Default value is FALSE. If TRUE a hashObject is produced. In this example, this is the first time that we are running the algorithm and our goal is to apply also a second one. Here mind that we are setting hash = TRUE, because we want that the univariate associations to be kept for next usage.

hashObject : A List with the hash objects generated in a previous run of SES. Each time SES runs with “hash=TRUE” it produces a list of hashObjects that can be re-used in order to speed up next runs of SES or MMPC. Here, since this is the first time that we run the analysis, we are not providing a hashObject.

ncores : How many cores to use. For more details, see the TIPS session. Here, since we have a small number of samples and features, we are using only one core.

### ~ ~ ~ Running SES For First Time ~ ~ ~ ###
ses_default_1st <- MXM::SES(target = wine_target,
                          dataset   = wine_dataset, 
                          max_k      = 3, 
                          threshold  = 0.05, 
                          test       = "testIndFisher",
                          ini        = NULL, 
                          wei        = NULL,
                          user_test  = NULL, 
                          hash       = TRUE, 
                          hashObject = NULL, 
                          ncores     = 1)

So, the algorithm run… Let's see what information we can take out of it.

Output

The main purpose of running SES algorithm is to see which variables should be selected as important. The indices of those variables are stored in selectedVars.

ses_default_1st@selectedVars
## [1]  4  5  7 11
SelectedVars_names<-colnames(wine_dataset[ses_default_1st@selectedVars])
SelectedVars_names
## [1] "Alcalinity" "Magnesium"  "Flavanoids" "Dilution"

Here, we see which are the important features.

But can we have them in an order, according to increasing p-value?

Yes, of course!

SelectedVars_indecies_Ordered<- ses_default_1st@selectedVarsOrder
SelectedVars_indecies_Ordered
## [1]  4  5  7 11

And what are the equivalent signatures?

ses_default_1st@signatures
##      Var1 Var2 Var3 Var4
## [1,]    4    5    7   11

But what is the difference of signatures and this:

ses_default_1st@queues
## [[1]]
## [1] 4
## 
## [[2]]
## [1] 5
## 
## [[3]]
## [1] 7
## 
## [[4]]
## [1] 11

Well, queues is a list containing a list (queue) of equivalent features for each variable included in selectedVars. An equivalent signature can be built by selecting one and only one feature from each queue. In our case there are no equivalent signatures, hence all variables have no equivalent variable.

What about the statistics corresponding to “p-values”? (Remember: higher values indicates higher association)

ses_default_1st@stats  
##  [1] 0.4599480 1.2376687 0.4068903 2.9376490 2.4364353 0.4707855 2.2854487
##  [8] 0.3049473 1.8515479 0.5539026 2.0418569 0.8210702

But what is the difference between @stats and @pvalues?

ses_default_1st@pvalues
##  [1] -0.4367580 -1.5255269 -0.3789372 -5.5846020 -4.1451887 -0.4488160
##  [7] -3.7504527 -0.2734232 -2.7215042 -0.5441125 -3.1541338 -0.8849622

Both lists are almost the same, however, in @pvalues, for each feature included in the dataset, this vector reports the strength of its association with the target in the context of all other variables. Particularly, this vector reports the max p-values found when the association of each variable with the target is tested against different conditional sets. Lower values indicate higher association.

And are all hyper-parameters that were used for the algorithm also stored in this Object?

ses_default_1st@max_k     # max_k option used in the current run
## [1] 3
ses_default_1st@threshold # threshold option used in the current run
## [1] -2.995732
ses_default_1st@test      # character name of the statistic test used
## [1] "testIndFisher"

Ok! All this is perfect, but can we see the number of univariate associations?

ses_default_1st@n.tests 
## [1] 55

Attention : If you have set hash = TRUE (as we did), then the number of tests performed by SES will be returned. So actually here we are seeing the number of tests performed.

Ok, but in the beginning we said that if we enable hash, then the univariate associations are also stored. How can we get them?

ses_default_1st@univ
## $stat
##  [1] -2.079719  3.992701  2.492686  5.014962 -3.467758 -6.410909 -7.953067
##  [8] -5.074785  1.851548 -3.557762 -7.324434 -4.260699
## 
## $pvalue
##  [1]  -3.243966  -9.251082  -4.297126 -13.558111  -7.322485 -20.461002
##  [7] -29.158820 -13.831160  -2.721504  -7.638713 -25.499442 -10.312081

And how quick has all this happened?

ses_default_1st@runtime 
##    user  system elapsed 
##    0.04    0.00    0.05

Running SES 2nd time

As mentioned before, we want to apply the analysis more than once. So, let's say that the goal is to apply the same analysis, but with a different threshold (0.1 instead of 0.05) and with a different max_k (4 instead of 3). Remember that during the first run, we kept hash = TRUE, therefore the information is stored in the created variable. ini is a supposed to be a list. After running SES (or MMPC) with some hyper-parameters you might want to run SES again with different hyper-parameters. To avoid calculating the univariate associations (first step of SES) again, you can extract them from the first run of SES and plug them here. This can speed up the second run (and subsequent runs of course) by 50%.

In addition, supplying the hashObject also saves computational time as no extra tests will be performed.

### ~ ~ ~ Running SES For First Time ~ ~ ~ ###
ses_default_2nd <- MXM::SES(target = wine_target,
                          dataset   = wine_dataset, 
                          max_k      = 4, 
                          threshold  = 0.1, 
                          test       = "testIndFisher",
                          ini        = ses_default_1st@univ, 
                          wei        = NULL,
                          user_test  = NULL, 
                          hash       = TRUE, 
                          hashObject = ses_default_1st@hashObject, 
                          ncores     = 1)

Creating the model

This is how we can create the Model including the significant variables. One or more regression models obtained from SES are returned.

### ~ ~ ~ glm() Model Estimates Using SES Feature Subset As Predictor Variables ~ ~ ~ ###

first_sign_ses<- ses.model(target = wine_target, 
                             dataset = as.matrix(wine_dataset),
                             wei = NULL, 
                             sesObject = ses_default_1st, # 
                             test = 'testIndFisher')

The model(s) created is/are stored in the \(mod\) variable. We can get the summary of the model (looks like the summary of many other linear models), or ask for the signatures ( \(signature\) ). The BIC score is also returned for each value of the signature.

ses_model_summary <- first_sign_ses$mod
signature          <- first_sign_ses$signature
signature
## Alcalinity  Magnesium Flavanoids   Dilution        bic 
##     4.0000     5.0000     7.0000    11.0000  -288.7666

SES for Categorical

On the example above, we run the analysis for a continuous variable (Nonflavanoids). What would happen if we choose to use as target the “Type” variable, which is the categorical variable referring to the three types of wine that are represented?

Running SES for Categorical

Selecting Appropriate Conditional Independence Test

Since the variable is categorical - and more specific it is a factor with more than two levels (unordered) -and the features are continuous, according to MXM's reference manual (p.21 “CondInditional independence tests” ) here: https://CRAN.R-project.org/package=MXM, we should use the Multinomial logistic regression ( 'testIndMultinom' ).

Creating Data & Target Matrices

In this step, we keep the whole dataset, in order to show how to use the algorithm also without subtracting the initial matrix.

### ~ ~ ~ Taking The Whole Dataset ~ ~ ~ ###

wine_dataset <- wine
head(wine_dataset)
##   Type Alcohol Malic  Ash Alcalinity Magnesium Phenols Flavanoids Nonflavanoids
## 1    1   14.23  1.71 2.43       15.6       127    2.80       3.06          0.28
## 2    1   13.20  1.78 2.14       11.2       100    2.65       2.76          0.26
## 3    1   13.16  2.36 2.67       18.6       101    2.80       3.24          0.30
## 4    1   14.37  1.95 2.50       16.8       113    3.85       3.49          0.24
## 5    1   13.24  2.59 2.87       21.0       118    2.80       2.69          0.39
## 6    1   14.20  1.76 2.45       15.2       112    3.27       3.39          0.34
##   Proanthocyanins Color  Hue Dilution Proline
## 1            2.29  5.64 1.04     3.92    1065
## 2            1.28  4.38 1.05     3.40    1050
## 3            2.81  5.68 1.03     3.17    1185
## 4            2.18  7.80 0.86     3.45    1480
## 5            1.82  4.32 1.04     2.93     735
## 6            1.97  6.75 1.05     2.85    1450

We will not create a different matrix for the target. As mentioned above, we are going to use the Type variable, but please… be patient…

Setting the Arguments

### ~ ~ ~ Running SES For Categorical Variable ~ ~ ~ ###
wine[, 1] <- as.factor(wine[, 1])
ses_default_1st <- MXM::SES(target = 1, ## Defining as target the 1st column
                          dataset   = wine, 
                          max_k      = 3, 
                          threshold  = 0.05, 
                          test       = "testIndMultinom",
                          ini        = NULL, 
                          wei        = NULL,
                          user_test  = NULL, 
                          hash       = TRUE, 
                          hashObject = NULL, 
                          ncores     = 1)

So, the algorithm run once again… Let's see what information we can take out of it.

Output

The main purpose of running SES algorithm is to see which variables should be selected as important. The indices of those variables are stored in selectedVars.

ses_default_1st@selectedVars
## [1]  2  4  8 11 13 14
SelectedVars_names<-colnames(wine_dataset[ses_default_1st@selectedVars])
SelectedVars_names
## [1] "Alcohol"    "Ash"        "Flavanoids" "Color"      "Dilution"  
## [6] "Proline"

We could again ask for each variable of the output, as in the first implementation of the algorithm. But we will concentrate on the returned signature.

SelectedVars_indecies_Ordered<- ses_default_1st@selectedVarsOrder
SelectedVars_indecies_Ordered
## [1]  8 14 11  2 13  4

And what are the equivalent signatures?

ses_default_1st@signatures
##      Var1 Var2 Var3 Var4 Var5 Var6
## [1,]    2    4    8   11   13   14

And how quick has all this happened?

ses_default_1st@runtime 
##    user  system elapsed 
##    3.79    0.00    3.80

Creating the model

This is how we can create the Model including the significant variables. One or more regression models obtained from SES are returned.

### ~ ~ ~ glm() Model Estimates Using SES Feature Subset As Predictor Variables ~ ~ ~ ###

second_sign_ses<- ses.model(target = as.matrix(wine_target), 
                             dataset = as.matrix(wine_dataset),
                             wei = NULL, 
                             sesObject = ses_default_2nd, # 
                             test = 'testIndFisher')

The model(s) created is/are stored in the \(mod\) variable. We can get the summary of the model (looks like the summary of many other linear models), or ask for the signatures ( \(signature\) ). The BIC score is also returned for each value of the signature.

ses_model_summary <- second_sign_ses$mod
signature          <- second_sign_ses$signature
signature
##        Ash Alcalinity    Phenols      Color        bic 
##      4.000      5.000      7.000     11.000   -265.697
str(signature)
##  Named num [1:5] 4 5 7 11 -266
##  - attr(*, "names")= chr [1:5] "Ash" "Alcalinity" "Phenols" "Color" ...

Permutation Based SES

If the analysis has to be done using permutations (when few observations are available), then this could be applied like this:

### ~ ~ ~ Permutations ~ ~ ~ ###
library('MXM')
permutation_ses_model <- MXM::perm.ses(wine_target,
                                         wine_dataset,  
                                         R = 999, # The number of permutations to use. The default value is 999
                                         max_k      = 3, 
                                         threshold  = 0.05, 
                                         test       = NULL, 
                                         ini        = NULL, 
                                         wei        = NULL,
                                         user_test  = NULL, 
                                         hash       = FALSE, 
                                         hashObject = NULL, 
                                         ncores     = 1)

There is a solution to avoid all permutations. As soon as the number of times the permuted test statistic is more than the observed test statistic is more than 50 (if threshold = 0.05 and R = 999), the p-value has exceeded the significance level (threshold value) and hence the predictor variable is not significant. There is no need to continue doing the extra permutations, as a decision has already been made.

Cross - Validation

Cross-validation (CV) is a technique for validating models which can assess the generalization of the results provided by a statistical analysis on independent datasets. It is a technique widely used in regression and classification approaches and provides a solution to the problem of choosing the right tuning-parameters in a prediction model. In the case of SES algorithm, the tuning parameters would be max_k and threshold. Therefore, in MXM package, a function named cv.ses() can be found. The function performs a k-fold cross-validation for identifying the best values for the SES tuning parameters. In case the user does not know what tuning parameters to use, we suggest the use of this function as Step Zero, before running the algorithm.

Creating Data & Target Matrices

For the purposes of this tutorial, we are going to use the same dataset and as target we will use again the Categorical variable “Type”, transforming our problem into a Classification problem by trying to predict the label of the each row, provided by the first column.

So, let's see how this algorithm works.

Function's Arguments

Before running the cv.ses() function we are going to explain what each Argument refers to:

target : The target or class variable as in SES. The difference is that it cannot accept a single numeric value, an integer indicating the column in the dataset. Here we choose wine$Type, which is a Categorical variable.

dataset : The dataset object as in SES. Here we choose the whole wine dataset.

wei : A vector of weights to be used for weighted regression.Here we choose the default value NULL.

kfolds : The number of the folds in the k-fold Cross Validation (integer). Here we choose 5. Since the dataset includes 178 observations, by splitting into 5 folds, each fold will have circa 35 rows.

folds : The folds of the data to use (a list generated by the function generateCVRuns TunePareto). If NULL the folds are created internally with the same function. Here we will let the algorithm run alone the internal function, by choosing NULL.

alphas : A vector of SES thresholds hyper parameters to be used in CV. Here choose the values 0.1, 0.05, 0.01.

max_ks : A vector of SES max_ks hyper parameters to be used in CV. Here we choose the values 3, 4, 5.

task : A character (“C”, “R” or “S”). It can be “C” for classification (logistic, multinomial or ordinal regression), “R” for regression (robust and non robust linear regression, median regression, (zero inflated) poisson and negative binomial regression, beta regression), “S” for survival regression (Cox, Weibull or exponential regression). Here we choose “C”, since we are applying a classification technique on the dataset, by placing as target the “Type” categorical variable.

metric : A metric function provided by the user. If NULL the following functions will be used: auc.mxm, mse.mxm, ci.mxm for classification, regression and survival analysis tasks, respectively. The metric functions that are currently supported are:

If you are certain about the meaning of one metric, then we suggest to use this one. Here we choose the “accuracy metric for multinomial” (acc_multinom.mxm).

modeler : A modeling function provided by the user. If NULL the following functions will be used: glm.mxm, lm.mxm, coxph.mxm for classification, regression and survival analysis tasks, respectively. The modelling functions that are currently supported are:

The predicted values are transformed into R using the logit transformation. This is so that the “mse.mxm” metric function can be used. In addition, this way the performance can be compared with the regression scenario, where the logit is applied and then a regression model is employed.

In case you decide not to choose one of the default modelers, we suggest to you to be certain about the way your chosen one works. Here we choose multinom.mxm, since we search for a multinomial regression model.

ses_test : A function object that defines the conditional independence test used in the SES function (see 3.1.1 Selecting Appropriate Conditional Independence Test). If NULL, “testIndFisher”, “testIndLogistic” and “censIndCR” are used for classification, regression and survival analysis tasks, respectively. Not all tests can be included here. “testIndClogit”, “testIndMVreg”, “testIndIG”, “testIndGamma”, “testIndZIP” and “testIndTobit” are not available yet.

Here we choose 'testIndMultinom'', as explained in 4.1.1 Selecting Appropriate Conditional Independence Test

robust : A boolean variable which indicates whether (TRUE) or not (FALSE) to use a robust version of the statistical test if it is available. It takes more time than a non robust version but it is suggested in case of outliers. Here, we set it as FALSE.

ncores : How many cores to use. For more details, see the TIPS session. Here, since we have a small number of samples and features, we are using only one core.

### ~ ~ ~ Cross-Validation ~ ~ ~ ### 
library('MXM')
cv_ses_model <- MXM::cv.ses(target   = wine[, 1], # Using the 1st column as target
                            dataset  = wine[, -1],  
                            wei      = NULL,
                            kfolds   = 5, 
                            folds    = NULL, 
                            alphas   = c(0.1, 0.05, 0.01), 
                            max_ks   = c(3, 4, 5), 
                            task     = "C", 
                            metric   = acc_multinom.mxm, # Note that we are passing it as a function and not as a character
                            modeler  = multinom.mxm,  # Note that we are passing it as a function and not as a character
                            ses_test = "testIndMultinom",
                            ncores = 1)

Output

The main purpose of running Cross - Validation SES algorithm is to see which hyperparameters should be selected for tuning the algorithm. The best configuration (pair of max_k and p-value) is stored under the variable $best_configuration.

cv_ses_model$best_configuration
## $id
## [1] 1
## 
## $a
## [1] 0.1
## 
## $max_k
## [1] 5
cv_ses_model$best_performance
## [1] 0.9379365

As we see, the id, the significance threshold (“a”) and the best max_k are returned.

But what if we want to see how good a specific configuration did during the Cross-Validation?
Yes yes, we can ask for the performance more specifically.
Here we check separately the results for the first configuration.

cv_ses_model$cv_results_all[[1]]$configuration #this configuration we are examining
## $id
## [1] 1
## 
## $a
## [1] 0.1
## 
## $max_k
## [1] 5
cv_ses_model$cv_results_all[[1]]$performances # those are the performances.
## [1] 0.9444444 0.9722222 0.9444444 1.0000000 0.8285714
cv_ses_model$cv_results_all[[1]]$signatures # signatures created by this configuration
## [[1]]
##      Var1 Var2 Var3 Var4
## [1,]    1    7   10   13
## 
## [[2]]
##      Var1 Var2 Var3 Var4
## [1,]    1    7   11   13
## 
## [[3]]
##      Var1 Var2 Var3 Var4
## [1,]    1    7   11   13
## 
## [[4]]
##      Var1 Var2 Var3 Var4
## [1,]    1    7   11   13
## [2,]   12    7   11   13
## [3,]    1    3   11   13
## [4,]   12    3   11   13
## 
## [[5]]
##      Var1 Var2 Var3 Var4
## [1,]    1    4    7   13

Note: By asking the best performance $best_performance, we get the mean value of all folds of the best configuration. If we ask for the performances of the best configuration (here it is the first configuration) $cv_results_all[[1]]$performances we get all performance values of each fold (here 5).

cv_ses_model$best_performance
## [1] 0.9379365
index<-cv_ses_model$best_configuration$id
cv_ses_model$cv_results_all[[index]]$performances
## [1] 0.9444444 0.9722222 0.9444444 1.0000000 0.8285714
mean(cv_ses_model$cv_results_all[[index]]$performances)  
## [1] 0.9379365

Tips

Choosing k_max

There is a rule of thumb that we are suggesting. If the sample size is small, take as max_k the smallest integer near N/10, where N is the number of samples.

#
N = dim(wine_dataset)[1]
suggested_max_k = floor(N/10)
N
## [1] 178
suggested_max_k
## [1] 17

However, an other approach would be to run the cross-validation function, described above. In that case, the best max_k will be returned together with the best threshold (see 6. Cross - Validation)

Choosing number of cores

This those kind of analyses may be applied on very big dataset, the SES algorithm provides the opportunity to be run on more than one cores. We suggest that you use more cores, when the datasets are big, but remember that the best run time will always be dependent on the architecture of the computer that is used.

Conclusion

Now you are ready to run your own analysis using MXM::SES algorithm!
Thank you for your attention.
Hope that you found this tutorial helpful.

Session Info {.unnumbered}

All analyses have been applied on:

sessionInfo()
## R version 4.0.3 (2020-10-10)
## Platform: i386-w64-mingw32/i386 (32-bit)
## Running under: Windows 10 x64 (build 19044)
## 
## Matrix products: default
## 
## locale:
## [1] LC_COLLATE=C                    LC_CTYPE=English_Jersey.1252   
## [3] LC_MONETARY=English_Jersey.1252 LC_NUMERIC=C                   
## [5] LC_TIME=English_Jersey.1252    
## 
## attached base packages:
## [1] stats     graphics  grDevices utils     datasets  methods   base     
## 
## other attached packages:
## [1] dplyr_1.0.7 MXM_1.5.5  
## 
## loaded via a namespace (and not attached):
##  [1] nlme_3.1-149        matrixStats_0.57.0  ordinal_2019.12-10 
##  [4] doParallel_1.0.16   RColorBrewer_1.1-2  R.cache_0.14.0     
##  [7] numDeriv_2016.8-1.1 tools_4.0.3         backports_1.2.0    
## [10] utf8_1.1.4          R6_2.5.0            rpart_4.1-15       
## [13] Hmisc_4.4-2         DBI_1.1.1           colorspace_2.0-0   
## [16] nnet_7.3-14         tidyselect_1.1.0    gridExtra_2.3      
## [19] compiler_4.0.3      coxme_2.2-16        quantreg_5.75      
## [22] htmlTable_2.1.0     SparseM_1.78        slam_0.1-48        
## [25] scales_1.1.1        checkmate_2.0.0     relations_0.6-9    
## [28] stringr_1.4.0       digest_0.6.27       foreign_0.8-80     
## [31] minqa_1.2.4         R.utils_2.10.1      base64enc_0.1-3    
## [34] jpeg_0.1-8.1        pkgconfig_2.0.3     htmltools_0.5.1.1  
## [37] lme4_1.1-26         Rfast2_0.1.3        htmlwidgets_1.5.3  
## [40] rlang_0.4.10        rstudioapi_0.13     visNetwork_2.0.9   
## [43] generics_0.1.0      energy_1.7-7        jsonlite_1.7.2     
## [46] R.oo_1.24.0         magrittr_2.0.1      Formula_1.2-4      
## [49] Matrix_1.2-18       Rcpp_1.0.7          munsell_0.5.0      
## [52] fansi_0.4.1         geepack_1.3-2       lifecycle_1.0.0    
## [55] RcppZiggurat_0.1.6  R.methodsS3_1.8.1   ucminf_1.1-4       
## [58] stringi_1.5.3       MASS_7.3-53         grid_4.0.3         
## [61] parallel_4.0.3      bdsmatrix_1.3-4     bigmemory.sri_0.1.3
## [64] crayon_1.4.1        lattice_0.20-41     splines_4.0.3      
## [67] knitr_1.36          pillar_1.6.1        boot_1.3-25        
## [70] markdown_1.1        codetools_0.2-16    glue_1.6.2         
## [73] evaluate_0.14       latticeExtra_0.6-29 data.table_1.13.4  
## [76] foreach_1.5.1       png_0.1-7           vctrs_0.3.8        
## [79] nloptr_1.2.2.2      MatrixModels_0.4-1  gtable_0.3.0       
## [82] RANN_2.6.1          purrr_0.3.4         tidyr_1.1.3        
## [85] assertthat_0.2.1    ggplot2_3.3.5       xfun_0.26          
## [88] Rfast_2.0.6         broom_0.7.6         survival_3.2-7     
## [91] tibble_3.0.4        conquer_1.0.2       iterators_1.0.13   
## [94] sets_1.0-18         cluster_2.1.0       statmod_1.4.35     
## [97] bigmemory_4.5.36    ellipsis_0.3.2      R.rsp_0.44.0

References {.unnumbered}

Lagani, V., Athineou, G., Farcomeni, A., Tsagris, M. & Tsamardinos, I. Feature Selection with the R Package MXM: Discovering Statistically Equivalent Feature Subsets. J. Stat. Softw. 80, 1-25 (2017).

I. Tsamardinos, V. Lagani and D. Pappas (2012). Discovering multiple, equivalent biomarker signatures. In proceedings of the 7th conference of the Hellenic Society for Computational Biology & Bioinformatics - HSCBB12.

Tsamardinos, Brown and Aliferis (2006). The max-min hill-climbing Bayesian network structure learning algorithm. Machine learning, 65(1), 31-78.

Brown, L. E., Tsamardinos, I., & Aliferis, C. F. (2004). A novel algorithm for scalable and accurate Bayesian network learning. Medinfo, 711-715.

Tsamardinos, I., Aliferis, C. F., & Statnikov, A. (2003). Time and sample efficient discovery of Markov blankets and direct causal relations. In Proceedings of the ninth ACM SIGKDD international conference on Knowledge discovery and data mining (pp. 673-678). ACM.