Partial Least Squares (PLS)

Authors: Golia Shafiei

This tutorial explains the concept of Partial Least Squares (PLS) analysis in general, which was first introduced to the neuroimaging community in 1996 (McIntosh et al., 1996). In addition, we illustrate how to use PLS process on a sample data in Brainstorm.

License

PLS is a free toolbox that is available at Baycrest (https://www.rotman-baycrest.on.ca/index.php?section=84). The PLS code is written entirely in MATLAB (Mathworks Inc) and can be downloaded from https://www.rotman-baycrest.on.ca/index.php?section=345. To cite PLS Toolbox, please see the “References” section of this tutorial.

Introduction

PLS analysis is explained in general in this section. However, this tutorial assumes that the users are already familiar with basics of PLS analysis and it does not include details about this technique. If PLS is new to you or if you want to read more about PLS and its applications in details, please refer to “References” and relevant articles introduced here.

Partial Least Squares (PLS) analysis is a multivariate statistical technique that is used to find the relationship between two blocks of variables. PLS that has various applications and types (Krishnan et al); however, the focus of this tutorial is on Mean-Centered PLS analysis, which is a common type of PLS while working with neuroimaging data. In this type of PLS analysis, one data block is neural activity (e.g. MEG measurements/source data here) while the other one is the experiment design (e.g. different groups/conditions).

PLS analysis is based on extracting the common information between the two data blocks by finding a correlation matrix and linear combinations of variables in both data blocks that have maximum covariance with one another. In the example provided here, we find a contrast (or an effect) between different conditions as well as patterns of brain activity that maximally covary with that specific contrast.

For this purpose, we take the neural activity as one data block, matrix X, where the rows of matrix X are observations (participants/trials) nested in conditions or groups, and the columns of X are variables that are arranged in a way that time scales are nested within sources. The other data block, matrix Y, is a matrix of dummy coding that is related to experimental design (different groups or conditions) (Krishnan et al).

PLS analysis first calculates a mean-centered matrix using matrices X and Y. Then, singular value decomposition (SVD) is applied on the mean-centered matrix. The outcome of PLS analysis is a set of latent variables that are in fact linear combinations of initial variables of the two data blocks that maximally covary with the resulting contrasts (419, Krishnan et al., 2011).

Finally, the statistical significance of a latent variable is defined by a p-value from permutation testing. In addition, bootstrapping is used to assess the reliability of each original variable (e.g. a source at a time point) that contributes to the latent variable. Bootstrap ratios are calculated for each original variable for this purpose. More specifically, each latent variable consists of a set of singular values that describe the effect size, as well as a set of singular vectors, or weights, that define the contribution of each initial variable to the latent variables. The ratio of these weights to the standard errors estimated from bootstrapping is called bootstrap ratio. Therefore, the larger the magnitude of a bootstrap ratio, the larger the weight (i.e. contribution to the latent variable) and the smaller the standard error (i.e. higher stability) (McIntosh and Lobaugh 2004, 419). Bootstrap ratio can be equivalent to a z-score if we have an approximately normal bootstrap distribution (Efron and Tibshirani, 1986).

PLS analysis was explained in general in this section. However, this tutorial assumes that the users are already familiar with basics of PLS analysis. If PLS is new to you or if you want to read more about PLS and its applications in details, please refer to the articles introduced in “References” section.

Download and installation

In order to run PLS process in Brainstorm, the PLS Toolbox must be downloaded from here and added to MATLAB pathway.

Data, Pre-Processing and Source Analysis

These two processes work in a similar way:

In the context of this tutorial, we have two condition types: faces, and scenes. We want to decode faces vs. scenes using 306 MEG channels. In the data, the faces are named as condition ‘201’; and the scenes are named as condition ‘203’.

Import the recordings

Select files

Select the Process2 tab at the bottom of the Brainstorm window.

Cross-validation

Cross-validation is a model validation technique for assessing how the results of our decoding analysis will generalize to an independent data set.

Permutation

This is an iterative procedure. The training and test data for the SVM/LDA classifier are selected in each iteration by randomly permuting the samples and grouping them into bins of size n (you can select the trial bin sizes). In each iteration two samples (one from each condition) are left out for test. The rest of the data are used to train the classifier with.

Acknowledgment

This work was supported by the McGovern Institute Neurotechnology Program to PIs: Aude Oliva and Dimitrios Pantazis. http://mcgovern.mit.edu/technology/neurotechnology-program

References

  1. Khaligh-Razavi SM, Bainbridge W, Pantazis D, Oliva A (2016)
    From what we perceive to what we remember: Characterizing representational dynamics of visual memorability. bioRxiv, 049700.

  2. Cichy RM, Pantazis D, Oliva A (2014)
    Resolving human object recognition in space and time, Nature Neuroscience, 17:455–462.





Feedback: Comments, bug reports, suggestions, questions
Email address (if you expect an answer):


GoliaShafiei (last edited 2017-01-09 21:30:08 by GoliaShafiei)