We are given a sample of n {\displaystyle n} paired observations ( x → i , y → i ) , i ∈ 1 , … , n {\displaystyle ({\vec {x}}_{i},{\vec {y}}_{i}),i\in {1,\ldots ,n}} . In the first step j = 1 {\displaystyle j=1} , the partial least squares regression searches for the normalized direction p → j {\displaystyle {\vec {p}}_{j}} , q → j {\displaystyle {\vec {q}}_{j}} that maximizes the covariance4
Note below, the algorithm is denoted in matrix notation.
The general underlying model of multivariate PLS with ℓ {\displaystyle \ell } components is
where
The decompositions of X and Y are made so as to maximise the covariance between T and U.
Note that this covariance is defined pair by pair: the covariance of column i of T (length n) with the column i of U (length n) is maximized. Additionally, the covariance of the column i of T with the column j of U (with i ≠ j {\displaystyle i\neq j} ) is zero.
In PLSR, the loadings are thus chosen so that the scores form an orthogonal basis. This is a major difference with PCA where orthogonality is imposed onto loadings (and not the scores).
A number of variants of PLS exist for estimating the factor and loading matrices T, U, P and Q. Most of them construct estimates of the linear regression between X and Y as Y = X B ~ + B ~ 0 {\displaystyle Y=X{\tilde {B}}+{\tilde {B}}_{0}} . Some PLS algorithms are only appropriate for the case where Y is a column vector, while others deal with the general case of a matrix Y. Algorithms also differ on whether they estimate the factor matrix T as an orthogonal (that is, orthonormal) matrix or not.5678910 The final prediction will be the same for all these varieties of PLS, but the components will differ.
PLS is composed of iteratively repeating the following steps k times (for k components):
PLS1 is a widely used algorithm appropriate for the vector Y case. It estimates T as an orthonormal matrix. (Caution: the t vectors in the code below may not be normalized appropriately; see talk.) In pseudocode it is expressed below (capital letters are matrices, lower case letters are vectors if they are superscripted and scalars if they are subscripted).
This form of the algorithm does not require centering of the input X and Y, as this is performed implicitly by the algorithm. This algorithm features 'deflation' of the matrix X (subtraction of t k t ( k ) p ( k ) T {\displaystyle t_{k}t^{(k)}{p^{(k)}}^{\mathrm {T} }} ), but deflation of the vector y is not performed, as it is not necessary (it can be proved that deflating y yields the same results as not deflating11). The user-supplied variable l is the limit on the number of latent factors in the regression; if it equals the rank of the matrix X, the algorithm will yield the least squares regression estimates for B and B 0 {\displaystyle B_{0}}
In 2002 a new method was published called orthogonal projections to latent structures (OPLS). In OPLS, continuous variable data is separated into predictive and uncorrelated (orthogonal) information. This leads to improved diagnostics, as well as more easily interpreted visualization. However, these changes only improve the interpretability, not the predictivity, of the PLS models.12 Similarly, OPLS-DA (Discriminant Analysis) may be applied when working with discrete variables, as in classification and biomarker studies.
The general underlying model of OPLS is
or in O2-PLS13
Another extension of PLS regression, named L-PLS for its L-shaped matrices, connects 3 related data blocks to improve predictability.14 In brief, a new Z matrix, with the same number of columns as the X matrix, is added to the PLS regression analysis and may be suitable for including additional background information on the interdependence of the predictor variables.
In 2015 partial least squares was related to a procedure called the three-pass regression filter (3PRF).15 Supposing the number of observations and variables are large, the 3PRF (and hence PLS) is asymptotically normal for the "best" forecast implied by a linear latent factor model. In stock market data, PLS has been shown to provide accurate out-of-sample forecasts of returns and cash-flow growth.16
A PLS version based on singular value decomposition (SVD) provides a memory efficient implementation that can be used to address high-dimensional problems, such as relating millions of genetic markers to thousands of imaging features in imaging genetics, on consumer-grade hardware.17
PLS correlation (PLSC) is another methodology related to PLS regression,18 which has been used in neuroimaging 192021 and sport science,22 to quantify the strength of the relationship between data sets. Typically, PLSC divides the data into two blocks (sub-groups) each containing one or more variables, and then uses singular value decomposition (SVD) to establish the strength of any relationship (i.e. the amount of shared information) that might exist between the two component sub-groups.23 It does this by using SVD to determine the inertia (i.e. the sum of the singular values) of the covariance matrix of the sub-groups under consideration.2425
Schmidli, Heinz (13 March 2013). Reduced Rank Regression: With Applications to Quantitative Structure-Activity Relationships. Springer. ISBN 978-3-642-50015-2. 978-3-642-50015-2 ↩
Wold, S; Sjöström, M.; Eriksson, L. (2001). "PLS-regression: a basic tool of chemometrics". Chemometrics and Intelligent Laboratory Systems. 58 (2): 109–130. doi:10.1016/S0169-7439(01)00155-1. S2CID 11920190. /wiki/Doi_(identifier) ↩
Abdi, Hervé (2010). "Partial least squares regression and projection on latent structure regression (PLS Regression)". WIREs Computational Statistics. 2: 97–106. doi:10.1002/wics.51. S2CID 122685021. https://wires.onlinelibrary.wiley.com/doi/epdf/10.1002/wics.51 ↩
See lecture https://www.youtube.com/watch?v=Px2otK2nZ1c&t=46s https://www.youtube.com/watch?v=Px2otK2nZ1c&t=46s ↩
Lindgren, F; Geladi, P; Wold, S (1993). "The kernel algorithm for PLS". J. Chemometrics. 7: 45–59. doi:10.1002/cem.1180070104. S2CID 122950427. /wiki/Doi_(identifier) ↩
de Jong, S.; ter Braak, C.J.F. (1994). "Comments on the PLS kernel algorithm". J. Chemometrics. 8 (2): 169–174. doi:10.1002/cem.1180080208. S2CID 221549296. /wiki/Doi_(identifier) ↩
Dayal, B.S.; MacGregor, J.F. (1997). "Improved PLS algorithms". J. Chemometrics. 11 (1): 73–85. doi:10.1002/(SICI)1099-128X(199701)11:1<73::AID-CEM435>3.0.CO;2-#. S2CID 120753851. /wiki/Doi_(identifier) ↩
de Jong, S. (1993). "SIMPLS: an alternative approach to partial least squares regression". Chemometrics and Intelligent Laboratory Systems. 18 (3): 251–263. doi:10.1016/0169-7439(93)85002-X. /wiki/Doi_(identifier) ↩
Rannar, S.; Lindgren, F.; Geladi, P.; Wold, S. (1994). "A PLS Kernel Algorithm for Data Sets with Many Variables and Fewer Objects. Part 1: Theory and Algorithm". J. Chemometrics. 8 (2): 111–125. doi:10.1002/cem.1180080204. S2CID 121613293. /wiki/Doi_(identifier) ↩
Abdi, H. (2010). "Partial least squares regression and projection on latent structure regression (PLS-Regression)". Wiley Interdisciplinary Reviews: Computational Statistics. 2: 97–106. doi:10.1002/wics.51. S2CID 122685021. /wiki/Doi_(identifier) ↩
Höskuldsson, Agnar (1988). "PLS Regression Methods". Journal of Chemometrics. 2 (3): 219. doi:10.1002/cem.1180020306. S2CID 120052390. /wiki/Doi_(identifier) ↩
Trygg, J; Wold, S (2002). "Orthogonal Projections to Latent Structures". Journal of Chemometrics. 16 (3): 119–128. doi:10.1002/cem.695. S2CID 122699039. /wiki/Doi_(identifier) ↩
Eriksson, S. Wold, and J. Tryg. "O2PLS® for improved analysis and visualization of complex data." https://www.dynacentrix.com/telecharg/SimcaP/O2PLS.pdf https://www.dynacentrix.com/telecharg/SimcaP/O2PLS.pdf ↩
Sæbøa, S.; Almøya, T.; Flatbergb, A.; Aastveita, A.H.; Martens, H. (2008). "LPLS-regression: a method for prediction and classification under the influence of background information on predictor variables". Chemometrics and Intelligent Laboratory Systems. 91 (2): 121–132. doi:10.1016/j.chemolab.2007.10.006. /wiki/Doi_(identifier) ↩
Kelly, Bryan; Pruitt, Seth (2015-06-01). "The three-pass regression filter: A new approach to forecasting using many predictors". Journal of Econometrics. High Dimensional Problems in Econometrics. 186 (2): 294–316. doi:10.1016/j.jeconom.2015.02.011. /wiki/Doi_(identifier) ↩
Kelly, Bryan; Pruitt, Seth (2013-10-01). "Market Expectations in the Cross-Section of Present Values". The Journal of Finance. 68 (5): 1721–1756. CiteSeerX 10.1.1.498.5973. doi:10.1111/jofi.12060. ISSN 1540-6261. /wiki/CiteSeerX_(identifier) ↩
Lorenzi, Marco; Altmann, Andre; Gutman, Boris; Wray, Selina; Arber, Charles; Hibar, Derrek P.; Jahanshad, Neda; Schott, Jonathan M.; Alexander, Daniel C. (2018-03-20). "Susceptibility of brain atrophy to TRIB3 in Alzheimer's disease, evidence from functional prioritization in imaging genetics". Proceedings of the National Academy of Sciences. 115 (12): 3162–3167. Bibcode:2018PNAS..115.3162L. doi:10.1073/pnas.1706100115. ISSN 0027-8424. PMC 5866534. PMID 29511103. https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5866534 ↩
Krishnan, Anjali; Williams, Lynne J.; McIntosh, Anthony Randal; Abdi, Hervé (May 2011). "Partial Least Squares (PLS) methods for neuroimaging: A tutorial and review". NeuroImage. 56 (2): 455–475. doi:10.1016/j.neuroimage.2010.07.034. PMID 20656037. S2CID 8796113. /wiki/Doi_(identifier) ↩
McIntosh, Anthony R.; Mišić, Bratislav (2013-01-03). "Multivariate Statistical Analyses for Neuroimaging Data". Annual Review of Psychology. 64 (1): 499–525. doi:10.1146/annurev-psych-113011-143804. ISSN 0066-4308. PMID 22804773. /wiki/Doi_(identifier) ↩
Beggs, Clive B.; Magnano, Christopher; Belov, Pavel; Krawiecki, Jacqueline; Ramasamy, Deepa P.; Hagemeier, Jesper; Zivadinov, Robert (2016-05-02). de Castro, Fernando (ed.). "Internal Jugular Vein Cross-Sectional Area and Cerebrospinal Fluid Pulsatility in the Aqueduct of Sylvius: A Comparative Study between Healthy Subjects and Multiple Sclerosis Patients". PLOS ONE. 11 (5): e0153960. Bibcode:2016PLoSO..1153960B. doi:10.1371/journal.pone.0153960. ISSN 1932-6203. PMC 4852898. PMID 27135831. https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4852898 ↩
Weaving, Dan; Jones, Ben; Ireton, Matt; Whitehead, Sarah; Till, Kevin; Beggs, Clive B. (2019-02-14). Connaboy, Chris (ed.). "Overcoming the problem of multicollinearity in sports performance data: A novel application of partial least squares correlation analysis". PLOS ONE. 14 (2): e0211776. Bibcode:2019PLoSO..1411776W. doi:10.1371/journal.pone.0211776. ISSN 1932-6203. PMC 6375576. PMID 30763328. https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6375576 ↩
Abdi, Hervé; Williams, Lynne J. (2013), Reisfeld, Brad; Mayeno, Arthur N. (eds.), "Partial Least Squares Methods: Partial Least Squares Correlation and Partial Least Square Regression", Computational Toxicology, vol. 930, Humana Press, pp. 549–579, doi:10.1007/978-1-62703-059-5_23, ISBN 9781627030588, PMID 23086857 9781627030588 ↩