Multivariate prediction of human behavior from resting state data is gaining increasing popularity in the neuroimaging community, with far-reaching translational implications in neurology and psychiatry. However, the high dimensionality of neuroimaging data increases the risk of overfitting, calling for the use of dimensionality reduction methods to build robust predictive models. In this work, we assess the ability of four dimensionality reduction techniques to extract relevant features from resting state functional connectivity matrices of stroke patients, which are then used to build a predictive model of the associated language deficits based on cross-validated regularized regression. Features extracted by Principal Component Analysis (PCA) were found to be the best predictors, followed by Independent Component Analysis (ICA), Dictionary Learning (DL) and Non-Negative Matrix Factorization. However, ICA and DL led to more parsimonious models. Overall, our findings suggest that the choice of the dimensionality reduction technique should not only be based on prediction/regression accuracy, but also on considerations about model complexity and interpretability.

A Systematic Assessment of Feature Extraction Methods for Robust Prediction of Neuropsychological Scores from Functional Connectivity Data

Testolin A.;De Filippo De Grazia M.;Zorzi M.
2020

Abstract

Multivariate prediction of human behavior from resting state data is gaining increasing popularity in the neuroimaging community, with far-reaching translational implications in neurology and psychiatry. However, the high dimensionality of neuroimaging data increases the risk of overfitting, calling for the use of dimensionality reduction methods to build robust predictive models. In this work, we assess the ability of four dimensionality reduction techniques to extract relevant features from resting state functional connectivity matrices of stroke patients, which are then used to build a predictive model of the associated language deficits based on cross-validated regularized regression. Features extracted by Principal Component Analysis (PCA) were found to be the best predictors, followed by Independent Component Analysis (ICA), Dictionary Learning (DL) and Non-Negative Matrix Factorization. However, ICA and DL led to more parsimonious models. Overall, our findings suggest that the choice of the dimensionality reduction technique should not only be based on prediction/regression accuracy, but also on considerations about model complexity and interpretability.
2020
Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
978-3-030-59276-9
978-3-030-59277-6
File in questo prodotto:
File Dimensione Formato  
_496162_1_En_3_Chapter_Author.pdf

non disponibili

Tipologia: Published (publisher's version)
Licenza: Accesso privato - non pubblico
Dimensione 1.16 MB
Formato Adobe PDF
1.16 MB Adobe PDF Visualizza/Apri   Richiedi una copia
Pubblicazioni consigliate

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11577/3355881
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 3
  • ???jsp.display-item.citation.isi??? 2
social impact