Next Article in Journal
A Wideband Cryogenic Readout Amplifier with Temperature-Insensitive Gain for SNSPD
Next Article in Special Issue
Dependency Factors in Evidence Theory: An Analysis in an Information Fusion Scenario Applied in Adverse Drug Reactions
Previous Article in Journal
Performance-Cost Trade-Off in Auto-Scaling Mechanisms for Cloud Computing
Previous Article in Special Issue
A Review on Planted (l, d) Motif Discovery Algorithms for Medical Diagnose
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Association of Neuroimaging Data with Behavioral Variables: A Class of Multivariate Methods and Their Comparison Using Multi-Task FMRI Data

by
M. A. B. S. Akhonda
1,*,
Yuri Levin-Schwartz
2,
Vince D. Calhoun
3 and
Tülay Adali
1
1
Department of Computer Science and Electrical Engineering, University of Maryland Baltimore County, Baltimore, MD 21250, USA
2
Department of Environmental Medicine and Public Health, Icahn School of Medicine at Mount Sinai, Manhattan, NY 10029, USA
3
Tri-Institutional Center for Translational Research in Neuroimaging and Data Science (TReNDS), Georgia State University, Georgia Institute of Technology, and Emory University, Atlanta, GA 30303, USA
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(3), 1224; https://doi.org/10.3390/s22031224
Submission received: 24 December 2021 / Revised: 28 January 2022 / Accepted: 28 January 2022 / Published: 5 February 2022

Abstract

:
It is becoming increasingly common to collect multiple related neuroimaging datasets either from different modalities or from different tasks and conditions. In addition, we have non-imaging data such as cognitive or behavioral variables, and it is through the association of these two sets of data—neuroimaging and non-neuroimaging—that we can understand and explain the evolution of neural and cognitive processes, and predict outcomes for intervention and treatment. Multiple methods for the joint analysis or fusion of multiple neuroimaging datasets or modalities exist; however, methods for the joint analysis of imaging and non-imaging data are still in their infancy. Current approaches for identifying brain networks related to cognitive assessments are still largely based on simple one-to-one correlation analyses and do not use the cross information available across multiple datasets. This work proposes two approaches based on independent vector analysis (IVA) to jointly analyze the imaging datasets and behavioral variables such that multivariate relationships across imaging data and behavioral features can be identified. The simulation results show that our proposed methods provide better accuracy in identifying associations across imaging and behavioral components than current approaches. With functional magnetic resonance imaging (fMRI) task data collected from 138 healthy controls and 109 patients with schizophrenia, results reveal that the central executive network (CEN) estimated in multiple datasets shows a strong correlation with the behavioral variable that measures working memory, a result that is not identified by traditional approaches. Most of the identified fMRI maps also show significant differences in activations across healthy controls and patients potentially providing a useful signature of mental disorders.

1. Introduction

The availability of multiple datasets that provide complementary information is becoming increasingly common in cognitive neuroimaging [1,2,3,4,5,6]. A significant challenge while analyzing such datasets is the development of fusion methods in which the datasets can fully interact with and inform each other when identifying useful features for further study [7,8,9]. Methods based on matrix and tensor decompositions use latent variable models and allow such interactions among the datasets [10,11,12]. Since these methods provide uniqueness guarantees with few assumptions, the estimated latent variables (components) are directly interpretable, that is, one can attach physical meaning to them and help explain the relationship of datasets and populations. An essential step following the extraction of such informative components that can be interpreted as putative biomarkers of neurological conditions, such as depression and schizophrenia, is associating those with neuropsychological variables (or broadly non-imaging) such as behavioral variables [13,14,15]. This is important not only for the explanation or prediction of the brain regions associated with the behavioral variables but also for the detection and identification of disease sub-types, among other tasks [16,17,18,19,20,21,22].
Independent component analysis (ICA), a popular matrix decomposition-based technique and our focus in this paper, provides an attractive foundation for fully multivariate data fusion [8,23,24]. The use of ICA and its extensions developed for the fusion of multiple datasets can help explain the underlying relationship across datasets by starting from the assumption of the independence of latent variables (components) and are successful when there is a good model match, that is, the model’s assumptions are satisfied [25,26,27,28,29]. One such extension, independent vector analysis (IVA) [23,30], generalizes ICA to multiple datasets by exploiting the available statistical information within the datasets as well as across the datasets, critical for multivariate data fusion. IVA allows the datasets to fully interact with each other in a symmetric manner by letting the datasets play a similar role in the decomposition and establish association only when it is available [23,30]. As such, IVA provides a fully multivariate approach for the analysis of multiple neuroimaging datasets [8,31].
While the fusion of imaging datasets continues to develop, the joint analysis of imaging and non-imaging datasets, a critical component to attribute meaning to neuroimaging fusion studies via links to behavior, is fairly new and has not yet received as much attention [22,32,33]. For example, identifying the association between imaging components and behavioral variables can be used not only to interpret the estimated neuroimaging results, that is, to explain the brain areas corresponding to the observed behaviors, but also to detect neuroimaging patterns that can be used to predict individuals cognitive or behavioral performance [19,20,34]. Furthermore, these imaging components and their corresponding subject co-variations can also be used to identify subtypes of disease and subject sub-groups, which is key to enabling personalized treatments [21,35,36].
Despite the potential uses, the identification of multivariate associations between imaging and behavioral datasets through a joint analysis poses multiple significant challenges. The primary source of these challenges stems from the significant difference in the nature and dimensionality of the datasets. Since the dimensionality of the behavioral datasets tends to be significantly smaller than the imaging datasets, for example, in the few hundreds compared to hundreds of thousands in imaging datasets, fusion results can be affected by this difference in dimensionality [37,38]. Furthermore, assumptions about the imaging features may not hold for the behavioral features due to differences in the nature of the data. Therefore, most of the previous work exploring the relationship between neuroimaging and behavioral datasets focus primarily on performing individual analyses on each neuroimaging dataset and conducting univariate comparisons or using one to predict another [39,40,41,42]. A few recently published methods use fusion-based methods on neuroimaging datasets and show improved performance compared with individual analysis methods [33,43,44]. However, in both the individual and fusion analysis-based approaches, simple correlation-based methods are used to identify associations between imaging and behavioral features [22,39,41,43]. Unfortunately, these pairwise analysis methods do not fully leverage the available information across datasets. One natural way to address this problem is to divide it into parts, fuse the imaging datasets first, obtain summary statistics, and then jointly analyze the behavioral variables with the summary statistics to identify multivariate associations across imaging and behavioral features, which we propose.
In this work, we focus on identifying multivariate associations between imaging and behavioral datasets and present a novel two-step approach based on IVA. We first make use of the strength of IVA to allow full interaction among the datasets to estimate joint independent components and their corresponding columns of mixing matrices, that is, subject co-variations. Subject co-variations provide an efficient way to summarize the imaging fusion results across multiple datasets, and at the same time, offer a natural way to discover associations with the behavioral variables [10]. Then in the second step, we demonstrate two unique ways to identify multivariate associations between the imaging data summary provided by the IVA and the behavioral variables, again by making use of IVA, but through two different approaches designed to discover multivariate associations across datasets. The proposed two-step approach alleviates the dimensionality, and the model mismatch issues of the imaging and the behavioral datasets by dividing the problem into parts and fusing subject co-variations with the behavioral variables. We demonstrate the performance advantages of our proposed method over existing ones first through carefully designed simulations and then using real-world data examples. With simulations, we show that our approach involving IVA improves the success rate of identifying the associated imaging and behavioral components in relation to accuracy and correlations with ground truths. We then demonstrate the successful application of our proposed scheme on real multitask functional magnetic resonance imaging (fMRI) data and behavioral variables collected from collected from 138 healthy controls and 109 patients with schizophrenia. The three tasks in this study involve an auditory oddball (AOD) task, a Sternberg item recognition paradigm (SIRP) task, and a sensorimotor (SM) task [45]. The cognitive assessments we use in this study are collected from the subjects using the Wechsler adult intelligence scale-third edition (WAIS-III) [46], Wechsler memory scale-third edition (WMS-III) [47], and Hopkins’s verbal learning test–revised (HVLT) [48]. We select these tests based on their ease of administration, good reliability of scoring, and ability to assess cognitive impairments more accurately in the schizophrenia patients [45]. We use the latter-number subtest of WAIS-III to measure the working memory, the face recognition subtest of WMS-III to measure the visual memory, and the logical memory test of WMS-III and HVLT to measure verbal memory and learning. We show that the proposed method identifies relationships across certain fMRI components and behavioral variables unidentified by the pairwise correlation technique. We identify profiles corresponding to the fMRI components that showing activation in the motor, sensory motor, dorsolateral prefrontal cortex (DLPFC) and lateral posterior parietal cortex (LPPC) is significantly correlated ( p < 0.05 ) with the behavioral variable responsible for measuring working memory. These areas have been linked with working memory deficits in schizophrenia in multiple prior reports [32,49,50], hence improving our confidence in the proposed method. The results also indicate that components that show a strong association with the behavioral variables in most cases show strong group differences between healthy controls and patients across multiple datasets, hence can be treated as candidate biomarkers of disease. The article is organized as follows. In Section 2, we introduce and discuss our new method and in Section 3, we discuss the implementation and results. We provide conclusions in the final section.

2. Materials and Methods

The joint decomposition of many neuroimaging datasets collected from a group of subjects is difficult. A significant issue is the high dimensionality of the datasets. Therefore, it is desirable to reduce each subject’s dataset into a feature and to obtain a lower-dimensional representation while keeping as much variability in the data as possible [51,52]. Let us consider X k R M × T , k = 1 , 2 , , K as a grouping of K neuroimaging datasets collected from M subjects, where the mth row of each dataset is formed by extracting one multivariate feature for each subject. Such reductions alleviate the high dimensionality problem and offer a natural way to discover association across these feature datasets, that is, the variations across subjects. We assume that the datasets are formed from a linear mixture of M independent components—referred to as sources—the noiseless ICA model for the kth dataset is formulated by
x k ( t ) = A k s k ( t ) , k = 1 , 2 , , K , t = 1 , 2 , , T ,
where A k R M × M , k = 1 , 2 , , K are invertible mixing matrices. Given the model, the ICA solution finds a demixing matrix W k such that the estimated components are statistically independent within the kth dataset and can be computed using u k ( t ) = W k x k ( t ) . For a given set of observations X k , the above equation can be written as U k = W k X k , where X k , U k R M × T and U k = [ u k [ 1 ] , u k [ 2 ] , , u k [ M ] ] T . The columns of estimated demixing matrices, A ^ k , k = 1 , 2 , , K are referred to as subject co-variations, or profiles as they quantify the contributions of subjects to the overall mixtures X k .
For K datasets, association across the datasets can be established in two different ways: either using the estimated components and/or the profiles. These associations across datasets are called links, which can either be hard or soft [10]. The hard link implies that the estimated components and/or profiles are identical—that is, common—across datasets, where the soft link implies the statistical dependence of components and/or profiles across datasets. If the datasets are all of the same type, dimension, and resolution, as in multiset data, we can use both components and profiles to establish links across datasets [53,54]. Examples of such multiset data are fMRI datasets from multiple subjects or tasks, or EEG data from multiple channels or frequencies [55,56,57,58]. However, if the datasets are different in type, dimension, or resolution, as in multimodal data, we can neither assume the components are identical nor statistically dependent across datasets [59,60,61]. Thus, the only possible way to link them is through their estimated profiles. Examples of multimodal data are fMRI, Electroencephalography (EEG), or structural MRI (sMRI) data from multiple subjects or tasks, or MRI and spectroscopy data from multiple subjects [62,63,64].
Now, let us consider y z R M , z = 1 , 2 , , Z to be the collection of Z behavioral features collected from M subjects using different cognitive tests. The behavioral dataset Y R M × Z is formed by putting all Z behavioral features side by side. A direct way to jointly analyze neuroimaging and behavioral features is by treating the behavioral dataset as another modality and jointly decompose it with the imaging modalities using fusion techniques [65]. However, such a direct approach is challenging due to multiple factors. An important one among those is the large dimensionality (sample) mismatch between the imaging and behavioral datasets. The contribution of the datasets to the fusion results can be significantly different due to the large difference in their sample sizes ( Z T ), which might bias the fusion results and affect the interpretability of the estimated results [37,38]. Furthermore, elements in each row of the behavioral dataset, y z [ m ] , z = 1 , 2 , , Z , m = 1 , 2 , , M , can be inherently very different as they come from different behavioral experiments. Thus, the same assumptions of imaging components cannot be placed on the behavioral components while decomposing the datasets jointly. Since the imaging and behavioral datasets are inherently very different, one natural way to establish a connection between these disparate datasets is through subject co-variations or profiles. This involves dividing the problem into parts, first, decompose the imaging datasets and obtain summary statistics through profiles, and second analyze the behavioral variables with the profiles obtained in the first step to identify the associated profiles and behavioral features, which we discuss next. For the simplicity of our discussion, we refer to the estimated subject co-variations or profiles a ^ k [ m ] ’s as imaging features and columns of behavioral dataset y z ’s as behavioral features.

2.1. Methods to Identify Associations between Imaging and Behavioral Features

In this section, we first introduce a general umbrella for the methods to identify associations between imaging and behavioral features.
Based on the number of variables used, associations between a ^ k [ m ] ’s and behavioral features y z ’s can be identified in three general ways: one-to-one, one-to-many, and many-to-many association techniques, as shown in Figure 1. We define the one-to-one technique as a pairwise analysis technique where associations between a ^ k [ m ] ’s and y z ’s are estimated in pairs. Examples of the one-to-one approach are different forms of correlation analysis. This is the most commonly used technique found in the literature when studying associations between imaging and behavioral information [41,66,67,68]. A one-to-many approach, on the other hand, identifies the association between a single behavioral feature ( y z ) and all imaging features ( A k ) or vice-versa, as shown in Figure 1b. This type of joint analysis is helpful for simultaneously identifying multiple imaging features associated with a single behavioral feature. Finally, the many-to-many approach, as shown in Figure 1c, takes all the available information, from both ( A ^ k ) and ( Y ), into account to identify the multivariate associations between imaging and behavioral features. The many-to-many approach facilitates the simultaneous identification of multiple brain areas associated with one or more behavioral features. Both one-to-many and many-to-many techniques use joint analysis and allow interaction among the features across imaging and behavioral datasets in a data-driven way. Hence, these approaches can identify multivariate associations and have the potential to discover relationships one cannot identify using the one-to-one approach.
To the best of our knowledge, this is the first work that aims to introduce the one-to-many and many-to-many techniques to identify the association between imaging and behavioral features in a fusion framework. A newly introduced method in [69] uses an indirect version of the one-to-many technique to estimate imaging components that share strong associations with the behavioral features. The method uses behavioral features as references to guide the decomposition of imaging datasets, thus constraining the process, potentially reducing the interpretability of the estimated imaging components. In this work, we are interested not only in accurately identifying the imaging components associated with the behavioral variables but also in the quality of the estimated components, that is, estimating components that are physically meaningful; hence we will not focus on the method introduced in [69]. Most of the other prior work which uses the one-to-one association technique use separate analyses of the individual imaging datasets, thus not allowing interaction among the datasets while estimating the imaging components [41,66,67,68]. A practical way to achieve both high quality imaging components and multivariate associations with behavioral features, as we propose in this paper, is to split the problem into parts, estimating the imaging components first through a proper fusion framework, and then identifying the imaging features associated with behavioral features using multivariate association techniques. The details about the proposed work and steps to perform are described next.

2.2. A Novel Two-Step Approach to Identify Multivariate Associations between Imaging and Behavioral Features Using IVA

2.2.1. Step 1: Estimation of the Imaging Features

The first step of the method, as shown in Figure 2, is to use IVA to jointly decompose the imaging datasets. IVA generalizes ICA to multiple datasets such that statistical dependence among the dataset can also be taken into account and is shown to provide an attractive solution for fusion of neuroimaging datasets [31,70]. Given the problem in (1), IVA solution finds K demixing matrices W k , k = 1 , 2 , , K such that source components from each dataset can be estimated jointly through u k ( t ) = W k x k ( t ) . The estimated components are independent within a dataset while maximally dependent on corresponding components across the datasets. This way, IVA takes the dependence among datasets into account, to obtain decompositions that fully leverage the commonalities across the datasets. This is done by modeling the source component vector (SCV), where the mth SCV is defined as:
s m ( t ) = [ s 1 [ m ] ( t ) , s 2 [ m ] ( t ) , , s K [ m ] ( t ) ] T R K , m = 1 , 2 , , M ,
that is, by concatenating the mth source components from each of the K datasets, where s k [ m ] R T is the mth source from the kth dataset. IVA thus maximizes independence across the SCVs by minimizing the mutual information [23]
J IVA ( W ) = m = 1 M k = 1 K H u m [ k ] I u m k = 1 K log det W [ k ] ,
where H u m [ k ] denotes the entropy of the mth source estimate for the kth dataset. I u m denotes the mutual information within the mth SCV. Here, the link across the datasets is established using a suitable multivariate probability density function (pdf) for the SCVs. Depending on the pdf used to model the SCVs, IVA can take either second or higher order statistics or both into account. In our formulation, we use IVA using multivariate Laplace distribution where both second-order and higher-order information are taken into account (IVA-L-SOS) to fuse the imaging datasets [71]. IVA-L-SOS provides a suitable model for neuroimaging datasets, especially for the fMRI datasets we use in this study. Since the source components estimated in fMRI applications correspond to brain regions with heavy-tailed distributions [72,73], an algorithm that takes both second and higher-order information into account, such as IVA-L-SOS, is preferable. In addition, the estimated subject profiles, a ^ k [ m ] , m = 1 , 2 , , M and k = 1 , 2 , , K explain each subject’s contribution to the components correlated across datasets, and hence provide a useful summary through fusion. Here, it is important to note that IVA-L-SOS provides a good match to the properties of the fMRI data, and hence is an attractive solution [71,74]. For simplicity, we do not take sample dependency into account in the rest of the article and consider only independent and identically distributed samples, thus dropping index t in (1).

2.2.2. Step 2: Association with Behavioral Variables

In the second step, we use the estimated profile matrices, A ^ k R M × M , k = 1 , 2 , , K , from the first step to identify multivariate associations with the behavioral features y z , z = 1 , 2 , , Z . Note that, The profiles provide an effective summary of the neuroimaging data. As shown in Figure 3, we propose two different techniques, ac-IVA and IVA-G, to identify the associations. Here, ac-IVA uses a behavioral feature at a time to constrain the imaging features so that the correlations with imaging features are maximized, thus using the one-to-many associations. In comparison, IVA-G simultaneously uses all the behavioral features and profile matrices and transforms them into a space where the correlation among all is maximized. Thus, using the many-to-many associations. Identifying the profiles associated with behavioral features will reveals the identities of the imaging components, s ^ k [ m ] R M , connected to the subjects’ behavioral aspects. We start with the ac-IVA [71] and then discuss the IVA-G [75] approach to identify multivariate associations between profiles and behavioral features. Since both methods use SCV correlation matrices to identify associated profiles and behavioral features, we will use correlation instead of association in the rest of the article for simplicity.
The ac-IVA algorithm, also known as parameter-tuned constrained IVA, is a semi-supervised technique that uses reference signals to influence the IVA estimated results [71]. The reference signals can be based on any prior information. In our formulation, as shown in Figure 3a, we use behavioral features y z , z = 1 , 2 , , Z as reference signals to constrain the decomposition of the profile matrices A ^ k , k = 1 , 2 , , K . The goal here is to use the estimated c k [ m ] ’s and their corresponding columns of B k to identify the a ^ k [ m ] ’s that are maximally correlated with referenced signals. ac-IVA does that by adding a constraint g ( a ^ k [ m ] , y z ) to the IVA cost function in (3) [71], where
g ( a ^ k [ m ] , y z ) = ρ m r ( a ^ k [ m ] , y z ) 0 .
Here, r ( · , · ) is the function that defines the similarities between the profiles and the behavioral features, and ρ m is the constraint parameter, selected adaptively from a set of possible values ranges from 0.1 to 0.99 . We use Pearson’s correlation as the similarity function r ( · , · ) . Thus, the absolute value of r ( · , · ) is bounded by 0 to 1 limiting the ρ m 1 . A higher the value of ρ m imposes a harder constraint on the decomposition, while a lower value reduces the power of the constraint. Thus, ρ m values indicate the strength of correlation of a behavioral feature over M subject profiles. This, along with the elements in the columns of B k in Figure 3a, can identify the profiles that are correlated with the zth behavioral feature. Indices of the correlated profiles reveal the identities of the corresponding imaging components that relate to the zth behavioral feature. We use a single behavioral feature at a time as a reference signal, thus using the one-to-many association technique, and repeat the experiment Z times for all Z behavioral features.
IVA-G, on the other hand, uses all the behavioral features and the profile matrices simultaneously, thus using the many-to-many association technique, to identify the correlated imaging and behavioral features. Generative model of IVA-G is shown in Figure 3b. IVA-G uses multivariate Gaussian distribution to model the SCVs, thus taking second-order information of the datasets into account [23,75]. Given A ^ k R M × M , k = 1 , 2 , , K and Y R M × Z , and M = Z , IVA-G jointly transforms the datasets into a space where the correlation between c k [ m ] ’s, where k = 1 , 2 , , K , K + 1 , are maximized. The elements of the corresponding columns of the B k ’s reveal the identities of the profiles in A ^ k ; thus, identities of the sources in U k that are strongly correlated to the behavioral features in Y . In the case of M Z , we generate M Z signals using standard Gaussian distribution and stack them with the behavioral matrix if M > Z or with profile matrices if M < Z to ensure equal dimensions of the datasets. We call the added standard Gaussian signals as padding signals and use those to make dimension of the datasets similar. It is important to note here that these additional signals do not change the results’ outcome but rather allow joint decomposition by alleviating the dimensionality mismatch issue in profile and behavioral matrices. Since IVA-G uses all the datasets simultaneously, it utilizes more information and provides a more accurate estimation of the identities of correlated profiles and behavioral features.
The proposed two-step method to jointly analyze imaging and behavioral datasets alleviates the dimensionality mismatch issue of the datasets. Its disjoint nature also gives us the flexibility to choose appropriate algorithms to decompose imaging datasets in step one and fuse imaging and behavioral features in step two. In the first step, where the dimensionality of the imaging datasets is large, using algorithms that take all order statistics into account, as in IVA-L-SOS, ensures proper estimation of interpretable brain maps and useful summary statistics across populations. Conversely, in the second step, where the number of samples is the number of subjects, and thus small, using flexible data-driven algorithms allows efficient estimation of the correlation information across the imaging and behavioral features. As we show in the next section, through simulations and real data analysis, the proposed approach can estimate correlated features more accurately and identify underlying relationships across imaging and behavioral datasets that cannot be identified by traditional techniques.

3. Results

3.1. Simulation Setup and Results

We perform two different simulation experiments. In the first set of experiments, we compare the performance of ICA and its three extensions for fusion, jICA [25], DS-ICA [76], and IVA [71], to estimate and identify the neuroimaging components associated with BVs using a simple one-to-one correlation technique. We show that fusion analysis provides a better estimation of correlated components than the separate analyses of individual datasets, and among the fusion methods, IVA provides better estimation performance. In the second set of experiments, we use IVA to estimate the neuroimaging components and compare the performance of one-to-one, one-to-many, and many-to-many techniques to identify the correlation between neuroimaging and behavioral features. We compare the performance of correlation, regression, ac-IVA, and IVA-G to accurately identify the correlated components. We show that ac-IVA and IVA-G identify correlated features more accurately than correlation and regression analysis techniques.
We generate simulated examples for K = 3 datasets. For each dataset, we generate N = 10 sources from the Laplacian distribution with T = 10,000 independent and identically distributed (i.i.d.) samples. We select the Laplacian distribution because it is a good match to fMRI data [71,77]. We introduce correlation to three sources across the datasets with correlation values randomly selected from 0.3 to 0.9 . These are the source that are correlated or common across datasets, and the seven other components are distinct. These sources are then linearly mixed with mixing matrices A R M × 10 with elements from a standard Gaussian distribution N ( 0 , 1 ) , where M is the number of subjects from two groups—a healthy control group and a patient group—each with M / 2 samples (subjects). To simulate group differences in the subjects, a step-type response with step height h is added to the first column of each mixing matrix or profile of all three datasets. Thus, we have profiles a k [ 1 ] ( m ) = v k [ 1 ] ( m ) + 1.5 u ( m ) , m = 1 , 2 , , M for k = 1 , 2 , 3 . Here, M is the even number of subjects, v i ( m ) ’s are generated using standard Gaussian distribution and u ( m ) is the step response with u ( m ) = 1 , m M / 2 0 , m > M / 2 . The standard deviations of other profiles are adjusted such that they match the standard deviation of the discriminant profiles for each case we consider. We refer to the components corresponding to these discriminative profiles as discriminative components. We use an order of 10 for dimensionality reduction resulting in X k R 10 × 10 , 000 for k = 1 , 2 , 3 . We generate B = 4 behavioral features from standard Gaussian distribution N ( 0 , 1 ) with M i.i.d. samples resulting in Y R 4 × M . We choose to generate four behavioral features to be consistent with the real data used in this study, which we will discuss later in the paper. A step-type response with step height h, similar to the one added to the discriminative profiles, is added to the first and fourth behavioral features. The purpose of this is to introduce correlation between behavioral features and discriminative profiles. We evaluate the performance of each method to identify the correlated profiles and behavioral features by either changing the number of subjects M or the step-height h used to distinguish two subject groups. In the first case, the number of subjects is changed from 50 to 500 with step-height fixed to 2, while in the second case, the step-height is changed from 0.5 to 3—resulting in correlation values in the range [0.3, 0.9]—with subject number fixed to 300.
For the first set of experiments, we evaluate the performance of individual ICAs, jICA, DS-ICA and IVA to estimate and accurately identify the correlated profiles and behavioral features. We estimate one-to-one correlations between the estimated profiles and the behavioral features and use p < 0.05 as a threshold to identify the correlated ones. We use the accuracy of the methods to correctly identify the target discriminative profiles, and average correlation between the true and the estimated discriminative components to evaluate each method’s performance. We estimate the accuracy using:
Accuracy = T P + T N N × 100 % ,
where T P ’s are the true positive number of profiles correlated with the behavioral features and T N ’s are the true negative number of profiles uncorrelated to the behavioral features. We use the same ICA by entropy bound minimization (ICA-EBM) [78] algorithm for individual ICAs, jICA, and DS-ICA, while using the IVA-L-SOS algorithm for IVA.
Figure 4 shows the performance comparison of the methods for (a) different numbers of subjects and (b) different step-heights. In Figure 4a,b, the accuracy of identifying the correlated profiles and the component correlation improve for all methods as we increase the subject’s number and the step-heights. The fusion methods—jICA, DS-ICA, and IVA—provide better performance than individual ICAs since they use complimentary information available across datasets. Among the fusion methods, IVA and DS-ICA provide better performance than jICA likely because these two methods are less constrained and allow better interaction among the datasets. In addition, IVA outperforms DS-ICA when the subject’s numbers are lower, or step-heights are small, indicating better flexibility of the IVA algorithm.
For the second set of experiments, we only use IVA to estimate the components and their corresponding profiles since it provides better performance than the other methods. Here, we evaluate the performance of four association identification techniques—correlation, regression, ac-IVA and IVA-G—to identify the profiles correlated with behavioral features. We use accuracy defined in (5) to evaluate the performance of these techniques.
Figure 5 compares the performance of the methods for (a) different numbers of subjects and (b) different step-heights. Th accuracy of identifying the target profiles improves for all techniques as we increase the subject’s number and the step-heights. IVA-G, ac-IVA, and regression analysis, which use either the many-to-many or the one-to-many analysis schemes, provide a better identification performance compared with the simple one-to-one correlation analysis technique. The two proposed techniques, IVA-G and ac-IVA, outperform regression analysis, and IVA-G provides a slightly better performance compared with ac-IVA to identify the correlated profiles and behavioral features. To summarize, IVA provides a better estimation performance among the ICA-based fusion methods in the first set of experiments. Meanwhile, IVA-G provides a better identification performance than correlation, regression, and ac-IVA in the second set of experiments.

3.2. FMRI Data and Extracted Features

The multitask fMRI datasets and the behavioral features used in this study are from the Mind Research Network Clinical Imaging Consortium Collection [45]. These datasets and the behavioral features are publicly available at https://coins.trendscenter.org/ (accessed on 23 December 2022). The fMRI datasets were collected from 247 subjects, 138 healthy individuals and 109 schizophrenia patients, while performing auditory oddball (AOD), Sternberg item recognition paradigm (SIRP), and sensory-motor (SM) tasks. In addition, the behavioral scores were collected from the same subjects using the Wechsler adult intelligence scale (WAIS), the Wechsler memory scale (WMS) and the Hopkins verbal learning test (HVLT). The imaging features are generated using the statistical parametric mapping (SPM) toolbox [79]. For each task, the voxels from each subject are analyzed utilizing simple linear regression using the SPM toolbox. Here, the regressors are created by convolving SPM’s hemodynamic response function (HRF) with desired predictors. The resultant regression coefficient maps are then used as imaging features for each subject and task. Details about the fMRI datasets, extracted features and behavioral tests are discussed below.

3.2.1. Auditory Oddball Task (AOD)

During the AOD task, subjects were required to listen to three different types of auditory stimuli—standard, novel and target—coming in a pseudo-random order and were required to press a button only when they heard the target stimuli [80]. The standard stimuli were frequent 1 kHz tones with a probability of occurrence of 0.82. The novel stimuli were infrequent computer-generated complex sounds with a probability of occurrence of 0.09. Finally, the target stimuli were infrequent 1.2 kHz tone with the same probability of occurrence as novel tones, except here, the subjects were required to press a button [45]. A regressor was created to model the target-related stimuli as a delta function convolved with the default SPM HRF [81] and subject averaged contrast images of target tones were used as the feature for this task.

3.2.2. Sternberg Item Recognition Paradigm Task (SIRP)

During this visual task, the subjects were asked to memorize a set of 1, 3, or 5 integer digits randomly selected from integers 0 to 9. The task paradigm (each WM block) lasted for a total of 46 s, including 1.5 s of learning (prompt condition), 0.5 s of a blank screen, 6 s of encoding (encode condition), and finally, 38 s of probing (probe condition). In the probing session, subjects were shown a sequence of integers and were required to press a button whenever a digit from the memorized set arrived [45]. Six 46 s WM blocks consisting of the prompt-encode-probe conditions were used in each scan. For this task, a regressor was created by convolving a three-digit probe response block with default SPM HRF [81], and the average map was used as the feature for this task.

3.2.3. Sensory Motor Task (SM)

The SM task involved subjects listening to a series of 16 different auditory tones [45]. Each tone lasted for 200 ms and was within the frequency range of 236 Hz to 1318 Hz. The tones were arranged in such a way that the first tone was set at the lowest pitch and each subsequent tone was at a higher pitch than the previous tone until the highest was reached. The order of the tones was reversed after the 16th tone. The subjects were required to press a button for each tonal change. Each run consisted of 15 increase-and-decrease blocks, alternated with 15 fixation blocks, with each block lasting for 16 s. For this task, a regressor was created by convolving the entire increase-and-decrease block with SPM HRF, and an average map was used as the feature for this task.

3.2.4. Behavioral Features

This study uses cognitive measures collected from the subjects using WAIS-III, WMS-III, and HVLT tests. The letter-number sequence subtest of WAIS-III was used to evaluate the working memory of the subjects. Visual memory of the subjects was evaluated using the face recognition subtest of WMS-III. Finally, verbal memory and learning of the subjects were evaluated using the logical memory test subset of WMS-III and HVLT. We use letter-number total raw score in WAIS-III, recall total score and recognition total scores in WMS-III, and total raw in HVLT as behavioral features in our analysis. In total, we have four behavioral features from 247 subjects. We perform a two-sample t-test on each feature. As shown in Figure 6, all four behavioral features show substantial group differences ( p < 0.05 ) between healthy controls and schizophrenia patients, with patients reporting lower scores (more cognitive deficits) than healthy controls. It is important to note that temporal synchronization of fMRI data and behavioral variables is not needed since the task-related associations are considered through the initial multivariate feature extraction step, and behavioral associations are at the group level.

3.3. Order and Algorithm Selection

For all three tasks, extracted features from all 247 subjects are concatenated vertically, healthy controls and then patients, to form the feature datasets each with T = 48,546 number of voxels. The dimension of the signal subspace is then estimated for each task feature dataset using an entropy-based method proposed in [82]. Determining the signal subspace order is critical for fMRI data analysis to avoid over-fitting due to the high noise level of medical imaging data. We use an order N = 25 for all three datasets resulting X k R 25 × 48 , 546 , k = 1 , 2 , 3 . A practical way to test the stability of the estimated order is to check the stability of the estimated components for different orders around that number [31]. We check the performance of the methods for a set of orders 15 , 20 , 25 , 30 , 35 , besides the estimated order of 25. The results are pretty similar in terms of activation areas for orders 20 and 30, whereas the stability of components started to change, either getting merged or split into two, for orders 15 and 35. Since there is no ground truth, we select the final order using the guidance of the selection methods and the quality and the stability of the estimated results [31,76].
We use IVA-L-SOS in the first step to decompose the task fMRI feature datasets and estimate interpretable components and their corresponding subject profiles. Since IVA-L-SOS is an iterative algorithm, it is crucial to run the algorithm multiple runs and evaluate the consistency of the estimated results to select the most consistent result. To accomplish this we run the algorithm multiple times and using cross intersymbol interference (Cross-ISI) [83] to select the most consistent run, leading to a reproducible decomposition. We use the ac-IVA [71] and regular IVA-G [75] to identify the associated profiles and behavioral features.

3.4. FMRI Data Results and Discussion

Since the total number of estimated components across three datasets is high (75), we visually check all the estimated components and use only the interpretable ones, that is, showing higher activation in physically meaningful brain areas and are hence useful. We select a total of 21, that is, seven components per dataset for the second step and use their corresponding profiles to estimate correlations with four behavioral features. These selected components are then thresholded at Z = 2.7 and shown in Figure 7. A two-sample t-test is used on the profiles to test for significant group differences between healthy subjects and patients. The components corresponding to those profiles that pass the significance tests ( p < 0.05 ) are referred to as discriminative components or putative biomarkers of disease. The colormap of the components is adjusted so that the colors red, orange and yellow indicate higher activation in healthy controls, and blue means higher activation in patients. Overall, the components in Figure 7 show higher activation in the angular gyrus (AG), visual, and motor cortex areas in patients with schizophrenia while showing higher activation in default mode network (DMN), auditory cortex, and frontoparietal (FP) regions in healthy controls. The FP component is estimated in two separate components, right and left frontoparietal (RFP and LFP). Both RFP and LFP contain the dorsolateral prefrontal cortex (DLPFC) and lateral posterior parietal cortex (LPPC), which are part of the central executive network (CEN). In Figure 7, components differentiating between healthy subjects and patients ( p < 0.05 ) show higher activations in DMN, CEN, motor, and auditory cortex in AOD and SM tasks; and visual, DMN, and AG regions in the SIRP task. All these areas are known to be associated with schizophrenia [49,50,84,85,86,87,88] and therefore indicate meaningful decomposition results.
Next, we use the profiles corresponding to these components and use ac-IVA and IVA-G to identify the profiles correlated with the behavioral features. We use one behavioral feature at a time to constrain ac-IVA decomposition to identify the associated components. After ac-IVA transformation, we identify that the behavioral features show significant correlations ( p < 0.05 ) with only one SCV with the strongest correlation structure. Since most of the profiles across datasets have step-type characteristics, meaning they show significant group differences between healthy controls and patients, ac-IVA is expected to estimate an SCV that summarizes that property of the profiles. We show the ac-IVA results in Figure 8. By analyzing the columns of the mixing matrices corresponding to the SCV with the strongest correlation structure, we identify that WAIS and WMS2 show significant correlations with components from the AOD dataset, and WMS1 shows a significant correlation with the components from the SIRP dataset. WAIS and WMS2 correlate with the profiles corresponding to DMN, right and left FPs, motor, and auditory cortex components in the AOD dataset, and WMS1 correlates with the profiles corresponding to the visual cortex and RFP components in the SIRP dataset. On the other hand, IVA-G uses all the behavioral features together and uses a single decomposition to analyze behavioral features and the profile matrices jointly. Since we have four behavioral features and seven profiles in each dataset, we generate three padding signals and stack them with the behavioral features to have similar dimensions as profile matrices. After the IVA-G transformation, we identify the SCVs within which the behavioral components show strong correlations with the imaging components. We show the components estimated by IVA-G as correlated with behavioral features in Figure 9. Again, by analyzing the columns of mixing matrices corresponding to these SCVs, we identify that WAIS and WMS1 show correlations to the profiles corresponding to the motor cortex, RFP, and DMN components in the AOD dataset; DMN, visual cortex, and AG components in SIRP dataset; and DMN, visual, and auditory cortex components in SM dataset. These results agree with the results estimated by the ac-IVA method, thus increasing our confidence about the estimated relationships between imaging components and behavioral variables.
In both ac-IVA and IVA-G results, the two behavioral features WAIS and WMS1 show significant correlation with the neuroimaging components. WAIS, which measures verbal working memory through the letter-number sequence subtest, correlates with the profiles corresponding to the motor cortex and the FP components across task datasets. The FP components, especially DLPFC and LPPC, are related to the central executive network of the brain and responsible for the retrieval of short-term memory, self-awareness, and language-based decoding [89]. From Figure 7, Figure 8 and Figure 9, DLPFC and LPPC show higher activation in healthy subjects and lower activation in schizophrenia patients. These areas have been linked with working memory deficits in schizophrenia in multiple reports [32,49,50]. WMS1, which measures visual memory using a face recognition subtest, shows an association with the visual cortex, DMN, and AG components across task datasets. These areas of the brain are responsible for vision, visual perception, memory retrieval, attention, spatial cognition, and reasoning, among other tasks and are found to be associated with schizophrenia in many prior studies [90,91,92,93,94]. These components explain the areas of the brain associated with the behavioral features and could be used as potential predictors for cognitive deficits in schizophrenia patients. Since the behavioral features are a strong indicator of an individual’s cognitive functions, the features or the neuroimaging components correlated to these features, separately or together, can be used to identify subject groups with heterogeneous cognitive profiles. Furthermore, components correlated to WAIS and WMS1 also show strong group differences between healthy controls and patients and hence can be used as putative biomarkers of disease.
Figure 10 shows the subject profiles estimated as correlated with behavioral features using the one-to-one correlation technique. Behavioral features significantly correlate with the DMN, LFP, motor, and auditory cortex components in the AOD dataset; visual cortex and AG components in the SIRP dataset; and motor and auditory cortex components in the SM dataset. Among the behavioral features, WAIS shows correlations with the largest number of subject profiles across datasets. Overall, these results agree with the results estimated by the ac-IVA and IVA-G methods, thus increasing our confidence in the proposed methods. However, behavioral features are correlated with more subject profiles from the AOD task dataset than the SIRP and SM task dataset. Especially, behavioral features are showing significant correlation, after a Bonferroni correction, with only one component in the SM dataset. On the other hand, ac-IVA and IVA-G highlight more components in the SM dataset correlated with behavioral features. A possible explanation for this is the fact that IVA incorporates common information across the datasets that helped ac-IVA and IVA-G to identify more correlated components in the SM dataset. However, IVA will be able to take advantage of common information across datasets only if it is available, and in the absence of it, will reduce to disjoint or one-to-one analyses of individual datasets. Though we use IVA in this work, other multivariate techniques, such as canonical correlation analysis (CCA) and multiset-CCA (MCCA), can also be used instead of IVA.

4. Conclusions

In this work, we present a novel two-step method to estimate and identify multivariate associations between imaging and behavioral features. In doing so, we also introduce a general framework for identifying correlated imaging and behavioral features. Finally, through simulation studies and real-world data, we show the performance advantage of the proposed method over existing ones. With simulation studies, we first show that fusion analysis provides better accuracy in identifying the correlated components than performing separate analyses on individual datasets, as typically done. Then, we show that joint analysis of the imaging and behavioral features using ac-IVA and IVA-G improves accuracy in identifying the underlying relationship between imaging and non-imaging datasets than traditional one-to-one correlation techniques. With real fMRI data collected from healthy controls and patients with schizophrenia, we show that both ac-IVA and IVA-G provide a complete picture of associations between the imaging and behavioral features compared with the existing approach. The results indicate that imaging components show higher activation in the motor cortex, FP, auditory and visual cortex, showing higher correlations with the behavioral variable responsible for measuring working memory. These areas are found to be associated with working memory deficits in many prior studies and hence can be used as predictors for cognitive impairments for schizophrenia patients.

Author Contributions

Conceptualization, M.A.B.S.A., Y.L.-S. and T.A.; methodology, M.A.B.S.A., Y.L.-S. and T.A.; software, M.A.B.S.A.; validation, M.A.B.S.A., Y.L.-S., V.D.C. and T.A.; formal analysis, M.A.B.S.A.; investigation, M.A.B.S.A. and T.A.; resources, M.A.B.S.A., V.D.C. and T.A.; data curation, M.A.B.S.A. and V.D.C.; writing—original draft preparation, M.A.B.S.A.; writing—review and editing, Y.L.-S., V.D.C. and T.A.; visualization, M.A.B.S.A. and T.A.; supervision, T.A.; project administration, T.A.; funding acquisition, V.D.C. and T.A. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the grants NSF-NCS 1631838, NSF 2112455, NIH R01 MH123610, NIH R01 MH118695, and NIH R01 AG073949.

Institutional Review Board Statement

The data is anonymized with no access to identifiers. The IRB determined this work was ‘not human subjects’, thus this is not applicable.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The datasets used in this study are publicly available at https://coins.trendscenter.org/ (accessed on 23 December 2022). The codes are available at https://mlsp.umbc.edu/resources.html (accessed on 23 December 2022).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Biessmann, F.; Plis, S.; Meinecke, F.C.; Eichele, T.; Muller, K.R. Analysis of Multimodal Neuroimaging Data. IEEE Rev. Biomed. Eng. 2011, 4, 26–58. [Google Scholar] [CrossRef] [PubMed]
  2. Greuel, A.; Trezzi, J.P.; Glaab, E.; Ruppert, M.C.; Maier, F.; Jäger, C.; Hodak, Z.; Lohmann, K.; Ma, Y.; Eidelberg, D.; et al. GBA variants in Parkinson’s disease: Clinical, metabolomic, and multimodal neuroimaging phenotypes. Mov. Disord. 2020, 35, 2201–2210. [Google Scholar] [CrossRef] [PubMed]
  3. Cole, J.H. Multimodality neuroimaging brain-age in UK biobank: Relationship to biomedical, lifestyle, and cognitive factors. Neurobiol. Aging 2020, 92, 34–42. [Google Scholar] [CrossRef] [PubMed]
  4. Niu, X.; Zhang, F.; Kounios, J.; Liang, H. Improved prediction of brain age using multimodal neuroimaging data. Hum. Brain Mapp. 2020, 41, 1626–1643. [Google Scholar] [CrossRef]
  5. Vidal-Ribas, P.; Janiri, D.; Doucet, G.E.; Pornpattananangkul, N.; Nielson, D.M.; Frangou, S.; Stringaris, A. Multimodal neuroimaging of suicidal thoughts and behaviors in a US population-based sample of school-age children. Am. J. Psychiatry 2021, 178, 321–332. [Google Scholar] [CrossRef]
  6. Zhang, Y.; Wang, S.; Xia, K.; Jiang, Y.; Qian, P. Alzheimer’s Disease Neuroimaging Initiative. Alzheimer’s disease multiclass diagnosis via multimodal neuroimaging embedding feature selection and fusion. Inf. Fusion 2021, 66, 170–183. [Google Scholar] [CrossRef]
  7. Uludağ, K.; Roebroeck, A. General overview on the merits of multimodal neuroimaging data fusion. NeuroImage 2014, 102, 3–10. [Google Scholar] [CrossRef]
  8. Lahat, D.; Adalı, T.; Jutten, C. Multimodal Data Fusion: An Overview of Methods, Challenges, and Prospects. Proc. IEEE 2015, 103, 1449–1477. [Google Scholar] [CrossRef] [Green Version]
  9. Dähne, S.; Bießmann, F.; Samek, W.; Haufe, S.; Goltz, D.; Gundlach, C.; Villringer, A.; Fazli, S.; Müller, K.R. Multivariate Machine Learning Methods for Fusing Multimodal Functional Neuroimaging Data. Proc. IEEE 2015, 103, 1507–1530. [Google Scholar] [CrossRef]
  10. Adalı, T.; Akhonda, M.A.B.S.; Calhoun, V.D. ICA and IVA for Data Fusion: An Overview and a New Approach Based on Disjoint Subspaces. IEEE Sens. Lett. 2018, 3, 1–4. [Google Scholar] [CrossRef]
  11. Karahan, E.; Rojas-López, P.A.; Bringas-Vega, M.L.; Valdés-Hernández, P.A.; Valdes-Sosa, P.A. Tensor Analysis and Fusion of Multimodal Brain Images. Proc. IEEE 2015, 103, 1531–1559. [Google Scholar] [CrossRef]
  12. Turk, M. Multimodal interaction: A review. Pattern Recognit. Lett. 2014, 36, 189–195. [Google Scholar] [CrossRef]
  13. Bigler, E.D. Neuroimaging and neuropsychology. In Physician’s Field Guide to Neuropsychology; Springer: Berlin/Heidelberg, Germany, 2019; pp. 421–434. [Google Scholar]
  14. Bilder, R.M. Neuropsychology 3.0: Evidence-based science and practice. J. Int. Neuropsychol. Soc. 2011, 17, 7–13. [Google Scholar] [CrossRef] [Green Version]
  15. Bigler, E.D. Structural neuroimaging in neuropsychology: History and contemporary applications. Neuropsychology 2017, 31, 934. [Google Scholar] [CrossRef]
  16. Young, L.; Koenigs, M. Investigating emotion in moral cognition: A review of evidence from functional neuroimaging and neuropsychology. Br. Med. Bull. 2007, 84, 69–79. [Google Scholar] [CrossRef]
  17. Hervais-Adelman, A.G.; Moser-Mercer, B.; Golestani, N. Executive control of language in the bilingual brain: Integrating the evidence from neuroimaging to neuropsychology. Front. Psychol. 2011, 2, 234. [Google Scholar] [CrossRef] [Green Version]
  18. Kringelbach, M.L.; Rolls, E.T. The functional neuroanatomy of the human orbitofrontal cortex: Evidence from neuroimaging and neuropsychology. Prog. Neurobiol. 2004, 72, 341–372. [Google Scholar] [CrossRef]
  19. Yarkoni, T.; Westfall, J. Choosing prediction over explanation in psychology: Lessons from machine learning. Perspect. Psychol. Sci. 2017, 12, 1100–1122. [Google Scholar] [CrossRef]
  20. Meng, X.; Jiang, R.; Lin, D.; Bustillo, J.; Jones, T.; Chen, J.; Yu, Q.; Du, Y.; Zhang, Y.; Jiang, T.; et al. Predicting individualized clinical measures by a generalized prediction framework and multimodal fusion of MRI data. NeuroImage 2017, 145, 218–229. [Google Scholar] [CrossRef] [Green Version]
  21. Karantonis, J.A.; Rossell, S.L.; Carruthers, S.P.; Sumner, P.; Hughes, M.; Green, M.J.; Pantelis, C.; Burdick, K.E.; Cropley, V.; Van Rheenen, T.E. Cognitive validation of cross-diagnostic cognitive subgroups on the schizophrenia-bipolar spectrum. J. Affect. Disord. 2020, 266, 710–721. [Google Scholar] [CrossRef]
  22. Zhu, J.; Wang, C.; Qian, Y.; Cai, H.; Zhang, S.; Zhang, C.; Zhao, W.; Zhang, T.; Zhang, B.; Chen, J.; et al. Multimodal neuroimaging fusion biomarkers mediate the association between gut microbiota and cognition. Prog. Neuro-Psychopharmacol. Biol. Psychiatry 2022, 113, 110468. [Google Scholar] [CrossRef] [PubMed]
  23. Adalı, T.; Anderson, M.; Fu, G. Diversity in Independent Component and Vector Analyses: Identifiability, algorithms, and applications in medical imaging. IEEE Signal Process. Mag. 2014, 31, 18–33. [Google Scholar] [CrossRef]
  24. Tang, F.; Yang, H.; Li, L.; Ji, E.; Fu, Z.; Zhang, Z. Fusion analysis of gray matter and white matter in bipolar disorder by multimodal CCA-joint ICA. J. Affect. Disord. 2020, 263, 80–88. [Google Scholar] [CrossRef] [PubMed]
  25. Calhoun, V.D.; Adalı, T.; Giuliani, N.R.; Pekar, J.J.; Kiehl, K.A.; Pearlson, G.D. Method for multimodal analysis of independent source differences in schizophrenia: Combining gray matter structural and auditory oddball functional data. Hum. Brain Mapp. 2006, 27, 47–62. [Google Scholar] [CrossRef]
  26. Liu, J.; Pearlson, G.; Windemuth, A.; Ruano, G.; Perrone-Bizzozero, N.I.; Calhoun, V.D. Combining fMRI and SNP data to investigate connections between brain function and genetics using parallel ICA. Hum. Brain Mapp. 2009, 30, 241–255. [Google Scholar] [CrossRef] [Green Version]
  27. Sui, J.; Adalı, T.; Pearlson, G.; Yang, H.; Sponheim, S.R.; White, T.; Calhoun, V.D. A CCA+ICA based model for multi-task brain imaging data fusion and its application to schizophrenia. NeuroImage 2010, 51, 123–134. [Google Scholar] [CrossRef] [Green Version]
  28. Groves, A.R.; Beckmann, C.F.; Smith, S.M.; Woolrich, M.W. Linked independent component analysis for multimodal data fusion. NeuroImage 2011, 54, 2198–2217. [Google Scholar] [CrossRef]
  29. Akhonda, M.A.B.S.; Levin-Schwartz, Y.; Bhinge, S.; Calhoun, V.D.; Adalı, T. Consecutive Independence and Correlation Transform for Multimodal Fusion: Application to EEG and fMRI Data. In Proceedings of the 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; pp. 2311–2315. [Google Scholar]
  30. Kim, T.; Attias, H.T.; Lee, S.Y.; Lee, T.W. Blind Source Separation Exploiting Higher-Order Frequency Dependencies. IEEE Trans. Audio Speech Lang. Process. 2007, 15, 70–79. [Google Scholar] [CrossRef]
  31. Adalı, T.; Levin-Schwartz, Y.; Calhoun, V.D. Multi-modal data fusion using source separation: Application to medical imaging. Proc. IEEE 2015, 103, 1494–1506. [Google Scholar] [CrossRef]
  32. Qi, S.; Calhoun, V.D.; van Erp, T.G.M.; Bustillo, J.; Damaraju, E.; Turner, J.A.; Du, Y.; Yang, J.; Chen, J.; Yu, Q.; et al. Multimodal Fusion With Reference: Searching for Joint Neuromarkers of Working Memory Deficits in Schizophrenia. IEEE Trans. Med. Imaging 2018, 37, 93–105. [Google Scholar] [CrossRef] [Green Version]
  33. Sui, J.; Li, X.; Bell, R.; Towe, S.; Gadde, S.; Chen, N.; Meade, C. Structural and Functional Brain Abnormalities in HIV Disease Revealed by Multimodal MRI Fusion: Association with Cognitive Function. Clin. Infect. Dis. 2020. Available online: https://www.natap.org/2020/HIV/ciaa1415.pdf (accessed on 8 May 2021).
  34. Sui, J.; Jiang, R.; Bustillo, J.; Calhoun, V. Neuroimaging-based Individualized Prediction of Cognition and Behavior for Mental Disorders and Health: Methods and Promises. Biol. Psychiatry 2020, 88, 818–828. [Google Scholar] [CrossRef]
  35. Iverson, G.L.; Brooks, B.L.; Langenecker, S.A.; Young, A.H. Identifying a cognitive impairment subgroup in adults with mood disorders. J. Affect. Disord. 2011, 132, 360–367. [Google Scholar] [CrossRef] [Green Version]
  36. Varo, C.; Solé, B.; Jiménez, E.; Bonnín, C.M.; Torrent, C.; Valls, E.; Lahera, G.; Martínez-Arán, A.; Carvalho, A.F.; Miskowiak, K.W.; et al. Identifying social cognition subgroups in euthymic patients with bipolar disorder: A cluster analytical approach. Psychol. Med. 2020, 52, 159–168. [Google Scholar] [CrossRef]
  37. Viswanath, S.E.; Tiwari, P.; Lee, G.; Madabhushi, A. Dimensionality reduction-based fusion approaches for imaging and non-imaging biomedical data: Concepts, workflow, and use-cases. BMC Med. Imaging 2017, 17, 2. [Google Scholar] [CrossRef] [Green Version]
  38. Schwolow, S.; Gerhardt, N.; Rohn, S.; Weller, P. Data fusion of GC-IMS data and FT-MIR spectra for the authentication of olive oils and honeys—Is it worth to go the extra mile? Anal. Bioanal. Chem. 2019, 411, 6005–6019. [Google Scholar] [CrossRef]
  39. August, S.M.; Kiwanuka, J.N.; McMahon, R.P.; Gold, J.M. The MATRICS Consensus Cognitive Battery (MCCB): Clinical and cognitive correlates. Schizophr. Res. 2012, 134, 76–82. [Google Scholar] [CrossRef] [Green Version]
  40. Rissling, A.J.; Makeig, S.; Braff, D.L.; Light, G.A. Neurophysiologic markers of abnormal brain activity in schizophrenia. Curr. Psychiatry Rep. 2010, 12, 572–578. [Google Scholar] [CrossRef] [Green Version]
  41. Choi, J.; Ku, B.; You, Y.G.; Jo, M.; Kwon, M.; Choi, Y.; Jung, S.; Ryu, S.; Park, E.; Go, H.; et al. Resting-state prefrontal EEG biomarkers in correlation with MMSE scores in elderly individuals. Sci. Rep. 2019, 9, 10468. [Google Scholar] [CrossRef] [Green Version]
  42. Tregellas, J.R.; Smucny, J.; Harris, J.G.; Olincy, A.; Maharajh, K.; Kronberg, E.; Eichman, L.C.; Lyons, E.; Freedman, R. Intrinsic hippocampal activity as a biomarker for cognition and symptoms in schizophrenia. Am. J. Psychiatry 2014, 171, 549–556. [Google Scholar] [CrossRef]
  43. Hirjak, D.; Rashidi, M.; Kubera, K.M.; Northoff, G.; Fritze, S.; Schmitgen, M.M.; Sambataro, F.; Calhoun, V.D.; Wolf, R.C. Multimodal magnetic resonance imaging data fusion reveals distinct patterns of abnormal brain structure and function in catatonia. Schizophr. Bull. 2020, 46, 202–210. [Google Scholar] [CrossRef] [Green Version]
  44. Qi, S.; Bustillo, J.; Turner, J.A.; Jiang, R.; Zhi, D.; Fu, Z.; Deramus, T.P.; Vergara, V.; Ma, X.; Yang, X.; et al. The relevance of transdiagnostic shared networks to the severity of symptoms and cognitive deficits in schizophrenia: A multimodal brain imaging fusion study. Transl. Psychiatry 2020, 10, 149. [Google Scholar] [CrossRef]
  45. Gollub, R.L.; Shoemaker, J.M.; King, M.D.; White, T.; Ehrlich, S.; Sponheim, S.R.; Clark, V.P.; Turner, J.A.; Mueller, B.A.; Magnotta, V.; et al. The MCIC collection: A shared repository of multi-modal, multi-site brain image data from a clinical investigation of schizophrenia. Neuroinformatics 2013, 11, 367–388. [Google Scholar] [CrossRef] [Green Version]
  46. Wechsler, D. WAIS-III Administration and Scoring Manual; The Psychological Corporation: San Antonio, TX, USA, 1997. [Google Scholar]
  47. Wechsler, D. WAIS-3, WMS-3: Wechsler Adult Intelligence Scale, Wechsler Memory Scale: Technical Manual; Psychological Corporation: San Antonio, TX, USA, 1997. [Google Scholar]
  48. Brandt, J. The Hopkins Verbal Learning Test: Development of a new memory test with six equivalent forms. Clin. Neuropsychol. 1991, 5, 125–142. [Google Scholar] [CrossRef]
  49. Borders, A. Rumination and Related Constructs: Causes, Consequences, and Treatment of Thinking Too Much; Academic Press: Cambridge, MA, USA, 2020. [Google Scholar]
  50. Glahn, D.C.; Ragland, J.D.; Abramoff, A.; Barrett, J.; Laird, A.R.; Bearden, C.E.; Velligan, D.I. Beyond hypofrontality: A quantitative meta-analysis of functional neuroimaging studies of working memory in schizophrenia. Hum. Brain Mapp. 2005, 25, 60–69. [Google Scholar] [CrossRef] [Green Version]
  51. Calhoun, V.D.; Adalı, T. Feature-Based Fusion of Medical Imaging Data. IEEE Trans. Inf. Technol. Biomed. 2009, 13, 711–720. [Google Scholar] [CrossRef] [Green Version]
  52. Atrey, P.K.; Hossain, M.A.; El Saddik, A.; Kankanhalli, M.S. Multimodal fusion for multimedia analysis: A survey. Multimed. Syst. 2010, 16, 345–379. [Google Scholar] [CrossRef]
  53. Guo, Y.; Pagnoni, G. A unified framework for group independent component analysis for multi-subject fMRI data. NeuroImage 2008, 42, 1078–1093. [Google Scholar] [CrossRef] [Green Version]
  54. Lee, J.H.; Lee, T.W.; Jolesz, F.A.; Yoo, S.S. Independent vector analysis (IVA): Multivariate approach for fMRI group study. Neuroimage 2008, 40, 86–109. [Google Scholar] [CrossRef]
  55. Polikar, R.; Topalis, A.; Parikh, D.; Green, D.; Frymiare, J.; Kounios, J.; Clark, C.M. An ensemble based data fusion approach for early diagnosis of Alzheimer’s disease. Inf. Fusion 2008, 9, 83–95. [Google Scholar] [CrossRef]
  56. Ramezani, M.; Marble, K.; Trang, H.; Johnsrude, I.S.; Abolmaesumi, P. Joint Sparse Representation of Brain Activity Patterns in Multi-Task fMRI Data. IEEE Trans. Med. Imaging 2015, 34, 2–12. [Google Scholar] [CrossRef]
  57. Gabrielson, B.; Akhonda, M.A.; Bhinge, S.; Brooks, J.; Long, Q.; Adali, T. Joint-IVA for identification of discriminating features in EEG: Application to a driving study. Biomed. Signal Process. Control 2020, 61, 101948. [Google Scholar] [CrossRef]
  58. Zhang, Y.; Xiao, L.; Zhang, G.; Cai, B.; Stephen, J.M.; Wilson, T.W.; Calhoun, V.D.; Wang, Y.P. Multi-Paradigm fMRI Fusion via Sparse Tensor Decomposition in Brain Functional Connectivity Study. IEEE J. Biomed. Health Inf. 2021, 25, 1712–1723. [Google Scholar] [CrossRef]
  59. Huster, R.J.; Debener, S.; Eichele, T.; Herrmann, C.S. Methods for simultaneous EEG-fMRI: An introductory review. J. Neurosci. 2012, 32, 6053–6060. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  60. Maglanoc, L.A.; Kaufmann, T.; Jonassen, R.; Hilland, E.; Beck, D.; Landrø, N.I.; Westlye, L.T. Multimodal fusion of structural and functional brain imaging in depression using linked independent component analysis. Hum. Brain Mapp. 2020, 41, 241–255. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  61. Liang, L.; Chen, Z.; Wei, Y.; Tang, F.; Nong, X.; Li, C.; Yu, B.; Duan, G.; Su, J.; Mai, W.; et al. Fusion Analysis of Gray Matter and White Matter in Subjective Cognitive Decline and Mild Cognitive Impairment by Multimodal CCA-joint ICA. NeuroImage Clin. 2021, 32, 102874. [Google Scholar] [CrossRef]
  62. Mangalathu-Arumana, J.; Liebenthal, E.; Beardsley, S.A. Optimizing Within-Subject Experimental Designs for jICA of Multi-Channel ERP and fMRI. Front. Neurosci. 2018, 12, 13. [Google Scholar] [CrossRef] [Green Version]
  63. Thomann, A.K.; Schmitgen, M.M.; Kmuche, D.; Ebert, M.P.; Thomann, P.A.; Szabo, K.; Gass, A.; Griebe, M.; Reindl, W.; Wolf, R.C. Exploring joint patterns of brain structure and function in inflammatory bowel diseases using multimodal data fusion. Neurogastroenterol. Motil. 2021, 33, e14078. [Google Scholar] [CrossRef]
  64. Hosseini, M.P.; Tran, T.X.; Pompili, D.; Elisevich, K.; Soltanian-Zadeh, H. Multimodal data analysis of epileptic EEG and rs-fMRI via deep learning and edge computing. Artif. Intell. Med. 2020, 104, 101813. [Google Scholar] [CrossRef]
  65. Yilmaz, Y.; Aktukmak, M.; Hero, A.O. Multimodal Data Fusion in High-Dimensional Heterogeneous Datasets Via Generative Models. IEEE Trans. Signal Process. 2021, 69, 5175–5188. [Google Scholar] [CrossRef]
  66. Morioka, H.; Calhoun, V.; Hyvärinen, A. Nonlinear ICA of fMRI reveals primitive temporal structures linked to rest, task, and behavioral traits. NeuroImage 2020, 218, 116989. [Google Scholar] [CrossRef]
  67. Smith, E.; Salat, D.; Jeng, J.; McCreary, C.; Fischl, B.; Schmahmann, J.; Dickerson, B.; Viswanathan, A.; Albert, M.; Blacker, D.; et al. Correlations between MRI white matter lesion location and executive function and episodic memory. Neurology 2011, 76, 1492–1499. [Google Scholar] [CrossRef] [Green Version]
  68. Xing, C.; Zhang, J.; Cui, J.; Yong, W.; Hu, J.; Yin, X.; Wu, Y.; Chen, Y.C. Disrupted Functional Network Connectivity Predicts Cognitive Impairment in Presbycusis Patients. Front. Aging Neurosci. 2020, 12, 246. [Google Scholar] [CrossRef]
  69. Sui, J.; Qi, S.; van Erp, T.G.; Bustillo, J.; Jiang, R.; Lin, D.; Turner, J.A.; Damaraju, E.; Mayer, A.R.; Cui, Y.; et al. Multimodal neuromarkers in schizophrenia via cognition-guided MRI fusion. Nat. Commun. 2018, 9, 3028. [Google Scholar] [CrossRef] [Green Version]
  70. Levin-Schwartz, Y.; Calhoun, V.D.; Adalı, T. Quantifying the Interaction and Contribution of Multiple Datasets in Fusion: Application to the Detection of Schizophrenia. IEEE Trans. Med. Imaging 2017, 36, 1385–1395. [Google Scholar] [CrossRef]
  71. Bhinge, S.; Mowakeaa, R.; Calhoun, V.D.; Adalı, T. Extraction of Time-Varying Spatiotemporal Networks Using Parameter-Tuned Constrained IVA. IEEE Trans. Med. Imaging 2019, 38, 1715–1725. [Google Scholar] [CrossRef]
  72. Long, Q.; Bhinge, S.; Levin-Schwartz, Y.; Boukouvalas, Z.; Calhoun, V.D.; Adalı, T. The role of diversity in data-driven analysis of multi-subject fMRI data: Comparison of approaches based on independence and sparsity using global performance metrics. Hum. Brain Mapp. 2019, 40, 489–504. [Google Scholar] [CrossRef] [Green Version]
  73. Calhoun, V.D.; Adalı, T. Multisubject Independent Component Analysis of fMRI: A Decade of Intrinsic Networks, Default Mode, and Neurodiagnostic Discovery. IEEE Rev. Biomed. Eng. 2012, 5, 60–73. [Google Scholar] [CrossRef] [Green Version]
  74. Bhinge, S.; Long, Q.; Calhoun, V.D.; Adali, T. Spatial dynamic functional connectivity analysis identifies distinctive biomarkers in schizophrenia. Front. Neurosci. 2019, 13, 1006. [Google Scholar] [CrossRef]
  75. Anderson, M.; Adalı, T.; Li, X. Joint Blind Source Separation With Multivariate Gaussian Model: Algorithms and Performance Analysis. IEEE Trans. Signal Process. 2012, 60, 1672–1683. [Google Scholar] [CrossRef]
  76. Akhonda, M.; Gabrielson, B.; Bhinge, S.; Calhoun, V.D.; Adali, T. Disjoint subspaces for common and distinct component analysis: Application to the fusion of multi-task FMRI data. J. Neurosci. Methods 2021, 358, 109214. [Google Scholar] [CrossRef]
  77. Lindquist, M.A. The statistical analysis of fMRI data. Stat. Sci. 2008, 2008, 439–464. [Google Scholar] [CrossRef]
  78. Li, X.; Adalı, T. A novel entropy estimator and its application to ICA. In Proceedings of the IEEE International Workshop on Machine Learning for Signal Processing (MLSP), Grenoble, France, 1–4 September 2009; pp. 1–6. [Google Scholar]
  79. SPM5. Statistical Parametric Mapping. 2011. Available online: http://www.fil.ion.ucl.ac.uk/spm/software/spm5 (accessed on 8 May 2021).
  80. Kiehl, K.A.; Liddle, P.F. An event-related functional magnetic resonance imaging study of an auditory oddball task in schizophrenia. Schizophr. Res. 2001, 48, 159–171. [Google Scholar] [CrossRef]
  81. Michael, A.M.; Baum, S.A.; Fries, J.F.; Ho, B.C.; Pierson, R.K.; Andreasen, N.C.; Calhoun, V.D. A method to fuse fMRI tasks through spatial correlations: Applied to schizophrenia. Hum. Brain Mapp. 2009, 30, 2512–2529. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  82. Fu, G.; Anderson, M.; Adalı, T. Likelihood Estimators for Dependent Samples and Their Application to Order Detection. IEEE Trans. Signal Process. 2014, 62, 4237–4244. [Google Scholar] [CrossRef]
  83. Long, Q.; Jia, C.; Boukouvalas, Z.; Gabrielson, B.; Emge, D.; Adalı, T. Consistent Run Selection for Independent Component Analysis: Application to fMRI Analysis. In Proceedings of the 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; pp. 2581–2585. [Google Scholar]
  84. Hu, M.; Zong, X.; Mann, J.J.; Zheng, J.; Liao, Y.; Li, Z.; He, Y.; Chen, X.; Tang, J. A Review of the Functional and Anatomical Default Mode Network in Schizophrenia. Neurosci. Bull. 2017, 33, 73–84. [Google Scholar] [CrossRef] [Green Version]
  85. Du, W.; Calhoun, V.; Li, H.; Ma, S.; Eichele, T.; Kiehl, K.; Pearlson, G.; Adalı, T. High Classification Accuracy for Schizophrenia with Rest and Task fMRI Data. Front. Hum. Neurosci. 2012, 6, 145. [Google Scholar] [CrossRef] [Green Version]
  86. Manoach, D.S.; Gollub, R.L.; Benson, E.S.; Searl, M.M.; Goff, D.C.; Halpern, E.; Saper, C.B.; Rauch, S.L. Schizophrenic subjects show aberrant fMRI activation of dorsolateral prefrontal cortex and basal ganglia during working memory performance. Biol. Psychiatry 2000, 48, 99–109. [Google Scholar] [CrossRef]
  87. Kim, D.; Manoach, D.; Mathalon, D.; Turner, J.; Mannell, M.; Brown, G.; Ford, J.; Gollub, R.; White, T.; Wible, C.; et al. Dysregulation of working memory and default-mode networks in schizophrenia during a Sternberg item recognition paradigm. Hum. Brain Mapp. 2009, 30, 3795. [Google Scholar] [CrossRef] [Green Version]
  88. Saldarini, F.; Gottlieb, N.; Stokes, P.R. Neural Correlates of Working Memory function in Euthymic People with Bipolar Disorder Compared to Healthy Controls: A Systematic Review and Meta-analysis. J. Affect. Disorders 2021, 297, 610–622. [Google Scholar] [CrossRef]
  89. Chan, Y.C.; Lavallee, J.P. Temporo-parietal and fronto-parietal lobe contributions to theory of mind and executive control: An fMRI study of verbal jokes. Front. Psychol. 2015, 6, 1285. [Google Scholar] [CrossRef] [Green Version]
  90. Silverstein, S.M.; Berten, S.; Essex, B.; Kovacs, I.; Susmaras, T.; Little, D.M. An fMRI examination of visual integration in schizophrenia. J. Integr. Neurosci. 2009, 8, 175–202. [Google Scholar] [CrossRef]
  91. Huo, L.; Li, R.; Wang, P.; Zheng, Z.; Li, J. The default mode network supports episodic memory in cognitively unimpaired elderly individuals: Different contributions to immediate recall and delayed recall. Front. Aging Neurosci. 2018, 10, 6. [Google Scholar] [CrossRef]
  92. Seghier, M.L. The angular gyrus: Multiple functions and multiple subdivisions. Neuroscientist 2013, 19, 43–61. [Google Scholar] [CrossRef] [Green Version]
  93. Öngür, D.; Lundy, M.; Greenhouse, I.; Shinn, A.K.; Menon, V.; Cohen, B.M.; Renshaw, P.F. Default mode network abnormalities in bipolar disorder and schizophrenia. Psychiatry Res. Neuroimaging 2010, 183, 59–68. [Google Scholar] [CrossRef] [Green Version]
  94. Whitfield-Gabrieli, S.; Thermenos, H.W.; Milanovic, S.; Tsuang, M.T.; Faraone, S.V.; McCarley, R.W.; Shenton, M.E.; Green, A.I.; Nieto-Castanon, A.; LaViolette, P.; et al. Hyperactivity and hyperconnectivity of the default network in schizophrenia and in first-degree relatives of persons with schizophrenia. Proc. Natl. Acad. Sci. USA 2009, 106, 1279–1284. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Three different ways to identify associations between imaging and behavioral features using (a) one-to-one, (b) one-to-many, and (c) many-to-many association techniques. Here, r ( · , · ) is the function that denotes the association between the profiles and behavioral features.
Figure 1. Three different ways to identify associations between imaging and behavioral features using (a) one-to-one, (b) one-to-many, and (c) many-to-many association techniques. Here, r ( · , · ) is the function that denotes the association between the profiles and behavioral features.
Sensors 22 01224 g001
Figure 2. Step 1: IVA step for the fusion of neuroimaging datasets. IVA estimates interpretable components independent within the datasets while maximally correlated with the same index components across the datasets. Subject profiles corresponding to these components represent each subject’s contribution to the components and can be used to identify multivariate associations with behavioral features.
Figure 2. Step 1: IVA step for the fusion of neuroimaging datasets. IVA estimates interpretable components independent within the datasets while maximally correlated with the same index components across the datasets. Subject profiles corresponding to these components represent each subject’s contribution to the components and can be used to identify multivariate associations with behavioral features.
Sensors 22 01224 g002
Figure 3. Step 2: Two different techniques to identify the association across imaging and behavioral features. Here, (a) represents the ac-IVA technique where a single behavioral feature is used to constrain imaging features (subjects profiles), and (b) represents the IVA-G approach where all the behavioral features are taken into account to identify the correlated imaging and behavioral features. In both cases, “black dot”s in the B k matrices represent the identities or indices of the correlated profiles or behavioral features in A ^ k T and Y T , which, in turn, represent indices of the components in U k ’s associated with behavioral features.
Figure 3. Step 2: Two different techniques to identify the association across imaging and behavioral features. Here, (a) represents the ac-IVA technique where a single behavioral feature is used to constrain imaging features (subjects profiles), and (b) represents the IVA-G approach where all the behavioral features are taken into account to identify the correlated imaging and behavioral features. In both cases, “black dot”s in the B k matrices represent the identities or indices of the correlated profiles or behavioral features in A ^ k T and Y T , which, in turn, represent indices of the components in U k ’s associated with behavioral features.
Sensors 22 01224 g003
Figure 4. Estimation performance of individual ICAs, jICA, DS-ICA, and IVA for different (a) number of subject and (b) Step-heights. Here, correlated profiles and behavioral features are identified using the one-to-one correlation analysis technique. All the results are averaged over 100 independent runs.
Figure 4. Estimation performance of individual ICAs, jICA, DS-ICA, and IVA for different (a) number of subject and (b) Step-heights. Here, correlated profiles and behavioral features are identified using the one-to-one correlation analysis technique. All the results are averaged over 100 independent runs.
Sensors 22 01224 g004
Figure 5. Accuracy performance of correlation, regression, ac-IVA, and IVA-G to identify the profiles correlated with the behavioral features for different (a) numbers of subjects and (b) step-heights. The components and their corresponding profiles are estimated using IVA. Performances of all the methods improve with the number of subjects and step-heights. All the results are averaged over 100 independent runs.
Figure 5. Accuracy performance of correlation, regression, ac-IVA, and IVA-G to identify the profiles correlated with the behavioral features for different (a) numbers of subjects and (b) step-heights. The components and their corresponding profiles are estimated using IVA. Performances of all the methods improve with the number of subjects and step-heights. All the results are averaged over 100 independent runs.
Sensors 22 01224 g005
Figure 6. The four behavioral features used in this study are collected using the letter-number sequence subtest of WAIS-III (WAIS), the face recognition subtest of WMS-III (WMS1), and the logical memory test of WMS-III (WMS2) and HVLT (HVLT). The behavioral feature WAIS measures the working memory, WMS1 measures the visual memory, and WMS2 and HVLT measure the verbal memory and learning. All four features show strong group differences ( p < 0.05 ) between healthy controls and patients with schizophrenia.
Figure 6. The four behavioral features used in this study are collected using the letter-number sequence subtest of WAIS-III (WAIS), the face recognition subtest of WMS-III (WMS1), and the logical memory test of WMS-III (WMS2) and HVLT (HVLT). The behavioral feature WAIS measures the working memory, WMS1 measures the visual memory, and WMS2 and HVLT measure the verbal memory and learning. All four features show strong group differences ( p < 0.05 ) between healthy controls and patients with schizophrenia.
Sensors 22 01224 g006
Figure 7. FMRI components estimated by the IVA-L-SOS algorithm. Here, the colormap of the components is adjusted so that the colors red, orange and yellow mean higher relative activation in controls and blue means higher relative activation in patients. The estimated components show significant group differences even after strict Bonferroni correction ( ( p × 25 ) 0.05 ) in DMN, FP, motor, and auditory cortex areas for AOD and SM tasks; and visual cortex and AG areas for the SIRP task.
Figure 7. FMRI components estimated by the IVA-L-SOS algorithm. Here, the colormap of the components is adjusted so that the colors red, orange and yellow mean higher relative activation in controls and blue means higher relative activation in patients. The estimated components show significant group differences even after strict Bonferroni correction ( ( p × 25 ) 0.05 ) in DMN, FP, motor, and auditory cortex areas for AOD and SM tasks; and visual cortex and AG areas for the SIRP task.
Sensors 22 01224 g007
Figure 8. Behavioral features and subject profiles identified as correlated by ac-IVA. Here, (a) WAIS is correlated with the profiles associated to motor, RFP, auditory, and DMN components in the AOD task; (b) WMS1 is correlated with the profiles associated with visual and RFP components in the SIRP task; and finally, (c) WMS2 is correlated with the profiles associated to motor, and FP components in the SM task.
Figure 8. Behavioral features and subject profiles identified as correlated by ac-IVA. Here, (a) WAIS is correlated with the profiles associated to motor, RFP, auditory, and DMN components in the AOD task; (b) WMS1 is correlated with the profiles associated with visual and RFP components in the SIRP task; and finally, (c) WMS2 is correlated with the profiles associated to motor, and FP components in the SM task.
Sensors 22 01224 g008
Figure 9. Behavioral features and subject profiles identified as correlated by IVA-G. Here, WAIS and WMS1 are correlated with the profiles associated to (a) motor cortex, RFP, and DMN components in AOD task; (b) DMN, visual cortex, and AG components in SIRP task; and (c) DMN, visual, and auditory cortex components in SM task.
Figure 9. Behavioral features and subject profiles identified as correlated by IVA-G. Here, WAIS and WMS1 are correlated with the profiles associated to (a) motor cortex, RFP, and DMN components in AOD task; (b) DMN, visual cortex, and AG components in SIRP task; and (c) DMN, visual, and auditory cortex components in SM task.
Sensors 22 01224 g009
Figure 10. Behavioral features and neuroimaging components identified as correlated using one-to-one Pearson correlation technique across (a) AOD, (b) SIRP, and (c) SM task datasets. Here, we are showing p-values of the correlation and use p < 0.05 as a threshold to identify the significantly correlated behavioral and neuroimaging features.
Figure 10. Behavioral features and neuroimaging components identified as correlated using one-to-one Pearson correlation technique across (a) AOD, (b) SIRP, and (c) SM task datasets. Here, we are showing p-values of the correlation and use p < 0.05 as a threshold to identify the significantly correlated behavioral and neuroimaging features.
Sensors 22 01224 g010
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Akhonda, M.A.B.S.; Levin-Schwartz, Y.; Calhoun, V.D.; Adali, T. Association of Neuroimaging Data with Behavioral Variables: A Class of Multivariate Methods and Their Comparison Using Multi-Task FMRI Data. Sensors 2022, 22, 1224. https://doi.org/10.3390/s22031224

AMA Style

Akhonda MABS, Levin-Schwartz Y, Calhoun VD, Adali T. Association of Neuroimaging Data with Behavioral Variables: A Class of Multivariate Methods and Their Comparison Using Multi-Task FMRI Data. Sensors. 2022; 22(3):1224. https://doi.org/10.3390/s22031224

Chicago/Turabian Style

Akhonda, M. A. B. S., Yuri Levin-Schwartz, Vince D. Calhoun, and Tülay Adali. 2022. "Association of Neuroimaging Data with Behavioral Variables: A Class of Multivariate Methods and Their Comparison Using Multi-Task FMRI Data" Sensors 22, no. 3: 1224. https://doi.org/10.3390/s22031224

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop