Next Article in Journal
High-Temperature Behavior of CaO-FeOx-Al2O3-SiO2-Rich Alkali Activated Materials
Next Article in Special Issue
A Deep Learning Ensemble Method to Visual Acuity Measurement Using Fundus Images
Previous Article in Journal
Effect of Fuel Composition on Carbon Black Formation Pathways
Previous Article in Special Issue
All You Need Is a Few Dots to Label CT Images for Organ Segmentation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Bayesian Linear Regression Model for the Analysis of Neuroimaging Data

by
Albert Belenguer-Llorens
1,
Carlos Sevilla-Salcedo
1,
Manuel Desco
2,3,4,5,
Maria Luisa Soto-Montenegro
3,4,* and
Vanessa Gómez-Verdejo
1,*
1
Department of Signal Processing and Communications, University Carlos III of Madrid Leganés, 28911 Leganés, Spain
2
Department of Bioengineering and Aerospace Engineering, University Carlos III of Madrid Leganés, 28911 Leganés, Spain
3
CIBER of Mental Health (CIBERSAM), 28029 Madrid, Spain
4
Instituto de Investigación Sanitaria Gregorio Marañón, 28007 Madrid, Spain
5
Centro Nacional de Investigaciones Cardiovasculares (CNIC), 28029 Madrid, Spain
*
Authors to whom correspondence should be addressed.
Submission received: 7 December 2021 / Revised: 8 February 2022 / Accepted: 21 February 2022 / Published: 1 March 2022
(This article belongs to the Special Issue Applications of Artificial Intelligence in Medicine Practice)

Abstract

:
In this paper, we propose a novel Machine Learning Model based on Bayesian Linear Regression intended to deal with the low sample-to-variable ratio typically found in neuroimaging studies and focusing on mental disorders. The proposed model combines feature selection capabilities with a formulation in the dual space which, in turn, enables efficient work with neuroimaging data. Thus, we have tested the proposed algorithm with real MRI data from an animal model of schizophrenia. The results show that our proposal efficiently predicts the diagnosis and, at the same time, detects regions which clearly match brain areas well-known to be related to schizophrenia.

1. Introduction

Neuroimaging has undergone a major breakthrough in recent years and has helped in the diagnosis, prognosis, and treatment monitoring of psychiatric disorders. The clinical diagnosis of these disorders is troublesome due to the lack of specific biomarkers [1] and to the fact that many of them share clinical features, thus hindering an accurate diagnosis. Specifically, schizophrenia is one of the most complex pathologies to diagnose [2] since it is commonly confused with other psychotic disorders in up to 20% of cases [3]. As consequence, new tools for the diagnosis of mental disorders are emerging [4,5].
Machine Learning (ML) techniques have emerged as a promising tool for the analysis of neuroimaging data. These algorithms are capable of analyzing any data source, either images (structural or functional), genetic information [6] or behavioral information [7], to carry out an automatic diagnosis of the pathology. Recent approaches based on Support Vector Machine algorithm (SVM) have been applied in Magnetic Resonance Imaging (MRI), showing great results in this field and detecting relevant brain areas involved in the pathology, as well as inferring new useful biomarkers for their diagnosis [8,9,10]. However, although these models have provided accurate results for automatic classification, the lack of interpretability in their results prevents the characterization of the pathology. In particular, in contexts where only a few features are relevant for the problem, it is advisable to detect the informative variables and eliminate the useless ones. For this reason, many authors combine ML models with Feature Selection (FS) approaches, such as the Recursive Feature Elimination (RFE) [11], consisting of the direct elimination of the less representative features, methods based on decision tree formulations, such as Random Forest Importance (RFI) [12,13], or embedded approaches which include L1 or L1–L2 regularizations to promote sparsity, such as Lasso and elastic-net algorithms [14,15]. Nevertheless, in neuroimaging, we have to deal with large datasets, where the number of cases is significantly smaller than the number of variables, and many of these approaches fail in this scenario, tending to over-fit. To avoid this problem, some authors propose Bayesian approaches but work over a reduced set of features [16,17,18], whereas others point to the use of more refined techniques better adapted to the problem needs [19,20,21].
To overcome these limitations, we present a novel formulation for the Bayesian Linear Regression model. Our proposal, called the Dual Bayesian Linear regression model with Feature Selection (DBL-FS), is formulated to work efficiently with a reduced number of samples characterized in high-dimensional spaces, e.g., neuroimaging data. For this purpose, the model is formulated in the dual space and simultaneously includes an Automatic Relevance Determination (ARD) prior over the primal weights to provide the model with FS capabilities so that it can remove irrelevant input features. Here, we have tested our formulation on rodent data in an animal model of schizophrenia that show similar brain anatomical deficits than patients with schizophrenia [22,23,24]. One advantage of using rodent data is a more solid knowledge of the ground truth due to the controlled experimental induction of the pathology.

2. Materials

Rodent MRI data were obtained from the Biomedical Imaging and Instrumentation Group (Biig) of the Gregorio Marañón Hospital. The dataset consisted of 53 rat brain MRI images divided into two groups: healthy rats (N = 24) and pathological rats (N = 29). Pathology was induced by the administration of the viral mimic polyriboinosinic-polyribocytidilic acid (poly I:C) in gestational day 15 to pregnant Wistar rats, since maternal immune stimulation (MIS) is associated with increased risk of onset of schizophrenia in the offspring, with behavioral abnormalities as well as neurophysiological and morphological traits. Model details can be found elsewhere [25,26,27].
All images were preprocessed following the standard preprocessing pipeline in neuroimaging research, using the processing toolbox of the Statistical Parametric mapping software (SPM12) [28], as shown in Figure 1. Output consisted of: White Matter (WM), Gray Matter (GM), and CerebroSpinal Fluid (CSF) regions, with 464,487, 582,467, and 30,702 voxels, respectively.

3. Methods

This section introduces the formulation of the proposed Dual Bayesian Linear regression model with Feature Selection (DBL-FS). Later, we also introduce some reference methods that we will use as baselines to show the advantages of the proposed approach together to the experimental setup.

3.1. A Dual Bayesian Linear Regression Model with Feature Selection

3.1.1. Model Definition

The proposed model borrows some ideas from the Bayesian Principal Component Analysis (BPCA) [30] and Bayesian Canonical Correlation Analysis (BCCA) [31] algorithms to endow a Bayesian Linear Regression (BLR) [32] with a dual formulation able to carry out automatic feature selection over the primal variables. This relies on including an ARD prior over the weight matrices to automatically infer the feature relevance in the input feature space by assigning higher/lower relevance values when there are more/less relevant features. Meanwhile, the model works with a formulation in the dual space. In turn, this allows the model to efficiently deal with large data problems by working in the data space while it applies a feature selection over the variable space. In addition, we can exploit the DBL-FS Bayesian formulation to facilitate including prior expert knowledge to guide the FS process. This way, we can guide the learning process and compensate the limited number of samples to train the model.
To define the model, let us consider X as the observation matrix with the MRI information of N subjects; this way, each row, x n , : for n = 1 , , N , is a D-dimensional vector containing the brain image of the n-th subject, and each column, x : , d for d = 1 , , D , contains the information of the d-th voxel over the N subjects. On the other hand, the column vector y represents the diagnosis labels (control or schizophrenic) for the N subjects under study. Although each label, y n , belongs to the set { 0 , 1 } (indicating the subject is control or not), for the model formulation, we consider y n , and thus, we will generalize the model for regression problems. Later, we will apply a threshold over the model output to classify each subject into one of two categories.

3.1.2. Generative Model

As the graphical model of Figure 2 shows, the generative model of DBL-FS considers that each datum, x n , : , is combined with a weight vector w plus some Gaussian noise to generate the output variable:
y n = x n , : w + η ,
where η is a Gaussian noise with zero mean and precision τ . In turn, the noise precision is modeled with a gamma distribution with parameters a 0 τ , b 0 τ :
τ Γ a 0 τ , b 0 τ
In addition, DBL-FS considers that the weight associated to the d-th input feature follows a normal distribution:
w d 𝒩 0 , α d 1 d = 1 , , D
where its precision, α d , is modeled with a gamma distribution as:
α d Γ a 0 α , b 0 α d = 1 , , D
This ARD prior over w d allows us to obtain the relevance over the elements of w , and therefore, DBL-FS is capable of automatically setting to zero the features that are irrelevant for the problem.
As the model will have to work with MRI data, composed by few samples (less than 100) and tens or hundreds of thousands of voxels, it is clear that working in the primal space is not the most efficient way to proceed. So, we propose to reformulate the model making use of the Representer Theorem [33] (RT). That is, as the RT states that the primal weights of any regression model resulting from minimizing an empirical error (risk) can be expressed as a linear combination of the input data and its equivalent dual coefficients, we can express w as:
w = X T a
where a is a vector of length N containing the dual variables. As we will see later (see Equation (19)), the lower bound that maximizes our variational inference is equivalent to minimizing an empirical cost. This way, the model can be formulated to work in the dual space as:
y n = k n , : a + η ,
where k n , : denotes to the n-th row of the linear kernel matrix K = X X T . This way, with the dual formulation, the target variables y n , for n = 1 , , N , are modeled as:
y n 𝒩 k n , : a , τ 1 n = 1 , , N .
With this new formulation, the model will be able to work in the space of a , where only N parameters have to be inferred. Thus, model complexity and overfitting risks are drastically reduced, as long as we are able to maintain the feature relevance determination over w , providing the model with feature selection capabilities.
Finally, it is important to note that the model formulation does not need to specifically include the distribution of a since the relation between w and a is deterministic, and therefore, the statistical characterization of w is also characterizing a .

3.1.3. Variational Inference

Once the generative model is defined, we should evaluate the posterior distribution of the variables to estimate their optimum values. Although, in this case, the posterior distribution is not tractable, we can use variational inference together with the mean-field technique [34] to find an approximation to this posterior q ( Θ ) p ( Θ | y , X ) , where Θ contains all model variables. Then, we can define a Lower Bound (LB) using the Kullback–Leibler divergence between the posterior and its approximation; so, maximizing this LB, we can obtain the optimum model parameters. Therefore, using the mean-field approximation to factorize over the posterior, we obtain:
p ( Θ | y , X ) q ( Θ ) = q w q ( a ) q α q ( τ ) ,
and we can determine each approximated distribution by calculating:
ln ( q j * ) = E q j [ ln ( p ( X , y , Θ ) ) ] + c o n s t ,
where E q j implies that we calculate the expectation over all random variables except the j-th variable, and p ( X , y , Θ ) is the joint probability.
Therefore, we can apply (9) to the joint probability for each random variable to obtain the model update rules. Firstly, the distribution of the dual weights a is:
q a = N ( a | a , Σ a ) ,
with mean and variance determined by:
a = τ Σ a K T y
Σ a 1 = X diag ( α ) X T + τ K T K ,
where diag ( α ) represents an identity matrix with vector α as the diagonal. The distribution of variable α is:
q α = Γ α | a α , b α ,
with parameters
a α = a 0 α + D 2
b α = b 0 α + 1 2 diag ( X T a a T X ) ,
where α 0 α and β 0 α are hyperparameters, and the operator diagonal returns a column vector formed by the main diagonal of the matrix. Moreover, the distribution of the noise precision τ is given by:
q τ = Γ τ | a τ , b τ ,
with parameters
a τ = a 0 τ + N 2
b τ = b 0 τ + 1 2 ( n = 1 N y n 2 2 T r { y T K a } + Tr { K T K aa T } ) ,
where α 0 τ and β 0 τ are hyperparameters, and T r { } ˙ is the trace operator. See Appendix A for the full development of these mean field distribution approximations.
Once we have defined the different distributions, the model updates the different random variables in an iterative coordinate-ascent-like optimization where the distribution of each factor is obtained using (10) to (22). This optimization process is guided by the LB cost function defined as:
L B = c o n s t + n = 1 N D 2 + a 0 α + 1 ln b α D 2 + a 0 τ + 1 ln b τ D 2 ln Σ a ,
where we analyze its convergence to stop the distribution parameters update. See Appendix B for the full development of the LB.
For an efficient optimization of the model, in practice, we will work in the dual space updating the Equations (10) to (18). However, when the model convergence is reached, we can obtain the approximate posterior distribution of w as:
q ( w ) = N ( w | w , Σ w ) ,
with parameters
w = a X
Σ w = X T Σ a X .
Once the model is trained, we can analyze the distribution of w and check which feature components are zero and, therefore, are eliminated, having an automatic selection of the relevant input voxels. This is due to the fact that, despite working in the dual space, the precision of w components, α , is considered in the distribution of a (see Equation (12)).
Moreover, the inclusion of a prior over α (see Equation (13)) in the generative model has an additional advantage since we can use it to adapt the prior distribution of w and include expert knowledge in the model. Thus, in case we want to add more relevance to a particular region (for instance, a neurobiologically meaningful Region of Interest (ROI)), we can initialize the parameters b α associated to the voxels of this region with higher values than the rest to promote that the distribution of w has also higher values for these voxels. Otherwise, if we do not want to include this expert knowledge, this variable will be uniformly initialized over all voxels.

3.2. Baselines

Here, we present the baseline methods used during the experimental section, whose performances will be compared with those of the proposed DBL-FS model. In particular, we considered three approaches, one baseline aimed to solve regression problems (as DBL-FS) and the other two methods specifically designed to solve classification tasks:
  • As the first baseline, we included a regression Gaussian Process (GP) [35], using the implementation provided by the GPy library (available at github accessed on 9 December 2021). We have selected this model since it allows us to define lineal kernels with ARD, so that we can work in the dual space and learn the relevance of the different input features.
  • Next, we included an SVM [36] with a linear kernel using the scikit-learn library [37] to also optimize the model in the dual space.
  • The last selected baseline is the recently proposed adaptation of Sparse Semi-supervised Heterogeneous Interbattery Bayesian Analysis (SSHIBA) [38] to work in the dual space, the Kernelized SSHIBA (KSSHIBA) [39] is available at github accessed on 29 March 2021. This algorithm can simultaneously combine different data sources or views (in our case, different tissues) in a common latent space providing a low-dimensional representation of the data. In addition, this model can also include an additional output view to categorically model the target variable (patient or control sample), as well as a linear kernel with ARD coefficients over the input features (equivalent to the GP configuration).
Both GP and KSSHIBA use an ARD to determine the relevance of the input features, but they do not have a prior distribution or constraint to force their input weights to be sparse and, therefore, obtain a real FS. Meanwhile, DBL-FS imposes sparsity with the Gamma prior to actually promote zero values in the model weights which, in turn, automatically eliminates the least relevant features.
Furthermore, it is important to mention that deep learning models are not included in this study, as these methods are severely limited by the sample size required to learn the model parameters. Therefore, although models such as convolutional neural networks have promising results in image analysis, they also pose serious challenges when working with datasets of small sample size. Furthermore, we have explored other baselines such as random forests but have not included the results due to their poor results. Nevertheless, all the methods under study will be evaluated with different configurations to be able to analyze different properties and, hence, to carry out an extensive study and analyze their advantages and disadvantages in comparison to our model.

3.3. Experimental Setup

MRI data were standardized to zero mean and unitary standard deviation. As we have a reduced number of subjects (only 53 samples), we have used a Leave-One-Out (LOO) framework to evaluate the model performance. This way, we have trained as many models as available samples, using in each training partition all the subjects except one, which was used afterwards for testing. Then, to evaluate the model performance, we present the results in terms of average accuracy, that is, the percentage of correctly classified test subjects computed over all LOO iterations. Furthermore, since the performance of some methods depends on their initialization we repeated the LOO process 10 times (with different initializations) and depicted the average accuracy over them in order to obtain more statically significant results.
To complete the performance analysis, the result table includes the final number of voxels selected by each model (and their percentage with respect to the total), computed as the average number of voxels used by each model for each LOO iteration and each run.
Regarding the different models under study, we considered several configurations to carry out a more comprehensive analysis and more adequate evaluation of the different methods.
For GPs, we have considered two versions: (1) the standard GP with a linear kernel, denoted as GP, and (2) the previous GP but including ARD capabilities and an FS stage. That is, we first trained a GP with ARD and analyzed the ARD coefficients to select the most relevant features, and then trained a standard GP with the chosen features. Thus, this two-step approach provided a GP with FS capabilities, denoted as GP+FS. For this pruning, we selected the 25% most relevant features in order to compare the performance of this method with DBL-FS. In addition, as both DBL-FS and GPs were formulated for regression problems and our predictive task is a binary classification (0 or 1), we set the threshold to 0.5.
We have implemented two different approaches for SVMs: (1) a standard SVM with linear kernel and (2) an SVM with a Multi-Kernel Learning (MKL) strategy, denoted as MKL-SVM. In the latter case, we independently considered the different tissues (GM, CSF, and WM) and a different linear kernels for each of them, and subsequently, the model learned the combination of these three kernels, including two parameters for their combination. These parameters were defined as scalars multiplying each kernel term, and a subsequent inner LOO was used to find their optimal values. Thus, the defined combinations coefficients gave more or less relevance to each kernel (therefore, to each tissue), providing additional flexibility to the model.
For KSSHIBA, we have included two versions, similarly to what was done in GP: (1) the standard KSSHIBA model and (2) a two-stage version of KSSHIBA (denoted as KSSHIBA+FS), in which KSSHIBA was first trained with ARD functionality, and subsequently, we selected the most relevant features to train the model using this subset of features. For these experiments, we initially had 1000 latent factors, from which the model will automatically prune the irrelevant ones. For FS, we kept the highest 25% of voxels equivalently to the number of selected features from the DBL-FS model.
Finally, we have also defined two approaches for the DBL-FS model, with and without expert knowledge. In the latter one, we have equally initialized the ARD prior for all voxels, setting the parameters a and b of random variable α to 2 and 1, respectively. In the expert knowledge case (denoted as DBL-FS+EK), we have initialized the parameters a and b in such a way that the areas of the prefrontal cortex, ventral hippocampus, and lateral ventricles (which are known to be more intensely affected [23]) had more relevance than the rest. In particular, parameter a was set to 50 and parameter b was fixed to either 1 or 0.001 , depending on whether the voxel belonged to the indicated ROIs or not.

4. Experimental Results

Table 1 shows the LOO accuracy results for the classification problem together with the number of selected voxels (the approaches without FS used 100 % ). Despite using different initializations in the evaluated models, the results were stable across them with a negligible standard deviation, showing that the initialization hardly influences the results. For this reason, we did not include the standard deviation in Table 1. The results show that GPs, KSSHIBA+FS, and MKL-SVM obtained the worst classification accuracies, while SVM and KSSHIBA achieved the best performances among the baselines. However, DBL-FS and DBL-FS+EK still obtained an improvement of 5.7% in terms of accuracy over the best baseline while learning the most restrictive selection mask. From these results, we need to highlight that (1) KSSHIBA obtained a predictive performance similar to that of SVM while summarizing the information of the original data (distributed in more than 10 6 voxels) in only nine latent variables, and (2) MKL-SVM showed worse results than the standard SVM, probably due to the higher number of hyperparameters it needed to learn in order to perform the MKL, which may be causing overfitting.
Figure 3 shows the brain areas selected by the methods with FS capabilities. As each voxel has an associated weight, the image masks represent the absolute value of these weight magnitudes for the selected of voxels as an indicator of the voxel relevance. In addition, since we have a model for each LOO iteration, Figure 3 displays the average values of these relevances (over all LOO iterations) and includes a normalization of their scales to the range ( 0 , 1 ) to simplify their analysis. As a result, we can observe that the GP-FS selected meaningless voxels in neurobiological terms while KSSHIBA detected well-defined areas corresponding to only WM tissues. Finally, the DBL-FS and DBL-FS+EK approaches obtained well-defined regions in the GM and WM tissues and the CSF, which are interpretable in neurobiological terms. Although both methods provided similar selections, DBL-FS+EK selected a reduced set of voxels.

5. Discussion

This study shows, for the first time, the great advantage of using DBL-FS for the detection and characterization of the morphometric brain changes in a rodent model of schizophrenia. This Bayesian model was adapted for neuroimaging data, characterized by a low sample-to-variable ratio (53 samples vs. 1,077,656 voxels in our case) relying on a dual formulation of the Bayesian Linear Regression model. Furthermore, as the main novelty of this proposal, we combine this dual formulation with a prior over the primal weights to learn the feature relevance over the input features, forcing an automatic FS. Finally, we can exploit the Bayesian nature of the model to include specific prior knowledge to guide the learning process and counterweight the limitations caused by the low sample size of the problem.
Thus, the comparison in terms of performance with the baselines provides clear evidence of the promising results of the proposed DBL-FS model in the characterization of neuroimaging data in mental disorders. Note that DBL-FS is able to largely outperform the baselines in prediction accuracy, showing an advantage of 5.7 % in terms of accuracy over the best baseline. In addition, DBL-FS is the only method capable of detecting regions within the three brain tissues that are known to be relevant in the biology of schizophrenia. In this sense, the relevance learned by the GPs is inconsistent between the different LOO iterations, generating a scattered voxel selection and, therefore, a non-localized, unreliable, and uninterpretable mask. On the other hand, KSSHIBA provides a consistent result but only finds relevant regions within WM tissue and, therefore, ignores relevant regions and includes some irrelevant areas.
Analyzing in detail the regions selected by the DBL-FS and DBL-FS+Ek models, we can verify that these areas belong to brain regions whose morphometric changes have been related to schizophrenia, based on the literature. First, as for CSF, the areas with the greatest weight were the most frontal areas of the lateral ventricles and the third ventricle. One of the morphometric hallmarks in schizophrenia is the enlargement of the ventricles [23,40], which is consistent with the learned selection. Second, regarding GM, our model clearly defined anatomical areas, such as the prefrontal cortex (PFC), hippocampus, amygdala, and septum, some of them in both hemispheres. Numerous studies have demonstrated the relevance of the morphological changes of these areas in mental disorders [41,42] together with the disconnection and lack of symmetry between both cerebral hemispheres [43,44]. Similar volumetric abnormalities have also been reported for the animal model used in this study [24,45]. In addition, the method also detected the medial septum, which plays a significant role in dopamine-related disorders such as schizophrenia [46,47] and addictions [48,49,50], which highlights the relevance of this structure in mental disorders. Regarding WM, the method found three well-defined brain areas, the frontal part of corpus callosum and WM tracts of the brainstem and the cerebellum [51,52].
Regarding the inclusion of expert knowledge by means of the α prior, it reveals two interesting behaviors. First, it demonstrates the robustness and potential of the standard DBL-FS since it is able to obtain similar accuracy and roughly similar brain masks to its DBL-FS+EK extension without the need for expert information. Second, the possibility of including expert knowledge makes the model converge faster, and it also refines the brain region selection. It is important to note that, although the expert knowledge guides the inference process, the model is also learning from the data, allowing it to redefine the initial expert knowledge into a specific set of voxels. For instance, looking at Figure 3, we can see that using expert knowledge, we obtain a higher relevance associated with the core of the WM brainstem and hippocampal areas.

6. Conclusions

This article shows a novel Bayesian approach using linear regression to characterize neuroimaging data, tested in an animal model of schizophrenia. The proposed DBL-FS+EK model allowed us to efficiently work with neuroimaging data, characterized by a low sample-to-variable ratio. This is achieved by taking advantage of its Bayesian formulation to work in the dual space while learning a voxel importance for feature selection. Furthermore, the use of a specific prior to force sparsity can be combined with expert knowledge to guide the model. The proposed model was analyzed using MRI data from a rodent model of a schizophrenia database and compared to different baselines. The results provided an outstanding classification performance of DBL-FS+EK, improving the accuracy of the second best classifier, SVM, in ∼6%. Furthermore, looking at the selected voxels and their associated relevance, we can confirm that the proposed model is able to detect biologically relevant areas for the characterization of this disease, as it clearly agrees with known literature.

Author Contributions

Conceptualization, V.G.-V. and M.L.S.-M.; methodology, V.G.-V. and C.S.-S.; software, A.B.-L. and C.S.-S.; validation, M.L.S.-M. and A.B.-L.; formal analysis, A.B.-L., V.G.-V. and C.S.-S.; investigation, V.G.-V., M.L.S.-M. and C.S.-S.; resources, M.L.S.-M.; data curation, V.G.-V. and M.L.S.-M.; writing—review and editing, all; visualization, all; supervision, V.G.-V., M.L.S.-M. and C.S.-S.; project administration, V.G.-V. and M.L.S.-M.; funding acquisition, V.G.-V., M.L.S.-M. and M.D. All authors have read and agreed to the published version of the manuscript.

Funding

This paper is part of the project PID2020-115363RB-I00 funded by MCIN/AEI/10.13039/ 501100011033. A.B.-L.’s work is funded by the Community of Madrid through the “Excellence of University Teaching Staff” line of the Multi-year Agreement with UC3M (EPUC3M27), within the framework of the V PRICIT. M.L.S.-M.’s was supported by Ministerio de Ciencia, Innovación y Universidades, Instituto de Salud Carlos III (project number PI17/01766, and grant number BA21/00030), co-financed by European Regional Development Fund (ERDF), “A way to make Europe”, CIBER de Salud Mental (project number CB07/09/0031), Delegación del Gobierno para el Plan Nacional sobre Drogas (project number 2017/085); Fundación Mapfre and Fundación Alicia Koplowitz. M.D.’s work was supported by Ministerio de Ciencia e Innovación (MCIN) and Instituto de Salud Carlos III (ISCIII) (PT20/00044). The CNIC is supported by the ISCIII, the MCIN and the Pro CNIC Foundation, and is a Severo Ochoa Center of Excellence (SEV-2015-0505).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The dataset generated and analyzed during the current study are available from the corresponding author on reasonable request.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. DBL-FS Variational Inference

This section explains in detail the development of the variational inference of the proposed DBL-FS indicated in the Methods section. In particular, here we present the calculation of the mean field approximation of the model parameters:
q ( Θ ) = q w q ( a ) q α q ( τ ) ,
where each term is calculated applying Equation (9) to the joint probability for each random variable to obtain the updated model rules.

Appendix A.1. Mean Field Approximation of a

Using the mean field approximation over variable a , we find that the logarithm of its approximate posterior is:
ln q a = E ln p y | X , a , τ + E ln p w | α , a + c o n s t .
If we develop the first term in the equation, we have:
ln p y | X , a , τ = n = 1 N ln p y n | x n , : , a , τ = n = 1 N ln N x n , : X T a , τ 1 = n = 1 N ( 1 2 ln τ τ 2 ( y n a T X x n , : T ) ( y n x n , : X T a ) ) + c o n s t = N 2 ln τ τ 2 n = 1 N y n 2 2 y n x n , : X T a + a T X x n , : T x n , : X T a + c o n s t ,
and, calculating the expectation of this expression, we obtain:
E τ ln p y | X , a , τ = N 2 ln ( τ ) + τ y K a τ 2 a T K T Ka + c o n s t .
Equivalently, the second term can be calculated as:
ln p w | α , a = d = 1 D ln p w d | α d , a = d = 1 D ln N 0 , α d 1 = d = 1 D 1 2 ln α d 1 2 a T x : , d α d x : , d T a + c o n s t = 1 2 d = 1 D ln α d 1 2 d = 1 D a T x : , d α d x : , d T a + c o n s t ,
and, we if we use the expectation, we have:
E α ln p w | α , a = 1 2 d = 1 D l n ( α d ) 1 2 a T X diag ( α ) X T a + c o n s t .
Now, joining Equations (A4) and (A6), we obtain:
ln q a = τ y T Ka τ 2 a T K T Ka 1 2 a T X diag ( α ) X T a + c o n s t .
Therefore, we can identify the parameters of the q distribution on this equation, having:
q a = N ( a | a , Σ a )
where the mean is:
a = τ Σ a K T y
and the variance is:
Σ a 1 = X diag ( α ) X T + τ K T K

Appendix A.2. Mean Field Approximation of α

Now, using the mean field approximation over variable α , we find that the logarithm of its approximate posterior is:
ln q α = E ln p w | α , a + E ln p α + c o n s t
Developing the first term, we obtain:
ln p w | α , a = 1 2 d = 1 D ln α d 1 2 d = 1 D Tr { a T x : , d α d x : , d T a } + c o n s t ,
and we can apply the expectation to obtain:
E a , τ p w | α , a = 1 2 d = 1 D ln α d 1 2 d = 1 D α d Tr { x : , d T aa T x : , d }
If we look at the second term, we have
ln p α = d = 1 D ( b 0 α α d + a 0 α 1 ln α d ) + c o n s t ,
where we can apply the expectation of the function as:
E ln p α = d = 1 D ln p α d = d = 1 D b 0 α α d + a 0 α 1 ln α d + c o n s t .
Now, if we sum Equations (A13) and (A15), we obtain:
ln q α = d = 1 D 1 2 + a 0 α 1 ln α d α d 2 ( Tr { x : , d T a a T x : , d } + 2 b 0 α ) + c o n s t .
Thus, if we identify terms on the variable distribution, we have:
q α d = d = 1 D Γ α d | a α d , b α d ,
where the first parameter is:
a α = 1 2 + a 0 α ,
and the second parameter can be expressed as:
b α = b 0 α + 1 2 diag ( X T aa T X ) .

Appendix A.3. Mean Field Approximation of τ

Following the same steps as in the two previous approaches, we use the mean field approximation over variable τ to obtain the logarithm of the approximate posterior:
ln q τ = E ln p y | X , a , τ + E ln p τ + c o n s t .
Therefore, the first term on this equation is:
ln p y | X , a , τ = N 2 ln τ τ 2 n = 1 N y n 2 2 Tr y T Ka + Tr K T Kaa T + c o n s t ,
and, applying the expectation we obtain:
E a ln p y | X , a , τ   = N 2 ln τ τ 2 n = 1 N y n 2 2 Tr y T K a + T r K T K aa T + c o n s t .
The second term is defined as:
E ln p τ = ln p τ = b 0 τ τ + a 0 τ 1 ln τ + c o n s t .
Now, if we sum Equations (A22) and (A23), we obtain:
ln q α = N 2 + a 0 τ 1 ln τ τ 2 n = 1 N y n 2 2 Tr y T K a + Tr K T K aa T + 2 b 0 τ + c o n s t .
Therefore, following the same procedure as with the previous variables, we identify terms from the distribution and obtain:
q τ = Γ τ | a τ , b τ ,
where the first parameter is:
a τ = N 2 + a 0 τ ,
and the second one is:
b τ = 1 2 ( n = 1 N y n 2 2 T r y T K a + T r K T K aa T ) + b 0 τ .

Appendix B. Lower Bound Inference

As mentioned in the Methods section, we use the Kullback–Leibler divergence to first determine the similarities between two distribution where, for any two probability density functions p ( x ) and q ( x ) , we have:
D K L = q ( x ) ln q ( x ) p ( x ) d x
In our case, if we particularize for the true posterior and the posterior approximation, the divergence can be expressed as:
D K L = q ( Θ ) q ( Θ ) p ( Θ | X ) d Θ = q ( Θ ) ln ( q ( Θ ) ) d Θ q ( Θ ) ln ( p ( Θ | X ) ) d Θ = E q [ ln ( q ( Θ ) ) ] E q [ ln ( p ( Θ | X ) ) ] .
Developing the conditional probability we obtain:
D K L = E q [ ln ( q ( Θ ) ) ] E q [ ln ( p ( Θ , X ) ) ] + ln ( p ( X ) ) .
Due to the impossibility of working with this distribution because the marginal distribution p ( X ) cannot be calculated, we use an Evidence Lower Bound (ELBO/LB) to this expression [34]. The LB is the divergence of negative KL plus ln ( p ( X ) ) ; therefore, the greatest similarity between the two functions is achieved by maximizing this new measure. We can calculate the LB as:
L q = q Θ ln q Θ p X , Θ d Θ = q Θ ln p X , Θ d Θ q Θ ln q Θ d Θ = E q ln p X , Θ E q ln q Θ
In order to easily calculate this lower bound, we will separately calculate the terms related to E q ln p X , Θ and to the entropy in the following subsections.

Appendix B.1. Terms Associated to E q ln p X , y , Θ

This first term of the lower bound would be composed by the following terms:
E q ln p X , y , Θ   = E q ln p X + E q [ ln ( p ( w | α , a ) ) ] + E q ln p α + E q [ ln ( p ( y | a , X , τ ) ) ] + E q ln p τ
This way, the different elements of this equation can be calculated as:
E q [ ln ( p ( w | α , a ) ) ]   = D 2 ln 2 π + D 2 d = 1 D ψ a f f d ln b f f d d = 1 D a f f d + b 0 α d = 1 D a f f d b f f d
E q ln p α   = a 0 α ln b 0 α ln Γ a 0 α + d = 1 D b 0 α a f f d b f f d + a 0 α 1 ψ a f f d ln b f f d
E q ln p w , α   = D 2 + a 0 α 1 d = 1 D ψ a f f d ln b f f d D 2 ln 2 π + a 0 α ln b 0 α ln Γ a 0 α d = 1 D a f f d
E q [ ln ( p ( y | a , X , τ ) ) ]   = N D 2 ln 2 π + D 2 n = 1 N E q ln τ 1 2 E q τ n = 1 N y n 2 2 T r y T K a + T r a , a T K T K = N D 2 ln 2 π + D 2 ψ a τ ln b τ a τ + a τ b τ b 0 τ
E q ln p τ = a 0 τ ln b 0 τ ln Γ a 0 α ln Γ a 0 τ b 0 τ a τ b τ + a 0 τ 1 ψ a τ ln b τ
E q [ ln ( p ( y , τ | a , X ) ) ]   = N D 2 ln 2 π a τ + a 0 τ ln b 0 τ ln Γ a 0 τ + D 2 + a 0 τ 1 ψ a τ ln b τ

Appendix B.2. Terms of Entropy

The second term in the LB expression is the entropy of the model parameters:
E q ln q Θ = E q ln q w + E q ln q α + E q ln q τ ,
where the entropy of these parameters is:
E q ln q w = D 2 ln 2 π e + D 2 ln Σ w
E q ln q α m   = k = 1 K c a ff k m + ln Γ a ff k m 1 a ff k m ψ a ff k m ln b ff k m
E q ln q ø m   =   a ø m + ln Γ a ø m 1 a ø m ψ a ø m ln b ø m

Appendix B.3. Complete Lower Bound

Finally, joining Equations (A32) and (A39), the complete lower bound is calculated as:
L q = D 2 + a 0 α 1 k = 1 K c ln b α k D 2 + a 0 τ 1 ln b τ D 2 ln Σ w + k = 1 K c ln b α k + ln b τ + const

References

  1. Carvalho, A.F.; Solmi, M.; Sanches, M.; Machado, M.O.; Stubbs, B.; Ajnakina, O.; Sherman, C.; Sun, Y.R.; Liu, C.S.; Brunoni, A.R.; et al. Evidence-based umbrella review of 162 peripheral biomarkers for major mental disorders. Transl. Psychiatry 2020, 10, 152. [Google Scholar] [CrossRef] [PubMed]
  2. Widing, L.; Simonsen, C.; Flaaten, C.B.; Haatveit, B.; Vik, R.K.; Wold, K.F.; Åsbø, G.; Ueland, T.; Melle, I. Symptom Profiles in Psychotic Disorder Not Otherwise Specified. Front. Psychiatry 2020, 11, 580444. [Google Scholar] [CrossRef] [PubMed]
  3. Correll, C.U.; Brevig, T.; Brain, C. Patient characteristics, burden and pharmacotherapy of treatment-resistant schizophrenia: Results from a survey of 204 US psychiatrists. BMC Psychiatry 2019, 19, 362. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Roberts, L.W.; Chan, S.; Torous, J. New tests, new tools: Mobile and connected technologies in advancing psychiatric diagnosis. NPJ Digit. Med. 2018, 1, 20176. [Google Scholar] [CrossRef] [PubMed]
  5. Li, Z.; Li, W.; Wei, Y.; Gui, G.; Zhang, R.; Liu, H.; Chen, Y.; Jiang, Y. Deep learning based automatic diagnosis of first-episode psychosis, bipolar disorder and healthy controls. Comput. Med. Imaging Graph. 2021, 89, 101882. [Google Scholar] [CrossRef] [PubMed]
  6. Trakadis, Y.J.; Sardaar, S.; Chen, A.; Fulginiti, V.; Krishnan, A. Machine learning in schizophrenia genomics, a case-control study using 5090 exomes. Am. J. Med. Genet. Part B Neuropsychiatr. Genet. 2019, 180, 103–112. [Google Scholar] [CrossRef]
  7. Hettige, N.C.; Nguyen, T.B.; Yuan, C.; Rajakulendran, T.; Baddour, J.; Bhagwat, N.; Bani-Fatemi, A.; Voineskos, A.N.; Chakravarty, M.M.; De Luca, V. Classification of suicide attempters in schizophrenia using sociocultural and clinical features: A machine learning approach. Gen. Hosp. Psychiatry 2017, 47, 20–28. [Google Scholar] [CrossRef]
  8. Xiao, Y.; Yan, Z.; Zhao, Y.; Tao, B.; Sun, H.; Li, F.; Yao, L.; Zhang, W.; Chandan, S.; Liu, J.; et al. Support vector machine-based classification of first episode drug-naïve schizophrenia patients and healthy controls using structural MRI. Schizophr. Res. 2019, 214, 11–17. [Google Scholar] [CrossRef]
  9. Guo, Y.; Qiu, J.; Lu, W. Support Vector Machine-Based Schizophrenia Classification Using Morphological Information from Amygdaloid and Hippocampal Subregions. Brain Sci. 2020, 10, 562. [Google Scholar] [CrossRef]
  10. Jahmunah, V.; Oh, S.L.; Rajinikanth, V.; Ciaccio, E.J.; Cheong, K.H.; Arunkumar, N.; Acharya, U.R. Automated detection of schizophrenia using nonlinear signal processing methods. Artif. Intell. Med. 2019, 100, 101698. [Google Scholar] [CrossRef]
  11. Brownlee, J. Recursive Feature Elimination (RFE) for Feature Selection in Python. Machine Learning Mastery. 2020. Available online: https://machinelearningmastery.com/rfe-feature-selection-in-python/ (accessed on 25 May 2020).
  12. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
  13. Li, X.; Chen, W.; Zhang, Q.; Wu, L. Building auto-encoder intrusion detection system based on random forest feature selection. Comput. Secur. 2020, 95, 101851. [Google Scholar] [CrossRef]
  14. Zou, H.; Hastie, T. Regularization and variable selection via the elastic net. J. R. Stat. Soc. Ser. (Stat. Methodol.) 2005, 67, 301–320. [Google Scholar] [CrossRef] [Green Version]
  15. Amini, F.; Hu, G. A two-layer feature selection method using genetic algorithm and elastic net. Expert Syst. Appl. 2021, 166, 114072. [Google Scholar] [CrossRef]
  16. Shen, L.; Qi, Y.; Kim, S.; Nho, K.; Wan, J.; Risacher, S.L.; Saykin, A.J. Sparse bayesian learning for identifying imaging biomarkers in AD prediction. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Beijing, China, 20–24 September 2010; pp. 611–618. [Google Scholar]
  17. Sabuncu, M.R.; Van Leemput, K. The relevance voxel machine (RVoxM): A self-tuning Bayesian model for informative image-based prediction. IEEE Trans. Med. Imaging 2012, 31, 2290–2306. [Google Scholar] [CrossRef] [Green Version]
  18. Sabuncu, M.R. A sparse Bayesian learning algorithm for longitudinal image data. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; pp. 411–418. [Google Scholar]
  19. Parrado-Hernández, E.; Gómez-Verdejo, V.; Martínez-Ramón, M.; Shawe-Taylor, J.; Alonso, P.; Pujol, J.; Menchón, J.M.; Cardoner, N.; Soriano-Mas, C. Discovering brain regions relevant to obsessive–compulsive disorder identification through bagging and transduction. Med. Image Anal. 2014, 18, 435–448. [Google Scholar] [CrossRef]
  20. Gómez-Verdejo, V.; Parrado-Hernández, E.; Tohka, J. Sign-consistency based variable importance for machine learning in brain imaging. Neuroinformatics 2019, 17, 593–609. [Google Scholar] [CrossRef] [Green Version]
  21. Sevilla-Salcedo, C.; Gómez-Verdejo, V.; Tohka, J. Regularized Bagged Canonical Component Analysis for Multiclass Learning in Brain Imaging. Neuroinformatics 2020, 18, 641–659. [Google Scholar] [CrossRef]
  22. Grimm, O.; Gass, N.; Weber-Fahr, W.; Sartorius, A.; Schenker, E.; Spedding, M.; Risterucci, C.; Schweiger, J.I.; Böhringer, A.; Zang, Z.; et al. Acute ketamine challenge increases resting state prefrontal-hippocampal connectivity in both humans and rats. Psychopharmacology 2015, 232, 4231–4241. [Google Scholar] [CrossRef]
  23. Hadar, R.; Soto-Montenegro, M.L.; Götz, T.; Wieske, F.; Sohr, R.; Desco, M.; Hamani, C.; Weiner, I.; Pascau, J.; Winter, C. Using a maternal immune stimulation model of schizophrenia to study behavioral and neurobiological alterations over the developmental course. Schizophr. Res. 2015, 166, 238–247. [Google Scholar] [CrossRef] [Green Version]
  24. Romero-Miguel, D.; Casquero-Veiga, M.; MacDowell, K.S.; Torres-Sanchez, S.; Garcia-Partida, J.A.; Lamanna-Rama, N.; Romero-Miranda, A.; Berrocoso, E.; Leza, J.C.; Desco, M.; et al. A Characterization of the Effects of Minocycline Treatment During Adolescence on Structural, Metabolic, and Oxidative Stress Parameters in a Maternal Immune Stimulation Model of Neurodevelopmental Brain Disorders. Int. J. Neuropsychopharmacol. 2021, 24, 734–748. [Google Scholar] [CrossRef] [PubMed]
  25. Ozawa, K.; Hashimoto, K.; Kishimoto, T.; Shimizu, E.; Ishikura, H.; Iyo, M. Immune activation during pregnancy in mice leads to dopaminergic hyperfunction and cognitive impairment in the offspring: A neurodevelopmental animal model of schizophrenia. Biol. Psychiatry 2006, 59, 546–554. [Google Scholar] [CrossRef] [PubMed]
  26. Zhu, F.; Zheng, Y.; Liu, Y.; Zhang, X.; Zhao, J. Minocycline alleviates behavioral deficits and inhibits microglial activation in the offspring of pregnant mice after administration of polyriboinosinic–polyribocytidilic acid. Psychiatry Res. 2014, 219, 680–686. [Google Scholar] [CrossRef] [PubMed]
  27. Meyer, U.; Feldon, J. To poly (I: C) or not to poly (I: C): Advancing preclinical schizophrenia research through the use of prenatal immune activation models. Neuropharmacology 2012, 62, 1308–1321. [Google Scholar] [CrossRef]
  28. Casquero-Veiga, M.; Garcia-Garcia, D.; MacDowell, K.S.; Perez-Caballero, L.; Torres-Sanchez, S.; Fraguas, D.; Berrocoso, E.; Leza, J.C.; Arango, C.; Desco, M.; et al. Risperidone administered during adolescence induced metabolic, anatomical and inflammatory/oxidative changes in adult brain: A pet and mri study in the maternal immune stimulation animal model. Eur. Neuropsychopharmacol. 2019, 29, 880–896. [Google Scholar] [CrossRef]
  29. Valdes Hernandez, P.A.; Sumiyoshi, A.; Nonaka, H.; Haga, R.; Aubert Vasquez, E.; Ogawa, T.; Iturria Medina, Y.; Riera, J.J.; Kawashima, R. An in vivo MRI template set for morphometry, tissue segmentation, and fMRI localization in rats. Front. Neuroinform. 2011, 5, 26. [Google Scholar]
  30. Bishop, C.M. Bayesian PCA. In Advances in Neural Information Processing Systems; The MIT Press: Cambridge, MA, USA, 1999; pp. 382–388. [Google Scholar]
  31. Klami, A.; Virtanen, S.; Kaski, S. Bayesian canonical correlation analysis. J. Mach. Learn. Res. 2013, 14, 965–1003. [Google Scholar]
  32. Bishop, C.M. Pattern recognition. Mach. Learn. 2006, 128, 1–39. [Google Scholar]
  33. Schölkopf, B.; Herbrich, R.; Smola, A.J. A generalized representer theorem. In Proceedings of the International Conference on Computational Learning Theory, Amsterdam, The Netherlands, 16–19 July 2001; pp. 416–426. [Google Scholar]
  34. Blei, D.M.; Kucukelbir, A.; McAuliffe, J.D. Variational Inference: A Review for Statisticians. J. Am. Stat. Assoc. 2017, 112, 859–877. [Google Scholar] [CrossRef] [Green Version]
  35. Rasmussen, C.E. Gaussian processes in machine learning. In Summer School on Machine Learning; Springer: Berlin/Heidelberg, Germany, 2003; pp. 63–71. [Google Scholar]
  36. Steinwart, I.; Christmann, A. Support Vector Machines; Springer Science & Business Media: New York, NY, USA, 2008. [Google Scholar]
  37. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-learn: Machine Learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  38. Sevilla-Salcedo, C.; Gómez-Verdejo, V.; Olmos, P.M. Sparse semi-supervised heterogeneous interbattery bayesian analysis. Pattern Recognit. 2021, 120, 108141. [Google Scholar] [CrossRef]
  39. Sevilla-Salcedo, C.; Guerrero-López, A.; Olmos, P.M.; Gómez-Verdejo, V. Bayesian Sparse Factor Analysis with Kernelized Observations. arXiv 2020, arXiv:2006.00968. [Google Scholar]
  40. Styner, M.; Lieberman, J.A.; McClure, R.K.; Weinberger, D.R.; Jones, D.W.; Gerig, G. Morphometric analysis of lateral ventricles in schizophrenia and healthy controls regarding genetic and disease-specific factors. Proc. Natl. Acad. Sci. USA 2005, 102, 4872–4877. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  41. Rapado-Castro, M.; Villar-Arenzana, M.; Janssen, J.; Fraguas, D.; Bombin, I.; Castro-Fornieles, J.; Mayoral, M.; González-Pinto, A.; de la Serna, E.; Parellada, M.; et al. Fronto-Parietal Gray Matter Volume Loss Is Associated with Decreased Working Memory Performance in Adolescents with a First Episode of Psychosis. J. Clin. Med. 2021, 10, 3929. [Google Scholar] [CrossRef]
  42. Wen, D.; Wang, J.; Yao, G.; Liu, S.; Li, X.; Li, J.; Li, H.; Xu, Y. Abnormality of subcortical volume and resting functional connectivity in adolescents with early-onset and prodromal schizophrenia. J. Psychiatr. Res. 2021, 140, 282–288. [Google Scholar] [CrossRef]
  43. Guo, S.; Kendrick, K.M.; Zhang, J.; Broome, M.; Yu, R.; Liu, Z.; Feng, J. Brain-wide functional inter-hemispheric disconnection is a potential biomarker for schizophrenia and distinguishes it from depression. Neuroimage Clin. 2013, 2, 818–826. [Google Scholar] [CrossRef] [Green Version]
  44. Boklage, C.E. Schizophrenia, brain asymmetry development, and twinning: Cellular relationship with etiological and possibly prognostic implications. Biol. Psychiatry 1977, 12, 19–35. [Google Scholar]
  45. Casquero-Veiga, M.; Romero-Miguel, D.; MacDowell, K.S.; Torres-Sanchez, S.; Garcia-Partida, J.A.; Lamanna-Rama, N.; Gómez-Rangel, V.; Romero-Miranda, A.; Berrocoso, E.; Leza, J.C.; et al. Omega-3 fatty acids during adolescence prevent schizophrenia-related behavioural deficits: Neurophysiological evidences from the prenatal viral infection with PolyI: C. Eur. Neuropsychopharmacol. 2021, 46, 14–27. [Google Scholar] [CrossRef]
  46. Bortz, D.M.; Grace, A.A. Medial septum activation produces opposite effects on dopamine neuron activity in the ventral tegmental area and substantia nigra in MAM vs. normal rats. NPJ Schizophr. 2018, 4, 17. [Google Scholar] [CrossRef] [Green Version]
  47. Takeuchi, Y.; Nagy, A.; Barcsai, L.; Li, Q.; Ohsawa, M.; Mizuseki, K.; Berényi, A. The medial septum as a potential target for treating brain disorders associated with oscillopathies. Front. Neural Circuits 2021, 15, 701080. [Google Scholar] [CrossRef] [PubMed]
  48. McGlinchey, E.M.; Aston-Jones, G. Dorsal hippocampus drives context-induced cocaine seeking via inputs to lateral septum. Neuropsychopharmacology 2018, 43, 987–1000. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  49. Pantazis, C.B.; Aston-Jones, G. Lateral septum inhibition reduces motivation for cocaine: Reversal by diazepam. Addict. Biol. 2020, 25, e12742. [Google Scholar] [CrossRef] [PubMed]
  50. Gárate-Pérez, M.F.; Méndez, A.; Bahamondes, C.; Sanhueza, C.; Guzmán, F.; Reyes-Parada, M.; Sotomayor-Zárate, R.; Renard, G.M. Vasopressin in the lateral septum decreases conditioned place preference to amphetamine and nucleus accumbens dopamine release. Addict. Biol. 2021, 26, e12851. [Google Scholar] [CrossRef] [PubMed]
  51. Yang, M.; Gao, S.; Zhang, X. Cognitive deficits and white matter abnormalities in never-treated first-episode schizophrenia. Transl. Psychiatry 2020, 10, 368. [Google Scholar] [CrossRef]
  52. Kim, S.E.; Jung, S.; Sung, G.; Bang, M.; Lee, S.H. Impaired cerebro-cerebellar white matter connectivity and its associations with cognitive function in patients with schizophrenia. NPJ Schizophr. 2021, 7, 38. [Google Scholar] [CrossRef]
Figure 1. MRI pipeline for data processing [28]. First, images were corrected for field homogeneity, resized by a factor of 10 and spatially normalized to create a custom template based on a Wistar rat brain template [29]. All images were resliced to this custom template and were segmented into gray matter (GM), white matter (WM), and cerebrospinal fluid (CSF). Later, all images were modulated using the Jacobian determinants and smoothed with a 10-mm FWHM Gaussian kernel. Finally, the segmented tissues were processed by the ML model to classify them into healthy and pathological subjects and to identify the brain areas relevant to this decision.
Figure 1. MRI pipeline for data processing [28]. First, images were corrected for field homogeneity, resized by a factor of 10 and spatially normalized to create a custom template based on a Wistar rat brain template [29]. All images were resliced to this custom template and were segmented into gray matter (GM), white matter (WM), and cerebrospinal fluid (CSF). Later, all images were modulated using the Jacobian determinants and smoothed with a 10-mm FWHM Gaussian kernel. Finally, the segmented tissues were processed by the ML model to classify them into healthy and pathological subjects and to identify the brain areas relevant to this decision.
Applsci 12 02571 g001
Figure 2. Plate diagram for the DBL-FS graphic model. Grey circles denote observed variables, white circles unobserved variables. Model hyperparameters do not have a circle.
Figure 2. Plate diagram for the DBL-FS graphic model. Grey circles denote observed variables, white circles unobserved variables. Model hyperparameters do not have a circle.
Applsci 12 02571 g002
Figure 3. Brain masks obtained by the FS of each model. Colors are defined in a linear scale and associated to the relevance of the voxel (white: more relevant; dark red: less relevant). GP-FS model yields meaningless results in neurobiological terms, where no anatomical regions can be identified. KSSHIBA-FS model only identifies brain areas related to WM deficits in schizophrenia. Both DBL-FS and DBL-FS+EK learn similar relevance in WM, GM, and CSF brain areas of interest in schizophrenia, such as the hippocampus (hipp), prefrontal cortex (PFC), amygdala (Aa), septum, lateral ventricles (LV), corpus callosum (cc), WM cerebellar (WM Cb), and WM brainstem (WM BS) fibers.
Figure 3. Brain masks obtained by the FS of each model. Colors are defined in a linear scale and associated to the relevance of the voxel (white: more relevant; dark red: less relevant). GP-FS model yields meaningless results in neurobiological terms, where no anatomical regions can be identified. KSSHIBA-FS model only identifies brain areas related to WM deficits in schizophrenia. Both DBL-FS and DBL-FS+EK learn similar relevance in WM, GM, and CSF brain areas of interest in schizophrenia, such as the hippocampus (hipp), prefrontal cortex (PFC), amygdala (Aa), septum, lateral ventricles (LV), corpus callosum (cc), WM cerebellar (WM Cb), and WM brainstem (WM BS) fibers.
Applsci 12 02571 g003
Table 1. Performance of the different methods under study showing the model accuracy and the number of selected voxels (with their percentages with respect to the total). In addition, models with the best accuracy have been highlighted in bold and placed at the bottom of the table, which corresponds to the proposed DBL-FS approaches.
Table 1. Performance of the different methods under study showing the model accuracy and the number of selected voxels (with their percentages with respect to the total). In addition, models with the best accuracy have been highlighted in bold and placed at the bottom of the table, which corresponds to the proposed DBL-FS approaches.
ExperimentAccuracy# Selected Voxels
GP67.9%1,077,656 (100%)
GP+FS67.9%269,414 (25%)
SVM71.6%1,077,656 (100%)
MKL-SVM67.9%1,077,656 (100%)
KSSHIBA69.8%1,077,656 (100%)
KSSHIBA+FS64.1%269,414 (25%)
DBL-FS77.3%287,996 (26.72%)
DBL-FS+EK77.3%242,754 (22.52%)
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Belenguer-Llorens, A.; Sevilla-Salcedo, C.; Desco, M.; Soto-Montenegro, M.L.; Gómez-Verdejo, V. A Novel Bayesian Linear Regression Model for the Analysis of Neuroimaging Data. Appl. Sci. 2022, 12, 2571. https://0-doi-org.brum.beds.ac.uk/10.3390/app12052571

AMA Style

Belenguer-Llorens A, Sevilla-Salcedo C, Desco M, Soto-Montenegro ML, Gómez-Verdejo V. A Novel Bayesian Linear Regression Model for the Analysis of Neuroimaging Data. Applied Sciences. 2022; 12(5):2571. https://0-doi-org.brum.beds.ac.uk/10.3390/app12052571

Chicago/Turabian Style

Belenguer-Llorens, Albert, Carlos Sevilla-Salcedo, Manuel Desco, Maria Luisa Soto-Montenegro, and Vanessa Gómez-Verdejo. 2022. "A Novel Bayesian Linear Regression Model for the Analysis of Neuroimaging Data" Applied Sciences 12, no. 5: 2571. https://0-doi-org.brum.beds.ac.uk/10.3390/app12052571

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop