Next Article in Journal
Strategies to Limit Cognitive Impairments under Sleep Restriction: Relationship to Stress Biomarkers
Previous Article in Journal
White Matter Microstructural Alterations in Newly Diagnosed Parkinson’s Disease: A Whole-Brain Analysis Using dMRI
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

fMRI Brain Decoding and Its Applications in Brain–Computer Interface: A Survey

1
School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing 100083, China
2
Department of Electronic Engineering, Tsinghua University, Beijing 100084, China
*
Author to whom correspondence should be addressed.
Brain Sci. 2022, 12(2), 228; https://doi.org/10.3390/brainsci12020228
Submission received: 20 December 2021 / Revised: 29 January 2022 / Accepted: 30 January 2022 / Published: 7 February 2022
(This article belongs to the Topic Human–Machine Interaction)

Abstract

:
Brain neural activity decoding is an important branch of neuroscience research and a key technology for the brain–computer interface (BCI). Researchers initially developed simple linear models and machine learning algorithms to classify and recognize brain activities. With the great success of deep learning on image recognition and generation, deep neural networks (DNN) have been engaged in reconstructing visual stimuli from human brain activity via functional magnetic resonance imaging (fMRI). In this paper, we reviewed the brain activity decoding models based on machine learning and deep learning algorithms. Specifically, we focused on current brain activity decoding models with high attention: variational auto-encoder (VAE), generative confrontation network (GAN), and the graph convolutional network (GCN). Furthermore, brain neural-activity-decoding-enabled fMRI-based BCI applications in mental and psychological disease treatment are presented to illustrate the positive correlation between brain decoding and BCI. Finally, existing challenges and future research directions are addressed.

1. Introduction

In recent years, the concept of the brain–computer interface has gradually entered the public’s field of vision and has become a hot topic in the field of brain research. Brain neural activity decoding is a key technology for the brain–computer interface. Therefore, this paper first surveys research in the field of visually decoding brain neuronal activity, explains the strengths and weaknesses of these studies, and updates recent research progress and potential clinical applications of brain–computer interfaces in psychotherapy. Finally, potential solutions are proposed for the problems existing in the current brain activity decoding models. Functional magnetic resonance imaging (fMRI) is a new neuroimaging method. Its principle is to use magnetic resonance imaging to measure the changes in hemodynamics caused by neuronal activity. From the perspective of neuroscience and neuroimaging, functional magnetic resonance imaging can be used to decode the perception and semantic information of the cerebral cortex in a non-invasive manner [1]. The spatial resolution of fMRI is high enough, while the noise contained in the measurement of the brain activity is relatively small [2]. The real-time fMRI monitors the activity state of the cerebral cortex through the blood-oxygen-level-dependent (BOLD) variation induced by brain neural activity and simultaneously collects and analyzes the BOLD signals of the brain [3,4]. In recent years, more and more researchers have adopted real-time fMRI to investigate the brain’s self-regulation and the connection among different brain regions [5,6,7]. Compared with Electroencephalogram (EEG) [2,8,9,10,11], fMRI has higher spatial resolution and suffers from lower noise. Therefore, decoding brain neural activity with fMRI data has become more popular [12,13,14,15,16]. Contrary to the visual encoding model [17,18,19,20], the decoding model predicts the visual stimuli through the neural responses in the brain. The current research on the decoding of visual stimuli can be divided into three categories: classification, recognition, and reconstruction.
The difficulty in brain decoding is the reconstruction of visual stimuli through the cognition model of the brain’s visual cortex, as well as learning algorithms [21,22,23,24]. The initial research on reconstructing visual stimuli used retinal shadows [25,26] or fMRI [12] to observe the response of the visual cortex and map a given visual stimuli to the brain neural activity linearly. These studies considered the reconstruction of the visual stimulus as a simple linear mapping problem between the monomer voxels and brain activities. Then, the researchers further used multi-voxel pattern analysis (MVPA) combined with the machine learning classification algorithms to decode the information from a large number of voxel BOLD signals [14,16,27,28,29]. For example, multi-voxel patterns were used to extract multi-scale information from fMRI signals to achieve a better performance of the reconstruction [29]. According to the sparsity of the human brain’s processing of external stimuli, ref. [14] added sparsity constraints to the multivariate analysis model of Bayesian networks to quantify the uncertainty of voxel features [16]. Because of the diversity of natural scenes and the limitations of recording neuronal activities in the brain, some studies try to include the a priori of natural images in multiple analysis models to reconstruct some simple images [23,30,31]. Using two different coding models to integrate the information of different visual areas in the cerebral cortex and based on a large number of image priors, ref. [31] reconstructed natural images from fMRI signals for the first time. In addition, the researchers also used machine learning [32] to decode the semantic information of human dreams [33] and facial images [34]. In order to trace the reversible mapping between visual images and corresponding brain activity, ref. [22] developed a Bayesian canonical correlation analysis (BCCA) model. However, its linear structure makes the model unable to present multi-level visual features, and its spherical covariance assumption cannot meet the correlation between fMRI voxels [21]. Ref. [24] introduced a Gaussian mixture model to perform percept decoding with the prior distribution of the image, which can infer high-level semantic categories from low-level image features by combining the prior distribution of different information sources [24]. Moreover, ref. [35] explored the impact of brain activity decoding in different cerebral cortex areas. Researchers believe that the high-level visual cortex, i.e., the ventral temporal cortex, is the key to decoding the semantic information and natural images from brain activity [31].
Significant achievements have been made in image restoration and image super-resolution with deep learning algorithms. The structure of deep neural networks (DNN) is similar to the feedforward of the human visual system [36], so it is not surprising that DNN can be used to decode the visual stimulus of the brain activity [30,37,38] mapped multi-level features of the human visual cortex to the hierarchical features of a pre-trained DNN, which can make use of the information from hierarchical visual features. Ref. [37] combined deep learning and probabilistic reasoning to reconstruct human facial images through nonlinear transformation from perceived stimuli to latent features with the adversarial training of convolutional neural networks.
Therefore, this paper surveys the deep neural networks used in visual stimulus decoding, including Convolutional Neural Networks (CNN), Recurrent Neural Networks (RNN), and Graph Convolutional Neural Networks (GCN). Deep neural networks, especially deep convolutional neural networks, have proven to be a powerful way to learn high-level and intermediate abstractions from low-level raw data [39]. Researchers used pre-trained CNN to extract features from fMRI data in the convolutional layer [28,38,40]. Furthermore, ref. [41] adopted an end-to-end approach and trained the generator directly using fMRI data. The feature extraction process of RNN is similar to that of CNN, but the process is not affected by the input time sequence [30]. The long short-term memory model (LSTM) is a typical RNN structure and is also used to decode brain activity [42,43,44]. However, RNN only considers the time series of the BOLD signal and ignores spatial correlations between different functional areas of the brain [30]. GCN can investigate the topological structure of the brain functional areas and then predict the cognitive state of the brain [45,46]. With limited fMRI and labeled image data, the GCN-based decoding model can provide an automated tool to derive the cognitive state of the brain [47].
Owing to the rapid adoption of deep generative models, especially Variational Auto-encoders (VAEs) and Generative Adversarial Networks (GANs), a large number of researchers have been encouraged to study non-natural image generation through VAEs and GANs [2,48].
VAE has a theoretically backed framework and is easy to train. Traditional methods usually separated encoding and decoding processes. The authors of [49] thought that the encoding and decoding models should not be mutually exclusive, and the prediction can be more accurate through an effective unified encoding and decoding framework. VAE is such a two-way model, which composes of an encoder, a decoder, and latent variables in between. VAE is trained to minimize the reconstruction error between the encoded–decoded data and the initial data [19,23,38]. Instead of taking an input as a single point, VAE takes the input as a distribution over the latent space [50] and naturally represents the latent space regularization. The encoder learns latent variables from the input, and the decoder generates an output based on samples of the latent variables. Given sufficient training data, the encoder and the decoder are trainable altogether by minimizing the reconstruction loss and the Kullback–Leibler (KL) divergence between the distributions of latent variables and independent normal distributions. However, the synthesized image derived from VAE lacks details [48]. At present, some researchers have tried to reconstruct the visual stimuli from the associated brain activities based on VAE [21,38,48,51].
GAN is widely used for the images synthesization [52,53] and image-to-image conversions [54,55]. The advantage of GAN is that the Markov chain is no longer needed, and the gradient can be obtained by back propagation. Various factors and interactions can be incorporated into the model during the learning process without inference [56]. GAN can synthesize relatively clear images, but the stability of training and the diversity of sampling are its major problems [48]. In brain decoding, researchers have successfully applied GAN-based models to decode natural images from human brain activity [57,58,59,60,61]. Although GAN can be used to generate very good images, the quality and variety of the images reconstructed from brain activity data are quite limited [62]. The significance of the generative models lies in the representation and manipulation of the high-dimensional probability distributions [63]. The generative model can be trained with lossy data (some samples are unlabeled) and be performed by semi-supervised learning, which reduces the difficulties of sampling the brain data [63].
Studying brain decoding intends to create brain–computer interfaces, which uses the brain neuron signals to control the device in real time [30], such as controlling the robotic arm or the voice generator. Another goal of studying brain neuron decoding is to better understand the cognitive structure of the brain by constructing a decoding model [57]. For people with mental illness, neuro-feedback therapy can help patients restore healthy cognition. The fMRI brain–computer interface (fMRI-BCI) collects signals in a single or multiple regions of interest (ROI) in the brain; researchers used a linear discriminant classifier [64] and a multi-layer neural network [14] to classify the scanned fMRI signals. Since the differences in mental activity and physical structure between people results in differences in the number and location of the active area of each person’s brain’s ROI, multivariate analysis is more appropriate for decoding brain activity [13,29,65]. The analysis of multi-voxel patterns and the development of neuro-imaging technology can be applied to the physical examination and diagnosis and rehabilitation training [13,65]. In order to decode the spiritual and psychological information in fMRI, ref. [66] used the Support Vector Machines (SVM) to classify and identify a variety of different emotional states. The participants’ brain state decoded from fMRI signals can be regarded as a BCI control signal, which provides the neuro-feedback to the participants [65]. However, many technical challenges still remain in fMRI brain decoding, i.e., the noise included in fMRI signals and the limited high-resolution fMRI signal data set [67].
After determining the topic of this survey, we first conduct a literature search. The database for the literature search is the Web of Science Core Collection, the keywords are fMRI, Brain decoding, BCI, and the time span is 1995 to 2020. Then, we performed literature screening. We imported the literature information retrieved in the previous step into CiteSpace (data mining software) for clustering analysis, and then, according to the high-frequency nodes (representing highly cited literature) in the clustering results and high betweenness centrality nodes (representing documents that form a co-citation relationship with multiple documents), we screened out 120 articles that met the requirements.
The rest of the paper is organized as follows. Section 2 introduces the basics of the brain activity decoding model. Section 3 introduces the hot research interests and specific examples in this area. Section 4 introduces the specific medical applications of fMRI-BCI. Section 5 analyzes the current challenges in using fMRI to decode brain activity and possible solutions. Section 6 is a summary of the survey.

2. Brain Decoding Based on Learning

Visual pathway is a typical illustration of brain encoding and decoding with regard to visual stimuli. Visual pathway processes visual information after the retina receiving a visual stimuli [68]. The optic nerve is responsible for transmitting the special sensory information for vision. The optic nerves from each eye unite to form the optic chiasm, then to the lateral geniculate body, which projects the visual information to the primary visual cortex. Figure 1 illustrates how the visual cortex of brain encodes the visual information and reflected by an fMRI. Additionally, the decoding process is the inverse of the encoding process. In this section, we elaborate fMRI brain decoding models based on machine learning.

2.1. The Relationship between Brain Encoding and Decoding

The encoding and decoding of brain neuron activity are the two fundamental aspects of human’s cognitive visual processing system [15,69]. The encoding process predicts brain activities based on visual stimuli aroused by the external environment [21]. Contrary to the encoding process, the decoding process analyzes brain activities to retrieve the associated visual stimuli [21]. The encoding process describes how to obtain information in voxels corresponding to one region, and once the encoding model is constructed, one can derive the decoding model using Bayesian derivation [15]. Moreover, decoding model can be applied to verify the integrity of encoding process [15]. The encoding model can be trained to predict the representations of neuronal activity, which in turn helps to extract features from the decoding model, thereby improving the decoding performance of brain activities [70]. Therefore, encoding and decoding models are not mutually exclusive because the prediction can be performed more accurately by an effective unified encoding and decoding framework. Theoretically, the general linear model (GLM) can be used to predict the voxel activity, so GLM can be regarded as an encoding model [49]. The review papers that survey brain activity encoding models are listed in Table 1.

2.2. Machine Learning and Deep Learning Preliminaries

We briefly describe the traditional machine learning and deep learning methods used in brain decoding models.

2.2.1. Machine Learning

The mapping of activities in the brain to cognitive states can be treated as a pattern recognition problem. In this regard, the fMRI data are a spatial pattern, and thus a statistical pattern recognition method (e.g., a machine learning algorithm) can be used to map these spatial patterns to instantaneous brain states. The machine learning algorithms are used to learn and classify multivariate data based on the statistical properties of the fMRI data set. Surveys on brain decoding models based on machine learning are listed in Table 2. In the surveyed literature, we summarized the traditional machine learning methods applied to brain decoding as follows:
(1) Support vector machine (SVM) is a linear classifier, which aims to find the largest margin hyperplane to classify the data [64]. The distance between the data of different categories is maximized through maximal margin. SVM can be extended to act as a nonlinear classifier by employing different kernel functions. However, fMRI data are highly dimensional with few data points, hence a linear kernel is sufficient to solve the problem.
(2) Naive Bayesian classifier (NBC) is a probabilistic classifier, which is based on Bayes theorem and strong independent assumptions between features. NBC is widely used in brain decoding models. When processing fMRI data, NBC assumes that each voxel’s fMRI is independent. For instance, the variance/covariance of the neuronal and hemodynamic parameters are estimated on a voxel-by-voxel basis using a Bayesian estimation algorithm, enabling population receptive fields (pRF) to be plotted while properly representing uncertainty about pRF size and location based on fMRI data [72]. In addition, NBC is often combined with other methods, such as Gaussian mixture models and convolutional neural networks.
(3) Principal Component Analysis (PCA) maps data to a new coordinate system through orthogonal linear transformation [73]. When processing high-dimensional fMRI data, PCA is used as a dimensionality reduction tool to remove noise and irrelevant features, thereby improving the data processing speed. In addition, the risk of over-fitting is avoided due to the reduction in training data in order to improve the generalization ability of the model [40].

2.2.2. Deep Learning

The hierarchical structure of DNN is similar to the feed-forward of the human visual system, and DNN can be used to encode external stimuli [36]. In particular, DNN is used to explore how the brain maps the complex external stimulus to corresponding areas in the cerebral cortex, which quantitatively shows that an explicit gradient for feature complexity exists in the ventral pathway of the human brain [74]. In the surveyed literature, DNN models are summarized as follows:
(1) Recurrent Neural Network (RNN) is a generalization of a feedforward neural network that has an internal memory. The hidden nodes of a RNN form a cyclic structure, and the state of the hidden layer at each stage depends on its past states. This structure allows RNN to save, memorize and process complex signals in the past for a long time. RNN can process variable length sequences of inputs, which occurs in neuroscience data [30]. The main advantage of using RNNs instead of standard neural networks is that the weights are shared between neurons in hidden layers across time. Therefore, RNN can process the previously input sequence, as well as the upcoming sequence to explore the temporal dynamic behavior. Therefore, RNNs can be applied to decode brain activity since they can process correlated data across time [57].
(2) Convolutional Neural Network (CNN) consists of three parts: convolution layers, pooling layers, and fully connected layers. CNNs can handle many different input and output formats: 1D sequences, 2D images, and 3D volumes. The most important task of the CNN is to iteratively calibrate the network weights through the training data, also known as the back propagation algorithm. For decoding different neurons activities, it can be realized via weights sharing across convolutional layers for learning a shared set of data-driven features [30]. Moreover, the generative models VAE and GAN based on CNN have attracted much attention recently. The deep generative models for brain decoding are listed in Table 3.
(3) Graph Convolutional Network (GCN): At present, most deep-learning-based brain decoding research has not considered the function correlations and dynamic temporal information between human brain regions [45]. For brain decoding, GCN can consider the connections among brain regions when performing brain decoding via fMRI [77]. Unlike CNN, GCN extracts features effectively, even with completely random initialized parameters [78]. Ref. [46] extracted fMRI data features with GCN to predict brain states. Given labeled data, compared to the other classifiers, GCN performs better on a limited data set. In this survey, GCN-based brain decoding models are listed in Table 4.

3. Brain Decoding Based on Deep Learning

In this section, the principles and specific applications of VAE, GAN, and GCN that have received the most attention in brain decoding are introduced.

3.1. VAE-Based Brain Decoding

VAE is a generative model that contains hidden variables, and it combines variational Bayesian estimation with a neural network, which learns variational deduction parameters and thus obtains the optimized inference through maximum likelihood (ML) or Max a posteriori (MAP) [32,79]. Let x represent the input image and z represent the latent variables. Latent variables can be considered as dependencies between features in multiple dimensions. Latent variables reconstruct the data x through a generative network, while generates new data that do not exist in the original data. For the image data set, latent variables z are the implicit factors that determine the features of an image x . The inference network in VAE infers hidden variables z . The process of variational inference is also a process of dimensionality reduction [79]. The generative network in VAE uses the randomly sampled latent variables z to generate an image x [32].
ϕ is an inferred model that x infers z; θ is a generative model where z generates x. The marginal likelihood of an input image x can be written as:
l o g p θ ( x ) = D K L [ q ϕ ( z | x ) p θ ( x | z ) ] + L ( θ , ϕ ; x ) .
Given x, the probability that ϕ infers z is q ϕ ( z | x ) ; given z, the probability that θ generates x is p θ ( x | z ) . The first term is a non-negative Kullback–Leibler (KL) difference, so L ( θ , ϕ ; x ) can be considered as the lower bound of Equation (1). The learning rules for VAE are to maximize Equation (1), which is equivalent to maximizing L ( θ , ϕ ; x ) , as shown in Equation (2):
L ( θ , ϕ ; x ) = D K L [ q ϕ ( z | x ) p θ ( z ) ] + E z q ϕ ( z | x ) [ log ( p θ ( x | z ) ) ]
In Equation (2), the first term is the Kullback–Leibler (KL) difference between the distribution of latent variable z inferred from x and the prior distribution of z; the second term is the expectation of the log-likelihood that the input image x is generated by z sampled from the inferred distribution q ϕ ( z | x ) . If q ϕ ( z | x ) follows a multivariate distribution with unknown expectation μ and variance σ , the objective function is differentiable with respect to ( θ , ϕ , μ , σ ) [32]. The parameters of VAE can be optimized by stochastic gradient descent [75]. The brain learns from experience in an unsupervised way. Likewise, VAE uses deep neural networks to learn representations from large amounts of data in such an unsupervised way as the brain does [50]. Furthermore, VAE-based brain decoding models can be found in [50,51,67].
Du et al. [51] proposed a structured deep generative neural decoding model, which consists of a structured multiple-output regression model (SMR) and an introspective conditional generation (ICG) model, to reconstruct visual images from brain activity, as shown in Figure 2. This conditional deep generative model (DGM) combines the advantages of VAE and GAN to reconstruct high-quality images with stable training. Specifically, SMR calculates the correlation between fMRI voxels on the CNN units, as well as the correlation between the output of each linear model. Then, the ICG model is proposed to decode these correlated CNN features extracted by SMR, inspired by the IntroVAE [48].
The key point of ICG is the IntroVAE that modifies the LOSS function of VAE and allows the encoder to perform a discriminator’s function in a GAN without introducing a new network. The ICG model can estimate the difference between the generated image and the training data in an introspective way. Similar to the original VAE, the training of ICG is to optimize encoder ϕ and decoder θ iteratively until convergence:
θ ^ = arg min θ [ L A E + α D K L ( q ϕ ( z | x f , h ) p ( z ) ) ] ,
ϕ ^ = arg min ϕ [ L A E + β D K L ( q ϕ ( z | x , h ) p ( z ) ) α D K L ( q ϕ ( z | x f , h ) p ( z ) ) ] .
where h represents the output CNN features from SMR, x represents the input training image, z represents the hidden variable, and z p ( z ) or z q ϕ ( z | x , h ) . x f is a fake image generated by p θ ( x | z , h ) . L A E is the error of VAE reconstruction. For real image data points, Equations (3) and (4) cooperate with each other rather than conflict with each other, which can be considered as a conditional variational autoencoder (CVAE) [80]. In this case, D K L ( q ϕ ( z | x f , h p ( z ) ) in Equations (3) and (4) does not work. The encoder and decoder cooperates to minimize the reconstruction loss L A E . Equation (4) regularizes the encoder by matching q ϕ ( z | x , h ) with p ( z ) . For fake image data points, Equations (3) and (4) confront with each other, which can be considered as a conditional generative adversarial network (CGAN) [56]. From Equations (3) and (4), we can see that D K L ( q ϕ ( z | x f , h ) p ( z ) ) are mutually exclusive during the training process. Specifically, Equation (3) hopes to match posterior q ϕ ( z | x f , h ) with prior distribution p ( z ) to minimize D K L ( q ϕ ( z | x f , h ) p ( z ) ) , while Equation (4) hopes to match posterior q ϕ ( z | x f , h ) with prior distribution p ( z ) to maximize D K L . Generally, Equation (3) (generative model) tends to generate a realistic image as much as possible so that Equation (4) (discrimination model) cannot discriminate its authenticity. The α and β in Equations (3) and (4) are parameters, which balance CVAE and CGAN.

3.2. GAN-Based Brain Decoding

GAN consists of a generative network and a discriminant network, and tends to generate high-quality pictures through adversarial training of the discriminator [81]. Specifically, the discriminator distinguishes whether the sample comes from the generated model or the probabilistic model of the training data. The generator is trained to generate realistic pictures for maximizing the error rate of the discriminator. The two models can theoretically achieve the Nash equilibrium (the probability out of the discriminator is 0.5 ) for creating a remarkably nature-like picture. Let D and G denote the discriminator and generator, respectively. The optimization objective function of GAN can be expressed by Equation (5):
min G max D ( D , G ) = E x p d a t a ( x ) [ log D ( x ) ] + E n p n ( n ) [ log ( 1 D ( G ( n ) ) ) ] .
The distribution of the real image data x is p d a t a ( x ) , and the distribution of the artificial noise variable n is p n ( n ) . The input variable of G is n, which outputs the distribution of the real data. The input of D is the real data x, which outputs a scalar D ( x ) , a probability of the data being real. The objective function maximizes D ( x ) by training D and minimizes log ( 1 D ( G ( n ) ) by training G. Because G is a differentiable function, the gradient-based optimization algorithm can be used to obtain a global optimal solution [82].
A shape-Semantic GAN is proposed in [76], which consists of a linear shape decoder, a semantic decoder based on DNN, and an image generator based on GAN. The fMRI signals are input into the shape decoder and semantic decoder. The shape decoder reconstructs the contour of the visual image from the lower visual cortex (V1, V2, V3), and the semantic decoder is responsible for extracting the semantic features of the visual image in the higher visual cortex (FFA, LOC, PPA). The output of the shape decoder is input to the GAN-based image generator, and the semantic features in GAN can compensate the details for the image to reconstruct high-quality images.
The linear shape decoder trains models for the activities of V1, V2, and V3 viewed from fMRI, and combines the predicted results linearly to reconstruct the shape of the natural image. The combined training model of the shape decoder is shown in Equation (6):
r s p ( i , j ) = w i j k p k * ( i , j ) , k = V 1 , V 2 , V 3 ,
where w i j k represents the weight of the image pixel at position ( i , j ) , and p k * ( i , j ) represents the predicted image pixel at position ( i , j ) . It can be seen from Equation (6) that the decoded image shape r s p is a linearly combination of image pixels p k * ( i , j ) . The semantic decoder, consisting of an input layer, two intermediate layers, and an output layer, is a lightweight DNN model. In the training phase, the high-level visual cortex activity recorded by fMRI is the input. The Tanh activation function is used to extract semantic features in the middle layer, and the sigmoid activation function is used to classify the images in the output layer.
In the final image generation stage, the original GAN can extract high-level semantic features. However, some low-level features (such as texture) may be lost in the process of encoding/decoding, which may cause the reconstructed image deformed and blurred. Therefore, U-Net [83] can be used as the generator, which consists of a pair of a symmetrical encoder and decoder. The U-Net breaks through the bottleneck of the structure. The U-Net does not require the low-level features necessarily to reconstruct the image, but it can extract the high-level features of the image. Moreover, the semantic features can be input to the U-Net decoder, which is equivalent to optimize a GAN generator with the constraints of the semantic features and shape features. In particular, the output of the shape decoder together with the output of the U-Net are input to the discriminator, which compares the high-frequency structures of the two and thus guides the generator training. Let G s represent a U-Net generator, and D t represent a discriminator. The conditional GAN model is shown in Equation (7):
L ( s , t ) = L a d v ( s , t ) + λ i m g L i m g ( s ) ,
where s and t represent the parameters of the generator and discriminator, respectively. L a d v ( s , t ) represents the adversarial loss, and L i m g ( s ) and λ i m g represent the image loss and weight, respectively. The conditional GAN discriminator is designed to model both low-frequency structures and high-frequency structures, thereby guiding the generator to synthesize images with more details.

3.3. Graph Convolutional Neural Networks

Unlike CNN and RNN, Graph Neural Networks (GNN) decode neuronal activity with brain connectivity [77]. Recently, some researchers have used Graph Convolutional Neural Networks (GCN) to predict [46] and annotate [47] brain consciousness and then predict the gender and age of the subjects [45]. When the training data set is quite limited, GCN shows better decoding performance than other classifiers.
GCN is first proposed in [78]. In the considered fMRI data, there are N nodes, and each node has d-dimensional features. The nodes’ features form an N × d matrix X. The relationships between the nodes form an N × N matrix A, known as an adjacency matrix. X and A are the inputs to the model. The propagation of GCN feature layers can be shown in Equation (8):
f ( H ( l ) , A ) = σ ( Q ^ 1 2 A ^ Q ^ 1 2 H ( l ) W ( l ) ) ,
where l is an identity matrix; H represents the feature matrix of each layer; W represents the initialized parameter matrix; and σ represents a nonlinear activation function. For the input layer, X is H. The diagonal elements of the relationship matrix A is 0, that is, when performing inner product with feature matrix H, the node’s own features will be ignored, which can be solved by A ^ = A + I . The key of the GCN is the symmetric normalized Laplacian matrix Q ^ 1 2 A ^ Q ^ 1 2 in Equation (8), where the degree matrix Q ^ of A ^ can be directly calculated. Therefore, when A ^ is multiplied by the feature matrix H, the distribution of the features will not be changed.
However, the methods based on GCN [47] and RNN [43,84] ignore the details of the fMRI signals and the dependence between different brain functional areas. Inspired by the work of spatio-temporal graph convolutional network (ST-GCN) developed in [85,86] to predict the graph structure, ref. [45] applied ST-GCN to brain decoding, where the temporal graph explores the dynamics brain activity, and the spatial graph explores the functional connectivity between different brain regions. The input of ST-GCN extracts d-dimensional spatio-temporal features u from N ROI regions of the brain, thus, u R N × d . ST-GCN is composed of three spatio-temporal convolution (ST-GC) layers. The time kernel size of each ST-GC layer is Γ = 11 ; the step size is 1; and the drop rate is 0.5. Each ST-GC layer outputs 64 channels. An edge importance matrix M R N × N is added between ST-GC layers. Specifically, when a certain node is spatially convoluted, the contribution of its adjacent nodes will be redefined by the row of the edge importance matrix where the node is located. Then, the 64-dimensional feature vector is generated through global average pooling and fed into the fully connected layer. The sigmoid activation function is then used to activate before the loss function calculated the classification probability. After back propagation, the stochastic gradient descent with a learning rate of 0.001 is used to optimize the weights in the ST-GC layer. Based on the ST-GCN model, the representation extracted from the fMRI signal can not only express the variable temporal brain activity but can also express the functional dependence between brain regions. The experimental results show that ST-GCN can predict the age and gender of a subject according to the BOLD signals. Although ST-GCN can consider both the temporal and spatial resolution of the fMRI signals, they are rarely used in brain decoding [30].

4. fMRI-BCI Application to Psychopsychiatric Treatment

It has greatly promoted the development of fMRI-BCI that rt-fMRI has been widely used to collect, analyze, and visualize the BOLD brain signals [3]. The fMRI-BCI outputs the real-time response of the target brain area to the devices to learn the brain areas’ regulations [87]. Different from EEG-BCI, fMRI-BCI can obtain both the temporal and spatial resolution of the brain neuron activity appropriately, then perform data analysis. Moreover, fMRI-BCI can calculate the correlation coefficients between voxels in the regions of interest (ROI) of the brain, as well as identify mental and psychological states [1]. The fMRI-BCI is a closed-loop system composed of four components: participants, signal acquisition, signal analysis, and signal feedback. In the signal acquisition stage, fMRI signals of subjects’ brain activity are obtained by a magnetic resonance scanning machine. The visual signal components perform fMRI data retrieval, correction processing, and statistical analysis, and the ROI that needs to be selected on the function diagram will be labeled. Afterwards, the ROI time series output of the visual signal component is imported into the signal feedback component, which provides a real-time feedback to the subjects in a certain way [87,88]. Since fMRI-BCI is able to target the diseased cerebral cortex, it can be used to treat certain physical and physiological diseases, such as: rehabilitation, mood disorders, and attention deficit and hyperactivity disorder (ADHD). Although Hinterberger et al. have proven that visual feedback is the best option to predict brain activity patterns, the BCI performance will not be affected by the type of feedback, whether it is touch or hearing [89,90].

4.1. Stroke Rehabilitation

For stroke and epilepsy patients, fMRI-BCI is a potential clinical application in rehabilitation treatment to recover patients’ motor and cognitive functions [87,91]. fMRI-BCI uses neurofeedback (NF) to learn the self-regulation ability of brain regions. The fMRI-BCI will transfer to the EEG-BCI [92,93], which is more flexible and lower cost. The EEG-BCI takes the patient’s EEG as the input to recover patients’ motion ability through real-time brain ROI’s feedback [9,11,94]. In particular, Pichiorri et al. combined BCI with Motor Imagery (MI) to help stroke rehabilitation using sensorimotor BCIs [95]. This combination allow one to control external devices through direct brain activity recognition by a computer, bypassing neuromuscular-based systems (voice, use of a mouse, or keyboard). Specifically, MI evokes an event-related desynchronization (i.e., a reduction in spectral power) that occurs within a certain EEG frequency band and primarily in sensorimotor cortical regions contralateral to the imagined part of the body. EEG signals of the two groups of subjects are recorded and analyzed with power spectral density, i.e., stronger desynchronization in the alpha and beta bands. The results showed the group of the subjects with BCI supported MI had more changes in EEG sensorimotor power spectra, occurring with greater involvement of the ipsilesional hemisphere, in response to MI of the paralyzed trained hand. Through functional measurements for stroke recovery, such as high-density EEG and single-pulse transcranial magnetic stimulation (TMS), visual guidance, motion-oriented, event-related synchronization, or desynchronization feature signals are analyzed as rehabilitation exercise control signals. In addition, the connection between post-stroke cerebral hemispheres in a certain frequency band is obtained by partial directed coherence. FMRI-BCI provides patients and therapists with a means to monitor and control MI, and through rich visual feedback consistent with the image content, it can promote their adherence to brain exercises and help patients restore motor function.

4.2. Chronic Pain Treatment

Many patients suffering from chronic pain felt that the original pain had not disappeared even after the disease was cured. This might be attributed to the function of memory. Once the memory nerve channel is formed, it is difficult to erase. Therefore, a psychological treatment plan must be designed to cope with the pain [96]. Rostral Anterior Cingulate Cortex (rACC) is the brain area responsible for the pain perception and regulation. The subjects can thus control the rACC activation through fMRI feedback learning [97]. Through fMRI feedback training, the subjects adjust the activity of voxels in rACC, and the degree of pain will also vary with the activation of the cerebral cortex. The experiments showed that when there was no fMRI feedback training, the pain of the subjects did not decrease. Guan et al. set up the comparative experiments to verify the adjustment effect of fMRI-BCI therapy on pain [98]. They divided the subjects into a testing group and a reference group. In the testing group, the subjects adjusted the activation of rACC through neurofeedback training, and the subjects in the reference group received neurofeedback in posterior cingulate cortex (PCC). The experimental results showed that the subjects in the neurofeedback training of the rACC area had reduced pain. Therefore, it can be concluded that neurofeedback training based on fMRI-BCI may be an effective way to relieve chronic clinical pain.

4.3. Emotional Disorders Treatment

Whether in studying, living, or working, one must master the ability to regulate emotions and respond to setbacks and failures with a positive attitude; otherwise, emotional disorders may occur. In [99], the participants can regulate BOLD-magnitude in the right anterior insular cortex. Two different control conditions of non-specific feedback and mental imagery were set up to confirm that rtfMRI feedback is area-specific. The experiment of [100] once again verified this conclusion. Twenty-seven participants were divided into three groups, equally, and the three groups received specific feedback, non-specific feedback, and no feedback, respectively, of BOLD in the insular area. After each group received several fMRI training sessions, each group evaluated the disgusting images. The results of the experiment showed that the participants who received the BOLD feedback in the insular area had the highest scores on the disgusting images, while the other two groups did not change significantly, so the mood did not change much. The insular area plays an important role in mood regulation and may be used in the treatment of clinical emotional disorders. In addition to the insular area related to emotion regulation, it has been proved that through behavioral tests, the activity of the amygdala in the brain is closely related to the generation, recognition, and regulation of emotions and can autonomously regulate the BOLD activity in the amygdala through fMRI neurofeedback learning [101,102]. In recent years, a connectivity-neurofeedback method based on fMRI has been proposed in [103]. After neurofeedback training, participants can automatically adjust the connectivity between the amygdala and other areas of the brain to enhance their ability of emotion regulating. Unlike the aforementioned reduction in negative emotions, the research in [103] aims to enhance positive emotions. Furthermore, fMRI neurofeedback is used to treat eight patients with major depression [104], which confirmed that fMRI-based neurofeedback was an effective aid to the current treatment of depression.

4.4. Criminal Psychotherapy

Although more and more researchers have achieved many promising results in the treatment of mental illness [103,104], however, in modern society, patients with criminal psychology have a much higher crime rates than mental patients [105]. The current neurobiology field cannot explain the neural mechanism of this disease [87]. It is believed that regulating the emotion-related areas (such as the amygdala) may cause changes in the patient’s mental state [106,107]. The development of the new fMRI-BCI to enable criminal mental patients to self-regulate BOLD activity in their cerebral cortex requires the joint efforts of neuroscience, psychology, and computer science.

5. Future Directions and Challenges

The challenges to brain decoding can be summarized in three aspects: 1. the ability of the mapping model between brain activity and visual stimuli; 2. not enough matching data between visual stimuli and brain activity; 3. fMRI signals interfered by noise [108]. Although previous studies on decoding visual stimuli of brain activity have made great achievements in classification and recognition, the performance of image reconstruction needs to be improved [21]. The following sections will indicate possible solutions and future development directions for the challenges.

5.1. Mapping Model Capabilities

5.1.1. Multi-Analysis Mode and Deep Learning

In recent years, the combination of multi-voxel analysis mode and deep learning has drawn much attention to identify brain states. By using deep neural networks, fMRI signals can be decoded to reveal brain activities. The performance of brain decoding has been further improved [30,47]. Although a decoding model based on MVPA has been proposed [65], the multi-voxel-based decoding model presents poor interpretability, especially when the decoding model uses linear kernels. In addition, this technique is susceptible to image artifacts, such as eye movement and cardiopulmonary artifacts. In addition, the speed of neuron vascular coupling, the sensitivity of BOLD activity, and the signal-to-noise ratio of fMRI signals should also be considered. In addition to the efficiency of the algorithm and the processing speed of the hardware, we should also consider the blood coupling delay in the brain [66]. The existing deep-learning-based decoding models [49] have achieved satisfactory results. Furthermore, ref. [109] correlate fMRI activity with multiple modalities, both visual and semantic features extracted from the viewed images. However, in order to obtain a higher-precision decoding model, there are still many challenges to reconstructing the corresponding visual stimuli from fMRI data by deep learning algorithms.

5.1.2. ROI and Feature Selection

The sample size of the fMRI signal and image pairing is small, while the dimensionality of the fMRI signal is higher. When the model is trained with limited high-dimensional data samples, it is easy to produce dimensionality [65]. The traditional methods are in danger of overfitting on small data sets [49]. The efficiency of deep-learning-based models depends on the number and reliability of training samples. Therefore, a large number of neural activities and corresponding images can improve the quality of the reconstructed image [77]. However, the running time of the experiment should be proportional to the efficiency, so it is particularly important to select the key features that contribute most to the image reconstruction and further to improve the feature extraction ability of the decoding model for neuroimaging data [30,62]. The visual attention [110] and axiomatic attribution [111] methods in computer vision can be utilized to determine which voxels of neurons contribute most to decoding visual stimuli.
In addition, the connection of the brain network of human cognition has become one of the major interests in neuroscience research [47]. The current decoding of brain activity is usually limited to the specific cognitive areas that humans can understand, and it takes a relatively long time to collect and record fMRI signals of brain activity in these areas. Moreover, most of the current deep learning-based research cannot simultaneously consider the functional dependence and time-variant information between different regions of the brain [45]. In order to use the dependency between the ROI regions of the brain to decode the brain activity, GCN is explored to predict or annotate the cognitive state of the brain [46,47]. In particular, based on the ST-GCN model, the representation extracted from the fMRI signal contains both the temporal information of brain activity and the functional dependence between the brain regions. This method of integrating edge importance and spatio-temporal map may have potential effects on the development of neuroscience [45].

5.1.3. Unsupervised Learning and Prior Knowledge

To learn an abstract representation of the brain activity in an unsupervised way should be fully studied in the future. Researchers’ exploration of the unsupervised learning methods led to the emergence of bidirectional generative models. For example, VAE is an unsupervised learning model. However, in the design of the VAE computing components, the encoder and the decoder are not related, but in the activities of the cerebral cortex, the feedforward and feedback processes are always related [50]. In addition, VAE does not have the ability to process dynamically and cyclically, but video information can be transmitted in both temporal and spatial space [50]. Reconstructing the dynamic features from brain activity is a huge challenge [21]. Some researchers used a large number of image priors to reconstruct visual stimuli [23]. When there is a priori knowledge, the decoder is a function of both brain neuron activity and prior knowledge. It is difficult to determine which information of the brain is decoded [30]. Therefore, the prior knowledge used in brain decoding still requires constant investigation of the researchers [31]. In recent years, an encoding model based on deep learning has emerged, which trains deep neural networks to perform representation learning that can predict neuronal activity [70]. Specifically, these deep learning based encoding models used visual stimuli to predict the neural response of the brain and served as a priori knowledge for the decoding model. The advancement of brain encoding has practical significance for brain enhanced communication, brain controlled machines, and disease state monitoring and diagnosis [31]. In summary, complementary model of encoding and decoding is a promising direction [30].

5.2. Limited fMRI and Image Data

5.2.1. Few-Shot Learning

Due to the high cost of fMRI research and the complicated research process, the collection of paired fMRI signals and image samples is almost an mission impossible, so the data amount is quite small [65,77,87]. Inspired by the field of computer vision, the few-shot learning for brain activity decoding is proposed in [77], which is promising in solving the data problem of neuroimaging. There are three ways to learn few-shot: the representation-based paradigm, the initialization-based paradigm, and the illusion-based paradigm.
The Representation-based paradigm aims to learn the representation of fMRI signals, which regards the first layer of the neural network as a feature extractor and the last layer as a classifier. A large amount of training data is used to train the neural network to extract relevant hidden representations and complete the training of the classifier. Later, when a limited number of training examples are available, the classifier extracts a small amount of data characterization and complete the classification of the new data.
The Initialization-based paradigm is also called meta-learning, the idea of which is to learn how to learn. This method aims to learn good initialization parameters so that the model can cope with various new data sets. In the process of meta-learning, the previous neural network can be understood as a low-level neural network, and the meta-learner is used to optimize the weights of the low-level neural network. The meta-learner inputs a list of samples and their corresponding labels. When training the meta-learner, the meta-loss (the error between the prediction and the target label) can be used to measure the performance of the meta-learner on the target task. Then, another meta-learner is needed to update the weight of the current meta-learner.
The Illusion-based paradigm is to perform a series of deformation operations such as rotation or combination of the samples in the original data set to increase training examples.

5.2.2. Transfer Learning

When the amount of data is limited and the prior knowledge is sufficient, sometimes the functions designed by hand are better than the neural network models learned from the data [30]. As the data amount in the fMRI data set continues to increase, manually designed functions will be replaced by the data-driven methods in the future. However, in neuroimaging, there is always a lack of data sets with large enough samples for specific experiments [112]. Most of the current transfer learning is to learn the data representation of the image in ImageNet and then build a model to adjust the medical image [73] or classify behavioral tasks from brain signals [113]. Although transfer learning can effectively make up for insufficient training data, natural images and medical images are quite different in nature [112]. For example, Gabor filters are often used for edge detection of natural images but have never been used in medical images. More and more studies have shown that the human cognitive system is a function of multiple functional areas of the brain [114]. Graph convolutional network models were trained with a a large number of medical imaging data sets in different experimental tasks and environments [112]. The experiments showed that the representation of the brain dynamics could be transferable between different ROI regions and different cognitive domains and even between different scanning sequences. Through fine-tuning, more can be learned about the high-level representation of brain functional areas, while preserving the low-level representation of the brain dynamics [47], which proved that transfer learning can not only improve decoding performance but also play a potential role in neuroimaging.

5.2.3. Graph Convolutional Networks

Ref. [115] used spatiotemporal convolution neural network to jointly extract spatiotemporal features from the target network. However, if CNNs are not trained, no effective features can be obtained at all. If GCNs are not trained and use completely randomly initialized parameters, the features extracted by GCN are still effective. If labeled information is given, GCNs will perform even better [78]. Compared with other classifiers, GCNs have better performance on a limited data set [77]. The fMRI signal can represent the spatial structure of brain activity, and GCNs can consider the connectivity of the brain to perform decoding, which has the potential to solve the problem of limited data [47,77,116,117].

5.3. fMRI Noise

5.3.1. Hemodynamic Delay

The spatial resolution of fMRI is very high, but its time resolution is relatively limited. It can only collect the average activities in about two seconds, and there is a certain delay in the detection of neural activity [3]. The fMRI signal contains the position information in the brain voxels, but due to its limited time resolution, sometimes the time series cannot be used to decode brain activity [30]. Because of the neurovascular coupling, the fMRI response is delayed after the neurological response [50]. Therefore, when fMRI signals are decoded into the latent variables of visual stimulation, the delay of the neurovascular dynamics should also be considered [50].

5.3.2. Brain Cognitive Limitation

FMRI-based BCI learns the self-regulation ability of ROI in the way of neural feedback. Due to the high cost of fMRI research and the complex research process, fMRI-based BCI can be transferred to the more flexible and lower cost EEG-BCI [92,93]. VAE and GAN can be combined to use fMRI data as the supplement of the EEG data, and then encode conditional vectors with less noise [2]. In addition to decoding low-level visual stimuli, researchers [118,119,120] decode the brain activity into low-level pixel space and high-level semantic space at the same time. Due to the inadequacy of human research on visual mechanisms, the current reconstruction methods are under exploration. In the reconstruction process, the decoded noise may be the true response of the brain’s visual cortex to the outside world, while the reconstructed clear image may be the noise [59].

6. Conclusions

This survey investigates brain activity decoding models based on machine learning and deep learning algorithms via fMRI. The relationship between the brain activity decoding model and the brain–computer interface is closely related to the development of the brain activity decoding model and promotes the development of fMRI-BCI. Furthermore, the specific application of fMRI-BCI in the treatments of mental and psychological diseases have been investigated. Finally, it outlines the current challenges in reconstructing visual stimuli from brain activity and potential solutions. With the advancement of brain signal measurement technology, the development of more complex encoding and decoding models, and better understanding of the brain structure, “mind reading” will become true soon.

Author Contributions

B.D., writing—review and editing, methodology, funding acquisition; X.C., writing, methodology; Y.D., Conceptualization, supervision; H.N., Conceptualization, supervision. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Nature Science Foundation of China U1633121, University of Science and Technology Course Fund KC2019SZ11 and 54896055.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Haynes, J.D.; Rees, G. Decoding mental states from brain activity in humans. Nat. Rev. Neurosci. 2006, 7, 523–534. [Google Scholar] [CrossRef] [PubMed]
  2. Kavasidis, I.; Palazzo, S.; Spampinato, C.; Giordano, D.; Shah, M. Brain2image: Converting brain signals into images. In Proceedings of the 25th ACM international conference on Multimedia, Mountain View, CA, USA, 23–27 October 2017; pp. 1809–1817. [Google Scholar]
  3. Cox, R.W.; Jesmanowicz, A.; Hyde, J.S. Real-time functional magnetic resonance imaging. Magn. Reson. Med. 1995, 33, 230–236. [Google Scholar] [CrossRef]
  4. Logothetis, N.K. The underpinnings of the BOLD functional magnetic resonance imaging signal. J. Neurosci. 2003, 23, 3963–3971. [Google Scholar] [CrossRef] [Green Version]
  5. Sulzer, J.; Haller, S.; Scharnowski, F.; Weiskopf, N.; Birbaumer, N.; Blefari, M.L.; Bruehl, A.B.; Cohen, L.G.; DeCharms, R.C.; Gassert, R.; et al. Real-time fMRI neurofeedback: Progress and challenges. Neuroimage 2013, 76, 386–399. [Google Scholar] [CrossRef] [Green Version]
  6. Thibault, R.T.; Lifshitz, M.; Raz, A. The climate of neurofeedback: scientific rigour and the perils of ideology. Brain 2018, 141, e11. [Google Scholar] [CrossRef] [Green Version]
  7. Yu, S.; Zheng, N.; Ma, Y.; Wu, H.; Chen, B. A novel brain decoding method: A correlation network framework for revealing brain connections. IEEE Trans. Cogn. Dev. Syst. 2018, 11, 95–106. [Google Scholar] [CrossRef] [Green Version]
  8. Arns, M.; Batail, J.M.; Bioulac, S.; Congedo, M.; Daudet, C.; Drapier, D.; Fovet, T.; Jardri, R.; Le-Van-Quyen, M.; Lotte, F.; et al. Neurofeedback: One of today’s techniques in psychiatry? L’Encéphale 2017, 43, 135–145. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  9. Felton, E.; Radwin, R.; Wilson, J.; Williams, J. Evaluation of a modified Fitts law brain—Computer interface target acquisition task in able and motor disabled individuals. J. Neural Eng. 2009, 6, 056002. [Google Scholar] [CrossRef]
  10. Spampinato, C.; Palazzo, S.; Kavasidis, I.; Giordano, D.; Souly, N.; Shah, M. Deep learning human mind for automated visual classification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 6809–6817. [Google Scholar]
  11. Wilson, J.A.; Walton, L.M.; Tyler, M.; Williams, J. Lingual electrotactile stimulation as an alternative sensory feedback pathway for brain—Computer interface applications. J. Neural Eng. 2012, 9, 045007. [Google Scholar] [CrossRef]
  12. Haxby, J.V.; Gobbini, M.I.; Furey, M.L.; Ishai, A.; Schouten, J.L.; Pietrini, P. Distributed and overlapping representations of faces and objects in ventral temporal cortex. Science 2001, 293, 2425–2430. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. Kamitani, Y.; Tong, F. Decoding the visual and subjective contents of the human brain. Nat. Neurosci. 2005, 8, 679–685. [Google Scholar] [CrossRef] [Green Version]
  14. Norman, K.A.; Polyn, S.M.; Detre, G.J.; Haxby, J.V. Beyond mind-reading: Multi-voxel pattern analysis of fMRI data. Trends Cogn. Sci. 2006, 10, 424–430. [Google Scholar] [CrossRef] [PubMed]
  15. Naselaris, T.; Kay, K.N.; Nishimoto, S.; Gallant, J.L. Encoding and decoding in fMRI. Neuroimage 2011, 56, 400–410. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Van Gerven, M.A.; Cseke, B.; De Lange, F.P.; Heskes, T. Efficient Bayesian multivariate fMRI analysis using a sparsifying spatio-temporal prior. NeuroImage 2010, 50, 150–161. [Google Scholar] [CrossRef] [PubMed]
  17. Huth, A.G.; De Heer, W.A.; Griffiths, T.L.; Theunissen, F.E.; Gallant, J.L. Natural speech reveals the semantic maps that tile human cerebral cortex. Nature 2016, 532, 453–458. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  18. Kay, K.N.; Winawer, J.; Rokem, A.; Mezer, A.; Wandell, B.A. A two-stage cascade model of BOLD responses in human visual cortex. PLoS Comput. Biol. 2013, 9, e1003079. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  19. Kay, K.N.; Naselaris, T.; Prenger, R.J.; Gallant, J.L. Identifying natural images from human brain activity. Nature 2008, 452, 352–355. [Google Scholar] [CrossRef]
  20. St-Yves, G.; Naselaris, T. The feature-weighted receptive field: An interpretable encoding model for complex feature spaces. NeuroImage 2018, 180, 188–202. [Google Scholar] [CrossRef]
  21. Du, C.; Du, C.; Huang, L.; He, H. Reconstructing perceived images from human brain activities with Bayesian deep multiview learning. IEEE Trans. Neural Netw. Learn. Syst. 2018, 30, 2310–2323. [Google Scholar] [CrossRef] [PubMed]
  22. Fujiwara, Y.; Miyawaki, Y.; Kamitani, Y. Modular encoding and decoding models derived from Bayesian canonical correlation analysis. Neural Comput. 2013, 25, 979–1005. [Google Scholar] [CrossRef]
  23. Nishimoto, S.; Vu, A.T.; Naselaris, T.; Benjamini, Y.; Yu, B.; Gallant, J.L. Reconstructing visual experiences from brain activity evoked by natural movies. Curr. Biol. 2011, 21, 1641–1646. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  24. Schoenmakers, S.; Güçlü, U.; Van Gerven, M.; Heskes, T. Gaussian mixture models and semantic gating improve reconstructions from human brain activity. Front. Comput. Neurosci. 2015, 8, 173. [Google Scholar] [CrossRef] [Green Version]
  25. Engel, S.A.; Glover, G.H.; Wandell, B.A. Retinotopic organization in human visual cortex and the spatial precision of functional MRI. Cereb. Cortex 1997, 7, 181–192. [Google Scholar] [CrossRef]
  26. Sereno, M.I.; Dale, A.; Reppas, J.; Kwong, K.; Belliveau, J.; Brady, T.; Rosen, B.; Tootell, R. Borders of multiple visual areas in humans revealed by functional magnetic resonance imaging. Science 1995, 268, 889–893. [Google Scholar] [CrossRef] [Green Version]
  27. Cox, D.D.; Savoy, R.L. Functional magnetic resonance imaging (fMRI) “brain reading”: Detecting and classifying distributed patterns of fMRI activity in human visual cortex. Neuroimage 2003, 19, 261–270. [Google Scholar] [CrossRef]
  28. Horikawa, T.; Kamitani, Y. Generic decoding of seen and imagined objects using hierarchical visual features. Nat. Commun. 2017, 8, 15037. [Google Scholar] [CrossRef]
  29. Miyawaki, Y.; Uchida, H.; Yamashita, O.; Sato, M.A.; Morito, Y.; Tanabe, H.C.; Sadato, N.; Kamitani, Y. Visual image reconstruction from human brain activity using a combination of multiscale local image decoders. Neuron 2008, 60, 915–929. [Google Scholar] [CrossRef] [Green Version]
  30. Livezey, J.A.; Glaser, J.I. Deep learning approaches for neural decoding: from CNNs to LSTMs and spikes to fMRI. arXiv 2020, arXiv:2005.09687. [Google Scholar]
  31. Naselaris, T.; Prenger, R.J.; Kay, K.N.; Oliver, M.; Gallant, J.L. Bayesian reconstruction of natural images from human brain activity. Neuron 2009, 63, 902–915. [Google Scholar] [CrossRef] [Green Version]
  32. Kingma, D.P.; Welling, M. Auto-encoding variational bayes. arXiv 2013, arXiv:1312.6114. [Google Scholar]
  33. Horikawa, T.; Tamaki, M.; Miyawaki, Y.; Kamitani, Y. Neural decoding of visual imagery during sleep. Science 2013, 340, 639–642. [Google Scholar] [CrossRef] [Green Version]
  34. Cowen, A.S.; Chun, M.M.; Kuhl, B.A. Neural portraits of perception: reconstructing face images from evoked brain activity. Neuroimage 2014, 94, 12–22. [Google Scholar] [CrossRef] [Green Version]
  35. Lee, H.; Kuhl, B.A. Reconstructing perceived and retrieved faces from activity patterns in lateral parietal cortex. J. Neurosci. 2016, 36, 6069–6082. [Google Scholar] [CrossRef] [PubMed]
  36. McCulloch, W.S.; Pitts, W. A logical calculus of the ideas immanent in nervous activity. Bull. Math. Biophys. 1943, 5, 115–133. [Google Scholar] [CrossRef]
  37. Güçlütürk, Y.; Güçlü, U.; Seeliger, K.; Bosch, S.; van Lier, R.; van Gerven, M.A. Reconstructing perceived faces from brain activations with deep adversarial neural decoding. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 4246–4257. [Google Scholar]
  38. Shen, G.; Horikawa, T.; Majima, K.; Kamitani, Y. Deep image reconstruction from human brain activity. PLoS Comput. Biol. 2019, 15, e1006633. [Google Scholar] [CrossRef] [Green Version]
  39. Huang, H.; Hu, X.; Zhao, Y.; Makkie, M.; Dong, Q.; Zhao, S.; Guo, L.; Liu, T. Modeling task fMRI data via deep convolutional autoencoder. IEEE Trans. Med. Imaging 2017, 37, 1551–1561. [Google Scholar] [CrossRef] [PubMed]
  40. Wen, H.; Shi, J.; Zhang, Y.; Lu, K.H.; Cao, J.; Liu, Z. Neural encoding and decoding with deep learning for dynamic natural vision. Cereb. Cortex 2018, 28, 4136–4160. [Google Scholar] [CrossRef] [PubMed]
  41. Shen, G.; Dwivedi, K.; Majima, K.; Horikawa, T.; Kamitani, Y. End-to-end deep image reconstruction from human brain activity. Front. Comput. Neurosci. 2019, 13, 21. [Google Scholar] [CrossRef]
  42. Anumanchipalli, G.K.; Chartier, J.; Chang, E.F. Speech synthesis from neural decoding of spoken sentences. Nature 2019, 568, 493–498. [Google Scholar] [CrossRef] [PubMed]
  43. Li, H.; Fan, Y. Brain decoding from functional MRI using long short-term memory recurrent neural networks. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Granada, Spain, 16–20 September 2018; pp. 320–328. [Google Scholar]
  44. Qiao, K.; Chen, J.; Wang, L.; Zhang, C.; Zeng, L.; Tong, L.; Yan, B. Category decoding of visual stimuli from human brain activity using a bidirectional recurrent neural network to simulate bidirectional information flows in human visual cortices. Front. Neurosci. 2019, 13, 692. [Google Scholar] [CrossRef] [PubMed]
  45. Gadgil, S.; Zhao, Q.; Adeli, E.; Pfefferbaum, A.; Sullivan, E.V.; Pohl, K.M. Spatio-Temporal Graph Convolution for Functional MRI Analysis. arXiv 2020, arXiv:2003.10613. [Google Scholar]
  46. Grigis, A.; Tasserie, J.; Frouin, V.; Jarraya, B.; Uhrig, L. Predicting Cortical Signatures of Consciousness using Dynamic Functional Connectivity Graph-Convolutional Neural Networks. bioRxiv 2020. [Google Scholar] [CrossRef]
  47. Zhang, Y.; Tetrel, L.; Thirion, B.; Bellec, P. Functional Annotation of Human Cognitive States using Deep Graph Convolution. bioRxiv 2020. [Google Scholar] [CrossRef]
  48. Huang, H.; Li, Z.; He, R.; Sun, Z.; Tan, T. Introvae: Introspective variational autoencoders for photographic image synthesis. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 3–8 December 2018; pp. 52–63. [Google Scholar]
  49. Du, C.; Li, J.; Huang, L.; He, H. Brain Encoding and Decoding in fMRI with Bidirectional Deep Generative Models. Engineering 2019, 5, 948–953. [Google Scholar] [CrossRef]
  50. Han, K.; Wen, H.; Shi, J.; Lu, K.H.; Zhang, Y.; Fu, D.; Liu, Z. Variational autoencoder: An unsupervised model for encoding and decoding fMRI activity in visual cortex. NeuroImage 2019, 198, 125–136. [Google Scholar] [CrossRef]
  51. Du, C.; Du, C.; Huang, L.; He, H. Conditional Generative Neural Decoding with Structured CNN Feature Prediction. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; pp. 2629–2636. [Google Scholar]
  52. Hong, S.; Yang, D.; Choi, J.; Lee, H. Inferring semantic layout for hierarchical text-to-image synthesis. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 7986–7994. [Google Scholar]
  53. Reed, S.; Akata, Z.; Yan, X.; Logeswaran, L.; Schiele, B.; Lee, H. Generative adversarial text to image synthesis. arXiv 2016, arXiv:1605.05396. [Google Scholar]
  54. Dong, H.; Neekhara, P.; Wu, C.; Guo, Y. Unsupervised image-to-image translation with generative adversarial networks. arXiv 2017, arXiv:1701.02676. [Google Scholar]
  55. Isola, P.; Zhu, J.Y.; Zhou, T.; Efros, A.A. Image-to-image translation with conditional adversarial networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1125–1134. [Google Scholar]
  56. Mirza, M.; Osindero, S. Conditional generative adversarial nets. arXiv 2014, arXiv:1411.1784. [Google Scholar]
  57. Huang, W.; Yan, H.; Wang, C.; Yang, X.; Li, J.; Zuo, Z.; Zhang, J.; Chen, H. Deep Natural Image Reconstruction from Human Brain Activity Based on Conditional Progressively Growing Generative Adversarial Networks. Neurosci. Bull. 2020, 37, 369–379. [Google Scholar] [CrossRef]
  58. Huang, W.; Yan, H.; Wang, C.; Li, J.; Zuo, Z.; Zhang, J.; Shen, Z.; Chen, H. Perception-to-Image: Reconstructing Natural Images from the Brain Activity of Visual Perception. Ann. Biomed. Eng. 2020, 48, 2323–2332. [Google Scholar] [CrossRef]
  59. Seeliger, K.; Güçlü, U.; Ambrogioni, L.; Güçlütürk, Y.; van Gerven, M.A. Generative adversarial networks for reconstructing natural images from brain activity. NeuroImage 2018, 181, 775–785. [Google Scholar] [CrossRef]
  60. St-Yves, G.; Naselaris, T. Generative adversarial networks conditioned on brain activity reconstruct seen images. In Proceedings of the 2018 IEEE International Conference on Systems, Man, and Cybernetics (SMC), Miyazaki, Japan, 7–10 October 2018; pp. 1054–1061. [Google Scholar]
  61. Lin, Y.; Li, J.; Wang, H. DCNN-GAN: Reconstructing Realistic Image from fMRI. In Proceedings of the 2019 16th International Conference on Machine Vision Applications (MVA), Tokyo, Japan, 27–31 May 2019; pp. 1–6. [Google Scholar]
  62. Hayashi, R.; Kawata, H. Image Reconstruction from Neural Activity Recorded from Monkey Inferior Temporal Cortex Using Generative Adversarial Networks. In Proceedings of the 2018 IEEE International Conference on Systems, Man, and Cybernetics (SMC), Miyazaki, Japan, 7–10 October 2018; pp. 105–109. [Google Scholar]
  63. Dosovitskiy, A.; Brox, T. Generating images with perceptual similarity metrics based on deep networks. arXiv 2016, arXiv:1602.02644. [Google Scholar]
  64. Mourao-Miranda, J.; Bokde, A.L.; Born, C.; Hampel, H.; Stetter, M. Classifying brain states and determining the discriminating activation patterns: support vector machine on functional MRI data. NeuroImage 2005, 28, 980–995. [Google Scholar] [CrossRef]
  65. LaConte, S.M. Decoding fMRI brain states in real-time. Neuroimage 2011, 56, 440–454. [Google Scholar] [CrossRef]
  66. Sitaram, R.; Caria, A.; Birbaumer, N. Hemodynamic brain—Computer interfaces for communication and rehabilitation. Neural Netw. 2009, 22, 1320–1328. [Google Scholar] [CrossRef] [Green Version]
  67. Du, C.; Du, C.; He, H. Sharing deep generative representation for perceived image reconstruction from human brain activity. In Proceedings of the 2017 International Joint Conference on Neural Networks (IJCNN), Anchorage, AK, USA, 14–19 May 2017; pp. 1049–1056. [Google Scholar]
  68. Awangga, R.; Mengko, T.; Utama, N. A literature review of brain decoding research. In Proceedings of the IOP Conference Series: Materials Science and Engineering, Chennai, India, 16–17 September 2020; Volume 830, p. 032049. [Google Scholar]
  69. Chen, M.; Han, J.; Hu, X.; Jiang, X.; Guo, L.; Liu, T. Survey of encoding and decoding of visual stimulus via FMRI: An image analysis perspective. Brain Imaging Behav. 2014, 8, 7–23. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  70. McIntosh, L.; Maheswaranathan, N.; Nayebi, A.; Ganguli, S.; Baccus, S. Deep learning models of the retinal response to natural scenes. In Proceedings of the Advances in Neural Information Processing Systems, Barcelona, Spain, 5–10 December 2016; pp. 1369–1377. [Google Scholar]
  71. Zhang, C.; Qiao, K.; Wang, L.; Tong, L.; Hu, G.; Zhang, R.Y.; Yan, B. A visual encoding model based on deep neural networks and transfer learning for brain activity measured by functional magnetic resonance imaging. J. Neurosci. Methods 2019, 325, 108318. [Google Scholar] [CrossRef] [PubMed]
  72. Zeidman, P.; Silson, E.H.; Schwarzkopf, D.S.; Baker, C.I.; Penny, W. Bayesian population receptive field modelling. NeuroImage 2018, 180, 173–187. [Google Scholar] [CrossRef] [PubMed]
  73. Tajbakhsh, N.; Shin, J.Y.; Gurudu, S.R.; Hurst, R.T.; Kendall, C.B.; Gotway, M.B.; Liang, J. Convolutional neural networks for medical image analysis: Full training or fine tuning? IEEE Trans. Med. Imaging 2016, 35, 1299–1312. [Google Scholar] [CrossRef] [Green Version]
  74. Güçlü, U.; van Gerven, M.A. Deep neural networks reveal a gradient in the complexity of neural representations across the ventral stream. J. Neurosci. 2015, 35, 10005–10014. [Google Scholar] [CrossRef] [Green Version]
  75. Rezende, D.J.; Mohamed, S.; Wierstra, D. Stochastic backpropagation and approximate inference in deep generative models. arXiv 2014, arXiv:1401.4082. [Google Scholar]
  76. Fang, T.; Qi, Y.; Pan, G. Reconstructing Perceptive Images from Brain Activity by Shape-Semantic GAN. In Proceedings of the Advances in Neural Information Processing Systems, Online, 6–12 December 2020; Volume 33. [Google Scholar]
  77. Bontonou, M.; Farrugia, N.; Gripon, V. Few-shot Learning for Decoding Brain Signals. arXiv 2020, arXiv:2010.12500. [Google Scholar]
  78. Kipf, T.N.; Welling, M. Semi-supervised classification with graph convolutional networks. arXiv 2016, arXiv:1609.02907. [Google Scholar]
  79. Doersch, C. Tutorial on variational autoencoders. arXiv 2016, arXiv:1606.05908. [Google Scholar]
  80. Sohn, K.; Lee, H.; Yan, X. Learning structured output representation using deep conditional generative models. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 7–12 December 2015; pp. 3483–3491. [Google Scholar]
  81. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 8–13 December 2014; pp. 2672–2680. [Google Scholar]
  82. Goodfellow, I. NIPS 2016 tutorial: Generative adversarial networks. arXiv 2016, arXiv:1701.00160. [Google Scholar]
  83. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  84. Dvornek, N.C.; Ventola, P.; Pelphrey, K.A.; Duncan, J.S. Identifying autism from resting-state fMRI using long short-term memory networks. In Proceedings of the International Workshop on Machine Learning in Medical Imaging, Quebec City, QC, Canada, 10 September 2017; pp. 362–370. [Google Scholar]
  85. Yu, B.; Yin, H.; Zhu, Z. Spatio-temporal graph convolutional networks: A deep learning framework for traffic forecasting. arXiv 2017, arXiv:1709.04875. [Google Scholar]
  86. Yan, S.; Xiong, Y.; Lin, D. Spatial temporal graph convolutional networks for skeleton-based action recognition. arXiv 2018, arXiv:1801.07455. [Google Scholar]
  87. Sitaram, R.; Caria, A.; Veit, R.; Gaber, T.; Rota, G.; Kuebler, A.; Birbaumer, N. FMRI brain-computer interface: a tool for neuroscientific research and treatment. Comput. Intell. Neurosci. 2007, 2007, 25487. [Google Scholar] [CrossRef]
  88. Paret, C.; Goldway, N.; Zich, C.; Keynan, J.N.; Hendler, T.; Linden, D.; Kadosh, K.C. Current progress in real-time functional magnetic resonance-based neurofeedback: Methodological challenges and achievements. Neuroimage 2019, 202, 116107. [Google Scholar] [CrossRef] [PubMed]
  89. Hinterberger, T.; Neumann, N.; Pham, M.; Kübler, A.; Grether, A.; Hofmayer, N.; Wilhelm, B.; Flor, H.; Birbaumer, N. A multimodal brain-based feedback and communication system. Exp. Brain Res. 2004, 154, 521–526. [Google Scholar] [CrossRef]
  90. Brouwer, A.M.; Van Erp, J.B. A tactile P300 brain-computer interface. Front. Neurosci. 2010, 4, 19. [Google Scholar] [CrossRef] [Green Version]
  91. Mohanty, R.; Sinha, A.M.; Remsik, A.B.; Dodd, K.C.; Young, B.M.; Jacobson, T.; McMillan, M.; Thoma, J.; Advani, H.; Nair, V.A.; et al. Machine learning classification to identify the stage of brain-computer interface therapy for stroke rehabilitation using functional connectivity. Front. Neurosci. 2018, 12, 353. [Google Scholar] [CrossRef]
  92. Hinterberger, T.; Veit, R.; Strehl, U.; Trevorrow, T.; Erb, M.; Kotchoubey, B.; Flor, H.; Birbaumer, N. Brain areas activated in fMRI during self-regulation of slow cortical potentials (SCPs). Exp. Brain Res. 2003, 152, 113–122. [Google Scholar] [CrossRef] [PubMed]
  93. Hinterberger, T.; Weiskopf, N.; Veit, R.; Wilhelm, B.; Betta, E.; Birbaumer, N. An EEG-driven brain-computer interface combined with functional magnetic resonance imaging (fMRI). IEEE Trans. Biomed. Eng. 2004, 51, 971–974. [Google Scholar] [CrossRef] [PubMed]
  94. De Kroon, J.; Van der Lee, J.; IJzerman, M.J.; Lankhorst, G. Therapeutic electrical stimulation to improve motor control and functional abilities of the upper extremity after stroke: A systematic review. Clin. Rehabil. 2002, 16, 350–360. [Google Scholar] [CrossRef]
  95. Pichiorri, F.; Morone, G.; Petti, M.; Toppi, J.; Pisotta, I.; Molinari, M.; Paolucci, S.; Inghilleri, M.; Astolfi, L.; Cincotti, F.; et al. Brain—Computer interface boosts motor imagery practice during stroke recovery. Ann. Neurol. 2015, 77, 851–865. [Google Scholar] [CrossRef] [PubMed]
  96. DeCharms, R.C.; Maeda, F.; Glover, G.H.; Ludlow, D.; Pauly, J.M.; Soneji, D.; Gabrieli, J.D.; Mackey, S.C. Control over brain activation and pain learned by using real-time functional MRI. Proc. Natl. Acad. Sci. USA 2005, 102, 18626–18631. [Google Scholar] [CrossRef] [Green Version]
  97. Maeda, F.; Soneji, D.; Mackey, S. Learning to explicitly control activation in a localized brain region through real-time fMRI feedback based training, with result impact on pain perception. In Proceedings of the Society for Neuroscience; Society for Neuroscience: Washington, DC, USA, 2004; pp. 601–604. [Google Scholar]
  98. Guan, M.; Ma, L.; Li, L.; Yan, B.; Zhao, L.; Tong, L.; Dou, S.; Xia, L.; Wang, M.; Shi, D. Self-regulation of brain activity in patients with postherpetic neuralgia: a double-blind randomized study using real-time FMRI neurofeedback. PLoS ONE 2015, 10, e0123675. [Google Scholar] [CrossRef] [PubMed]
  99. Caria, A.; Veit, R.; Sitaram, R.; Lotze, M.; Weiskopf, N.; Grodd, W.; Birbaumer, N. Regulation of anterior insular cortex activity using real-time fMRI. Neuroimage 2007, 35, 1238–1246. [Google Scholar] [CrossRef]
  100. Caria, A.; Sitaram, R.; Veit, R.; Begliomini, C.; Birbaumer, N. Volitional control of anterior insula activity modulates the response to aversive stimuli. A real-time functional magnetic resonance imaging study. Biol. Psychiatry 2010, 68, 425–432. [Google Scholar] [CrossRef]
  101. Anders, S.; Lotze, M.; Erb, M.; Grodd, W.; Birbaumer, N. Brain activity underlying emotional valence and arousal: A response-related fMRI study. Hum. Brain Mapp. 2004, 23, 200–209. [Google Scholar] [CrossRef] [PubMed]
  102. Zotev, V.; Krueger, F.; Phillips, R.; Alvarez, R.P.; Simmons, W.K.; Bellgowan, P.; Drevets, W.C.; Bodurka, J. Self-regulation of amygdala activation using real-time fMRI neurofeedback. PLoS ONE 2011, 6, e24522. [Google Scholar] [CrossRef] [PubMed]
  103. Koush, Y.; Meskaldji, D.E.; Pichon, S.; Rey, G.; Rieger, S.W.; Linden, D.E.; Van De Ville, D.; Vuilleumier, P.; Scharnowski, F. Learning control over emotion networks through connectivity-based neurofeedback. Cereb. Cortex 2017, 27, 1193–1202. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  104. Linden, D.E.; Habes, I.; Johnston, S.J.; Linden, S.; Tatineni, R.; Subramanian, L.; Sorger, B.; Healy, D.; Goebel, R. Real-time self-regulation of emotion networks in patients with depression. PLoS ONE 2012, 7, e38115. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  105. Viding, E. On the nature and nurture of antisocial behavior and violence. Ann. N. Y. Acad. Sci. 2004, 1036, 267–277. [Google Scholar] [CrossRef]
  106. Birbaumer, N.; Veit, R.; Lotze, M.; Erb, M.; Hermann, C.; Grodd, W.; Flor, H. Deficient fear conditioning in psychopathy: A functional magnetic resonance imaging study. Arch. Gen. Psychiatry 2005, 62, 799–805. [Google Scholar] [CrossRef] [Green Version]
  107. Veit, R.; Flor, H.; Erb, M.; Hermann, C.; Lotze, M.; Grodd, W.; Birbaumer, N. Brain circuits involved in emotional learning in antisocial behavior and social phobia in humans. Neurosci. Lett. 2002, 328, 233–236. [Google Scholar] [CrossRef]
  108. Cheng, X.; Ning, H.; Du, B. A Survey: Challenges and Future Research Directions of fMRI Based Brain Activity Decoding Model. In Proceedings of the 2021 International Wireless Communications and Mobile Computing (IWCMC), Harbin, China, 28 June–2 July 2021; pp. 957–960. [Google Scholar]
  109. Akamatsu, Y.; Harakawa, R.; Ogawa, T.; Haseyama, M. Multi-view bayesian generative model for multi-subject fmri data on brain decoding of viewed image categories. In Proceedings of the ICASSP 2020—2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain, 4–8 May 2020; pp. 1215–1219. [Google Scholar]
  110. Xu, K.; Ba, J.; Kiros, R.; Cho, K.; Courville, A.; Salakhudinov, R.; Zemel, R.; Bengio, Y. Show, Attend and Tell: Neural Image Caption Generation with Visual Attention. In Proceedings of the 32nd International Conference on Machine Learning, Lille, France, 6–11 July 2015; Volume 23. [Google Scholar]
  111. Sundararajan, M.; Taly, A.; Yan, Q. Axiomatic attribution for deep networks. arXiv 2017, arXiv:1703.01365. [Google Scholar]
  112. Zhang, Y.; Bellec, P. Transferability of brain decoding using graph convolutional networks. bioRxiv 2020. [Google Scholar] [CrossRef]
  113. Gao, Y.; Zhang, Y.; Wang, H.; Guo, X.; Zhang, J. Decoding behavior tasks from brain activity using deep transfer learning. IEEE Access 2019, 7, 43222–43232. [Google Scholar] [CrossRef]
  114. Shine, J.M.; Bissett, P.G.; Bell, P.T.; Koyejo, O.; Balsters, J.H.; Gorgolewski, K.J.; Moodie, C.A.; Poldrack, R.A. The dynamics of functional brain networks: integrated network states during cognitive task performance. Neuron 2016, 92, 544–554. [Google Scholar] [CrossRef] [Green Version]
  115. Zhao, Y.; Li, X.; Huang, H.; Zhang, W.; Zhao, S.; Makkie, M.; Zhang, M.; Li, Q.; Liu, T. Four-Dimensional Modeling of fMRI Data via Spatio–Temporal Convolutional Neural Networks (ST-CNNs). IEEE Trans. Cogn. Dev. Syst. 2019, 12, 451–460. [Google Scholar] [CrossRef]
  116. Kazeminejad, A.; Sotero, R.C. Topological properties of resting-state fMRI functional networks improve machine learning-based autism classification. Front. Neurosci. 2019, 12, 1018. [Google Scholar] [CrossRef] [PubMed]
  117. Ktena, S.I.; Parisot, S.; Ferrante, E.; Rajchl, M.; Lee, M.; Glocker, B.; Rueckert, D. Distance metric learning using graph convolutional networks: Application to functional brain networks. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Quebec, QC, Canada, 11–13 September 2017; pp. 469–477. [Google Scholar]
  118. Huth, A.G.; Nishimoto, S.; Vu, A.T.; Gallant, J.L. A continuous semantic space describes the representation of thousands of object and action categories across the human brain. Neuron 2012, 76, 1210–1224. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  119. Huth, A.G.; Lee, T.; Nishimoto, S.; Bilenko, N.Y.; Vu, A.T.; Gallant, J.L. Decoding the semantic content of natural movies from human brain activity. Front. Syst. Neurosci. 2016, 10, 81. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  120. Walther, D.B.; Caddigan, E.; Fei-Fei, L.; Beck, D.M. Natural scene categories revealed in distributed patterns of activity in the human brain. J. Neurosci. 2009, 29, 10573–10581. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Visual pathway.
Figure 1. Visual pathway.
Brainsci 12 00228 g001
Figure 2. Structured deep generative neural decoding model [51].
Figure 2. Structured deep generative neural decoding model [51].
Brainsci 12 00228 g002
Table 1. Brain encoding models.
Table 1. Brain encoding models.
LiteratureObjectiveModel/MethodExplanation
[71]Predict cortical responsesA pre-trained DNNTrain a nonlinear mapping from visual features to brain activity with a pre-trained DNN (i.e., AlexNet) using transfer learning technique.
[15]Representation of information in the visual cortexGLMA systematic modeling method is proposed to estimate an encoding model for each voxel and then to perform decoding with the estimated encoding model.
[18]Predict responses to a wide range of stimuli of the input imagesTwo-stage cascade modelThis encoding model is a two-stage cascade architecture of a linear stage and nonlinear stage. The linear stage involves calculations of local filters and division normalization. The nonlinear stage involves compressive spatial summation and a second-order contrast.
[72]Predict the response of a single voxel or brain neurons in a region of interest in any dimensional space of the stimulusReceptive Field (rPF)The encoding model quantifies the uncertainty of neuron parameters, rPF size, and location by estimating the covariance of the parameters.
[20]Map the brain activity to natural scenesFeature-weighted
Receptive field
This method converts visual stimuli to corresponding visual features and assumes that spatial features are separable and uses visual feature maps to train deep neural networks. The pre-trained deep neural network weights the contribution of each feature map to voxel activity in brain regions.
Table 2. Brain decoding model based on machine learning.
Table 2. Brain decoding model based on machine learning.
LiteratureObjectiveModel/MethodExplanation
[12]Classification of visual stimuli decoded from brain activityGLMFor different types of stimuli (objects and pictures), the cerebral cortex has different response patterns through the fMRI of the abdominal temporal cortex.
[27]fMRI and brain signals classificationMVPAFor fMRI in the pre-defined ROI of the cerebral cortex, the activated mode of fMRI is classified by the multivariate statistical pattern recognition.
[64]fMRI signals classificationSVMSVM classifier finds the best area in the cerebral cortex that can distinguish the brain state, and then SVM is trained to predict brain state through fMRI.
[14]Classify fMRI activity patternsMVPA
Bayesian method
The Gaussian Naive Bayes classifier has high classification accuracy. Because it assumes that the importance of each voxel is the same and it does not consider the sparsity constraint, its interpretability is poor.
[29]Reconstruct geometric images from brain activityMVPABased on a modular modeling method, the multi-voxel pattern of fMRI signals and multi-scale vision are used to reconstruct geometric image stimuli that composes of flashing checkerboard patterns.
[31]Reconstruct the structure and semantic content of natural imagesBayesian methodUse Bayes’ theorem to combine encoding model and prior information of natural images to calculate the probability of a measured brain response due to the visual stimuli of each image. However, only a simple correlation between the reconstructed image and the training image can be established.
[16]Improve fMRI Bayesian classifier accuracyMVPA
Bayesian method
The sparsity constraint is added to the multivariate analysis model of the Bayesian network to quantify the uncertainty of voxel features.
[23]Reconstruct spatio-temporal stimuli using image priorsBayesian methodThis method used a large amount of videos as a priori information and combines the videos with a Bayesian decoder to reconstruct visual stimuli from fMRI signals.
[33]Decoding human dreamsSVMSVM classifier is trained to map natural images to brain activities, and a vocabulary database is used to label the images with semantic tags to decode the semantic content of dreams.
[22]Decode the reversible mapping between brain activity and visual imagesBCCAThe encoding and decoding network is composed of generated multi-view models. The disadvantage is that its linear structure makes the model unable to express the multi-level visual features of the image, and its spherical covariance assumption cannot understand the correlation between fMRI voxels, making it more susceptible to noise.
[34]Dimensionality reduction of high-dimensional fMRI dataPCAPCA reduces the dimensionality of the facial training data set, and the partial least squares regression algorithm maps the fMRI activity pattern to the dimensionality-reduced facial features.
[24]Infer the semantic category of the reconstructed imageBayesian methodPropose a mixed Bayesian network based on the Gaussian mixture model. The Gaussian mixture model represents the prior distribution of the image and can infer high-order semantic categories from low-order image features through combining the prior distributions of different information sources.
[28]Predict object categories in dreamsMVPA
CNN
Based on CNN, train a decoder with the data set of the normal visual perceptions, and decode the neural activity to the object category. This process involves two parts: 1. map the fMRI signal to the feature space; 2. using correlation analysis to infer the object category based on the feature space.
[44]Decode visual stimuli from human brain activityRNN
CNN
Use CNN to select a set of small fMRI voxel signals as the input and then use RNN to classify the selected fMRI voxels.
[41]Capture the direct mapping between brain activity and perceptionCNNThe generator is directly trained with fMRI data by an end-to-end approach.
[40]Reconstruct dynamic video stimuliCNN
PCA
The CNN-based coding model extracts the linear combination of the input video features and then uses PCA to reduce the dimensionality of the extracted high-dimensional feature space while retaining the variance of 99% of the principal components.
Table 3. Deep generative models based on VAE or GAN.
Table 3. Deep generative models based on VAE or GAN.
LiteratureObjectiveModel/MethodExplanation
[21]Reconstruct perception images from brain activityDeep Generative Multiview Model (DGMM)DGMM first uses DNN to extract the image’s hierarchical features. Based on the fact that the human brain’s processing model for external stimuli is sparse, a sparse linear model is used to avoid over-fitting of fMRI data. The statistical relationships between the visual stimuli and the evoked fMRI data are modeled by using two view-specific generators with a shared latent space to obtain multiple correspondences between fMRI voxel patterns and image pixel patterns. DGMM can be optimized with an automatically encoded Bayesian model [32,75].
[37]Reconstruct facial imagesDeep Adversarial Neural Decoding (DAND)DAND uses the maximum posterior estimation to transform brain activity linearly to the hidden features. Then, the pre-trained CNN and adversarial training are used to transform the hidden features nonlinearly to reconstruct human facial images. DAND showed good performance in reconstructing the details of the face’s gender, skin color, and facial expressions.
[48]Improve the quality of reconstructed imagesIntrospective Variational Autoencoders (IntroVAE)IntroVAE generator and inference model can be jointly trained in a self-assessment manner. The generator takes the output of the inference model noise as the input to generate the image. The inference model not only learns the potential popular structure of the input image but also classifies the real image and the generative image, which is similar to GAN’s adversarial learning.
[59]Reconstruct natural images from brain activityDeep Convolution Generative Adversarial Network (DCGAN)DCGAN uses a large natural image data set to train a deep convolutional generation confrontation network in an unsupervised manner, and learn the potential space of stimuli. This DCGAN is used to generate arbitrary images from the stimulus domain.
[60]Reconstruct the visual stimuli of brain activityGANThey used an encoding model to create surrogate brain activity samples, with which the generative adversarial networks (GANs) are trained to learn a generative model of images and then generalized to real fRMI data measured during the perception of images. The basic outline of the stimuli can finally be reconstructed.
[50]Reconstruct visual stimuli (video) of brain activityVAEVAE is trained with a five-layer encoder and a five-layer decoder to learn visual representations from a diverse set of unlabeled images in an unsupervised way. VAE first converts the fMRI activity to the latent variables and then converts the latent variables to the reconstructed video frames through the VAE’s decoder. However, VAE could only provide relatively lower accuracy in higher-order visual areas compared to CNN.
[38]Reconstruct color images and simple gray-scale imagesGANA pre-trained DNN decodes the measured fMRI patterns into the hierarchical features that can represent the human visual layering mechanism. The DNN network extracts image features, and then compares them with the decoded human brain activity features, which guides the deep generator network (DGN) to reconstruct images and iteratively minimizes the errors of the two. A natural image prior introduced by an enhanced DGN semantically details to the reconstructions, which improves the visual quality of generated images.
[51]Reconstruct the visual image from brain activityA structured multi-output regression (SMR) model and Introspective Conditional Generation (ICG)Decodes the brain activity to the intermediate CNN features and then maps these intermediate features to visual images. Combining maximum likelihood estimation and adversarial learning, ICG model uses divergence and reconstruction error for adversarial optimization, which can evaluate the difference between the generated image and the real image.
[76]Use semantic features to add details to the generated imageShape-Semantic GANThis framework consists of a linear shape decoder, a semantic decoder based on DNN, and an image generator based on GAN. The output of the shape decoder and the semantic decoder are input to the GAN-based image generator, and the semantic features in GAN are used as a supplement to the image details to reconstruct high quality images.
[57]Reconstruct natural images from brain activityProgressively Growing GAN (PG-GAN)This model adds a priori knowledge of potential features to (PG-GAN). The decoder decodes the measured response of the cerebral cortex into the latent features of the natural image and then reconstruct the natural image through the generator.
[58]Reconstruct natural images from brain activitySimilarity-conditions generative adversarial network (SC-GAN)SC-GAN not only extracts the response patterns of the cerebral cortex to natural images but also captures the high-level semantic features of natural images. The captured semantic features is input to GAN to reconstruct natural images.
Table 4. GCN-based brain decoding.
Table 4. GCN-based brain decoding.
LiteratureObjectiveModel/MethodExplanation
[45]Localize brain regions and functional connectionsSpatio-Temporal Graph Convolution Networks (ST-GCN)Based on ST-GCN, the representation extracted from the fMRI data expresses both temporal dynamic information of brain activity and functional dependence between brain regions. Through training ST-GCN, this method can learn the edge importance matrix on short sub-sequences of BOLD time series to improve the prediction accuracy and interpretability of the model.
[46]Decode the consciousness level from cortical activity recordingBrainNetCNNBrainNetCNN is a GCN-based decoding model, which uses multi-layer non-linear units to extract features and predict brain consciousness states.
[47]Predict human brain cognitive stateGCNThe brain annotation model uses six graph convolutional layers as feature extractors and two fully connected layers as classifiers to decode the cognitive state of the brain, taking a short series of fMRI data as the input, spreading the information in the annotation model network, and generating high-level domain-specific graph representations to predict the brain cognitive state.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Du, B.; Cheng, X.; Duan, Y.; Ning, H. fMRI Brain Decoding and Its Applications in Brain–Computer Interface: A Survey. Brain Sci. 2022, 12, 228. https://doi.org/10.3390/brainsci12020228

AMA Style

Du B, Cheng X, Duan Y, Ning H. fMRI Brain Decoding and Its Applications in Brain–Computer Interface: A Survey. Brain Sciences. 2022; 12(2):228. https://doi.org/10.3390/brainsci12020228

Chicago/Turabian Style

Du, Bing, Xiaomu Cheng, Yiping Duan, and Huansheng Ning. 2022. "fMRI Brain Decoding and Its Applications in Brain–Computer Interface: A Survey" Brain Sciences 12, no. 2: 228. https://doi.org/10.3390/brainsci12020228

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop