Next Article in Journal
Connection between Variations of the Probability Distribution of the Recurrence Time and Phases of the Seismic Activity
Previous Article in Journal
A Wigner Quasiprobability Distribution of Work
Previous Article in Special Issue
Multimodal Attention Dynamic Fusion Network for Facial Micro-Expression Recognition
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

A Survey of Deep Learning-Based Multimodal Emotion Recognition: Speech, Text, and Face

1
Key Laboratory of Child Development and Learning Science (Ministry of Education), Southeast University, Nanjing 210000, China
2
School of Information Science and Engineering, Southeast University, Nanjing 210000, China
3
School of Biological Science and Medical Engineering, Southeast University, Nanjing 210000, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Entropy 2023, 25(10), 1440; https://doi.org/10.3390/e25101440
Submission received: 3 August 2023 / Revised: 20 September 2023 / Accepted: 9 October 2023 / Published: 12 October 2023
(This article belongs to the Special Issue Machine and Deep Learning for Affective Computing)

Abstract

:
Multimodal emotion recognition (MER) refers to the identification and understanding of human emotional states by combining different signals, including—but not limited to—text, speech, and face cues. MER plays a crucial role in the human–computer interaction (HCI) domain. With the recent progression of deep learning technologies and the increasing availability of multimodal datasets, the MER domain has witnessed considerable development, resulting in numerous significant research breakthroughs. However, a conspicuous absence of thorough and focused reviews on these deep learning-based MER achievements is observed. This survey aims to bridge this gap by providing a comprehensive overview of the recent advancements in MER based on deep learning. For an orderly exposition, this paper first outlines a meticulous analysis of the current multimodal datasets, emphasizing their advantages and constraints. Subsequently, we thoroughly scrutinize diverse methods for multimodal emotional feature extraction, highlighting the merits and demerits of each method. Moreover, we perform an exhaustive analysis of various MER algorithms, with particular focus on the model-agnostic fusion methods (including early fusion, late fusion, and hybrid fusion) and fusion based on intermediate layers of deep models (encompassing simple concatenation fusion, utterance-level interaction fusion, and fine-grained interaction fusion). We assess the strengths and weaknesses of these fusion strategies, providing guidance to researchers to help them select the most suitable techniques for their studies. In summary, this survey aims to provide a thorough and insightful review of the field of deep learning-based MER. It is intended as a valuable guide to aid researchers in furthering the evolution of this dynamic and impactful field.

1. Introduction

Emotion recognition is an important area of research because it enables computers to accurately comprehend human emotions and provide intelligent responses to meet human requirements. In educational settings, emotions can play a critical role in student performance and well-being. Emotion recognition technology can be utilized to monitor students’ emotional states, allowing educators to gain a deeper understanding of their academic progress and overall condition. Emotion recognition has potential in the healthcare sector as it can assist doctors in understanding the emotional states of their patients, thus enabling the provision of individualized and tailored medical services. Emotion recognition also has the capacity to enhance intelligent customer service systems by detecting the emotional needs of users and offering more personalized services. As a result, emotion recognition technologies have emerged as a significant research area in artificial intelligence, attracting a great deal of attention for their potential to revolutionize how humans and computers interact.
During the initial phases of emotion recognition research, researchers primarily concentrated on the recognition of single modalities, e.g., speech emotion recognition [1,2,3,4,5] text emotion recognition [6,7], and facial expression recognition [8,9,10,11]. However, the limited accuracy of single-modal emotional judgments caused by incomplete data and susceptibility to noise led to an increase in the utilization of multiple modalities by researchers in making emotional evaluations. In light of this, multimodal emotion recognition (MER) was developed. Moreover, MER can leverage the maximal mutual information to quantify the dependency between different modality features, extracting the most informative and discriminative features from each modality. This technique further enhances the model’s capacity to make accurate emotional distinctions. As such, MER has captured the attention of many researchers, aiming to combine information from various modalities, which can corroborate and complement each other, thus providing more comprehensive and accurate information for emotional judgment and significantly improving the performance of emotional judgments [12,13,14,15].
With the rapid development of deep learning techniques, MER based on deep learning has emerged as one of the most prominent research topics in the field of emotion recognition [16]. Central to this is the design of specific network architectures and associated loss functions. Many current designs of loss functions draw from principles related to entropy, with cross-entropy loss functions being a predominant example. Essentially, a vast array of deep learning architectures optimize their models by either maximizing or minimizing certain entropy-associated loss functions [17,18,19,20]. This flexibility suggests that deep learning techniques can be adeptly customized to harness the interactions between various modalities, thereby gleaning features rich in emotional information for accurate emotion identification. An increasing number of researchers are adopting deep learning methodologies for MER, yielding commendable outcomes. Given this burgeoning interest, a comprehensive survey of the most recent methodologies and a forecast of future research directions have become imperative. Gu et al. [21] provided a comprehensive analysis of the framework and research methods applicable to video and music emotion recognition, with particular emphasis on deep learning techniques. However, this survey is deficient in its detailed description of existing multimodal databases and feature specifications. Koromilas et al. [22] exhaustively introduced an intermediate fusion method based on a deep multimodal learning architecture, but failed to mention early fusion and late fusion techniques. Zhang et al. [23] thoroughly depicted single-modal emotion recognition technologies, such as facial expression recognition, speech emotion recognition, and text emotion recognition. For MER, only feature-level and decision-level fusion methods were summarized. Given these observations, it becomes clear that existing surveys have not fully encapsulated the advancements in the deep domain of MER. Consequently, there is an urgent need for a comprehensive survey article on deep multimodal emotion recognition, which will facilitate a well-rounded understanding of the latest technological developments in the field.
Our survey aims to fill the existing gap in the field of deep learning-based MER. Unlike existing surveys, our work offers a more comprehensive exploration of the deep learning-based MER domain. We delve not only into different fusion strategies (early fusion, late fusion, hybrid fusion, simple concatenation fusion, utterance-level interaction fusion, and fine-grained interaction fusion) but also provide an extensive analysis of multimodal datasets and emotional feature extraction methods. This comprehensive analysis sets our research apart from prior surveys, which often concentrated on singular aspects of MER. Specifically, we first introduce widely employed MER datasets, conducting a systematic analysis of their inherent characteristics and potential challenges. We also delineate the merits and demerits of these datasets, aiding researchers in selecting the most suitable datasets for their specific studies. Secondly, we meticulously explore various approaches to multimodal emotional feature extraction, emphasizing the strengths and weaknesses of each method. Finally, we provide a comprehensive analysis of MER algorithms, with a focus on early, late, hybrid, and intermediate layer fusion strategies. We evaluate the pros and cons of these fusion strategies, offering valuable insights to assist researchers in selecting the most appropriate fusion techniques. In summary, this survey offers a thorough and insightful comprehension of the present state of the art and future research avenues in the field of MER. We hope that our research will fill the current gap in the field and serve as a valuable reference for researchers aiming to acquire a comprehensive understanding of the most recent advancements and forthcoming research prospects in the realm of MER.
Figure 1 illustrates the comprehensive framework of a typical deep learning-based MER, which also delineates the structure of this survey. The organization of this paper is as follows: Section 2 presents widely used MER datasets and their characteristics and challenges. Section 3 delves into multimodal emotional feature extraction methods. Section 4 details MER algorithms and analyzes their advantages, disadvantages, and scope of application. Section 5 elucidates evaluation metrics prevalently employed in multimodal emotion recognition. Section 6 analyzes the results of existing research. Finally, Section 7 provides a summation of the paper’s main content and conclusions, highlighting future challenges in MER.

2. Popular Datasets in Multimodal Emotion Recognition

In the field of MER, several key datasets have been developed to facilitate research and experimental progress. While the existing literature presents a wide range of datasets, each with its unique advantages and limitations, there are still areas that warrant further attention. Here, we aim to provide an in-depth review of these existing resources. Table 1 presents the most popular datasets used for MER (including speech, text, and facial expression information).

2.1. IEMOCAP

The Interactive Emotional Dyadic Motion Capture (IEMOCAP) dataset, assembled by the Signal Analysis and Interpretation Lab (SAIL) at the University of Southern California, serves as a prominent multimodal resource for emotion recognition research [24]. The dataset amalgamates diverse modalities such as video, speech, facial motion capture, and text data, enabling comprehensive examination of emotional states in interactive circumstances. Ten actors, an equal mix of males and females, contributed to the data collection. These actors, paired by gender and divided into five groups, performed both scripted and improvised dialogues. This assortment of dialogues enriches the dataset with a variety of emotional content, enhancing its representativeness of real-world emotional communication. IEMOCAP comprises 4784 improvised and 5255 scripted conversations, providing a wide spectrum of emotional contexts for analysis. The dialogues span nine discrete emotions (happiness, sadness, anger, surprise, fear, disgust, frustration, excitement, and neutrality), in addition to continuous emotional dimensions like activation, arousal, and dominance, facilitating a more dynamic emotional analysis. IEMOCAP’s chief strength resides in its authenticity, as the emotions captured are natural, not simulated, thereby helping deep learning models to discern genuine emotional cues from various modalities. Nonetheless, IEMOCAP is not without its limitations. Despite its rich modalities and emotional labels, the dataset size is comparatively small, potentially limiting its suitability for deep learning models that require larger data volumes. The language constraint to English reduces the model’s versatility across different languages and cultures. The controlled laboratory setting for data collection might also affect the authenticity of emotions, potentially impairing the model’s generalization performance in authentic settings. Furthermore, the dataset manifests a severe class imbalance issue, potentially impacting the model’s accuracy for less represented emotions.

2.2. Youtube Dataset

The YouTube dataset was first introduced by Morency et al. in 2011 [25]. The dataset comprises 47 videos, with 20 videos featuring female subjects and the remaining 27 highlighting male perspectives. The dataset further extends its diversity with an age range spanning from 14 to 60, depicting emotive expressions across various stages of life. Although the participants originate from a variety of cultural backgrounds, each video within the dataset is characterized by participants expressing their views in English. This complexity is crucial to the development of models that can robustly function in real-world scenarios where noise is omnipresent. The emotional labeling of the videos in the dataset further augments its value. Each of the 47 videos has been tagged with either positive, negative, or neutral emotion tags. These categorical annotations provide a solid foundation for supervised learning approaches in deep learning. The YouTube dataset contains diverse scenes and environmental noise, which makes it advantageous for improving model generalization. Note that due to the early collection time of the data, the impact of new cultural, social, and technological developments on emotional expression may not be fully reflected in the dataset. Nonetheless, it has served as a cornerstone in the multimodal sentiment analysis field, paving the way for more comprehensive, dynamic, and contemporaneous datasets.

2.3. MOUD

The Multimodal Opinion Utterances Dataset (MOUD) was pioneered by Perez-Rosas et al. in 2013 [26]. The dataset offers a unique perspective by focusing on opinion-based utterances, which are intrinsically rich in emotional cues. The MOUD consists of 498 utterances, all painstakingly labeled as either positive, negative, or neutral in emotion. An intriguing aspect of MOUD is its source; the dataset comprises videos curated from YouTube, one of the world’s largest platforms for user-generated content. This provides the advantage of real-world, unscripted emotional expressions, a factor that significantly boosts the applicability and robustness of models trained on this data. The dataset features speakers whose ages range between 20 and 60 years, ensuring a good representation of emotional expressions across different age groups. It includes 15 female speakers, further enhancing its gender diversity. All speakers express their opinions in Spanish, focusing on product reviews. The average duration of the speeches in the dataset is approximately 5 s, which is a typical duration for a single utterance. The dataset includes 182 positive, 231 negative, and 85 neutral utterances. While the MOUD has made significant contributions to the field of MER, it is worth noting that it contains a relatively limited number of samples, with just 498 utterances. This may potentially limit the diversity and the breadth of the emotional expressions it can capture. Nevertheless, the dataset has served as a valuable resource for emotion recognition research, particularly in opinion-based expressions.

2.4. ICT-MMMO

The Institute for Creative Technologies’ Multi-Modal Movie Opinion (ICT-MMMO) database, established by Wollmer et al. in 2013 [27], marks a significant contribution to the realm of MER datasets. It assembles videos from two disparate platforms, namely YouTube and ExpoTV. Totaling 370 videos, the dataset leverages the wealth of emotionally charged user-generated content that is ubiquitous on platforms like YouTube. ICT-MMMO sets itself apart with its focus on sentiment analysis of genuine user comments. This real-world context is indispensable for constructing resilient deep learning models. The database houses 228 positive, 57 negative, and 23 neutral YouTube videos, in conjunction with 62 negative ExpoTV movie review videos. Each video in the ICT-MMMO database is classified under one of five emotional states: strongly positive, weakly positive, neutral, strongly negative, or weakly negative. This refined sentiment categorization offers an extension beyond the conventional positive–negative–neutral spectrum, furnishing researchers with a more sophisticated understanding of emotional subtleties and intensities. Nonetheless, it is crucial to recognize the constraints of the ICT-MMMO database, such as its comparably small sample size of 370 videos. Furthermore, it exclusively comprises negative samples from ExpoTV movie reviews. These restrictions could possibly confine the array of emotional expressions the dataset can capture, thereby potentially influencing the broad applicability of models trained on these data.

2.5. CMU-MOSI

The Multimodal Opinion-level Sentiment Intensity (CMU-MOSI) dataset, introduced by Zadeh et al. in 2016 [28], marks a significant milestone in multimodal sentiment analysis databases due to its unique characteristic of incorporating subjective sentiment and emotional intensity annotations. The dataset includes 93 randomly selected videos from YouTube, involving 89 different speakers—41 females and 48 males. All these videos were recorded in different settings, some of which used high-tech microphones and cameras while others used less professional recording equipment. In addition, the distance between the users and the camera, as well as the background and lighting conditions, were different. The original quality of the videos remains unaltered, preserving their fidelity. This authentic approach guarantees the data accurately reflect the varying audio-visual quality in user-generated content, thereby providing a robust training set for real-world applications. The CMU-MOSI dataset covers a wide spectrum of topics, including movie and book reviews and product evaluations. The videos are segmented into 2199 clips, each rated on an emotional polarity scale, from +3 (signifying strongly positive sentiment) to −3 (signifying strongly negative sentiment). This scale was developed based on the annotations of five separate annotators, with the average of these five scores serving as the definitive emotional polarity for each clip. This methodology effectively distills the sentiments down to two categories, positive and negative, simplifying the complexity of emotions and providing a clear benchmark for training sentiment analysis models. In summary, the CMU-MOSI dataset’s contribution to the field of MER is invaluable, providing a complex, rich, and diverse set of data for the development and testing of robust, real-world applicable models.

2.6. NNIME

The NTHU-NTUA Chinese Interactive Emotion Corpus (NNIME) [29] is a groundbreaking initiative aimed at creating an interactive multimodal corpus for Chinese emotional expressions. This accomplishment is the product of a successful collaboration between Tsinghua University and the National Taiwan University of Arts in Taiwan. The dataset leverages dyadic spoken interactions to evoke genuine emotional responses. Distinctly, the database features spontaneous dyadic spoken interactions involving 44 participants. These interactions have been assiduously captured, producing roughly 11 h of continuous and synchronized data across audio, video, and electrocardiogram modalities. The multimodal character of this dataset extends the scope of sentiment analysis by merging physiological indicators with audio–visual data, setting a precedent for more detailed and holistic emotion assessments. One of the salient characteristics of the NNIME dataset is its extensive annotation process. A robust team of 49 annotators meticulously annotated the data, yielding valuable emotional insights. The annotation procedure includes four distinctive perspectives: peer reports, director reports, self-reports, and observer reports. Such a diversified approach ensures a comprehensive view of the emotional domain, offering researchers profound and expansive insights into the affective phenomena being analyzed. Structurally, the NNIME dataset comprises 6701 sentences and includes both discrete and continuous emotion labels. Discrete labels encompass emotions like anger, sadness, joy, surprise, neutrality, and happiness. Conversely, continuous labels entail the scales of valence and arousal. This two-pronged labeling system captures the subtlety of emotional expression, facilitating a more precise exploration of emotional complexities and variations. Overall, the rigorous process of data collection and annotation employed in creating the NNIME dataset renders it a valuable resource for quantifying and investigating various facets of affective phenomena and human communication. This database’s rich, multimodal nature makes it particularly suitable for developing and validating deep learning models for emotion recognition, thereby pushing the frontiers of MER.

2.7. CMU-MOSEI

The CMU Multimodal Opinion Sentiment and Emotion Intensity (CMU-MOSEI) dataset [30] is a crucial resource in multimodal sentiment analysis and emotion recognition, distinguished by its scale, which is unmatched in the field. This dataset encompasses 3837 videos collected from over 1000 distinctive YouTube speakers, with a reasonably balanced gender distribution of 57% male and 43% female speakers, highlighting the dataset’s focus on inclusivity. One of CMU-MOSEI’s unique characteristics is its annotations at the utterance level, with 23,259 samples annotated in this manner. The data collection process incorporated face detection techniques to verify a single speaker’s presence in each video, reinforcing the dataset’s monologic characteristics. This characteristic facilitates individual emotional expression study, thus amplifying the reliability of subsequent sentiment analysis and emotion recognition research. The thematic diversity of the CMU-MOSEI dataset is notable, with a spread across 250 different themes. The three predominant themes are comments (16.2%), debates (2.9%), and consultations (1.8%), while the rest of the themes are nearly evenly distributed. Such thematic richness encourages a comprehensive contextual understanding, enabling a more in-depth exploration of sentiment and emotional expressions across varied discussion scenarios. Although CMU-MOSEI is recognized for its substantial scale, diverse contexts, and inherent authenticity and complexity, it is worth acknowledging the dataset’s limitation—the sole use of English. This constraint could potentially reduce the applicability of models trained on this dataset to other languages. Despite this, the CMU-MOSEI dataset remains a vital resource, offering a broad base for developing and validating models for multimodal sentiment analysis and emotion recognition, thereby advancing research frontiers in this area.

2.8. OMG

The One-Minute Gradual-Emotion Recognition (OMG) dataset [31], which incorporates videos from a multitude of YouTube channels, provides a comprehensive view of various emotional behaviors within different contexts. The unique selection process of these videos, centered around the term “monologue”, is designed to depict the gradual evolution of emotions. This approach is integral to the dataset, as it facilitates the modeling of both immediate and long-term emotional trajectories. Every video within the OMG dataset is segmented based on speech, and each of these segments is annotated by at least five independent individuals. This multi-annotator methodology enhances the reliability of the annotations by minimizing individual bias and subjectivity. As part of their assignment, annotators watch the video segments continuously and annotate each one according to the valence/arousal scale and Ekman’s theory of universal emotions. The annotation strategy incorporates both discrete (anger, disgust, fear, happiness, sadness, surprise, and neutral) and continuous (valence and arousal) measures. This combination allows for a holistic examination of emotional states and fosters in-depth investigation and analysis. Comprising 7371 annotated monologue-based speeches, the OMG dataset serves as a robust platform for researchers exploring the dynamism of emotional behaviors within singular narratives. However, the construction of the dataset necessitates certain considerations. The exhibited emotions, influenced primarily by the performers’ interpretive choices, may not genuinely reflect natural emotional expressions. Moreover, the dataset’s reliance on YouTube videos may induce cultural bias skewed towards particular demographics. Despite these potential constraints, the OMG dataset is a formidable resource for the progression of deep learning-based models in multimodal sentiment analysis and emotion recognition. Its focus on gradual emotional change, alongside its detailed annotations, offers researchers an unparalleled opportunity to investigate the complex dynamics of emotional evolution, furthering the boundaries of emotion recognition research.

2.9. MELD

In the domain of MER, the Multimodal EmotionLines Dataset (MELD) [32] offers a unique perspective by focusing on the emotional intricacies present in multi-participant dialogue. The dataset is an organized collection of dialogues from the popular American television series “Friends”, featuring 1433 conversations that account for a total of 13,708 utterances. Each utterance within the MELD dataset is annotated with one of seven emotional labels: anger, disgust, sadness, joy, neutrality, surprise, or fear, providing a diverse emotional spectrum for analysis. In addition to these discrete emotion labels, each utterance also includes an emotional classification of positive, negative, or neutral, thereby facilitating a broader understanding of the sentiment behind each statement. The dataset’s design aligns with its primary objective: to furnish training data that are conducive for developing and refining contextual models in dialogue-based emotion recognition. This dataset, therefore, serves as a remarkable tool for investigating emotional dynamics within multi-party dialogue scenarios, a crucial aspect of human communication that remains underexplored in existing datasets. However, the dataset’s realism and cultural applicability must be considered when utilizing it for research or model development. Since the dialogues originate from a scripted television series, they may not accurately reflect the nuances of spontaneous, real-life conversations, thereby limiting the dataset’s authenticity. Moreover, given the American origin of the television series, the dialogues predominantly represent American cultural norms and idiomatic expressions. Consequently, models trained on this dataset may exhibit limitations when deployed in other cultural contexts, highlighting the importance of cultural diversity in emotion recognition research. Despite these limitations, the MELD dataset undeniably provides an enriching foundation for the study of emotion recognition within conversational contexts. It effectively bridges a gap in the literature by focusing on multi-participant dialogues, and can thus guide the development of more sophisticated and context-aware emotion recognition models in the future.

2.10. SEWA

The SEWA dataset [33], a significant contribution to the field of MER, comprises a diverse array of audio–visual data gathered using web cameras and microphones. It features over 2000 min of footage capturing the emotional responses of 398 individuals across six distinct cultures, thus providing a comprehensive perspective on multicultural emotional expression. The dataset boasts a gender-balanced pool of participants, with 197 females contributing. The participants span a wide age range from 18 to 65 years, offering valuable insights into the variations in emotional expression and recognition across different age groups. They were recorded while either viewing commercials or discussing them in video chats, settings designed to elicit natural emotional reactions and conversations in response to a variety of stimuli. Such diverse emotional contexts enrich the dataset’s content considerably. The SEWA dataset includes 2562 utterances, each annotated with continuous emotion metrics, i.e., valence and arousal. The multicultural and multilingual nature of the dataset represents one of its primary strengths. By incorporating participants from diverse nations and cultural backgrounds, the dataset broadens its applicability, extending its use to a global scale in the realm of emotion recognition research. However, this cultural diversity poses certain challenges. Variations in emotional expression and interpretation due to cultural differences may result in inconsistencies in the emotion annotations. Such discrepancies could potentially cast doubt on the reliability and consistency of these annotations. These variations underscore the complexity of interpreting emotional behavior across cultures, emphasizing the need for culturally aware emotion recognition systems. Despite these challenges, the SEWA dataset provides a unique opportunity for exploring cultural differences in emotional expressions, as well as for training and validating culturally diverse emotion recognition models.

2.11. CH-SIMS

The Chinese single- and multimodal sentiment analysis (CH-SIMS) dataset [34], developed by Yu et al. in 2020, stands as a distinctive contribution to the realm of MER. Consisting of 2281 meticulously refined “in-the-wild” video clips, the average length of each utterance is approximately 3.67 s. In each video clip, only the speaker’s face is visible, thereby eliminating potential distractions or influences from other facial expressions. This unique aspect of the dataset ensures that the emotional state attributed to each utterance pertains exclusively to the speaker, bolstering the accuracy of emotion recognition. The CH-SIMS dataset provides four separate annotations for each utterance: a multimodal annotation and three unimodal annotations. These layered annotations facilitate versatile research avenues, empowering researchers to examine the interplay between different modalities or to conduct unimodal emotion recognition using the distinct unimodal annotations. Emotion categorization within the CH-SIMS dataset is streamlined into five distinct categories: negative, weakly negative, neutral, weakly positive, and positive. This grading system provides a nuanced spectrum of emotional states, further enriching the dataset’s potential for complex emotion recognition tasks. However, despite the advantages offered by the CH-SIMS dataset, certain limitations must be acknowledged. Primarily, the amount of video and audio data compiled within the dataset is relatively limited. Additionally, the dataset exclusively contains Chinese language data, which may inadvertently introduce cultural biases or limitations. While this focus provides an in-depth exploration of Chinese cultural and linguistic emotional expression, it does potentially limit the dataset’s applicability across broader, multicultural emotion recognition research.

2.12. CH-SIMS V2.0

In this study, Liu et al. [35] constructed the CH-SIMS v2.0 dataset, which is an enhanced and extended version of the original CH-SIMS dataset. The dataset now includes 4402 supervised data samples, each annotated with unimodal annotations, along with an impressive assortment of over 10,000 unsupervised data samples. A distinctive characteristic of the CH-SIMS v2.0 dataset lies in its diverse array of contexts for data collection. The inclusion of emotional dramas, interviews, modern television shows, talk shows, video blogs, movies, period dramas, variety shows, and other scenarios has ensured that the dataset mirrors the wide-ranging and dynamic environments in which emotions are expressed in real life. This extensive scope enables the dataset to encapsulate a broad spectrum of emotional expressions across various contexts and scenarios, thereby further enriching its potential for intricate emotion recognition research. The primary objective of the CH-SIMS v2.0 dataset is to emulate realistic human–computer interaction scenarios. This goal reflects a critical application of emotion recognition, as understanding users’ emotions can significantly improve the interaction quality and user experience of embodied conversational agents (ECAs).
In conclusion, although existing datasets have undoubtedly propelled the field of MER, there is a need for more comprehensive and diverse datasets to accommodate an increasingly global and multicultural landscape. Continual updates and expansions of these datasets are crucial to align with evolving cultural, social, and technological trends that shape emotional expression. Moreover, the incorporation of varied modalities and contexts in future datasets is recommended to foster more robust and generalizable MER models. Such developments will lay the groundwork for enhanced human–computer interaction, fostering more empathetic and understanding Artificial Intelligence (AI) systems.

3. Feature Extraction

In the diverse landscape of MER, three prominent channels of emotional expression take center stage: speech, face, and text. Each channel, or modality, offers a unique perspective on emotion, often carrying nuanced emotional cues that are integral to comprehensive emotion recognition. However, unlocking these cues requires the extraction of key emotional features—a pivotal task in MER that has been greatly refined and evolved with the advent of deep learning techniques. In this section, we delve into the realm of feature extraction, exploring the nuances of each modality and the potent methodologies developed to harness their full potential. Our discussion spans the intricacies of textual, facial, and auditory emotional expressions, spotlighting popular and effective methods that have surfaced since the technology’s widespread proliferation.

3.1. Speech Feature Extraction

The emotion-laden speech presents an opportunity to mine acoustic features indicative of the speaker’s emotional state, with the quality of these features exerting a significant influence on the efficacy of emotion recognition. The careful selection of effective speech features that accurately reflect emotional fluctuations is a pivotal challenge within the field of MER. Predominantly, the speech emotion features used can be categorized into two key types: hand-crafted features and deep speech emotion features. The latter are derived using cutting-edge deep learning technologies. This discourse will delve deeper into these categories, shedding light on their unique merits, potential limitations, and the creative methodologies that researchers are employing in the pursuit of augmenting the accuracy of emotion recognition.

3.1.1. Hand-Crafted Feature

Hand-crafted features refer to the features used to describe speech signals that are designed by people based on prior knowledge and professional experience. These features provide us with invaluable insights into the emotional nuances encapsulated within speech, assisting in the complex task of emotion recognition. They are traditionally organized into three main categories: prosodic features, voice quality features, and spectral features.
(1)
Prosodic features: Emotions precipitate physiological changes that directly impact speech production and, thus, significantly affect speech intonation [36]. Prosodic features, capturing variations in pitch, tone, rhythm, and stress in speech, serve as robust indicators of these emotional shifts. As a result, these features have become indispensable in the field of emotion recognition and are widely acknowledged by researchers [37,38]. The array of prosodic features contributing to the depiction of emotions in speech encompasses, among others, fundamental frequency (F0), energy, and speech rate [39,40,41,42]. Among them, F0 is considered one of the most crucial features to predict emotional states in speech. Relevant studies have shown that people’s F0 tends to be higher with a wider range when they are in an excited emotional state such as anger, lower and with a narrower range when they are in a depressed emotional state such as sadness, and relatively stable when they are in a calm emotional state [42]. In addition to F0, energy (i.e., intensity) in speech is also a vital feature. Scherer et al. [38] found that high-frequency energy increases when people experience fear, happiness, and anger, while high-frequency energy decreases during times of sadness. Furthermore, Nwe et al. [43] pointed out that emotions with higher activation levels, such as anger, surprise, and happiness, generally have higher intensity, while emotions like fear, sadness, and disgust have lower intensity. Furthermore, speech rate, the speed at which words are spoken, greatly affects emotion detection. Research indicates that the speech rate tends to be faster for fear, disgust, anger, and happiness; normal for surprise; and slower for sadness [44]. Consequently, fluctuations in the speech rate serve as a vital cue in distinguishing different emotional states. In conclusion, the careful analysis and extraction of prosodic features offer a comprehensive, although not exhaustive, approach to understanding and recognizing emotional states through speech. They form a crucial part of hand-crafted features used in MER, with each contributing uniquely to the richness of emotion recognition. Nevertheless, it is important to remember that prosodic features form just one aspect of emotional recognition, and should be used in concert with other features for an all-encompassing, nuanced understanding of emotions.
(2)
Voice quality features: Voice quality features are primarily utilized to assess the purity, clarity, and intelligibility of speech. These are affected by acoustic performances such as wheezing, choking, and tremors, which frequently arise when the speaker is emotionally excited or has difficulty calming down. Voice quality features include formant, harmonic-to-noise ratio (HNR), jitter, and shimmer. The formant reflects the characteristics of the vocal tract, which is a significant parameter impacted by emotional speech. According to previous research findings, compared with neutral speech, the first formant frequency of happy and angry speech is slightly higher, while the first formant frequency of sad speech is significantly lower [36,42]. Another crucial voice quality feature is the HNR, which represents the ratio of harmonic components to noise components in speech. It is worth noting that the noise in this context is not environmental noise, but glottal noise arising from incomplete closure of the vocal folds during speech production. Jitter is a term for the degree of rapid and repetitive changes in the measured fundamental frequency, which mainly indicates roughness in the sound and, to a lesser extent, the degree of hoarseness commonly observed in anxious speech. Shimmer mainly demonstrates the degree of hoarseness by reflecting changes in the amplitude of voice between adjacent vibratory cycles. In conclusion, voice quality features play an indispensable role in extracting valuable insights from speech for emotion recognition. The concurrent use of these features substantially enhances the breadth and depth of emotion recognition, thus facilitating a more sophisticated comprehension of emotional states in speech.
(3)
Spectral features: Spectrum-based correlation features are usually short-term representations of speech signals, as speech signals are generated from the coordinated movement of multiple articulators, and the physical characteristics of the articulators determine that speech is difficult to change significantly in a short period of time. Typically, speech signals ranging from 5–50 ms are relatively stable. When recognizing emotions in speech, the most important perceptual properties are reflected in the power spectrum. Emotional expression in speech has a great impact on the distribution of spectral energy in different frequency bands. Spectral features encompass linear and cepstral features, wherein the former involves Linear Prediction Coefficients (LPC), and Log-Frequency Power Coefficients (LFPC), and the latter involves Mel Frequency Cepstral Coefficients (MFCC) and Linear Prediction Cepstral Coefficients (LPCC). Among them, MFCCs, arguably the most commonly used feature in speech signal processing, capture the power spectrum of speech on a nonlinear mel scale, which approximates the human auditory system’s response. The use of the mel scale makes MFCCs particularly valuable when capturing perceptually relevant spectral shapes, thereby enabling effective emotion recognition. To conclude, spectral feature extraction and analysis present a robust methodology for comprehending and recognizing emotional states via speech. Each feature type provides unique insights, collectively contributing significantly to the richness of emotion recognition in multimodal emotion recognition frameworks. However, it is pivotal to remember that spectral features represent merely one facet of emotion recognition and should be employed in combination with other features for a comprehensive and nuanced emotional understanding.
In current MER, it is common practice to use OpenSMILE [45], COVAREP [46], OpenEAR [47], and librosa [48] to extract the aforementioned acoustic features that represent the emotional characteristics of the speech modality. Furthermore, the robustness and generalizability of these features may vary across different speech contexts and individual speakers, owing to the subjective and dynamic nature of emotional expression. Therefore, while hand-crafted features provide us with a powerful tool to harness emotional cues from speech, they should be used in conjunction with other methodologies and modalities to attain a comprehensive understanding of emotional states. This serves to enhance the richness and precision of emotion recognition in a multimodal setting.

3.1.2. Deep Feature

Although hand-crafted features can describe the basic acoustic properties and emotional information of speech signals well, this extraction method requires prior knowledge and professional experience, and it is difficult to cover all speech emotion features. As a result, an increasing number of researchers are turning to deep learning technology to extract speech features automatically in the current field of SER. These methods typically involve using deep learning models to learn how to represent speech signals and generate a set of feature vectors that represent their deep features. Currently, a common approach involves using self-supervised training methods to learn high-quality representations from unlabeled data, and then using these representations to replace hand-crafted features to complete downstream tasks. This is a research hotspot in the field of speech emotion [49,50,51,52,53]. For example, Schneider et al. [54] proposed an unsupervised speech pre-training method named wav2vec that only uses unlabeled speech data for self-supervised training to obtain generalized speech features that can be applied to multiple tasks. Note that wav2vec directly trains on speech’s raw waveform to extract deep features that represent speech signals. Subsequently, Baevski et al. [55] developed a new self-supervised learning framework for speech named wav2vec 2.0. Empirical evidence underscored the superior performance and generalization of wav2vec 2.0 over its predecessor. Hsu et al. [56] introduced a novel self-supervised learning framework for speech representation learning named HuBERT, which employs a more direct predictive loss by separating the acoustic unit discovery step from the masked prediction representation learning phase than wav2vec 2.0. Most recently, Chen et al. introduced WavLM [57], the first universal speech pre-training model capable of being applied to a broad range of tasks, including speaker recognition, speech recognition, and emotion recognition. Experimental results indicate that WavLM outperforms all preceding models on these tasks, pointing to the promise and potential of deep learning in enhancing SER. Despite these advancements, there are challenges inherent to deep feature extraction in SER. Deep learning models require large volumes of data for training, which may not always be available or feasible to obtain in specific contexts. Additionally, the self-supervised learning models, though efficient, still struggle with differentiating finer emotional nuances in speech. Finally, despite the high performance of models like WavLM, they are computationally intensive and demand substantial resources, posing a challenge for real-time applications. Future research must aim to address these challenges to further enhance the effectiveness and applicability of deep feature extraction in SER.

3.2. Textual Feature Extraction

Textual emotions are typically conveyed through textual information, and extracting emotional features from text is critical to recognizing emotions in text. Traditional methods for extracting emotional features use the bag-of-words (BoW) model [58,59]. Although the BoW model has straightforward processing steps that are easy to understand, it neglects word order and grammar relationships, resulting in a loss of contextual information. As a consequence, it is unable to capture the connections and structured information between words. Word embedding techniques have been proposed to overcome these limitations. Word embedding is a feature extraction technique that uses neural networks to learn representations of text, such that words with similar meanings have similar representations. Pre-trained deep learning-based word embedding models have been widely applied in text emotion recognition tasks. Earlier word embedding models, e.g., word2vec [60] and GloVe [61], predominantly relied on syntactic contexts for training. These models were trained on vast amounts of unlabeled data, aiming to capture intricate grammatical and semantic rules. Pre-trained word embedding models generally perform better than those initialized randomly. However, a limitation of these early models is their assumption that every word has a unique vector representation, irrespective of its monosemous or polysemous nature. This limitation overlooks diverse contextual information, compromising the models’ effectiveness.
To mitigate the highlighted limitations, Peters et al. introduced the ELMo (Embeddings from Language Models) [62] method. It produces deep contextualized word embeddings by capturing semantic variations based on context. Distinct from conventional language models, ELMo adopts a profound bidirectional architecture, accounting for both left and right contexts within a sentence. This structure can capture intricate word features, such as synonyms and polysemy, across different linguistic contexts. Later, Devlin et al. introduced BERT (Bidirectional Encoder Representations from Transformers) [63], a transformer-based encoder with bidirectional contextualized word representations. Unlike traditional language models, BERT simultaneously considers both left and right contexts to predict the next word in a sentence. It pre-trains deep, bidirectional contextualized word representations from large amounts of unmarked text through unsupervised learning.
Nonetheless, it is imperative to note that the models mentioned above do not explicitly account for emotional information. Addressing this gap, Tang et al. [64] pioneered the Sentiment-Specific Word Embeddings (SSWE). Empirical results have shown its superior efficacy over other embeddings in integrating emotion labels at the word level for emotion-centric tasks. Following this, Xu et al. [65] unveiled Emo2Vec, embedding emotional semantics into fixed-size vector representations at the word level. Experiments revealed that Emo2Vec surpasses conventional word embedding techniques in numerous emotion-driven tasks. Moreover, Emo2Vec’s training methodology bears significant relevance for feature extraction in diverse domains.

3.3. Facial Feature Extraction

Facial expressions have always been essential to recognizing the emotions of the speakers in MER. Therefore, obtaining features that represent facial emotional information for affective discrimination is imperative. Traditional facial feature extraction methods include Local Binary Pattern (LBP) [66], Active Appearance Model (AAM) [67], Active Shape Model (ASM) [68], Scale Invariant Feature Transform (SIFT) [69], Histograms of Oriented Gradient (HOG) [70], and Gabor Wavelet Transform [71]. However, traditional facial feature extraction methods require manual feature extraction, which is time-consuming and labor-intensive. Because facial semantic information is more diverse than that of other images, manual feature extraction can overlook significant semantic information. Deep learning models, especially convolutional neural networks (CNN), can automatically learn the optimal features without prior intervention, eliminating the tedious process of manual feature extraction. Furthermore, deep learning frameworks can perform robust and accurate feature learning in supervised and unsupervised settings, and generate optimal performance in various applications, including digital recognition, image classification, and feature learning. Consequently, an increasing number of researchers have adopted deep learning algorithms to extract facial features in affective computing tasks. For instance, Tran et al. [72] proposed a simple and efficient three-dimensional convolutional neural network (3D-CNN) for extracting spatiotemporal features. This model can be used for multiple tasks, such as action recognition and dynamic scene recognition. Yu et al. [73] proposed a new architecture called spatio-temporal convolutional features with nested LSTM (STC-LSTM), which is based on three deep learning subnetworks: 3DCNN, T-LSTM, and C-LSTM. 3DCNN is used for extracting spatiotemporal features, T-LSTM is used for maintaining temporal dynamics, and C-LSTM is used for modeling multi-level features. At present, commonly utilized libraries consist of OKAO, Computer Expression Recognition Toolbox (CERT) [74], OpenFace [75], and FACET. OKAO Vision detects and extracts facial features from individual frames, providing measures of smile intensity (ranging from 0 to 100) and eye gaze direction. The CERT can automatically extract smiles and estimates head poses and facial action units. The OpenFace 2.0 toolkit provides data on 68 facial landmarks, 17 facial action units, head poses, head orientation, and eye gaze. Finally, the FACET library extracts a multitude of visual features including facial action units, facial landmarks, head poses, gaze tracking, and HOG features [76].

4. Multimodal Fusion Methods

Emotion, as a complex interplay of psychological and physiological processes, operates largely via non-linguistic means, thereby adding a layer of complexity to its recognition. Unimodal information, limited in expressiveness by nature, often falls short of capturing the full spectrum of emotional states. This limitation emphasizes the need for the integration of multiple modalities of information, leading to a more comprehensive understanding of emotions and subsequently improving emotion recognition efficacy. Various experimental studies have attested to the superior performance of MER models over unimodal ones, placing MER at the vanguard of research within affective computing [77,78,79,80]. The crux of MER lies in its fusion process, which amalgamates varied types of signals to derive complementary information, which is mutually supportive across the modalities. This crucial step significantly contributes to improving the overall performance of the MER model. The fusion of different types of signals to obtain mutually supportive complementary information is central to improving model performance in MER. We categorize the current deep learning-based MER fusion methods into two main types. The first type is model-agnostic fusion methods, where the fusion process is not directly dependent on a specific deep learning model. The second type is fusion methods based on the intermediate layers of the deep models (i.e., intermediate layer fusion), which performs the fusion within the deep network. For an overview of these methods, please refer to Figure 2.

4.1. Model-Agnostic Fusion

Model-agnostic fusion methods are sub-categorized as early fusion, which integrates immediately after feature extraction, usually by simply connecting various feature representations; late fusion, which consolidates decisions (like classification or regression) subsequent to individual modality decision-making; and hybrid fusion, an amalgamation of both early and late fusion techniques. Although model-agnostic fusion methods can be implemented using nearly any single-modality classifier or regressor, they are not typically crafted for multi-modal data [81]. In the subsequent sections, we will delve deeper into these fusion techniques.

4.1.1. Early Fusion

In the realm of MER, early fusion, synonymous with feature-level fusion, holds significant prominence. Figure 3 shows the framework of multimodal emotion recognition based on early fusion. It is characterized by its simplicity and lower computational complexity. In this approach, features from various modalities, e.g., text, speech, and vision, are combined through a concatenation operation. These amalgamated features are then utilized as inputs for a Deep Neural Network (DNN) to undergo classification training. Poria et al. [82] were pioneers in the implementation of this fusion method. They extracted feature vectors from the speech, vision, and text modalities, amalgamated them into a comprehensive feature vector, and then utilized this unified vector as an input for their classification model. This laid the groundwork for further explorations in early fusion. In a following study, Huang et al. [83] enhanced data richness to evaluate the efficacy of early fusion. This was achieved during the 2018 Audio/Visual Emotion Challenge (AVEC), in which the researchers augmented the training data by replacing the original training samples with shorter overlapping samples extracted from them, thus multiplying the number of training samples. This amplification of training data diversity, coupled with the implementation of early fusion, illuminated the potential of this method. Further progress in early fusion was seen in the work of Williams et al. [84].They constructed a multimodal fusion system that discerns emotions and their intensities. By carrying out feature fusion at the input level and employing a Bidirectional Long Short-Term Memory (BLSTM) model for sequence learning, they underscored that early fusion can effectively predict the presence of multi-label emotions and their coarse-grained intensities. The primary advantage of early fusion is its ability to enable early correlations among distinct multimodal features, thereby significantly improving the model’s emotion recognition performance. This advantage has solidified the role of early fusion in MER applications. Nevertheless, the simplicity of early fusion also presents certain limitations. The mere concatenation of different modal data might result in the model’s inability to process the unique features of each modality independently. As a result, vital emotional information from certain modalities may be overlooked. Additionally, this straightforward concatenation method cannot effectively filter out conflicting or redundant information from different modalities. Additionally, early fusion faces a critical challenge in addressing the time synchronization issue across various data sources. Given that features harvested from different modalities have varying time scales, measurement levels, and temporal structures, achieving harmony among them poses a complex task. In summary, while early fusion has proven instrumental in MER, researchers must address its limitations to optimize this method further. This ongoing exploration offers a fascinating area of focus in the continuing evolution of MER.

4.1.2. Late Fusion

Late fusion methods, also referred to as decision-level fusion methods, are widely employed in multimodal affect recognition. The basic idea is to independently extract features and train models for each modality, and then combine their prediction results through averaging, weighted sum, majority voting, or deep neural networks to obtain the final emotion category. We show late fusion in Figure 4. Due to its feature-independent fusion process and uncorrelated mistakes from different classifiers, this fusion method has garnered significant attention. Several researchers have demonstrated the application of this method. Poria et al. [85], for example, developed separate unimodal affect recognition models for speech, text, and visual modalities in 2016. Following the training phase, these models produced prediction probabilities for each modality, which were then integrated through weighted fusion to produce the final prediction. Furthermore, Huang et al. [86] adopted Long Short-Term Memory (LSTM) networks to create affect prediction models that fully leverage affective information from each modality. They formulated prediction models for various feature types within each modality and used Support Vector Regression (SVR) as the fusion model in the decision layer. In an attempt to augment performance, Su et al. [87] proposed a multi-level segmented decision-level fusion model for affect recognition. Their model incorporated Bidirectional LSTM (BLSTM) as a multi-level segmental affect feature learning model and SVR as the fusion model at the decision level. Notably, the BLSTM was capable of effectively modeling different forms of affective information, considering the influence of previous and subsequent affective features on the current outcome. Meanwhile, SVR assisted in compensating for redundant information in affect recognition. Recently, Sun et al. [88,89] employed LSTM and a self-attention mechanism to capture complex temporal dependencies in feature sequences. They then fused the prediction results of several unimodal affect recognition models using a second-level LSTM as the fusion model. Compared to feature-level fusion, decision-level fusion is simpler, more flexible, and does not require the temporal synchronization of modality features. Additionally, during the late fusion process, each modality can utilize the most suitable classifier or model to learn its features, thus leading to improved local decision outcomes. However, late fusion presumes that each modality operates independently, thereby ignoring the interaction information between different modalities and potentially limiting the accuracy of the final affect prediction. Future investigations should explore fusion methods that preserve the advantages of late fusion while incorporating inter-modality interactions to enhance affect recognition accuracy.

4.1.3. Hybrid Fusion

Unlike the previously discussed early, late, and intermediate fusion methods, hybrid multimodal fusion provides a distinct approach, integrating the strengths of both early and late fusion strategies, as depicted in Figure 5. The goal of this fusion method is to maximize the utilization of emotional information extracted from multimodal data, while also considering the interplay and synergy between different modalities. For example, Wollmer et al. [27] proposed a hybrid fusion method that uses Bidirectional Long Short-Term Memory (BLSTM) to merge audio and visual features at the feature level. Following this, they integrated the fused outcomes with the predictions from a text classifier using decision-level fusion, a strategy typically found in late fusion methods. By using this hybrid approach, their method benefits from both the high-level abstraction of feature interactions provided by early fusion and the modality-specific expertise of separate classifiers provided by late fusion. Within the field of Music Emotion Recognition (MER), Nemati et al. [90] introduced a comparable hybrid fusion methodology. Initially, they mapped the speech and visual modalities onto a shared latent space using feature-level fusion, a technique similar to early fusion methodologies. These resultant latent features are subsequently utilized for emotion classification. Concurrently, they train an independent supervised classification model using text features. In the final stage, they apply a decision-level fusion method, underpinned by the Dempster–Shafer theory, to amalgamate the outcomes of the text and audio–visual classifications. This hybrid strategy enables the utilization of inter-modality interactions at an early phase while preserving the flexibility of independent modality classification at a later stage. To conclude, hybrid multimodal fusion techniques endeavor to exploit the unique advantages of both early and late fusion methods, aiming to deliver a more resilient and comprehensive affect recognition system. Nonetheless, the complexity of these methods also presents challenges, particularly when it comes to determining the optimal blend of early and late fusion strategies and managing the escalated computational demands. Consequently, ongoing research and development in hybrid fusion methods remain a promising and critical direction for the advancement of the field of multimodal affect recognition.

4.2. Intermediate Layer Fusion

Diverging from the previously mentioned early, late, and hybrid fusion techniques, intermediate fusion designates a distinctive approach that focuses on the fusion of modalities within the network layer of deep neural network models. This technique bestows upon multimodal recognition models the capacity for comprehensive, end-to-end emotion recognition, while fully modeling the intra- and inter-modality interactions. The strength of intermediate fusion lies in its ability to capitalize on the strengths of deep learning, namely its capacity to learn intricate patterns and represent high-dimensional data effectively. This stands in contrast with early fusion methods, which may limit the capability to model complex interactions due to preprocessing steps, and late fusion techniques, which often lack the capacity to capture detailed inter-modality relationships due to the segregation of modality training. Notably, the application of intermediate fusion is not a one-size-fits-all method. The focus of fusion methods, based on deep model intermediate layers, is the modeling of interactions between modalities. These methods can be further categorized based on the type of interaction between modalities as follows: simple concatenated fusion, which does not explicitly model inter-modality interactions; utterance-level interaction fusion, which models inter-modality interactions based on utterance-level features; and fine-grained interaction fusion, which is based on sequence features of different modalities to complete the inter-modality interactions.

4.2.1. Simple Concatenation Fusion

Simple concatenation fusion based on deep neural networks refers to the process in which MER models independently analyze the input features of different modalities and directly concatenate these features as the input of the next layer network to learn the interaction information between modalities. Nguyen et al. [91] proposed a novel multimodal sentiment recognition model that utilizes three convolutional neural networks to separately process the low-level features of speech, text, and visual modalities. The aim is to attain high-level emotional feature representations. Subsequently, the feature vectors from each modality are merged into a single vector and injected into the emotion classification network for sentiment evaluation. Similarly, Tripathi et al. [92] adopt the optimal feature extraction network for each modality and execute concatenation fusion in the final layer for classification. Experimental evidence has proven that this approach can create more robust and precise emotion detection models. Ortega et al. [93] proposed a new deep neural network for multimodal fusion emotion recognition in audio, video, and text modalities. This deep neural network architecture includes independent and shared layers, aiming to learn the representations of each modality as well as the optimal combined representation to achieve the best prediction performance. Experimental results achieved using the AVEC wild emotion analysis dataset indicate that the proposed deep neural network model outperforms early and late fusion approaches. Liang et al. [94] postulate that despite the potential misalignment of emotional expressions at the frame level across different modalities, the overall emotional states should exhibit similarity at a coarse-grained discourse level. Leveraging this assumption, they train the encoder via semi-supervised learning, using unlabeled video data to secure improved encoder representations. Given that the representation vectors obtained for each modality are similar, the authors simply concatenate them and conduct direct classification. Yu et al. [95] proposed a self-supervised multi-task multimodal sentiment analysis network (Self-MM). In the Self-MM model, representations from different modalities are fused using the simplest concatenation method. Despite the apparent similarity between simple concatenation fusion and early fusion, noteworthy distinctions exist between the two. Furthermore, Han et al. [96] pointed out that the aforementioned methods lack control in the information flow from raw input to fused embedding, which may lead to information loss and an increased risk of unintended noise. To address this issue, they introduced the concept from information theory—mutual information (MI)—and proposed a hierarchical mutual information maximization framework for MER (i.e., MultiModal InfoMax, MMIM). In MMIM, MI maximization is performed at both the input and fusion layers. By enhancing the MI between multi-modal inputs, it can effectively filter out modality-specific noise irrelevant to the task while preserving modality-invariant content across all modalities. MMIM also applies MI maximization between the fusion result and input modalities again, ensuring that the fusion output sufficiently captures modality-invariant clues among the modalities. Like Han et al. [96], Zheng et al. [97] also introduced a multimodal representation model, MMMIE, grounded in the principles of mutual information maximization, minimization, and identity embedding. This model aims to maximize the mutual information between modalities while minimizing the mutual information between input data and its features, extracting modality-invariant and task-related information. Moreover, by incorporating identity embedding, MMMIE bolsters the context-aware capabilities of the downstream network. Subsequently, Mai et al. [98] aimed to minimize redundancy in the fusion process by introducing the Multimodal Information Bottleneck (MIB) for MER. This method provides a non-redundant, robust multimodal representation for MER. Drawing inspiration from the general Information Bottleneck (IB) principle, the MIB is designed to learn the minimal sufficient representation for a given task. It accomplishes this by maximizing the mutual information between the representation and the target while simultaneously constraining the mutual information between the representation and the input data. In simple fusion, the model conducts modality-specific modeling on each modality’s features prior to executing feature concatenation fusion. Therefore, simple concatenation fusion not only addresses the issue of modality-specific modeling, which cannot be attained via early fusion, but also eradicates the need to consider the temporal synchronization of different modalities’ features. However, simple concatenation fusion might not fully learn the interaction information between modalities, posing a limitation. This overview of simple concatenation fusion underscores its utility in MER while also acknowledging its potential limitations. As research in this field advances, the exploration of other fusion methods and their combination with simple concatenation fusion may hold promising potential for further enhancing emotion recognition systems.

4.2.2. Utterance-Level Interaction Fusion

The technique of utterance-level interaction fusion explicitly models the features at the utterance-level across different modalities instead of merely concatenating them. Importantly, utterance-level features represent the emotional information of each modality by utilizing a single feature vector. One of the pioneers in this area is Zadeh et al. [99], who proposed the Tensor Fusion Network (TFN) for MER. The TFN consists of three primary components: modality embedding subnetworks, a tensor fusion layer, and a sentiment inference subnetwork. In the first stage, modality embedding subnetworks model intra-modality dynamics, extracting a rich modality embedding, namely, the utterance-level features. Next, the tensor fusion layer models unimodal, bimodal, and trimodal interactions using a three-fold Cartesian product derived from utterance-level features. Lastly, the sentiment inference subnetwork conducts sentiment inference. Continuing the exploration in this field, Liu et al. [100] proposed Low-rank Multimodal Fusion (LMF) to achieve the interaction between modalities based on utterance-level features. The experiment confirmed that the LMF effectively improves the training and testing efficiency compared to TFN. Following the inception of TFN and LMF, various other methodologies were proposed to further improve TFN. These include Temporal Tensor Fusion Network (T2FN) [101], Hierarchical Feature Fusion Network (HFFN) [102], Modality-based Redundancy Reduction Fusion (MRRF) [103], and Semi-Tensor Product (STP) [104]. Each of these methodologies contributes unique perspectives and solutions to MER. Notably, Mai et al. [105] introduced a pioneering adversarial encoder–decoder classifier framework aimed at acquiring a modality-invariant embedding space. Considering the distinct distributions across modalities, the authors employed adversarial training to align the distribution of the source modality with that of the target modality. Moreover, the embedding space was subject to additional constraints by incorporating reconstruction loss and classification loss. Lastly, a hierarchical graph neural network was employed to integrate the encoded representations, explicitly exploring the interactions between unimodal, bimodal, and trimodal information across multiple stages. Additionally, Wu et al. [106] developed a fusion network named Bimodal Information-augmented Multi-Head Attention (BIMHA), which relies on multi-head attention and comprises a total of four layers. The first layer captures modality-specific dynamics within a single view. The second layer represents cross-view dynamics. In the third layer, bimodal interaction information is extracted by utilizing a multi-head attention mechanism, which calculates bimodal attention for weight allocation in feature attention. Lastly, the independent modality embeddings are concatenated with bimodal attention to form a multimodal representation used for predicting the emotion of each utterance. Recently, Liu et al. [107] devised a multimodal fusion network with complementarity and importance for emotion recognition. This network takes into account the varying importance of different modalities and employs an importance attention network to assign weights accordingly. Additionally, given the complementary relationship between modalities, this approach constructs a complementary attention network to enhance feature fusion and achieve effective interaction among multimodal features. In summary, utterance-level interaction fusion has proven to be an effective approach for MER, with significant strides made in the development of various network structures and fusion methods. As the field continues to progress, further exploration and refinement of these methodologies is expected to improve both the efficiency and accuracy of MER.

4.2.3. Fine-Grained Interaction Fusion

Although multimodal feature fusion methods based on utterance-level features can use global features for prediction, they ignore fine-grained interactions between modalities in the modeling process, which may fail to capture local information [108]. Hence, another option is multimodal feature interaction methods based on fine-grained features. Multimodal streams collected from different modal sequences often exhibit inherent asynchrony due to variations in their receiving frequencies. Since the fine-grained inter-modality interaction needs to match the alignment relationship between modalities, the fine-grained features are often divided into aligned word-level features and unaligned features. Aligned word-level features refer to features that have been forced to align before interacting. To extract word-level features, the first step is to apply forced alignment to obtain timestamps for each word, including start and end times. The utterance is then divided into segments according to the timestamps. Finally, word-level acoustic or visual features are obtained by averaging frame-level features of speech segments. Based on word-level features, many methods for word-level multimodal MSA have been proposed. To the best of our knowledge, Chen et al. [109] first based the idea of word-level modality fusion to propose a gated multimodal embedding LSTM with temporal attention (GME-LSTM(A)) model, which consists of two modules, i.e., gated multimodal embedding and LSTM with temporal attention. The gated multimodal embedding alleviated the difficulty of fusion when noisy modalities were present. The LSTMs with temporal attention performed word-level fusion between input modalities at a finer interaction and focused on the most significant time steps. Zadeh et al. [110] proposed a novel network for understanding human communication based on word-level features called Multi-Attention Recurrent Network (MARN). MARN includes two key components: long-short-term mixed (LSTHM) memory and multi-attention blocks (MAB). Hybrid memory LSTHM was an extension of LSTM that carries mixed information by reconstructing memory components. MAB was used to discover cross-view dynamics across different modalities. LSTHM was expanded at the word level at time steps, while MAB was applied at each time step to achieve fine-grained interactions between modalities. Subsequently, Zadeh et al. [111] proposed a memory fusion network (MFN) for multi-view sequential learning, which also used word-level features as interactive features. The MFN first used the LSTMs module system to model the intra-modality interaction, then used the delta-memory attention network (DMAN) module to achieve fine-grained inter-modality interactions. Zadeh et al. [30] introduced a novel multimodal fusion technique known as Dynamic Fusion Graph (DFG) to investigate the interplay between modalities in human multimodal language. DFG is an improved version of the MFN in which the original fusion method, DMAN, has been replaced with DFG. The resulting model is named the Graph Memory Fusion Network (Graph-MFN). Furthermore, there are several other enhanced models based on MFN, e.g., M3ER [112]. Liang et al. [113] presented Recurrent Multistage Fusion Network (RMFN), which decomposes the fusion problem into multiple stages. Each stage focuses on a subset of multimodal signals to achieve specialized and effective fusion. This multi-stage fusion approach models cross-modal interactions by leveraging the intermediate representations from previous stages. Integrating this fusion method with a recurrent neural network system enables the simulation of temporal and intra-modal interactions.
According to the above description, the aforementioned studies based on word-level features require forced alignment of text and speech, which is time-consuming and labor-intensive. To handle this problem, researchers have begun to study fine-grained interactions between modalities based on unidentified features. In this case, the core problem is how to match the alignment relationship between modalities automatically. Since the attention mechanism can measure the similarity between features, the feature alignment between modalities can be achieved based on the attention mechanism. Currently, multimodal MER based on attention mechanism has attracted widespread attention. For example, Tsai et al. [114] proposed a multimodal Transformer (MULT) for unaligned multimodal language sequences. The core of the Mult model is the crossmodal attention, which attends to fine-grained interactions between multimodal sequences across distinct time steps and latently adapts streams from one modality to another. Although the multimodal Transformer (MulT) method extends the self-attention mechanism of the original Transformer network to learn cross-modal dependencies between elements, directly copying the self-attention is influenced by the mismatch between different modal features, leading to potentially unreliable cross-modal dependencies. Based on this observation, Liang et al. [115] introduced the Modality-Invariant Cross-Modal Attention (MICA) method, which learns cross-modal interactions in a modality-invariant space and effectively solves the sequence matching problem in unaligned features. The effectiveness of the MICA method has been validated through experiments on multiple datasets. Additionally, to tackle the problem of cross-modal asynchrony in multimodal sequences, Lv et al. [116] introduced the Progressive Modality Reinforcement (PMR) method, which relies on cross-modal Transformers. This method incorporates a message hub that facilitates information exchange among modalities. The message hub sends mutually shared messages to each modality, reinforcing their features through cross-modal attention. Simultaneously, the message hub accumulates reinforced features from each modality and leverages them to generate enhanced common messages. This iterative process enables the gradual complementary integration of shared information and modality-specific features. Ultimately, the enhanced features are employed for sentiment prediction tasks. Tzirakis et al. [117] presented a textual architecture based on Transformers and an attention-based fusion strategy to effectively integrate diverse modal features and enhance sentiment recognition performance. The proposed textual model employs multilinear projection and context-aware feature generators to capture sentence semantics. Moreover, the proposed fusion strategy achieves superior balance among the relationships across different modalities compared to a straightforward concatenation approach, resulting in enhanced recognition performance. In summary, fine-grained interaction fusion methods provide an innovative avenue for capturing nuanced interactions between modalities, proving to be pivotal in advancing MER. These methods ensure that both global and local information is integrated for precise emotion recognition. While advancements have been made in this sphere, continuous research is warranted to refine these methods further and develop more robust systems for MER.

5. Evaluation Metrics

When we delve into deep MER models, we find that their methods of handling continuous emotions and discrete emotions differ. For the prediction of discrete emotions, the model typically employs the cross-entropy loss function, which is a commonly used loss function in classification problems, aiming to minimize the difference between the probability distribution predicted by the model and the probability distribution of the true labels.
L = i = 1 N y i log ( p i )
where N is the number of classes, y i is the true label, and p i is the probability predicted by the model for that category.
As for the prediction of continuous emotions, due to their nature leaning more toward regression problems, the model tends to use MSE (Mean Squared Error) and MAE (Mean Absolute Error) as loss functions.
MSE = 1 M i = 1 M ( y i y ^ i ) 2 , MAE = 1 M i = 1 M | y i y ^ i |
where M is the total number of samples, y i is the true continuous emotion value, and y ^ i is the predicted value by the model.
As a result, in this survey, we summarize two primary types of evaluation metrics for emotion recognition models: one for classification models, denoted as classification evaluation metrics, and another for continuous emotion prediction models, referred to as regression evaluation metrics.

5.1. Classification Evaluation Metrics

5.1.1. Weighted Average Accuracy (ACC)

The ACC is a widely adopted evaluation metric for emotion classification tasks. It considers the imbalanced distribution of samples across various categories and computes the weighted accuracy for each category based on its sample proportion to acquire the overall accuracy. The calculation formula is as follows [118]:
ACC = 1 N i = 1 N w i · T P i + T N i T P i + T N i + F P i + F N i
where w i is the weight of class i , T P i is the number of true positives in class i , T N i is the number of true negatives in class i , F P i is the number of false positives in class i, and F N i is the number of false negatives in class i.

5.1.2. Unweighted Average Accuracy (UACC):

The UACC is calculated by computing the mean of accuracies for different emotion categories while disregarding any imbalances between these categories. The following formula represents the calculation process [118]:
UACC = 1 N i = 1 N T P i + T N i T P i + T N i + F P i + F N i

5.1.3. Weighted Average F1 (F1)

The F1 score is a widely employed evaluation metric in the field of emotion classification. It is calculated by combining precision and recall according to the formula provided below [118]:
F 1 = 1 N i = 1 N w i · 2 · Precision i · Recall i Precision i + Recall i
where Precision i = T P i T P i + F P i is the precision of class i, and Recall i = T P i T P i + F N i is the recall of class i.

5.1.4. Unweighted Average F1 (UF1)

The UF1 score computes the arithmetic average of F1 scores for each emotion category. Its calculation formula is shown below [118].
UF 1 = 1 N i = 1 N 2 · Precision i · Recall i Precision i + Recall i

5.2. Regression Evaluation Metrics

5.2.1. Mean Squared Error (MSE)

The MSE is a commonly employed evaluation metric for regression tasks. It calculates the average of the squared differences between the predicted values and the true labels. For continuous emotion prediction tasks, a lower MSE signifies that the model’s predictions align more closely with the true emotion values [119].

5.2.2. Root Mean Squared Error (RMSE)

The RMSE is obtained by calculating the square root of the MSE. A lower RMSE indicates fewer prediction errors in the model. The calculation process is outlined as follows [119]:
RMSE = MSE

5.2.3. Pearson Correlation Coefficient (PCC)

The PCC quantifies the linear relationship between the predicted values of a model and the actual emotional values. PCC values fall within the range of −1 to 1, where values near 1 signify a strong positive correlation, values near −1 denote a significant negative correlation, and values near 0 signify no correlation. The PCC is calculated as follows [120]:
PCC = i = 1 N ( y i μ y ) ( y ^ i μ y ^ ) i = 1 M ( y i μ y ) 2 · i = 1 M ( y ^ i μ y ^ ) 2
where μ y and μ y ^ are the means of actual and predicted emotion scores.

5.2.4. Concordance Correlation Coefficient (CCC)

The CCC combines the advantages of PCC and MSE. It not only captures the co-variation relationship between predictions and ground truth but also reflects its deviation. Therefore, it provides a better reflection of the alignment between predictions and ground truth, making it a widely utilized performance evaluation metric for continuous-dimensional sentiment prediction. Higher CCC values indicate strong performance in terms of consistency and accuracy. The specific calculation process for CCC is as follows [119]:
CCC = 2 · ρ · σ y · σ y ^ σ y 2 + σ y ^ 2 + ( μ y μ y ^ ) 2
where ρ is the Pearson correlation coefficient and σ y and σ y ^ are the standard deviations of actual and predicted emotion scores.

6. Analysis of Research Results

Multimodal emotion recognition aims to integrate diverse modalities, including voice, text, and facial expressions, to precisely recognize and classify emotions. Table 2 provides an overview of recent seminal studies in this domain, highlighting 17 advanced multimodal emotion recognition techniques based on features, fusion strategies, datasets, and evaluation metrics. An analysis of this table reveals the following insights:
  • Vocally, most studies prefer feature extraction using the COVAREP and openSMILE tools, which are open-source and adept at extracting a wide range of voice-related features. However, there is a growing inclination towards deep learning-based feature extraction, as illustrated by the significant outcomes achieved by Siriwardhana and colleagues using the Wav2Vec approach on the IEMOCAP and MELD datasets.
  • In the textual domain, initial studies primarily relied on word vector models such as word2vec and Glove. With the evolution of deep learning, recent methods show a preference for sophisticated models like BERT and RoBERTa, which excel at capturing textual contextual nuances.
  • Techniques for extracting facial expression features are diverse, encompassing methods like 3D-CNN, FACET, and OpenFace. Additionally, some studies are delving into technologies such as DenseNet, MTCNN, and Fabnet.
  • Concerning fusion strategies, Fine-grained Interaction Fusion remains a popular choice, owing to its ability to foster detailed interactions between modalities based on nuanced features.
  • Regarding datasets, IEMOCAP emerges as the predominant choice, likely due to the accuracy of its emotion labels and its longstanding reputation. Nonetheless, emerging datasets like MELD and YouTube are gaining traction, possibly to explore emotion recognition challenges in multifaceted and authentic contexts.
In conclusion, the field of multimodal emotion recognition has witnessed remarkable progress in recent years, particularly in the realms of feature extraction and data fusion. By integrating diverse modalities, these techniques offer enhanced accuracy when it comes to understanding and representing human emotions, providing substantial support for applications in human–machine interaction and emotion surveillance.

7. Conclusions and Future Challenges

In this comprehensive survey, we have detailed the progression and current state of deep neural network-based MER technology. We have examined popular feature extraction methods, considered significant multimodal emotion databases, and evaluated the influence of deep learning technology in this rapidly developing field. Crucially, we have also explored representative fusion methods, such as early fusion, late fusion, hybrid fusion, and intermediate layer fusion, which are critical to the effective integration of modal information in deep neural networks. Despite these strides, numerous challenges persist in the field of MER, especially with regard to its application to deep learning.
  • Scale, annotation, and diversity of datasets: Datasets are crucial to deep learning models’ performance and generalization ability. An ideal dataset should exhibit representativeness, diversity, and sufficient scale while maintaining high-quality labels. Datasets enable models to learn patterns through sample observation, while diverse data provide learning opportunities in different contexts. Large-scale datasets can mitigate overfitting issues, and high-quality labels can provide accurate supervision signals. Consequently, constructing a large-scale and diverse dataset becomes imperative in the MER domain. Nevertheless, the annotation of multimodal data requires professionals to subjectively evaluate text, speech, and images, which is both time-consuming and expensive. Thus, the construction of a challenging task lies in developing a high-quality, large-scale, and diverse MER dataset.
  • Multimodal Fusion: The task of fusing data from disparate modalities for emotion recognition is another predicament. Temporal misalignment and heterogeneities among features across different modalities can make the fusion process complex. This complexity often leads to models not fully utilizing the supplementary information from various modalities, which in turn undermines their performance in emotion recognition. In this context, information theory and entropy offer profound insights. By quantifying the information content in various modalities through entropy, it becomes feasible to evaluate the uncertainty or predictability of each data source. This approach can pinpoint modalities that provide substantial information while also identifying those that introduce noise or redundancy. Additionally, employing information theory concepts, such as mutual information, can elucidate the extent of interdependence between modalities. Such insights would facilitate a more harmonious and informed fusion process, ensuring that the relationships and synergies between modalities are optimally leveraged. Therefore, harnessing information theory to craft a potent strategy for amalgamating textual, facial, and auditory features—and seamlessly incorporating this multimodal information into a comprehensive model—is of paramount importance.
  • Generalization Ability of Models: Despite the proposal of numerous excellent MER models, they are usually trained on specific datasets that rely on non-realistic scenarios, making them difficult to adapt to industrial applications. Therefore, MER models need to possess good generalization capability to be applicable to different scenarios and tasks. However, due to limitations in datasets and the complexity of models, the generalization capability of models with regard to new domains or unseen data remains a challenge. Addressing the construction of more universal and transferable models is one of the challenges that must be tackled.
  • Leveraging Large-scale Training and Computational Resources: To attain greater accuracy and improved outcomes, deep learning models commonly necessitate abundant data and computational resources for training purposes. Nevertheless, acquiring extensive annotated data and conducting model training in the context of MER presents a formidable and expensive endeavor. Furthermore, ample computational resources and storage space are indispensable for large-scale training. Consequently, effectively utilizing limited data and computational resources, as well as accelerating the training process, is a challenge that needs to be addressed.
  • Ethical and Privacy Issues: MER is a technology that extensively analyzes various data modalities, including speech, text, and facial cues, to discern individuals’ emotional states. This in-depth analysis often involves the processing of highly personal and intimate details about a person’s emotions and experiences. In practical applications, this level of analysis results in a direct interaction with users’ private information, creating a significant risk of potential privacy breaches. Therefore, striking a balance between ensuring model performance, protecting user privacy, ensuring reasonable data usage, and safeguarding user rights is an important challenge.
In conclusion, while considerable strides have been made in the realm of deep neural network-based MER, many obstacles remain. Overcoming these challenges will not only push the boundaries of what is achievable in this field but will also enable more robust, efficient, and ethically sound systems for emotion recognition. The future of MER is one filled with exciting opportunities for advancement and innovation. As we move forward, the focus will be on creating solutions that can overcome these hurdles and pave the way for a new era in emotion recognition technology.

Author Contributions

Conceptualization and methodology, H.L., C.L., S.L., Y.Z. (Yan Zhao), C.T. and Y.Z. (Yuan Zong); validation, H.L. and C.L.; writing—original draft preparation, H.L., C.L. and S.L.; writing—review and editing, C.T. and Y.Z. (Yan Zhao); funding acquisition, H.L. and Y.Z.(Yuan Zong). All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Key R & D Project under the Grant 2022YFC2405600, in part by the Zhishan Young Scholarship of Southeast University, in part by the Postdoctoral Scientific Research Foundation of Southeast University 4007032320, and in part by the Jiangsu Province Excellent Postdoctoral Program.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. El Ayadi, M.; Kamel, M.S.; Karray, F. Survey on speech emotion recognition: Features, classification schemes, and databases. Pattern Recognit. 2011, 44, 572–587. [Google Scholar] [CrossRef]
  2. Swain, M.; Routray, A.; Kabisatpathy, P. Databases, features and classifiers for speech emotion recognition: A review. Int. J. Speech Technol. 2018, 21, 93–120. [Google Scholar] [CrossRef]
  3. Zong, Y.; Lian, H.; Chang, H.; Lu, C.; Tang, C. Adapting Multiple Distributions for Bridging Emotions from Different Speech Corpora. Entropy 2022, 24, 1250. [Google Scholar] [CrossRef]
  4. Fu, H.; Zhuang, Z.; Wang, Y.; Huang, C.; Duan, W. Cross-Corpus Speech Emotion Recognition Based on Multi-Task Learning and Subdomain Adaptation. Entropy 2023, 25, 124. [Google Scholar] [CrossRef] [PubMed]
  5. Lu, C.; Tang, C.; Zhang, J.; Zong, Y. Progressively Discriminative Transfer Network for Cross-Corpus Speech Emotion Recognition. Entropy 2022, 24, 1046. [Google Scholar] [CrossRef] [PubMed]
  6. Wu, C.H.; Chuang, Z.J.; Lin, Y.C. Emotion recognition from text using semantic labels and separable mixture models. ACM Trans. Asian Lang. Inf. Process. (TALIP) 2006, 5, 165–183. [Google Scholar] [CrossRef]
  7. Shaheen, S.; El-Hajj, W.; Hajj, H.; Elbassuoni, S. Emotion recognition from text based on automatically generated rules. In Proceedings of the 2014 IEEE International Conference on Data Mining Workshop, Shenzhen, China, 14 December 2014; IEEE: New York, NY, USA, 2014; pp. 383–392. [Google Scholar]
  8. Zhang, L.; Tjondronegoro, D. Facial expression recognition using facial movement features. IEEE Trans. Affect. Comput. 2011, 2, 219–229. [Google Scholar] [CrossRef]
  9. Valstar, M.F.; Jiang, B.; Mehu, M.; Pantic, M.; Scherer, K. The first facial expression recognition and analysis challenge. In Proceedings of the 2011 IEEE International Conference on Automatic Face & Gesture Recognition (FG), Santa Barbara, CA, USA, 21–25 March 2011; IEEE: New York, NY, USA, 2011; pp. 921–926. [Google Scholar]
  10. Li, S.; Deng, W. Deep facial expression recognition: A survey. IEEE Trans. Affect. Comput. 2020, 13, 1195–1215. [Google Scholar] [CrossRef]
  11. Yang, H.; Xie, L.; Pan, H.; Li, C.; Wang, Z.; Zhong, J. Multimodal Attention Dynamic Fusion Network for Facial Micro-Expression Recognition. Entropy 2023, 25, 1246. [Google Scholar] [CrossRef]
  12. Zeng, J.; Liu, T.; Zhou, J. Tag-assisted Multimodal Sentiment Analysis under Uncertain Missing Modalities. In Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval, Madrid, Spain, 11–15 July 2022; pp. 1545–1554. [Google Scholar]
  13. Shou, Y.; Meng, T.; Ai, W.; Yang, S.; Li, K. Conversational emotion recognition studies based on graph convolutional neural networks and a dependent syntactic analysis. Neurocomputing 2022, 501, 629–639. [Google Scholar] [CrossRef]
  14. Li, Y.; Wang, Y.; Cui, Z. Decoupled Multimodal Distilling for Emotion Recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 6631–6640. [Google Scholar]
  15. Liu, F.; Chen, J.; Tan, W.; Cai, C. A multi-modal fusion method based on higher-order orthogonal iteration decomposition. Entropy 2021, 23, 1349. [Google Scholar] [CrossRef] [PubMed]
  16. Liu, F.; Shen, S.Y.; Fu, Z.W.; Wang, H.Y.; Zhou, A.M.; Qi, J.Y. Lgcct: A light gated and crossed complementation transformer for multimodal speech emotion recognition. Entropy 2022, 24, 1010. [Google Scholar] [CrossRef] [PubMed]
  17. Li, Q.; Liu, Y.; Liu, Q.; Zhang, Q.; Yan, F.; Ma, Y.; Zhang, X. Multidimensional Feature in Emotion Recognition Based on Multi-Channel EEG Signals. Entropy 2022, 24, 1830. [Google Scholar] [CrossRef] [PubMed]
  18. Chang, H.; Zong, Y.; Zheng, W.; Xiao, Y.; Wang, X.; Zhu, J.; Shi, M.; Lu, C.; Yang, H. EEG-based major depressive disorder recognition by selecting discriminative features via stochastic search. J. Neural Eng. 2023, 20, 026021. [Google Scholar] [CrossRef]
  19. Sapiński, T.; Kamińska, D.; Pelikant, A.; Anbarjafari, G. Emotion recognition from skeletal movements. Entropy 2019, 21, 646. [Google Scholar] [CrossRef] [PubMed]
  20. Chang, H.; Liu, B.; Zong, Y.; Lu, C.; Wang, X. EEG-Based Parkinson’s Disease Recognition Via Attention-based Sparse Graph Convolutional Neural Network. IEEE J. Biomed. Health Inform. 2023. [Google Scholar] [CrossRef]
  21. Gu, X.; Shen, Y.; Xu, J. Multimodal Emotion Recognition in Deep Learning: A Survey. In Proceedings of the 2021 International Conference on Culture-oriented Science & Technology (ICCST), Beijing, China, 18–21 November 2021; IEEE: New York, NY, USA, 2021; pp. 77–82. [Google Scholar]
  22. Koromilas, P.; Giannakopoulos, T. Deep multimodal emotion recognition on human speech: A review. Appl. Sci. 2021, 11, 7962. [Google Scholar] [CrossRef]
  23. Zhang, T.; Tan, Z. Deep Emotion Recognition Using Facial, Speech and Textual Cues: A Survey. TechRxiv. Preprint. 2015. [Google Scholar] [CrossRef]
  24. Busso, C.; Bulut, M.; Lee, C.C.; Kazemzadeh, A.; Mower, E.; Kim, S.; Chang, J.N.; Lee, S.; Narayanan, S.S. IEMOCAP: Interactive emotional dyadic motion capture database. Lang. Resour. Eval. 2008, 42, 335–359. [Google Scholar] [CrossRef]
  25. Morency, L.P.; Mihalcea, R.; Doshi, P. Towards multimodal sentiment analysis: Harvesting opinions from the web. In Proceedings of the 13th International Conference on Multimodal Interfaces, Alicante, Spain, 14–18 November 2011; pp. 169–176. [Google Scholar]
  26. Pérez-Rosas, V.; Mihalcea, R.; Morency, L.P. Utterance-level multimodal sentiment analysis. In Proceedings of the 51st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Sofia, Bulgaria, 4–9 August 2013; pp. 973–982. [Google Scholar]
  27. Wöllmer, M.; Weninger, F.; Knaup, T.; Schuller, B.; Sun, C.; Sagae, K.; Morency, L.P. Youtube movie reviews: Sentiment analysis in an audio-visual context. IEEE Intell. Syst. 2013, 28, 46–53. [Google Scholar] [CrossRef]
  28. Zadeh, A.; Zellers, R.; Pincus, E.; Morency, L.P. Mosi: Multimodal corpus of sentiment intensity and subjectivity analysis in online opinion videos. arXiv 2016, arXiv:1606.06259. [Google Scholar]
  29. Chou, H.C.; Lin, W.C.; Chang, L.C.; Li, C.C.; Ma, H.P.; Lee, C.C. NNIME: The NTHU-NTUA Chinese interactive multimodal emotion corpus. In Proceedings of the 2017 Seventh International Conference on Affective Computing and Intelligent Interaction (ACII), San Antonio, TX, USA, 23–26 October 2017; IEEE: New York, NY, USA, 2017; pp. 292–298. [Google Scholar]
  30. Zadeh, A.B.; Liang, P.P.; Poria, S.; Cambria, E.; Morency, L.P. Multimodal language analysis in the wild: Cmu-mosei dataset and interpretable dynamic fusion graph. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Melbourne, Australia, 15–20 July 2018; pp. 2236–2246. [Google Scholar]
  31. Barros, P.; Churamani, N.; Lakomkin, E.; Siqueira, H.; Sutherland, A.; Wermter, S. The OMG-emotion behavior dataset. In Proceedings of the 2018 International Joint Conference on Neural Networks (IJCNN), Rio de Janeiro, Brazil, 8–13 July 2018; IEEE: New York, NY, USA, 2018; pp. 1–7. [Google Scholar]
  32. Poria, S.; Hazarika, D.; Majumder, N.; Naik, G.; Cambria, E.; Mihalcea, R. MELD: A Multimodal Multi-Party Dataset for Emotion Recognition in Conversations. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Florence, Italy, 28 July–2 August 2019; pp. 527–536. [Google Scholar]
  33. Kossaifi, J.; Walecki, R.; Panagakis, Y.; Shen, J.; Schmitt, M.; Ringeval, F.; Han, J.; Pandit, V.; Toisoul, A.; Schuller, B.; et al. Sewa db: A rich database for audio-visual emotion and sentiment research in the wild. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 43, 1022–1040. [Google Scholar] [CrossRef] [PubMed]
  34. Yu, W.; Xu, H.; Meng, F.; Zhu, Y.; Ma, Y.; Wu, J.; Zou, J.; Yang, K. Ch-sims: A chinese multimodal sentiment analysis dataset with fine-grained annotation of modality. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Online, 5–10 July 2020; pp. 3718–3727. [Google Scholar]
  35. Liu, Y.; Yuan, Z.; Mao, H.; Liang, Z.; Yang, W.; Qiu, Y.; Cheng, T.; Li, X.; Xu, H.; Gao, K. Make Acoustic and Visual Cues Matter: CH-SIMS v2.0 Dataset and AV-Mixup Consistent Module. In Proceedings of the 2022 International Conference on Multimodal Interaction, Bengaluru, India, 7–11 November 2022; pp. 247–258. [Google Scholar]
  36. Lin, J.C.; Wu, C.H.; Wei, W.L. Error weighted semi-coupled hidden Markov model for audio-visual emotion recognition. IEEE Trans. Multimed. 2011, 14, 142–156. [Google Scholar] [CrossRef]
  37. Murray, I.R.; Arnott, J.L. Toward the simulation of emotion in synthetic speech: A review of the literature on human vocal emotion. J. Acoust. Soc. Am. 1993, 93, 1097–1108. [Google Scholar] [CrossRef]
  38. Scherer, K.R. Vocal communication of emotion: A review of research paradigms. Speech Commun. 2003, 40, 227–256. [Google Scholar] [CrossRef]
  39. Wu, S.; Falk, T.H.; Chan, W.Y. Automatic speech emotion recognition using modulation spectral features. Speech Commun. 2011, 53, 768–785. [Google Scholar] [CrossRef]
  40. Zeng, Z.; Pantic, M.; Roisman, G.I.; Huang, T.S. A survey of affect recognition methods: Audio, visual and spontaneous expressions. In Proceedings of the 9th International Conference on Multimodal Interfaces, Nagoya, Japan, 12–15 November 2007; pp. 126–133. [Google Scholar]
  41. Kwon, O.W.; Chan, K.; Hao, J.; Lee, T.W. Emotion recognition by speech signals. In Proceedings of the Eighth European Conference on Speech Communication and Technology, Geneva, Switzerland, 1–4 September 2003. [Google Scholar]
  42. Morrison, D.; Wang, R.; De Silva, L.C. Ensemble methods for spoken emotion recognition in call-centres. Speech Commun. 2007, 49, 98–112. [Google Scholar] [CrossRef]
  43. Nwe, T.L. Analysis and Detection of Human Emotion and Stress from Speech Signals. Ph.D. Thesis, National University of Singapore, Singapore, 2004. [Google Scholar]
  44. Nwe, T.L.; Foo, S.W.; De Silva, L.C. Speech emotion recognition using hidden Markov models. Speech Commun. 2003, 41, 603–623. [Google Scholar] [CrossRef]
  45. Eyben, F.; Wöllmer, M.; Schuller, B. Opensmile: The munich versatile and fast open-source audio feature extractor. In Proceedings of the 18th ACM international conference on Multimedia, Firenze, Italy, 25–29 October 2010; pp. 1459–1462. [Google Scholar]
  46. Degottex, G.; Kane, J.; Drugman, T.; Raitio, T.; Scherer, S. COVAREP—A collaborative voice analysis repository for speech technologies. In Proceedings of the 2014 IEEE International Conference on Acoustics, Speech and signal Processing (ICASSP), Florence, Italy, 4–9 May 2014; IEEE: New York, NY, USA, 2014; pp. 960–964. [Google Scholar]
  47. Eyben, F.; Wöllmer, M.; Schuller, B. OpenEAR—Introducing the Munich open-source emotion and affect recognition toolkit. In Proceedings of the 2009 3rd International Conference on Affective Computing and Intelligent Interaction and Workshops, Amsterdam, The Netherlands, 10–12 September 2009; IEEE: New York, NY, USA, 2009; pp. 1–6. [Google Scholar]
  48. McFee, B.; Raffel, C.; Liang, D.; Ellis, D.P.; McVicar, M.; Battenberg, E.; Nieto, O. Librosa: Audio and music signal analysis in python. In Proceedings of the 14th Python in Science Conference, Austin, TX, USA, 11–17 July 2015; Volume 8, pp. 18–25. [Google Scholar]
  49. Baevski, A.; Schneider, S.; Auli, M. vq-wav2vec: Self-Supervised Learning of Discrete Speech Representations. arXiv 2019, arXiv:1910.05453. [Google Scholar]
  50. Lu, C.; Zheng, W.; Lian, H.; Zong, Y.; Tang, C.; Li, S.; Zhao, Y. Speech Emotion Recognition via an Attentive Time–Frequency Neural Network. IEEE Trans. Comput. Soc. Syst. 2022, 1–10. [Google Scholar] [CrossRef]
  51. Lu, C.; Lian, H.; Zheng, W.; Zong, Y.; Zhao, Y.; Li, S. Learning Local to Global Feature Aggregation for Speech Emotion Recognition. arXiv 2023, arXiv:2306.01491. [Google Scholar]
  52. Zhao, Y.; Wang, J.; Zong, Y.; Zheng, W.; Lian, H.; Zhao, L. Deep Implicit Distribution Alignment Networks for cross-Corpus Speech Emotion Recognition. In Proceedings of the ICASSP 2023—2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Rhodes Island, Greece, 4–10 June 2023; pp. 1–5. [Google Scholar] [CrossRef]
  53. Zhao, Y.; Wang, J.; Ye, R.; Zong, Y.; Zheng, W.; Zhao, L. Deep Transductive Transfer Regression Network for Cross-Corpus Speech Emotion Recognition. Proc. Interspeech 2022, 2022, 371–375. [Google Scholar] [CrossRef]
  54. Schneider, S.; Baevski, A.; Collobert, R.; Auli, M. wav2vec: Unsupervised pre-training for speech recognition. arXiv 2019, arXiv:1904.05862. [Google Scholar]
  55. Baevski, A.; Zhou, Y.; Mohamed, A.; Auli, M. wav2vec 2.0: A framework for self-supervised learning of speech representations. Adv. Neural Inf. Process. Syst. 2020, 33, 12449–12460. [Google Scholar]
  56. Hsu, W.N.; Bolte, B.; Tsai, Y.H.H.; Lakhotia, K.; Salakhutdinov, R.; Mohamed, A. Hubert: Self-supervised speech representation learning by masked prediction of hidden units. IEEE/ACM Trans. Audio Speech Lang. Process. 2021, 29, 3451–3460. [Google Scholar] [CrossRef]
  57. Chen, S.; Wang, C.; Chen, Z.; Wu, Y.; Liu, S.; Chen, Z.; Li, J.; Kanda, N.; Yoshioka, T.; Xiao, X.; et al. Wavlm: Large-scale self-supervised pre-training for full stack speech processing. IEEE J. Sel. Top. Signal Process. 2022, 16, 1505–1518. [Google Scholar] [CrossRef]
  58. Soumya George, K.; Joseph, S. Text classification by augmenting bag of words (BOW) representation with co-occurrence feature. IOSR J. Comput. Eng. 2014, 16, 34–38. [Google Scholar] [CrossRef]
  59. Zhao, R.; Mao, K. Fuzzy bag-of-words model for document representation. IEEE Trans. Fuzzy Syst. 2017, 26, 794–804. [Google Scholar] [CrossRef]
  60. Mikolov, T.; Sutskever, I.; Chen, K.; Corrado, G.S.; Dean, J. Distributed representations of words and phrases and their compositionality. Adv. Neural Inf. Process. Syst. 2013, 26, 3111–3119. [Google Scholar]
  61. Pennington, J.; Socher, R.; Manning, C.D. Glove: Global vectors for word representation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), Doha, Qatar, 25–29 October 2014; pp. 1532–1543. [Google Scholar]
  62. Peters, M.E.; Neumann, M.; Iyyer, M.; Gardner, M.; Clark, C.; Lee, K.; Zettlemoyer, L. Deep contextualized word representations. arXiv 2018, arXiv:1802.05365. [Google Scholar]
  63. Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv 2018, arXiv:1810.04805. [Google Scholar]
  64. Tang, D.; Wei, F.; Yang, N.; Zhou, M.; Liu, T.; Qin, B. Learning sentiment-specific word embedding for twitter sentiment classification. In Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Baltimore, MA, USA, 22–27 June 2014; pp. 1555–1565. [Google Scholar]
  65. Xu, P.; Madotto, A.; Wu, C.S.; Park, J.H.; Fung, P. Emo2vec: Learning generalized emotion representation by multi-task training. arXiv 2018, arXiv:1809.04505. [Google Scholar]
  66. Chen, J.; Liu, X.; Tu, P.; Aragones, A. Learning person-specific models for facial expression and action unit recognition. Pattern Recognit. Lett. 2013, 34, 1964–1970. [Google Scholar] [CrossRef]
  67. Lanitis, A.; Taylor, C.J.; Cootes, T.F. Automatic face identification system using flexible appearance models. Image Vis. Comput. 1995, 13, 393–401. [Google Scholar] [CrossRef]
  68. Cootes, T.F.; Taylor, C.J.; Cooper, D.H.; Graham, J. Active shape models-their training and application. Comput. Vis. Image Underst. 1995, 61, 38–59. [Google Scholar] [CrossRef]
  69. Chu, W.S.; De la Torre, F.; Cohn, J.F. Selective transfer machine for personalized facial expression analysis. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 529–545. [Google Scholar] [CrossRef] [PubMed]
  70. Baltrušaitis, T.; Mahmoud, M.; Robinson, P. Cross-dataset learning and person-specific normalisation for automatic action unit detection. In Proceedings of the 2015 11th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (FG), Ljubljana, Slovenia, 4–8 May 2015; IEEE: New York, NY, USA, 2015; Volume 6, pp. 1–6. [Google Scholar]
  71. Ahsan, T.; Jabid, T.; Chong, U.P. Facial expression recognition using local transitional pattern on Gabor filtered facial images. IETE Tech. Rev. 2013, 30, 47–52. [Google Scholar] [CrossRef]
  72. Tran, D.; Bourdev, L.; Fergus, R.; Torresani, L.; Paluri, M. Learning spatiotemporal features with 3d convolutional networks. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 4489–4497. [Google Scholar]
  73. Yu, Z.; Liu, G.; Liu, Q.; Deng, J. Spatio-temporal convolutional features with nested LSTM for facial expression recognition. Neurocomputing 2018, 317, 50–57. [Google Scholar] [CrossRef]
  74. Littlewort, G.; Whitehill, J.; Wu, T.; Fasel, I.; Frank, M.; Movellan, J.; Bartlett, M. The computer expression recognition toolbox (CERT). In Proceedings of the 2011 IEEE International Conference on Automatic Face & Gesture Recognition (FG), Santa Barbara, CA, USA, 21–25 March 2011; IEEE: New York, NY, USA, 2011; pp. 298–305. [Google Scholar]
  75. Baltrusaitis, T.; Zadeh, A.; Lim, Y.C.; Morency, L.P. Openface 2.0: Facial behavior analysis toolkit. In Proceedings of the 2018 13th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2018), Xi’an, China, 15–19 May 2018; IEEE: New York, NY, USA, 2018; pp. 59–66. [Google Scholar]
  76. Zhu, L.; Zhu, Z.; Zhang, C.; Xu, Y.; Kong, X. Multimodal sentiment analysis based on fusion methods: A survey. Inf. Fusion 2023, 95, 306–325. [Google Scholar] [CrossRef]
  77. D’mello, S.K.; Kory, J. A review and meta-analysis of multimodal affect detection systems. ACM Comput. Surv. (CSUR) 2015, 47, 1–36. [Google Scholar] [CrossRef]
  78. Poria, S.; Cambria, E.; Bajpai, R.; Hussain, A. A review of affective computing: From unimodal analysis to multimodal fusion. Inf. Fusion 2017, 37, 98–125. [Google Scholar] [CrossRef]
  79. Shoumy, N.J.; Ang, L.M.; Seng, K.P.; Rahaman, D.M.; Zia, T. Multimodal big data affective analytics: A comprehensive survey using text, audio, visual and physiological signals. J. Netw. Comput. Appl. 2020, 149, 102447. [Google Scholar] [CrossRef]
  80. Abdullah, S.M.S.A.; Ameen, S.Y.A.; Sadeeq, M.A.; Zeebaree, S. Multimodal emotion recognition using deep learning. J. Appl. Sci. Technol. Trends 2021, 2, 52–58. [Google Scholar] [CrossRef]
  81. Baltrušaitis, T.; Ahuja, C.; Morency, L.P. Multimodal machine learning: A survey and taxonomy. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 41, 423–443. [Google Scholar] [CrossRef] [PubMed]
  82. Poria, S.; Cambria, E.; Hussain, A.; Huang, G.B. Towards an intelligent framework for multimodal affective data analysis. Neural Netw. 2015, 63, 104–116. [Google Scholar] [CrossRef]
  83. Huang, J.; Li, Y.; Tao, J.; Lian, Z.; Niu, M.; Yang, M. Multimodal continuous emotion recognition with data augmentation using recurrent neural networks. In Proceedings of the 2018 on Audio/Visual Emotion Challenge and Workshop, Seoul, Korea, 22 October 2018; pp. 57–64. [Google Scholar]
  84. Williams, J.; Kleinegesse, S.; Comanescu, R.; Radu, O. Recognizing emotions in video using multimodal dnn feature fusion. In Proceedings of the Grand Challenge and Workshop on Human Multimodal Language (Challenge-HML), Melbourne, Australia, 20 July 2018; pp. 11–19. [Google Scholar]
  85. Poria, S.; Cambria, E.; Howard, N.; Huang, G.B.; Hussain, A. Fusing audio, visual and textual clues for sentiment analysis from multimodal content. Neurocomputing 2016, 174, 50–59. [Google Scholar] [CrossRef]
  86. Huang, J.; Li, Y.; Tao, J.; Lian, Z.; Wen, Z.; Yang, M.; Yi, J. Continuous multimodal emotion prediction based on long short term memory recurrent neural network. In Proceedings of the 7th Annual Workshop on Audio/Visual Emotion Challenge, Mountain View, CA, USA, 23 October 2017; pp. 11–18. [Google Scholar]
  87. Su, H.; Liu, B.; Tao, J.; Dong, Y.; Huang, J.; Lian, Z.; Song, L. An Improved Multimodal Dimension Emotion Recognition Based on Different Fusion Methods. In Proceedings of the 2020 15th IEEE International Conference on Signal Processing (ICSP), Beijing, China, 6–9 December 2020; IEEE: New York, NY, USA, 2020; Volume 1, pp. 257–261. [Google Scholar]
  88. Sun, L.; Lian, Z.; Tao, J.; Liu, B.; Niu, M. Multi-modal continuous dimensional emotion recognition using recurrent neural network and self-attention mechanism. In Proceedings of the 1st International on Multimodal Sentiment Analysis in Real-life Media Challenge and Workshop, Seattle, VA, USA, 16 October 2020; pp. 27–34. [Google Scholar]
  89. Sun, L.; Xu, M.; Lian, Z.; Liu, B.; Tao, J.; Wang, M.; Cheng, Y. Multimodal emotion recognition and sentiment analysis via attention enhanced recurrent model. In Proceedings of the 2nd on Multimodal Sentiment Analysis Challenge, Online, 24 October 2021; pp. 15–20. [Google Scholar]
  90. Nemati, S.; Rohani, R.; Basiri, M.E.; Abdar, M.; Yen, N.Y.; Makarenkov, V. A hybrid latent space data fusion method for multimodal emotion recognition. IEEE Access 2019, 7, 172948–172964. [Google Scholar] [CrossRef]
  91. Nguyen, T.L.; Kavuri, S.; Lee, M. A multimodal convolutional neuro-fuzzy network for emotion understanding of movie clips. Neural Netw. 2019, 118, 208–219. [Google Scholar] [CrossRef] [PubMed]
  92. Tripathi, S.; Tripathi, S.; Beigi, H. Multi-Modal Emotion Recognition on Iemocap with Neural Networks. arXiv 2018, arXiv:1804.05788. [Google Scholar]
  93. Ortega, J.D.; Senoussaoui, M.; Granger, E.; Pedersoli, M.; Cardinal, P.; Koerich, A.L. Multimodal fusion with deep neural networks for audio-video emotion recognition. arXiv 2019, arXiv:1907.03196. [Google Scholar]
  94. Liang, J.; Li, R.; Jin, Q. Semi-supervised multi-modal emotion recognition with cross-modal distribution matching. In Proceedings of the 28th ACM International Conference on Multimedia, Seattle, WA, USA, 12–16 October 2020; pp. 2852–2861. [Google Scholar]
  95. Yu, W.; Xu, H.; Yuan, Z.; Wu, J. Learning modality-specific representations with self-supervised multi-task learning for multimodal sentiment analysis. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 2–9 February 2021; Volume 35, pp. 10790–10797. [Google Scholar]
  96. Han, W.; Chen, H.; Poria, S. Improving Multimodal Fusion with Hierarchical Mutual Information Maximization for Multimodal Sentiment Analysis. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, Punta Cana, Dominican Republic, 7–11 November 2021; pp. 9180–9192. [Google Scholar]
  97. Zheng, J.; Zhang, S.; Wang, X.; Zeng, Z. Multimodal Representations Learning Based on Mutual Information Maximization and Minimization and Identity Embedding for Multimodal Sentiment Analysis. arXiv 2022, arXiv:2201.03969. [Google Scholar]
  98. Mai, S.; Zeng, Y.; Hu, H. Multimodal information bottleneck: Learning minimal sufficient unimodal and multimodal representations. IEEE Trans. Multimed. 2022. [Google Scholar] [CrossRef]
  99. Zadeh, A.; Chen, M.; Poria, S.; Cambria, E.; Morency, L.P. Tensor Fusion Network for Multimodal Sentiment Analysis. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, Copenhagen, Denmark, 9–11 September 2017; pp. 1103–1114. [Google Scholar]
  100. Liu, Z.; Shen, Y.; Lakshminarasimhan, V.B.; Liang, P.P.; Zadeh, A.B.; Morency, L.P. Efficient Low-rank Multimodal Fusion With Modality-Specific Factors. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Melbourne, Australia, 15–20 July 2018; pp. 2247–2256. [Google Scholar]
  101. Liang, P.P.; Liu, Z.; Tsai, Y.H.H.; Zhao, Q.; Salakhutdinov, R.; Morency, L.P. Learning representations from imperfect time series data via tensor rank regularization. arXiv 2019, arXiv:1907.01011. [Google Scholar]
  102. Mai, S.; Hu, H.; Xing, S. Divide, conquer and combine: Hierarchical feature fusion network with local and global perspectives for multimodal affective computing. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Florence, Italy, 28 July–2 August 2019; pp. 481–492. [Google Scholar]
  103. Barezi, E.J.; Fung, P. Modality-based factorization for multimodal fusion. arXiv 2018, arXiv:1811.12624. [Google Scholar]
  104. Liu, F.; Chen, J.; Li, K.; Tan, W.; Cai, C.; Ayub, M.S. A Parallel Multi-Modal Factorized Bilinear Pooling Fusion Method Based on the Semi-Tensor Product for Emotion Recognition. Entropy 2022, 24, 1836. [Google Scholar] [CrossRef] [PubMed]
  105. Mai, S.; Hu, H.; Xing, S. Modality to modality translation: An adversarial representation learning and graph fusion network for multimodal fusion. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 164–172. [Google Scholar]
  106. Wu, T.; Peng, J.; Zhang, W.; Zhang, H.; Tan, S.; Yi, F.; Ma, C.; Huang, Y. Video sentiment analysis with bimodal information-augmented multi-head attention. Knowl. Based Syst. 2022, 235, 107676. [Google Scholar] [CrossRef]
  107. Liu, S.; Gao, P.; Li, Y.; Fu, W.; Ding, W. Multi-modal fusion network with complementarity and importance for emotion recognition. Inf. Sci. 2023, 619, 679–694. [Google Scholar] [CrossRef]
  108. Wu, Y.; Lin, Z.; Zhao, Y.; Qin, B.; Zhu, L.N. A text-centered shared-private framework via cross-modal prediction for multimodal sentiment analysis. In Proceedings of the Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021, Online, 1–6 August 2021; pp. 4730–4738. [Google Scholar]
  109. Chen, M.; Wang, S.; Liang, P.P.; Baltrušaitis, T.; Zadeh, A.; Morency, L.P. Multimodal sentiment analysis with word-level fusion and reinforcement learning. In Proceedings of the 19th ACM International Conference on Multimodal Interaction, Glasgow, Scotland, 13–17 November 2017; pp. 163–171. [Google Scholar]
  110. Zadeh, A.; Liang, P.P.; Poria, S.; Vij, P.; Cambria, E.; Morency, L.P. Multi-attention recurrent network for human communication comprehension. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; Volume 32. [Google Scholar]
  111. Zadeh, A.; Liang, P.P.; Mazumder, N.; Poria, S.; Cambria, E.; Morency, L.P. Memory fusion network for multi-view sequential learning. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; Volume 32. [Google Scholar]
  112. Mittal, T.; Bhattacharya, U.; Chandra, R.; Bera, A.; Manocha, D. M3er: Multiplicative multimodal emotion recognition using facial, textual, and speech cues. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 1359–1367. [Google Scholar]
  113. Liang, P.P.; Liu, Z.; Zadeh, A.B.; Morency, L.P. Multimodal Language Analysis with Recurrent Multistage Fusion. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium, 31 October–4 November 2018; pp. 150–161. [Google Scholar]
  114. Tsai, Y.H.H.; Bai, S.; Liang, P.P.; Kolter, J.Z.; Morency, L.P.; Salakhutdinov, R. Multimodal transformer for unaligned multimodal language sequences. Proc Conf Assoc Comput Linguist Meet. 2019, 2019, 6558. [Google Scholar]
  115. Liang, T.; Lin, G.; Feng, L.; Zhang, Y.; Lv, F. Attention is not enough: Mitigating the distribution discrepancy in asynchronous multimodal sequence fusion. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 8148–8156. [Google Scholar]
  116. Lv, F.; Chen, X.; Huang, Y.; Duan, L.; Lin, G. Progressive modality reinforcement for human multimodal emotion recognition from unaligned multimodal sequences. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–25 June 2021; pp. 2554–2562. [Google Scholar]
  117. Tzirakis, P.; Chen, J.; Zafeiriou, S.; Schuller, B. End-to-end multimodal affect recognition in real-world environments. Inf. Fusion 2021, 68, 46–53. [Google Scholar] [CrossRef]
  118. Lian, Z.; Liu, B.; Tao, J. CTNet: Conversational transformer network for emotion recognition. IEEE/ACM Trans. Audio Speech Lang. Process. 2021, 29, 985–1000. [Google Scholar] [CrossRef]
  119. Nicolaou, M.A.; Gunes, H.; Pantic, M. Continuous prediction of spontaneous affect from multiple cues and modalities in valence-arousal space. IEEE Trans. Affect. Comput. 2011, 2, 92–105. [Google Scholar] [CrossRef]
  120. Ringeval, F.; Schuller, B.; Valstar, M.; Jaiswal, S.; Marchi, E.; Lalanne, D.; Cowie, R.; Pantic, M. Av+ ec 2015: The first affect recognition challenge bridging across audio, video, and physiological data. In Proceedings of the 5th International Workshop on Audio/Visual Emotion Challenge, Brisbane, Australia, 26–23 October 2015; pp. 3–8. [Google Scholar]
  121. Poria, S.; Cambria, E.; Hazarika, D.; Mazumder, N.; Zadeh, A.; Morency, L.P. Multi-level multiple attentions for contextual multimodal sentiment analysis. In Proceedings of the 2017 IEEE International Conference on Data Mining (ICDM), New Orleans, LA, USA, 18–21 November 2017; IEEE: New York, NY, USA, 2017; pp. 1033–1038. [Google Scholar]
  122. Pham, H.; Manzini, T.; Liang, P.P.; Poczós, B. Seq2Seq2Sentiment: Multimodal Sequence to Sequence Models for Sentiment Analysis. arXiv 2018, arXiv:1807.03915. [Google Scholar]
  123. Poria, S.; Majumder, N.; Hazarika, D.; Cambria, E.; Gelbukh, A.; Hussain, A. Multimodal sentiment analysis: Addressing key issues and setting up the baselines. IEEE Intell. Syst. 2018, 33, 17–25. [Google Scholar] [CrossRef]
  124. Zhang, K.; Zhang, Z.; Li, Z.; Qiao, Y. Joint face detection and alignment using multitask cascaded convolutional networks. IEEE Signal Process. Lett. 2016, 23, 1499–1503. [Google Scholar] [CrossRef]
  125. Majumder, N.; Hazarika, D.; Gelbukh, A.; Cambria, E.; Poria, S. Multimodal sentiment analysis using hierarchical fusion with context modeling. Knowl.-Based Syst. 2018, 161, 124–133. [Google Scholar] [CrossRef]
  126. Tsai, Y.H.H.; Liang, P.P.; Zadeh, A.; Morency, L.P.; Salakhutdinov, R. Learning Factorized Multimodal Representations. In Proceedings of the International Conference on Representation Learning, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  127. Wang, Y.; Shen, Y.; Liu, Z.; Liang, P.P.; Zadeh, A.; Morency, L.P. Words can shift: Dynamically adjusting word representations using nonverbal behaviors. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 7216–7223. [Google Scholar]
  128. Pham, H.; Liang, P.P.; Manzini, T.; Morency, L.P.; Póczos, B. Found in translation: Learning robust joint representations by cyclic translations between modalities. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 6892–6899. [Google Scholar]
  129. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–27 July 2017; pp. 4700–4708. [Google Scholar]
  130. Wang, Z.; Wan, Z.; Wan, X. Transmodality: An end2end fusion method with transformer for multimodal sentiment analysis. In Proceedings of the Web Conference 2020, Taipei, Taiwan, 20–24 April 2020; pp. 2514–2520. [Google Scholar]
  131. Sun, Z.; Sarma, P.; Sethares, W.; Liang, Y. Learning relationships between text, audio, and video via deep canonical correlation for multimodal language analysis. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 8992–8999. [Google Scholar]
  132. Siriwardhana, S.; Kaluarachchi, T.; Billinghurst, M.; Nanayakkara, S. Multimodal emotion recognition with transformer-based self supervised feature fusion. IEEE Access 2020, 8, 176274–176285. [Google Scholar] [CrossRef]
  133. Koepke, A.S.; Wiles, O.; Zisserman, A. Self-supervised learning of a facial attribute embedding from video. In Proceedings of the BMVC, Newcastle, UK, 3–6 September 2018; p. 302. [Google Scholar]
Figure 1. General framework of a typical multimodal emotion recognition system.
Figure 1. General framework of a typical multimodal emotion recognition system.
Entropy 25 01440 g001
Figure 2. Multimodal fusion methods for multimodal emotion recognition.
Figure 2. Multimodal fusion methods for multimodal emotion recognition.
Entropy 25 01440 g002
Figure 3. Multimodal emotion recognition framework based on early fusions.
Figure 3. Multimodal emotion recognition framework based on early fusions.
Entropy 25 01440 g003
Figure 4. Multimodal emotion recognition framework based on late fusions.
Figure 4. Multimodal emotion recognition framework based on late fusions.
Entropy 25 01440 g004
Figure 5. Multimodal emotion recognition framework based on hybrid fusions.
Figure 5. Multimodal emotion recognition framework based on hybrid fusions.
Entropy 25 01440 g005
Table 1. Statistics on popular multimodal emotion recognition datasets. “N/A” indicates that the corresponding reference does not provide the relevant value.
Table 1. Statistics on popular multimodal emotion recognition datasets. “N/A” indicates that the corresponding reference does not provide the relevant value.
DatasetSamplesLanguageTypeRecording
Environment
EmotionsData SourcesSpeakers
IEMOCAP (2008) [24]10,039EnglishActed
and Natural
In the labDiscrete: anger, sadness,
happiness, disgust, fear,
surprise, frustration, excited,
and neutral states;
Continuous: VAD (activation,
valence, and dominance)
Recored10
Youtube dataset (2011) [25]47EnglishNaturalIn the wildDiscrete: positive, negative,
and neutral
Youtube47
MOUD (2013) [26]498SpanishNaturalIn the wildDiscrete: positive, negative,
and neutral
Youtube80
ICT-MMMO (2013) [27]370EnglishNaturalIn the wildDiscrete: strongly negative,
weakly negative, neutral,
weakly positive,
and strongly positive
Youtube
and ExpoTV
370
CMU-MOSI (2016) [28]2199EnglishNaturalIn the wildDiscrete: positive, negativeYoutube89
NNIME (2017) [29]6701ChineseNaturalIn the labDiscrete: anger, sadness,
happiness, frustration,
neutral, and surprise;
Continuous:
VA (valence and arousal)
Record44
CMU-MOSEI (2018) [30]23,453EnglishNaturalIn the wildDiscrete: anger, happiness,
disgust, sadness, fear,
and surprise
Youtube1000
OMG(2018) [31]7371EnglishActed
and Natural
In the wildDiscrete: anger, happiness,
disgust, sadness, fear,
surprise, and neutral;
Continuous:
VA (valence and arousal)
YoutubeN/A
MELD (2019) [32]13,708EnglishActedIn the wildDiscrete: (1) joy, sadness,
anger, fear, disgust,
surprise, and neutral;
(2) positive, negative, neutral
TV series
Friends
407
SEWA (2019) [33]2562Chinese,
English,
German,
Greek,
Hungarian,
and Serbian
NaturalIn the wildContinuous:
VA (valence and arousal)
Record398
CH-SIMS (2020) [34]2281ChineseActedIn the wildDiscrete: negative,
weakly negative, neutral,
weakly positive, and positive
Movies,
TV series
N/A
CH-SIMS V2.0 (2022) [35]4402ChineseActedIn the wildDiscrete: negative,
weakly negative, neutral,
weakly positive, and positive
Movies,
TV series,
etc.
N/A
Table 2. Summary of the state of the art in Multimodal Emotion Recognition. Here, S represents speech, T represents text, and F represents facial expressions.
Table 2. Summary of the state of the art in Multimodal Emotion Recognition. Here, S represents speech, T represents text, and F represents facial expressions.
MethodFeature ExtractionFusion MethodDatasetsPerformance
Poria et al. [121]S: openSMILE
T: word2vec
F: 3D-CNN
Early FusionCMU-MOSIACC: 0.813 (2 classes)
Liu et al. [100]S: COVAREP
T: Glove
F: FACET
Utterance-level
Interaction Fusion
IEMOCAPUF1: 0.831 (4 classes)
Zadeh et al. [111]S: COVAREP
T: Glove
F: FACET
Fine-grained
Interaction Fusion
a. Youtube
b. MOUD
c. IEMOCAP
a. ACC: 0.610, F1: 0.607
(3 classes);
b. ACC: 0.811, F1: 0.804
(2 classes);
c. ACC: 0.365, F1: 0.349
(9 classes)
Pham et al. [122]S: MFCC
T: Glove
F: FACET and OpenFace
Fine-grained
Interaction Fusion
CMU-MOSIACC: 0.765, F1: 0.734 (2 classes)
Poria et al. [123]S: openSMILE
T: CNN
F: 3D-CNN
Early Fusiona. IEMOCAP;
b. MOUD;
c. CMU-MOSI;
a. ACC: 0.716 (4 classes);
b. ACC: 0.679 (2 classes);
c. ACC: 0.767 (2 classes)
Zadeh et al. [30]S: COVARE
T: Glove
F: MTCNN [124]
Fine-grained
Interaction Fusion
CMU-MOSEIUACC: 0.624, UF1: 0.763 (6 classes)
Majumber et al. [125]S: openSMILE
T: word2vec
F: 3D-CNN
Fine-grained
Interaction Fusion
a. CMU-MOSI
b. IEMOCAP
a. ACC: 0.800 (2 classes);
b. ACC: 0.765 (4 classes)
Tsai et al. [126]S: COVAREP
T: Glove
F: FACET
Fine-grained
Interaction Fusion
a. ICT-MMMO
b. Youtube
c. MOUD
d.IEMOCAP
a. ACC: 0.813, F1: 0.792
(2 classes);
b. ACC: 0.533, F1: 0.524
(3 classes);
c. ACC: 0.821,
F1: 0.817 (2 classes);
d.UACC: 0.848, UF1: 0.814
(6 classes)
Wang et al. [127]S: COVAREP
T: Glove
F: FACET
Fine-grained
Interaction Fusion
IEMOCAPUACC: 0.819, UF1: 0.812 (4 classes)
Tsai et al. [114]S: COVAREP
T: Glove
F: FACET
Fine-grained
Interaction Fusion
IEMOCAPUACC: 0.747, UF1: 0.715 (4 classes)
Pham et al. [128]S: COVAREP
T: Glove
F: FACET
Fine-grained
Interaction Fusion
a. ICT-MMMO
b. Youtube
a. ACC: 0.813, F1: 0.808 (2 classes);
b. ACC: 0.517, F1: 0.524 (2 classes)
Liang et al. [94]S: OpenSMILE
T: BERT
F: DenseNet [129]
Simple Concatenation
Fusion
a. IEMOCAP
b. MELD
a. ACC: 0.756, UACC: 0.745 (4 class);
b. F1: 0.571 (7 classes)
Mittal et al. [112]S: 12 MFCC, pitch, voiced/unvoiced
segmentation features, glottal source
parameters, among others.
T: Glove
F: Facial recognition models,
facial action units,
and facial landmarks.
Fine-grained
Interaction Fusion
a. IEMOCAP
b. CMU-MOSEI
a. ACC:0.827, F1: 0.824 (4 classes);
b. ACC: 0.890, F1: 0.802 (6 classes)
Wang et al. [130]S: OpenSMILE
T: CNN
F: 3D-CNN
Fine-grained
Interaction Fusion
a. IEMOCAP
b. CUM-MOSI
c. MELD
d.CMU-MOSI
a. ACC: 0.608 (6 classes);
b. ACC: 0.827 (2 classes);
c. ACC:0.620 (7 classes);
d.ACC:0.827 (2 classes)
Sun et al. [131]S: COVAREP
T: BERT
F: FACET
Fine-grained
Interaction Fusion
IEMOCAPUACC: 0.830, UF1: 0.818 (4 classes)
Siriwardhana et al. [132]S: Wav2Vec
T: RoBERTa
F: Fabnet [133]
Fine-grained
Interaction Fusion
a. IEMOCAP
b. MELD
a. UACC: 0.847 UF1: 0.842 (4 classes);
b. ACC: 0.643, F1: 0.639 (7 classes)
Lv et al. [116]S: COVAREP
T: BERT
F: FACET
Fine-grained
Interaction Fusion
a. IEMOCAP
b. CMU-MOSI
a. UACC: 0.851, UF1: 0.838 (4 classes);
b. ACC: 0.836, F1: 0.834 (2 classes)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lian, H.; Lu, C.; Li, S.; Zhao, Y.; Tang, C.; Zong, Y. A Survey of Deep Learning-Based Multimodal Emotion Recognition: Speech, Text, and Face. Entropy 2023, 25, 1440. https://doi.org/10.3390/e25101440

AMA Style

Lian H, Lu C, Li S, Zhao Y, Tang C, Zong Y. A Survey of Deep Learning-Based Multimodal Emotion Recognition: Speech, Text, and Face. Entropy. 2023; 25(10):1440. https://doi.org/10.3390/e25101440

Chicago/Turabian Style

Lian, Hailun, Cheng Lu, Sunan Li, Yan Zhao, Chuangao Tang, and Yuan Zong. 2023. "A Survey of Deep Learning-Based Multimodal Emotion Recognition: Speech, Text, and Face" Entropy 25, no. 10: 1440. https://doi.org/10.3390/e25101440

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop