Open Access
ARTICLE
Hierarchical Joint Cross-Modal Attention and Gating Mechanism for Multimodal Sentiment Analysis
School of Computer Science and Engineering, Chongqing University of Technology, Chongqing, China
* Corresponding Author: Yahui Liu. Email:
(This article belongs to the Special Issue: Sentiment Analysis for Social Media Data: Lexicon-Based and Large Language Model Approaches)
Computers, Materials & Continua 2026, 88(1), 43 https://doi.org/10.32604/cmc.2026.077982
Received 21 December 2025; Accepted 06 March 2026; Issue published 08 May 2026
Abstract
Multimodal sentiment analysis aims to accurately identify emotional states by comprehensively utilizing information from multiple sources such as text, audio, and visual data. However, semantic heterogeneity and temporal differences exist between different modalities, limiting the effectiveness of feature fusion. To address this issue, this paper proposes a hierarchical joint cross-modal attention and gating mechanism (HJCAG) for multimodal sentiment analysis. This method introduces a hierarchical structure, dividing modal interactions into bimodal and trimodal layers to progressively model the semantic relevance between modalities. First, deep features are extracted from text, audio, and visual modalities using pre-trained models to obtain high-dimensional representations of semantics, speech, and facial expressions, which are then aligned to a unified feature space. Second, a joint cross-modal attention module is designed at the bimodal and trimodal levels, calculating cross-attention weights based on the correlation between the joint feature representation and individual modal representations. Explicit modeling of multimodal interaction relationships and semantic alignment fully leverages the complementary information of different modalities. Furthermore, this paper introduces a gating mechanism to adaptively control the contribution weights of each modal feature, reducing redundant information interference and improving the discriminativeness of the fused representation. Finally, the fused global features are input into the emotion classifier to identify emotional states. The proposed method achieves 75.47Keywords
With the rapid development of social media, short video platforms, and human-computer interaction technologies, affective computing [1] has gradually become one of the important research directions in the field of artificial intelligence. Sentiment analysis, as a core task within this field, aims to identify and understand the emotional states conveyed by humans in multimodal signals such as language, speech, and facial expressions. Compared to traditional unimodal sentiment analysis, multimodal sentiment analysis [2] (MSA) utilizes multi-source information such as text, audio, and visual data simultaneously, thus enabling a more comprehensive and accurate characterization of human emotional features. Therefore, it has broad application prospects in areas such as intelligent customer service, mental health monitoring, virtual human interaction [3], and film and television content analysis.
However, multimodal sentiment analysis still faces many challenges. First, data from different modalities differ significantly in their representation and semantic space. Text modality primarily expresses semantic information, while audio and visual modalities more directly reflect emotional intensity and non-verbal features [4]. This modal heterogeneity complicates feature alignment and semantic fusion. Second, issues such as inconsistent time steps, missing information, and noise interference between modalities lead to insufficient cross-modal information interaction, affecting the accuracy of sentiment recognition. Traditional feature concatenation or simple weighted fusion methods [5] are insufficient to effectively capture inter-modal dependencies and semantic complementarity. Therefore, designing a fusion mechanism that can adaptively model inter-modal correlations and suppress redundant information has become crucial for research.
In recent years, attention mechanisms [6] have demonstrated outstanding performance in natural language processing and multimodal learning. Their core idea is to selectively focus information by dynamically assigning weights to highlight key information and weaken irrelevant features. In multimodal sentiment analysis, cross-modal attention is widely used for modeling intermodal interactions, achieving finer semantic alignment by learning the correlations between text, audio, and visual features [7]. However, most existing studies still have two shortcomings in their design: first, multimodal interactions are usually limited to single-layer or symmetrical structures, failing to distinguish and model the relationships between different levels (such as bimodal and trimodal); second, existing methods generally perform bidirectional attention calculations directly on modal pairs (e.g., text-visual, text-audio), lacking modeling of the interactions between joint modal features and single-modal features, thus limiting the model’s ability to capture complex semantic dependencies.
To address the aforementioned issues, this paper proposes a multimodal sentiment analysis framework based on Hierarchical Joint Cross-modal Attention with Gating Mechanism (HJCAG). This method is designed to fully consider the hierarchical nature of intramodal feature aggregation and intermodal semantic interaction. Specifically, it introduces a hierarchical modeling strategy, performing feature fusion at both bimodal (A-V, T-A, T-V) and trimodal (T-A-V) layers. Through this hierarchical design, the model first captures the local semantic dependencies between bimodalities, and then aggregates the global semantic information of the trimodal layers at a higher level, thus achieving a progressive information integration from local to global. In the fusion process at each layer, this paper proposes a Joint Cross-modal Attention (JCA) mechanism. Unlike traditional cross-modal attention, JCA uses joint modal features (such as A-V joint or T-A-V joint) as queries and individual modal features as keys/values, calculating the relevance weights between them to achieve dynamic attention and information interaction between joint features and individual modal features. This design enables the model to more effectively capture high-order semantic dependencies between modalities, strengthening semantic consistency and complementarity. Furthermore, to avoid some modal noise or redundant information affecting the overall discrimination performance, this paper introduces a gating mechanism to adaptively adjust the contributions of each modality, ensuring the effectiveness and stability of the fused features in local and global sentiment representations.
In summary, the main contributions of this paper include the following four aspects:
• A hierarchical multimodal fusion framework is proposed, which models the semantic dependencies between modalities step by step from the bimodal level to the trimodal level.
• A joint cross-modal attention mechanism is designed, with the interaction between joint features and single-modal features as the core, to achieve deep semantic alignment and information enhancement.
• A gating mechanism is introduced to achieve adaptive adjustment of modal weights, effectively suppressing the interference of noisy modalities on feature fusion and improving the stability and generalization performance of the model.
• Experimental results on two widely used benchmark datasets show that the proposed method consistently outperforms the mainstream baseline models, improving accuracy by 1.52% and 1.70%, respectively, and weighted F1 scores by 2.76% and 2.67%, respectively, compared to the closest methods.
Modal interaction and feature fusion are core components of multimodal sentiment analysis, and their performance directly impacts sentiment recognition accuracy. Based on the timing and method of fusion, existing methods are mainly divided into early fusion and late fusion. Early fusion [8] directly concatenates representations of different modalities at the feature layer, resulting in a simple structure. However, due to significant differences in feature distribution between modalities, it is prone to introducing noise and lacks semantic alignment constraints. Late fusion [9] fuses prediction results after each modality independently completes sentiment discrimination. While this alleviates the feature mismatch problem, it struggles to fully exploit collaborative information between modalities. Therefore, in recent years, researchers have proposed more complex and efficient fusion mechanisms.
For example, The Tensor Fusion Network (TFN) introduced by Zadeh et al. [10] employs third-order tensor representations to explicitly model full interaction patterns among modalities. To alleviate the computational burden of high-order tensors, Liu et al. [11] developed the Low-rank Multimodal Fusion (LMF) framework based on tensor decomposition techniques. Building upon Transformer architectures, the MulT model introduced by Tsai et al. [12] models inter-modal dependencies through directional cross-modal attention and incorporates a self-attention mechanism to handle asynchronous multimodal sequences. However, subsequent analysis by Ma et al. [13] pointed out that MulT primarily captures interactions at a single feature scale, without sufficiently exploiting fine-grained semantic information. To address this limitation, they further proposed the MCMulT, enabling semantic modeling across hierarchical feature representations. Considering heterogeneity in multimodal sequences, Sun et al. [14] introduced the EMT-DLFR framework to integrate global contextual information for enhanced global–local interaction modeling. In addition, UniMSE designed by Hu et al. [15] performs multimodal fusion at both syntactic and semantic levels and integrates contrastive learning strategies to refine fused feature representations. Shi and Huang [16] proposed MultiEMO, which effectively integrates multimodal cues by capturing cross-modal mapping relationships between text, audio, and visual modalities through a bidirectional multi-head cross-attention layer.
Furthermore, gating mechanisms play a crucial role in modality weight control. By learning the contribution of each modality to the final emotion decision through gating units, the model can suppress irrelevant or noisy features, thereby achieving adaptive fusion. For example, Arevalo et al. [17] proposed a gated multimodal unit (GMU) to balance the information flow between different modalities through gating functions. Rahman et al. [18] proposed MAG-BERT, which further introduced a modality adaptive gating strategy into the BERT framework, achieving superior fusion performance.
Inspired by the above-mentioned research, this paper proposes the hierarchical joint cross-modal attention and gating mechanism model. This model achieves improvements over existing work in hierarchical structure modeling, cross-modal semantic association capture, and adaptive control of modal weights, providing a more efficient and robust fusion strategy for multimodal sentiment analysis. Details will be presented in the Section 3.
The overall structure of the HJCAG proposed in this paper is shown in Fig. 1. The design principles and implementation methods of each part will be introduced below.

Figure 1: Overview of the proposed framework.
3.1 Unimodal Feature Extraction
This paper uses the pre-trained language model RoBERTa [19] as the feature extractor for text modalities. RoBERTa is based on a multi-layer Transformer encoder structure and is optimized on the basis of the BERT model, which can more fully model the semantic information of the text. To adapt to the sentiment recognition task, we fine-tune RoBERTa end-to-end and use the hidden state corresponding to the [CLS] label in its last layer as the sentiment representation of the text. The final text feature vector has a dimension of 1024.
In terms of acoustic modal feature extraction, this paper adopts the openSMILE toolkit [20]. openSMILE is a general feature extraction framework for audio signal processing, which supports flexible configuration of various modular feature extraction components through a scriptable command line. Based on the original acoustic features extracted by openSMILE, the dimensionality is further compressed through a fully connected layer. The acoustic feature dimension of the IEMOCAP dataset is reduced to 1582, and the feature dimension of the MELD dataset is reduced to 300.
For visual modal feature extraction, this paper employs the ResNet-50 network as the visual feature extractor. ResNet-50 [21] is a convolutional neural network based on a deep residual learning mechanism. By introducing residual connections, it effectively alleviates the gradient degradation problem in deep network training, thereby improving feature representation capabilities. This paper utilizes a pre-trained ResNet-50 to encode visual information in video frames and extracts its high-level semantic features as visual representations with a dimension of 512, which are then used for subsequent multimodal fusion and emotion recognition tasks.
Although we adopt widely-used feature extractors such as ResNet-50 and openSMILE to ensure fair comparison with existing multimodal sentiment analysis methods, the proposed HJCAG framework is backbone-agnostic. It can be readily extended to incorporate modern self-supervised representations.
3.2 Hierarchical Joint Cross-Modal Attention
3.2.1 Bimodal Joint Interaction
While bimodal feature fusion can be achieved through unified multimodal training, studies have found that multimodal performance typically degrades compared to single-modal fusion [22]. This is attributed to a number of factors, such as the differences in learning dynamics between T and V modalities [22], different noise topologies, the inclusion of more or less task-related information in certain modal flows, and specialized input representations [23]. Therefore, we train DL models separately for each modality to extract single-modal features, which are then further fed into the HJCA fusion model for pairwise modality fusion. To reliably fuse these modalities by effectively leveraging the complementary relationships between bimodalities, we rely on a cross-attention-based fusion mechanism to efficiently encode intermodal information while preserving intramodal features. While cross-attention is typically applied to features of individual modalities, we explore it within a joint framework. Specifically, our joint feature representation is obtained by concatenating pairwise modal features to focus on joint single-modal features, for example, concatenating A and V features to focus on individual A and V features. By using joint representations, each modality’s features focus on itself and other modalities, thus helping to capture semantic intermodal relationships between the two modalities. Furthermore, using combined feature representations in the cross-attention module can significantly reduce heterogeneity between the two modalities, thereby further improving system performance.
Given an input video subsequence S of L frames, the audio, visual, and text data are preprocessed and input into the corresponding encoders, as shown in Fig. 1. The feature representations of the audio, visual, and text modalities are represented as
where
Although the concatenated features are transformed through a fully connected layer, this operation does not perform dimensional compression. Instead, it serves as a modality alignment transformation in a shared embedding space, The projection matrix preserves the original dimensions of the concatenated features, ensuring that no explicit information reduction is introduced.
Subsequently, the joint features
where
where
where

Figure 2: Joint cross-modal attention blocks of text, audio, and visual modalities.
The obtained joint cross-correlation matrix for a single modality is used to calculate attention weights, thereby capturing semantic relevance within the same modality and between different modalities. The joint cross-correlation matrix quantifies the correlation between specific modal feature representations and the global multimodal context. Higher correlation coefficients indicate stronger semantic consistency between corresponding features in the global multimodal context. Based on the joint cross-correlation matrix, the attention map estimates the relative importance of specific modal features. Specifically, for the text modality, the joint correlation matrix
where
where
where
3.2.2 Trimodal Joint Interaction
After bimodal joint interaction, we obtained the relevance and information supplementation of a single modality to the other two modalities. Subsequently, the weighted fusion features were further combined to obtain a text-audio-visual modal joint feature representation containing global semantic information, as shown below:
where
Subsequently, the joint feature
where
Similarly, the joint relevance matrix
here
where
This hierarchical design enables the model to first capture the local semantic dependencies between the two modalities, and then fuse the global sentiment information of the three modalities at a higher level, effectively mitigating the negative impact of modal heterogeneity and semantic inconsistency.
By inputting the features of interest from each modality into the multimodal gating fusion module, the contribution of each modality to the final prediction is dynamically adjusted, preventing the model from being overly biased towards a single modality and ensuring balanced integration of multimodal information. This design enhances the robustness of multimodal fusion and improves the model’s generalization performance in fine-grained category sentiment classification tasks. As shown below:
Finally, the multimodal feature representation
where C represents the number of categories, N represents the number of samples,
IEMOCAP [24] is an audiovisual corpus designed for emotion recognition research, comprising approximately 12 h of dyadic conversational recordings. The dataset includes 151 dialogue sessions with a total of 7433 annotated utterances. Each utterance is manually assigned to one of six emotional categories: happy, sad, neutral, angry, excited, and frustrated.
MELD [25] is a multimodal, multi-party conversational emotion dataset collected from the television series Friends. It consists of 1433 dialogue segments and 13,708 utterances. Each utterance is annotated with one of seven emotion labels, including anger, disgust, fear, joy, neutral, sadness, and surprise.
The statistics of different emotion types in the IEMOCAP and MELD datasets are shown in Table 1.

Both datasets feature sentiment annotation at the utterance-level. Therefore, we perform utterance-level alignment between the two datasets. For each utterance U, corresponding to a video subsequence S, for the text modality, we first extract lexical embeddings and then pad or crop them to a fixed sequence length L. For the visual modality, we uniformly sample L frames from S and encode them using a visual backbone network. For the audio modality, we copy or broadcast the extracted descriptors along the temporal dimension to align them temporally with the visual and text sequences to match the length L. This strategy ensures consistency in temporal granularity across modalities.
We adopt the RoBERTa model as the textual feature encoder and fine-tune it using the AdamW optimization strategy. During training, the learning rate is initialized to 2e−5, with a batch size of 16 and a maximum input length of 128 tokens. A linear learning rate schedule with warm-up is employed, where the warm-up phase occupies the first 10% of the total optimization steps. To enhance training robustness and mitigate overfitting, the fine-tuning process is repeated three times, and early stopping is applied based on the validation loss.
Given the pronounced differences in class distribution between the MELD and IEMOCAP datasets, distinct batch sizes are employed to better accommodate their respective imbalance characteristics. Specifically, the batch size is set to 64 for IEMOCAP and 100 for MELD. The model is trained for a total of 200 epochs using the Adam optimizer, with momentum parameters configured as
Evaluation metrics: Following previous work [7,26], we report overall accuracy and weighted average F1 score to measure overall performance, and provide F1 score for each emotion category. Furthermore, all reported results were obtained by averaging five-fold cross-validation.
All comparison models used the same backbone architecture and experimental settings.
Experimental evaluations on the IEMOCAP and MELD benchmarks, as reported in Tables 2 and 3, along with comparisons against representative existing approaches. As can be observed, our model obtains the highest weighted-average F1 scores on both datasets and consistently outperforms competing methods across the majority of emotion classes. Such improvements can be mainly due to the proposed architecture, which learns semantic dependencies between modalities in a progressively deeper manner. The bimodal layer focuses on capturing local emotional associations between modal pairs, helping to uncover complementary information from different modalities at a fine-grained level; the trimodal layer further integrates global semantic representations of text, audio, and visual modalities, thereby achieving holistic modeling of complex emotional states and effectively mitigating the impact of modal heterogeneity and semantic inconsistency.
On the IEMOCAP dataset, HJCAG achieves the highest Acc and W-F1 scores, demonstrating overall superior performance. The model shows clear advantages in Sad, Anger, Excited, and Frustrated categories, which are typically minority or semantically overlapping emotions. Although slightly underperforming EmoCaps [31] and SDT [7] in the Happy and Neutral categories, respectively, HJCAG maintains competitive results and exhibits a more balanced performance across emotion classes, indicating reduced bias toward majority categories. On the MELD dataset, HJCAG outperforms all baseline methods in both Acc and W-F1, confirming its robustness under severe class imbalance. Notable gains are observed in low-frequency emotions such as Fear and Disgust, while strong performance is also maintained for high-frequency categories like Neutral and Surprise. Despite slightly lower performance than CFN-ESA [33] on Joy, HJCAG consistently surpasses other methods across the remaining emotion classes.
We further examine the confusion matrices obtained on both datasets, as illustrated in Fig. 3. The analysis reveals several noteworthy observations. First, the proposed model still encounters difficulties in distinguishing semantically similar emotion pairs, such as happiness vs. excitement and anger vs. frustration in IEMOCAP, as well as surprise vs. joy in MELD. Second, on the MELD dataset, a considerable number of samples from other emotion categories are misclassified as neutral, largely due to the dominance of this class in the dataset. Third, emotions with extremely limited training samples, particularly fear and disgust in MELD, remain challenging to recognize reliably, which hinders effective model learning. Overall, these results indicate that emotion confusion caused by semantic similarity and severe class imbalance continues to be a major obstacle in multimodal emotion recognition.

Figure 3: (a) Confusion matrix on IEMOCAP dataset (b) confusion matrix on MELD.
We observed that the proposed model performs poorly in scenarios involving irony, where the text conveys a positive emotion but its underlying intention is negative. Furthermore, the model occasionally misclassifies samples with subtle emotional differences (e.g., neutral vs. slightly positive), indicating that fine-grained emotional boundaries remain challenging. These observations point to potential directions for future research, such as introducing pragmatic reasoning mechanisms or ambiguity perception fusion strategies.
To investigate the effectiveness of our proposed hierarchical structure and gating mechanism, we conducted ablation experiments. Table 4 shows the performance under different ablation conditions.

Bimodal joint layers, trimodal joint layers, and multimodal gating mechanisms are the three key components of our proposed hierarchical joint cross-modal attention. We evaluate the effectiveness of a component by removing only one component at a time. In Table 4, we observe that: (1) Overall, each ablation method led to a performance decrease to varying degrees, fully demonstrating the necessity of each component of the model. (2) The bimodal joint layer has a significant impact on both datasets. On the IEMOCAP dataset, removing this component leads to a 3.02% decrease in overall recognition accuracy and a 4.46% decrease in the weighted F1 score. On the MELD dataset, removing this component leads to a 3.91% decrease in accuracy and a 3.85% decrease in the weighted F1 score. These results indicate that the feature interactions between the two modalities play a fundamental role in capturing cross-modal correlations and are a core support for subsequent high-level fusion. (3) The trimodal integration layer plays a crucial role. On the IEMOCAP dataset, removing this component leads to a 2.60% decrease in accuracy and a 3.69% decrease in the weighted F1 score. On the MELD dataset, removing this component leads to a 2.68% decrease in accuracy and a 2.47% decrease in the weighted F1 score. This indicates that integrating information from three modalities simultaneously yields more complete emotional cues than any two modalities. (4) Removing the multimodal gating mechanism leads to performance degradation. On the IEMOCAP dataset, accuracy drops by 2.33%, and the weighted F1 score drops by 3.63%. On the MELD dataset, accuracy drops by 1.80%, and the weighted F1 score drops by 1.34%. This verifies that the mechanism can effectively filter noise from different modalities and adaptively adjust its contribution. Therefore, dynamic weight allocation is crucial to ensuring model stability in the presence of noise, inconsistencies in context, or modal conflicts.
Furthermore, removing the bimodal or trimodal joint layer reduces the model to a single-layer structure. As shown in Table 4, the hierarchical structure performs better than the single-layer structure. This is because, although single-layer trimodal attention theoretically possesses strong expressive power, directly modeling the complete trimodal interaction within a single stage may introduce excessive high-order coupling and modality dominance effects into the shared attention space. Although hierarchical interaction uses more interaction attention, increasing model complexity, it decomposes trimodal learning into two progressive stages, achieving local cross-modal alignment before global fusion. This structured design improves representation efficiency and optimization stability.
Effects of different modalities: To demonstrate the effects of different modalities, we remove one or two modalities at a time. From Table 4, we observe: (1) From the results of the single-modal analysis, the text modality performed the best, while the audio and visual single-modal performances were lower, mainly due to the influence of factors such as background noise, changes in speaking style, and blurred facial expressions. (2) Any combination of two modalities significantly outperforms the corresponding single-modal results. For example, both Text+Audio and Audio+Visual show significant improvements, indicating that audio and visual signals can supplement emotional details that are difficult to capture in the text modality. (3) The performance of the bimodal combination is still lower than that of the complete trimodal model, further demonstrating that the three modalities are complementary in emotion expression and that trimodal interaction can provide the most comprehensive emotional information.
The proposed hierarchical joint cross-modal attention and gating (HJCAG) module mainly consists of lightweight fully connected layers, attention projections, and gating networks. Compared with the backbone encoders (e.g., RoBERTa and ResNet-50), the additional parameters introduced by HJCAG are relatively small.
In summary, the ablation experiments fully demonstrate that both bimodal and trimodal joint interaction modules play a crucial role in performance improvement, and the multimodal gating mechanism effectively enhances the model’s adaptability to contributions from different modalities. The HJCAG model, through its hierarchical cross-modal attention mechanism and dynamic weight allocation, is able to capture more comprehensive and robust emotional representations.
This paper proposes a multimodal sentiment analysis method based on hierarchical joint cross-modal attention and gating mechanisms. This method models the semantic relationships between modalities at bimodal and trimodal levels and explicitly characterizes the high-order interaction relationships between joint features and single-modal features using a joint cross-modal attention mechanism, thereby achieving more effective semantic alignment and information fusion. Simultaneously, the introduced gating mechanism adaptively adjusts the contributions of different modalities, improving the discriminativeness and stability of the fused features. In real-world scenarios, the proposed framework may be challenged by noisy audio, subtle or ambiguous emotional expressions (e.g., sarcasm), and imperfect modality alignment. These factors highlight potential limitations and motivate future work on robustness enhancement and adaptive fusion strategies.
Acknowledgement: The authors would like to thank all contributors and institutions that supported this research.
Funding Statement: This research was supported by the Chongqing Basic Research and Frontier Exploration Project (Chongqing Natural Science Foundation) under Grant No. CSTB2022NSCQ-MSX0918, and the Science and Technology Research Project (Youth) of Chongqing Municipal Education Commission under Grant No. KJQN202301122.
Author Contributions: The authors confirm contribution to the paper as follows: Conceptualization, Shuqiu Tan; methodology, Chunsheng Tan; data curation, Chunsheng Tan; writing—original draft preparation, Chunsheng Tan; writing—review and editing, Shuqiu Tan and Chunsheng Tan; visualization, Chunsheng Tan; supervision, Shuqiu Tan and Yahui Liu; funding acquisition, Shuqiu Tan. All authors reviewed and approved the final version of the manuscript.
Availability of Data and Materials: The data presented in this study are openly available in (IEMOCAP) at (https://doi.org/10.1007/s10579-008-9076-6) and (MELD) at (https://doi.org/10.48550/arXiv.1810.02508) reference numbers [24,25].
Ethics Approval: Not applicable.
Conflicts of Interest: The authors declare no conflicts of interest.
References
1. Poria S, Cambria E, Bajpai R, Hussain A. A review of affective computing: from unimodal analysis to multimodal fusion. Inf Fusion. 2017;37:98–125. [Google Scholar]
2. Song H, Li J, Xia Z, Yang Z, Du X. Multimodal sentiment analysis based on pre-LN transformer interaction. In: Proceedings of the 6th IEEE Information Technology and Mechatronics Engineering Conference; 2022 Mar 4–6; Chongqing, China. p. 1609–13. [Google Scholar]
3. Zhang Y, Rong L, Song D, Zhang P. A survey on multimodal sentiment analysis. Pattern Recognit Artif Intell. 2020;33(5):426–38. doi:10.1007/s10462-023-10555-8. [Google Scholar] [CrossRef]
4. Shivappa ST, Trivedi MM, Rao BD. Audiovisual information fusion in human–computer interfaces and intelligent environments: a survey. Proc IEEE. 2010;98(10):1692–715. doi:10.1109/jproc.2010.2057231. [Google Scholar] [CrossRef]
5. Poria S, Cambria E, Hazarika D, Majumder N, Zadeh A, Morency LP. Context-dependent sentiment analysis in user-generated videos. In: Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers); 2017 Jul 30–Aug 4; Vancouver, BC, Canada. Stroudsburg, PA, USA: ACL. p. 873–83. [Google Scholar]
6. Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, et al. Attention is all you need. In: Advances in neural information processing systems. Red Hook, NY, USA: Curran Associates Inc.; 2017. [Google Scholar]
7. Ma H, Wang J, Lin H, Zhang B, Zhang Y, Xu B. A transformer-based model with self-distillation for multimodal emotion recognition in conversations. IEEE Trans Multimedia. 2024;26(11):776–88. doi:10.1109/tmm.2023.3271019. [Google Scholar] [CrossRef]
8. Poria S, Chaturvedi I, Cambria E, Hussain A. Convolutional MKL-based multimodal emotion recognition and sentiment analysis. In: Proceedings of the 16th IEEE International Conference on Data Mining; 2016 Dec 12–15; Barcelona, Spain. p. 439–48. [Google Scholar]
9. Nojavanasghari B, Gopinath D, Koushik J, Baltrušaitis T, Morency LP. Deep multimodal fusion for persuasiveness prediction. In: Proceedings of the ACM International Conference on Multimodal Interaction; 2016 Nov 12–16; Tokyo, Japan. p. 284–88. [Google Scholar]
10. Zadeh A, Chen M, Poria S, Cambria E, Morency LP. Tensor fusion network for multimodal sentiment analysis. In: Proceedings of the Conference on Empirical Methods in Natural Language Processing; 2017 Sep 9–11; Copenhagen, Denmark. p. 1103–14. [Google Scholar]
11. Liu Z, Shen Y, Lakshminarasimhan VB, Liang PP, Zadeh AB, Morency LP. Efficient low-rank multimodal fusion with modality-specific factors. In: Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Long Papers); 2018 Jul 15–20; Melbourne, Australia. Stroudsburg, PA, USA: ACL; 2018. p. 2247–56. [Google Scholar]
12. Tsai Y-HH, Bai S, Liang PP, Kolter JZ, Morency LP, Salakhutdinov R. Multimodal transformer for unaligned multimodal language sequences. In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics; 2019. p. 6558–69. [Google Scholar]
13. Ma L, Yao Y, Liang T, Liu T. Multi-scale cooperative multimodal transformers for multimodal sentiment analysis in videos. In: Proceedings of the Australasian Joint Conference on Artificial Intelligence; 2024 Nov 25–29; Melbourne, Australia. p. 281–97. [Google Scholar]
14. Sun L, Lian Z, Liu B, Tao J. Efficient multimodal transformer with dual-level feature restoration for robust multimodal sentiment analysis. IEEE Trans Affect Comput. 2023;15(1):309. doi:10.1109/taffc.2023.3274829. [Google Scholar] [CrossRef]
15. Hu G, Lin TE, Zhao Y, Lu G, Wu Y, Li Y. UniMSE: towards unified multimodal sentiment analysis and emotion recognition. arXiv:2211.11256. 2022. doi:10.48550/arXiv.2211.11256. [Google Scholar] [CrossRef]
16. Shi T, Huang S-L. MultiEMO: an attention-based correlation-aware multimodal fusion framework for emotion recognition in conversations. In: Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers); 2023 Jul 9–14; Toronto, ON, Canada. Stroudsburg, PA, USA: ACL; 2023. p. 14752–66. [Google Scholar]
17. Arevalo J, Solorio T, Montes-y-Gómez M, González F. Gated multimodal units for information fusion. arXiv:1702.01992. 2017. doi:10.48550/arXiv.1702.01992. [Google Scholar] [CrossRef]
18. Rahman W, Hasan MK, Lee S, Bagher Zadeh A, Mao C, Morency LP, et al. Integrating multimodal information in large pretrained transformers. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics; 2020 Jul 5–10; Online. p. 2359–69. [Google Scholar]
19. Liu Y, Ott M, Goyal N, Du J, Joshi M, Chen D, et al. RoBERTa: a robustly optimized BERT pretraining approach. arXiv:1907.11692. 2019. doi:10.48550/arXiv.1907.11692. [Google Scholar] [CrossRef]
20. Eyben F, Weninger F, Gross F, Schuller B. Recent developments in openSMILE, the Munich open-source multimedia feature extractor. In: Proceedings of the 21st ACM International Conference on Multimedia; 2013 Oct 21–25; Barcelona, Spain. p. 835–8. [Google Scholar]
21. He K, Zhang X, Ren S, Sun J. Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition; 2016 Jun 27–30; Las Vegas, NV, USA. p. 770–8. [Google Scholar]
22. Wang W, Tran D, Feiszli M. What makes training multi-modal classification networks hard? In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition; 2020 Jun 13–19; Seattle, WA, USA. p. 12692–702. [Google Scholar]
23. Nagrani A, Yang S, Arnab A, Jansen A, Schmid C, Sun C. Attention bottlenecks for multimodal fusion. In: Advances in neural information processing systems. Red Hook, NY, USA: Curran Associates Inc.; 2021. p. 14200–13. [Google Scholar]
24. Busso C, Bulut M, Lee CC, Kazemzadeh A, Mower E, Kim S, et al. IEMOCAP: interactive emotional dyadic motion capture database. Lang Resour Eval. 2008;42(4):335–59. [Google Scholar]
25. Poria S, Hazarika D, Majumder N, Naik G, Cambria E, Mihalcea R. MELD: a multimodal multi-party dataset for emotion recognition in conversations. In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics; 2019 Jul 28–Aug 2; Florence, Italy. p. 527–36. [Google Scholar]
26. Qi Y, Ibrayim M, Tohti T. Contextual xLSTM-based multimodal fusion for conversational emotion recognition. Pattern Anal Applic. 2025;28(3):132. doi:10.1007/s10044-025-01508-8. [Google Scholar] [CrossRef]
27. Majumder N, Poria S, Hazarika D, Mihalcea R, Gelbukh A, Cambria E. DialoguERN: an attentive RNN for emotion detection in conversations. In: Proceedings of the AAAI Conference on Artificial Intelligence; 2019 Jan 29–31; Honolulu, HI, USA. p. 6818–25. [Google Scholar]
28. Yi Z, Zhao Z, Shen Z, Zhang T. Multimodal fusion via hypergraph autoencoder and contrastive learning for emotion recognition in conversation. In: Proceedings of the 32nd ACM International Conference on Multimedia; 2024 Oct 28–Nov 1; Melbourne, Australia. p. 4341–8. [Google Scholar]
29. Tu G, Xie T, Liang B, Wang H, Xu R. Adaptive graph learning for multimodal conversational emotion detection. Proc AAAI Conf Artif Intell. 2024;38(17):19089–97. doi:10.1609/aaai.v38i17.29876. [Google Scholar] [CrossRef]
30. Zou S, Huang X, Shen X, Liu H. Improving multimodal fusion with main modal transformer for emotion recognition in conversation. Knowl Based Syst. 2022;258(4):109978. doi:10.1016/j.knosys.2022.109978. [Google Scholar] [CrossRef]
31. Shou Y, Liu H, Cao X, Meng D, Dong B. A low-rank matching attention based cross-modal feature fusion method for conversational emotion recognition. IEEE Trans Affect Comput. 2024;16(2):1177–89. doi:10.1109/taffc.2024.3498443. [Google Scholar] [CrossRef]
32. Yang J, Dong X, Du X. SMFNM: semi-supervised multimodal fusion network with main-modal for real-time emotion recognition in conversations. J King Saud Univ Comput Inf Sci. 2023;35(9):101791. doi:10.1016/j.jksuci.2023.101791. [Google Scholar] [CrossRef]
33. Li J, Wang X, Liu Y, Zeng Z. CFN-ESA: a cross-modal fusion network with emotion-shift awareness for dialogue emotion recognition. IEEE Trans Affect Comput. 2024;15(4):1919–33. [Google Scholar]
34. Mao Y, Liu G, Wang X, Gao W, Li X. DialogueTRM: exploring multi-modal emotional dynamics in a conversation. In: Findings of the Association for Computational Linguistics: EMNLP. Stroudsburg, PA, USA: ACL; 2021. p. 2694–704. [Google Scholar]
35. Hu J, Liu Y, Zhao J, Jin Q. MMGCN: multimodal fusion via deep graph convolution network for emotion recognition in conversation. In: Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics (Long Papers); 2021 Aug 1–6; Virtual. Stroudsburg, PA, USA: ACL; 2021. p. 5666–75. [Google Scholar]
Cite This Article
Copyright © 2026 The Author(s). Published by Tech Science Press.This work is licensed under a Creative Commons Attribution 4.0 International License , which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.


Submit a Paper
Propose a Special lssue
View Full Text
Download PDF

Downloads
Citation Tools