Multimodal fusion emotion recognition algorithm based on face image and heart rate variability
Author:
Affiliation:

1.School of Computer Science,Nanjing University of Information Science and Technology;2.Nanjing University of Information Science and Technology;3.Nanjing University of Information Science and Technology Metax Research Institute;4.School of Artificial Intelligence,Nanjing University of Information Science and Technology

  • Article
  • | |
  • Metrics
  • |
  • Reference [26]
  • | |
  • Cited by
  • | |
  • Comments
    Abstract:

    Aiming at the problem of ineffective emotion recognition in virtual sports, this study proposes a multimodal fusion emotion recognition algorithm based on facial images and heart rate variability (MFER-FIHRV). By capturing and analyzing data from both facial image and heart rate variability (HRV) modalities, the proposed approach enables a fine-grained perception of users’ emotional changes, thereby facilitating a personalized human-computer interaction (HCI) experience tailored to their current emotional state. The algorithm first employs a multimodal fusion Transformer to achieve complementary learning between facial images and HRV data. Then, a multimodal feature fusion strategy is utilized to concatenate fused features with the original feature representations. Additionally, a lightweight self-attention mechanism is introduced to capture high-level semantic representations within the multimodal data. Extensive experiments were conducted on two public datasets, demonstrating the superior performance of the proposed method. The results confirm its effectiveness and robustness, providing valuable insights for the design and development of user experience systems.

    Reference
    [1] Maithri M ,U. R ,Anjan G , et al.Automated Emotion Recognition: Current Trends and Future Perspectives[J].Computer Methods and Programs in Biomedicine,2022,215(prepublish):106646-.
    [2] Chowdary K M ,Nguyen N T ,Hemanth J D .Deep learning-based facial emotion recognition for human–computer interaction applications[J].Neural Computing and Applications, 2021, (prepublish):1-18.
    [3] 杨春霞,韩煜,陈启岗,等.基于BERT与注意力机制的方面级隐式情感分析模型[J].南京信息工程大学学报(自然科学版),2023,15(05):551-560.Yang Chunxia, Han Yu, Chen Qigang, et al. An Aspect - level Implicit Sentiment Analysis Model Based on BERT and Attention Mechanism[J]. Journal of Nanjing University of Information Science Technology (Natural Science Edition), 2023, 15(05): 551-560.
    [4] Bisogni C, Castiglione A, Hossain S, et al. Impact of deep learning approaches on facial expression recognition in healthcare industries[J]. IEEE Transactions on Industrial Informatics, 2022, 18(8): 5619-5627.
    [5] Anam M ,Farhan A ,Zeeshan A , et al.Emotion recognition framework using multiple modalities for an effective human–computer interaction[J].The Journal of Supercomputing,2023,79(8):9320-9349.
    [6] 岳丽欣,刘自强,朱承宁.突发自然灾害事件短视频用户情感时空演化分析方法研究[J].图书情报工作,2024,68(24):77-91.Yue Lixin, LIU Ziqiang, ZHU Chengning. Research on temporal and spatial evolution of user emotion in short video of sudden natural disaster events [J]. Library and Information Work, 2019,68(24):77-91. (in Chinese)
    [7] Ayetiran E F. Attention-based aspect sentiment classification using enhanced learning through CNN-BiLSTM networks[J]. Knowledge-Based Systems, 2022, 252: 109409.
    [8] 任楚岚,仇全涛,劣思敏.融合TCN和BiLSTM的文本情感分析[J].计算机工程与设计,2024,45(07):2090-2096.Ren Chulan, Qiu Quantao, Lie Simin. Text Sentiment Analysis Combining TCN and BiLSTM [J]. Computer Engineering and Design,2024,45(07): 2090-2096.
    [9] Shiqing Z ,Yijiao Y ,Chen C , et al.Deep learning-based multimodal emotion recognition from audio, visual, and text modalities: A systematic review of recent advancements and future prospects[J].Expert Systems With Applications,2024,237(PC):121692-.
    [10] Hao S, Zhou Y, Liu P, et al. Bi-syntax guided transformer network for aspect sentiment triplet extraction[J]. Neurocomputing, 2024, 594: 127880.
    [11] Somarathna R, Bednarz T, Mohammadi G. Virtual reality for emotion elicitation–a review[J]. IEEE Transactions on Affective Computing, 2022.
    [12] Pan B, Hirota K, Dai Y, et al. Adaptive key-frame selection-based facial expression recognition via multi-cue dynamic features hybrid fusion[J]. Information Sciences, 2024, 660: 120138.
    [13] 冉瑞生,石凯,江小鹏,等.基于双注意力CrossViT的微表情识别方法[J].南京信息工程大学学报(自然科学版),2023,15(05):541-550.Ran Ruisheng, Shi Kai, Jiang Xiaopeng, et al. Micro - expression Recognition Method Based on Dual - attention CrossViT [J]. Journal of Nanjing University of Information Science Technology (Natural Science Edition), 2023,15(05):541-550.
    [14] Thuseethan S, Rajasegarar S, Yearwood J. Deep3DCANN: A Deep 3DCNN-ANN framework for spontaneous micro-expression recognition[J]. Information Sciences, 2023, 630: 341-355.
    [15] Suryady Z, Rahman A W A, Oktarina D. Human emotion classification based on phonocardiography signals (PCG)[C]//AIP Conference Proceedings. AIP Publishing, 2024, 2729(1).
    [16] Govarthan P K, Peddapalli S K, Ganapathy N, et al. Emotion classification using electrocardiogram and machine learning: A study on the effect of windowing techniques[J]. Expert Systems with Applications, 2024: 124371.
    [17] Wang X, Zhang J, He C, et al. A novel emotion recognition method based on the feature fusion of single-lead EEG and ECG signals[J]. IEEE Internet of Things Journal, 2023.
    [18] Mellouk W, Handouzi W. CNN-LSTM for automatic emotion recognition using contactless photoplythesmographic signals[J]. Biomedical Signal Processing and Control, 2023, 85: 104907.
    [19] Fu Z, Liu F, Xu Q, et al. LMR-CBT: Learning modality-fused representations with CB-transformer for multimodal emotion recognition from unaligned multimodal sequences[J]. Frontiers of Computer Science, 2024, 18(4): 184314.
    [20] Huang J, Zhou J, Tang Z, et al. TMBL: Transformer-based multimodal binding learning model for multimodal sentiment analysis[J]. Knowledge-Based Systems, 2024, 285: 111346.
    [21] Yang K, Wang C, Gu Y, et al. Behavioral and physiological signals-based deep multimodal approach for mobile emotion recognition[J]. IEEE Transactions on Affective Computing, 2021, 14(2): 1082-1097.
    [22] Ghosh R, Sinha D. Human emotion recognition by analyzing facial expressions, heart rate and blogs using deep learning method[J]. Innovations in Systems and Software Engineering, 2024, 20(3): 499-507.
    [23] Koelstra S ,Mühl C ,0001 S M , et al.DEAP: A Database for Emotion Analysis; Using Physiological Signals. [J].IEEE Trans. Affective Computing,2012,3(1):18-31.
    [24] Abadi K M ,Subramanian R ,Kia M S , et al.DECAF: MEG-Based Multimodal Database for Decoding Affective Physiological Responses.[J].IEEE Trans. Affective Computing, 2015, 6(3): 209-222.
    [25] Huang Q, Huang C, Wang X, et al. Facial expression recognition with grid-wise attention and visual transformer[J]. Information Sciences, 2021, 580: 35-54.
    [26] Wang X, Zhang J, He C, et al. A novel emotion recognition method based on the feature fusion of single-lead EEG and ECG signals[J]. IEEE Internet of Things Journal, 2023.
    Related
    Cited by
    Comments
    Comments
    分享到微博
    Submit
Get Citation
Share
Article Metrics
  • Abstract:17
  • PDF: 0
  • HTML: 0
  • Cited by: 0
History
  • Received:February 13,2025
  • Revised:March 17,2025
  • Adopted:March 18,2025
Article QR Code

Address:No. 219, Ningliu Road, Nanjing, Jiangsu Province

Postcode:210044

Phone:025-58731025