Deep Multimodal CNN Fusion Scheme for Accurate Stress Identification

  • Sarala Patchala Associate Professor, Department of ECE, KKR & KSR Institute of Technology and Sciences, Guntur, Andhra Pradesh, India, Andhra Pradesh, India
  • Banda Snv Ramana Murthy Assistant Professor, Department of CSE AIML, ADITYA University, Surampalem http://orcid.org/0009-0003-8371-1691
  • Haritha Tummala Assistant Professor, MIC college of Engineering, Vijayawada, Andhra Pradesh https://orcid.org/0000-0001-6534-4929
  • Bandla Srinivasa Rao Dept of Computer Science Engineering, Koneru Lakshmaiah Education Foundation, Vaddeswaram, AP http://orcid.org/0000-0002-2512-1023
  • V.V. Jaya Rama Krishnaiah Professor of CSE, Teegala Krishna Reddy Engineering College,Telangana
  • Vullam Nagagopiraju Professor, Department of Computer Science and Engineering, Chalapathi Institute of Engineering and Technology, Guntur https://orcid.org/0009-0008-4894-375X
  • Suneetha Jalli Associate Professor, Department of ECE, KKR & KSR Institute of Technology and Sciences, Guntur, Andhra Pradesh, India, Andhra Pradesh, India
  • Inakoti Ramesh Raja Associate Professor, Department of Electronics & Communication Engineering, Aditya University Surampalem, A.P., India https://orcid.org/0009-0006-6806-2605

Abstract

Stress is a major issue in today’s life. It harms health and lower work output. People sometimes do not notice when under stress. That is why early stress detection is important. This paper uses two types of body signals: ECG (Electrocardiogram) and EDA (Electrodermal Activity). Both are physiological signals and help measure stress levels. A deep learning model named CNN (Convolutional Neural Network) is used. CNN has many layers. Each layer captures different kinds of features—low-level, mid-level and high-level. These features are useful in identifying stress. Instead of using features from only one level, this paper combines all three levels. This process is named as hierarchical feature fusion. It helps in creating a strong and rich representation of the signals. The features from ECG and EDA are first extracted at different CNN layers. Then, a module termed MMTM (Multimodal Transfer Module) is used. This module helps combine features from both signals. It improves the way the model learns from the data. The model is tested using both raw data and features from selected frequency bands. Results show that using features from all three CNN levels gives better performance. The proposed model performs better than existing models when using frequency band features. This shows that combining low, mid and high-level CNN features with multimodal fusion is helpful. It improves the accuracy and generalization of stress detection. This method works better across different datasets and different people. The proposed system is a useful tool in real-world stress detection systems.

 

Downloads

Download data is not yet available.
Published
2026-04-28
How to Cite
Patchala, S., Ramana Murthy, B. S., Tummala, H., Srinivasa Rao, B., Rama Krishnaiah, V. J., Nagagopiraju, V., Jalli, S., & Ramesh Raja, I. (2026). Deep Multimodal CNN Fusion Scheme for Accurate Stress Identification. ITEGAM-JETIA, 12(58), 1245-1255. https://doi.org/10.5935/jetia.v12i58.3304
Section
Articles