Evaluation of electrocardiogram: numerical vs. image data for emotion recognition system.
Autor: | Sayed Ismail SNM; Faculty of Information Science & Technology, Multimedia University, Bukit Beruang,, Melaka, 75450, Malaysia., Ab Aziz NA; Faculty of Engineering, Multimedia University, Bukit Beruang, Melaka, 75450, Malaysia., Ibrahim SZ; Faculty of Information Science & Technology, Multimedia University, Bukit Beruang,, Melaka, 75450, Malaysia., Nawawi SW; School of Electrical Engineering, Faculty of Engineering, Universiti Teknologi Malaysia, Skudai, Johor Bahru, 81310, Malaysia., Alelyani S; Center for Artificial Intelligence, King Khalid University, Abha, 61421, Saudi Arabia.; College of Computer Science, King Khalid University, Abha, 61421, Saudi Arabia., Mohana M; Center for Artificial Intelligence, King Khalid University, Abha, 61421, Saudi Arabia., Chia Chun L; Hexon Data Sdn Bhd, Kuala Lumpur, 59200, Malaysia. |
---|---|
Jazyk: | angličtina |
Zdroj: | F1000Research [F1000Res] 2021 Nov 04; Vol. 10, pp. 1114. Date of Electronic Publication: 2021 Nov 04 (Print Publication: 2021). |
DOI: | 10.12688/f1000research.73255.2 |
Abstrakt: | Background: The electrocardiogram (ECG) is a physiological signal used to diagnose and monitor cardiovascular disease, usually using 2- D ECG. Numerous studies have proven that ECG can be used to detect human emotions using 1-D ECG; however, ECG is typically captured as 2-D images rather than as 1-D data. There is still no consensus on the effect of the ECG input format on the accuracy of the emotion recognition system (ERS). The ERS using 2-D ECG is still inadequately studied. Therefore, this study compared ERS performance using 1-D and 2-D ECG data to investigate the effect of the ECG input format on the ERS. Methods: This study employed the DREAMER dataset, which contains 23 ECG recordings obtained during audio-visual emotional elicitation. Numerical data was converted to ECG images for the comparison. Numerous approaches were used to obtain ECG features. The Augsburg BioSignal Toolbox (AUBT) and the Toolbox for Emotional feature extraction from Physiological signals (TEAP) extracted features from numerical data. Meanwhile, features were extracted from image data using Oriented FAST and rotated BRIEF (ORB), Scale Invariant Feature Transform (SIFT), KAZE, Accelerated-KAZE (AKAZE), Binary Robust Invariant Scalable Keypoints (BRISK), and Histogram of Oriented Gradients (HOG). Dimension reduction was accomplished using linear discriminant analysis (LDA), and valence and arousal were classified using the Support Vector Machine (SVM). Results: The experimental results show 1-D ECG-based ERS achieved 65.06% of accuracy and 75.63% of F1 score for valence, and 57.83% of accuracy and 44.44% of F1-score for arousal. For 2-D ECG-based ERS, the highest accuracy and F1-score for valence were 62.35% and 49.57%; whereas, the arousal was 59.64% and 59.71%. Conclusions: The results indicate that both inputs work comparably well in classifying emotions, which demonstrates the potential of 1-D and 2-D as input modalities for the ERS. Competing Interests: No competing interests were disclosed. (Copyright: © 2022 Sayed Ismail SNM et al.) |
Databáze: | MEDLINE |
Externí odkaz: |