Autor: |
Avola D; Department of Computer Science, Sapienza University of Rome, Via Salaria 113, Rome 00198, Italy., Cinque L; Department of Computer Science, Sapienza University of Rome, Via Salaria 113, Rome 00198, Italy., Mambro AD; Department of Computer Science, Sapienza University of Rome, Via Salaria 113, Rome 00198, Italy., Fagioli A; Department of Computer Science, Sapienza University of Rome, Via Salaria 113, Rome 00198, Italy., Marini MR; Department of Computer Science, Sapienza University of Rome, Via Salaria 113, Rome 00198, Italy., Pannone D; Department of Computer Science, Sapienza University of Rome, Via Salaria 113, Rome 00198, Italy., Fanini B; Institute of Heritage Science, National Research Council, Area della Ricerca Roma 1, SP35d, 9, Montelibretti 00010, Italy., Foresti GL; Department of Computer Science, Mathematics and Physics, University of Udine, Via delle Scienze 206, Udine 33100, Italy. |
Abstrakt: |
Emotion recognition plays an essential role in human-human interaction since it is a key to understanding the emotional states and reactions of human beings when they are subject to events and engagements in everyday life. Moving towards human-computer interaction, the study of emotions becomes fundamental because it is at the basis of the design of advanced systems to support a broad spectrum of application areas, including forensic, rehabilitative, educational, and many others. An effective method for discriminating emotions is based on ElectroEncephaloGraphy (EEG) data analysis, which is used as input for classification systems. Collecting brain signals on several channels and for a wide range of emotions produces cumbersome datasets that are hard to manage, transmit, and use in varied applications. In this context, the paper introduces the Empátheia system, which explores a different EEG representation by encoding EEG signals into images prior to their classification. In particular, the proposed system extracts spatio-temporal image encodings, or atlases, from EEG data through the Processing and transfeR of Interaction States and Mappings through Image-based eNcoding (PRISMIN) framework, thus obtaining a compact representation of the input signals. The atlases are then classified through the Empátheia architecture, which comprises branches based on convolutional, recurrent, and transformer models designed and tuned to capture the spatial and temporal aspects of emotions. Extensive experiments were conducted on the Shanghai Jiao Tong University (SJTU) Emotion EEG Dataset (SEED) public dataset, where the proposed system significantly reduced its size while retaining high performance. The results obtained highlight the effectiveness of the proposed approach and suggest new avenues for data representation in emotion recognition from EEG signals. |