Emotion Dependent Facial Animation from Affective Speech

Autor: Sasan Asadiabadi, Engin Erzin, Rizwan Sadiq
Přispěvatelé: Sadiq, Rizwan, Asadiabadi, Sasan, Erzin, Engin (ORCID 0000-0002-2715-2368 & YÖK ID 34503), Graduate School of Sciences and Engineering, College of Engineering, Department of Electrical and Electronics Engineering, Department of Computer Engineering
Rok vydání: 2020
Předmět:
FOS: Computer and information sciences
Sound (cs.SD)
Mean squared error
Speech recognition
ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION
02 engineering and technology
Universal model
Computer Science - Sound
Audio and Speech Processing (eess.AS)
FOS: Electrical engineering
electronic engineering
information engineering

0202 electrical engineering
electronic engineering
information engineering

Computer facial animation
ComputingMethodologies_COMPUTERGRAPHICS
Landmark
business.industry
Deep learning
Estimator
020206 networking & telecommunications
020207 software engineering
Animation
Multimedia (cs.MM)
Multimedia signal processing
Training
Visualization
Shape
Training data
Facial animation
Speech processing
Artificial intelligence
business
Psychology
Computer Science - Multimedia
Electrical Engineering and Systems Science - Audio and Speech Processing
Zdroj: MMSP
2020 IEEE 22nd International Workshop on Multimedia Signal Processing (MMSP)
DOI: 10.1109/mmsp48831.2020.9287086
Popis: In human-to-computer interaction, facial animation in synchrony with affective speech can deliver more naturalistic conversational agents. In this paper, we present a two-stage deep learning approach for affective speech driven facial shape animation. In the first stage, we classify affective speech into seven emotion categories. In the second stage, we train separate deep estimators within each emotion category to synthesize facial shape from the affective speech. Objective and subjective evaluations are performed over the SAVEE dataset. The proposed emotion dependent facial shape model performs better in terms of the Mean Squared Error (MSE) loss and in generating the landmark animations, as compared to training a universal model regardless of the emotion.
Scientific and Technological Research Council of Turkey (TÜBİTAK)
Databáze: OpenAIRE