Zobrazeno 1 - 10
of 31
pro vyhledávání: '"Kevin El Haddad"'
Publikováno v:
Informatics, Vol 8, Iss 4, p 84 (2021)
In this paper, we study the controllability of an Expressive TTS system trained on a dataset for a continuous control. The dataset is the Blizzard 2013 dataset based on audiobooks read by a female speaker containing a great variability in styles and
Externí odkaz:
https://doaj.org/article/ce613f2836204276a2564a149c22b163
Autor:
Kevin El Haddad, Yara Rizk, Louise Heron, Nadine Hajj, Yong Zhao, Jaebok Kim, Trung Ngô Trọng, Minha Lee, Marwan Doumit, Payton Lin, Yelin Kim, Hüseyin Çakmak
Publikováno v:
Journal of Science and Technology of the Arts, Vol 10, Iss 2 (2018)
In this work, we established the foundations of a framework with the goal to build an end-to-end naturalistic expressive listening agent. The project was split into modules for recognition of the user’s paralinguistic and nonverbal expressions, pre
Externí odkaz:
https://doaj.org/article/fa1ac0b904974532b6abf0ac58fa1261
Publikováno v:
2022 10th International Conference on Affective Computing and Intelligent Interaction (ACII).
Publikováno v:
INTERSPEECH
Despite the growing interest for expressive speech synthesis, synthesis of nonverbal expressions is an under-explored area. In this paper we propose an audio laughter synthesis system based on a sequence-to-sequence TTS synthesis system. We leverage
Externí odkaz:
https://explore.openaire.eu/search/publication?articleId=doi_dedup___::f610e83ed9ceef71f532145a205faeae
http://arxiv.org/abs/2008.09483
http://arxiv.org/abs/2008.09483
Publikováno v:
HRI (Companion)
State of the art in speech synthesis considerably reduced the gap between synthetic and human speech on the perception level. However the impact of a speech style control on the perception is not well known. In this paper, we propose a method to anal
Publikováno v:
Computers & Electrical Engineering. 62:588-600
In this paper, we present our work on analysis and classification of smiled vowels, chuckling (or shaking) vowels and laughter syllables. This work is part of a larger framework that aims at assessing the level of amusement in speech using the audio
As part of the Human-Computer Interaction field, Expressive speech synthesis is a very rich domain as it requires knowledge in areas such as machine learning, signal processing, sociology, and psychology. In this chapter, we will focus mostly on the
Externí odkaz:
https://explore.openaire.eu/search/publication?articleId=doi_dedup___::dc8bfd02cc65328c1a59047e1dbe764f
http://www.intechopen.com/articles/show/title/the-theory-behind-controllable-expressive-speech-synthesis-a-cross-disciplinary-approach
http://www.intechopen.com/articles/show/title/the-theory-behind-controllable-expressive-speech-synthesis-a-cross-disciplinary-approach
Publikováno v:
ICMI
Smiles and laughs have been the subject of many studies over the past decades, due to their frequent occurrence in interactions, as well as their social and emotional functions in dyadic conversations. In this paper we push forward previous work by p
Publikováno v:
INTERSPEECH
The field of Text-to-Speech has experienced huge improvements last years benefiting from deep learning techniques. Producing realistic speech becomes possible now. As a consequence, the research on the control of the expressiveness, allowing to gener
Publikováno v:
ACII Workshops
In this work we present an open-source avatar project aiming to be used in Human-Agent Interaction systems. We demonstrate here a system that is portable, works in real-time and implemented in a modular way. This is an attempt to respond to the need