Zobrazeno 1 - 10
of 21
pro vyhledávání: '"Vincent Colotte"'
Publikováno v:
Neural Networks
Neural Networks, Elsevier, 2021, 141, pp.315-329. ⟨10.1016/j.neunet.2021.04.021⟩
Neural Networks, 2021, 141, pp.315-329. ⟨10.1016/j.neunet.2021.04.021⟩
Neural Networks, Elsevier, 2021, 141, pp.315-329. ⟨10.1016/j.neunet.2021.04.021⟩
Neural Networks, 2021, 141, pp.315-329. ⟨10.1016/j.neunet.2021.04.021⟩
International audience; Great improvement has been made in the field of expressive audiovisual Text-to-Speech synthesis (EAVTTS) thanks to deep learning techniques. However, generating realistic speech is still an open issue and researchers in this a
Externí odkaz:
https://explore.openaire.eu/search/publication?articleId=doi_dedup___::fb49718d2c957d8ff1ca49d1428e6a35
https://hal.inria.fr/hal-03204193
https://hal.inria.fr/hal-03204193
Publikováno v:
Language Resources and Evaluation
Language Resources and Evaluation, Springer Verlag, 2020, ⟨10.1007/s10579-020-09500-w⟩
Language Resources and Evaluation, 2020, ⟨10.1007/s10579-020-09500-w⟩
Language Resources and Evaluation, Springer Verlag, 2020, ⟨10.1007/s10579-020-09500-w⟩
Language Resources and Evaluation, 2020, ⟨10.1007/s10579-020-09500-w⟩
International audience; In this paper, we present a multimodal acquisition setup that combines different motion-capture systems. This system is mainly aimed for recording expressive audiovisual corpus in the context of audiovisual speech synthesis. W
Externí odkaz:
https://explore.openaire.eu/search/publication?articleId=doi_dedup___::58b1f70a3ea06c35ca558f46b21b004b
https://hal.archives-ouvertes.fr/hal-02907046
https://hal.archives-ouvertes.fr/hal-02907046
Publikováno v:
INTERSPEECH 2020
INTERSPEECH 2020, Oct 2020, Shanghai / Virtual, China
INTERSPEECH
INTERSPEECH 2020, Oct 2020, Shanghai / Virtual, China
INTERSPEECH
International audience; In this paper, we present a novel deep metric learning architecture along with variational inference incorporated in a paramet-ric multispeaker expressive text-to-speech (TTS) system. We proposed inverse autoregressive flow (I
Externí odkaz:
https://explore.openaire.eu/search/publication?articleId=doi_dedup___::66f3de9ad2a382e71bb19836d45dc112
https://hal.inria.fr/hal-02572106v3
https://hal.inria.fr/hal-02572106v3
Publikováno v:
Multimedia Tools and Applications
Multimedia Tools and Applications, Springer Verlag, 2020, ⟨10.1007/s11042-020-09901-7⟩
Multimedia Tools and Applications, 2020, ⟨10.1007/s11042-020-09901-7⟩
Multimedia Tools and Applications, Springer Verlag, 2020, ⟨10.1007/s11042-020-09901-7⟩
Multimedia Tools and Applications, 2020, ⟨10.1007/s11042-020-09901-7⟩
International audience; Sound duration is responsible for rhythm and speech rate. Furthermore , in some languages phoneme length is an important phonetic and prosodic factor. For example, in Arabic, gemination and vowel quantity are two important cha
Externí odkaz:
https://explore.openaire.eu/search/publication?articleId=doi_dedup___::968dc91fd96a538bf9a30740d13cad76
https://hdl.handle.net/11574/212842
https://hdl.handle.net/11574/212842
Publikováno v:
Statistical Language and Speech Processing ISBN: 9783030594299
SLSP
SLSP 2020-8th International Conference on Statistical Language and Speech Processing
SLSP 2020-8th International Conference on Statistical Language and Speech Processing, Oct 2020, Cardiff / Virtual, United Kingdom
SLSP
SLSP 2020-8th International Conference on Statistical Language and Speech Processing
SLSP 2020-8th International Conference on Statistical Language and Speech Processing, Oct 2020, Cardiff / Virtual, United Kingdom
In this paper, we propose an approach relying on multiclass N-pair loss based deep metric learning in recurrent conditional variational autoencoder (RCVAE). We used RCVAE for implementation of multispeaker expressive text-to-speech (TTS) system. The
Externí odkaz:
https://explore.openaire.eu/search/publication?articleId=doi_dedup___::56935c6443b934c97f625ca3326142e6
https://doi.org/10.1007/978-3-030-59430-5_13
https://doi.org/10.1007/978-3-030-59430-5_13
Publikováno v:
INTERSPEECH 2019-20th Annual Conference of the International Speech Communication Association
INTERSPEECH 2019-20th Annual Conference of the International Speech Communication Association, Sep 2019, Graz, Austria
INTERSPEECH
INTERSPEECH 2019-20th Annual Conference of the International Speech Communication Association, Sep 2019, Graz, Austria
INTERSPEECH
International audience; In recent years, the performance of speech synthesis systems has been improved thanks to deep learning-based models, but generating expressive audiovisual speech is still an open issue. The variational auto-encoders (VAE)s are
Externí odkaz:
https://explore.openaire.eu/search/publication?articleId=doi_dedup___::1ae9273155b76bb8f695011c91835036
https://hal.inria.fr/hal-02175776/document
https://hal.inria.fr/hal-02175776/document
Publikováno v:
Proceedings of the International Neural Networks Society ISBN: 9783030168407
INNSBDDL
INNSBDDL
Deep neural networks (DNN) are gaining increasing interest in speech processing applications, especially in text-to-speech synthesis. Actually state-of-the-art speech generation tools, like MERLIN and WAVENET are totally DNN-based. However, every lan
Externí odkaz:
https://explore.openaire.eu/search/publication?articleId=doi_________::521a4d197ee8f01fc5f25aa4e8db21a5
https://doi.org/10.1007/978-3-030-16841-4_20
https://doi.org/10.1007/978-3-030-16841-4_20
Publikováno v:
International Journal of Speech Technology
International Journal of Speech Technology, Springer Verlag, 2018, pp.1-12. ⟨10.1007/s10772-018-09558-6⟩
International Journal of Speech Technology, 2018, pp.1-12. ⟨10.1007/s10772-018-09558-6⟩
International Journal of Speech Technology, Springer Verlag, 2018, pp.1-12. ⟨10.1007/s10772-018-09558-6⟩
International Journal of Speech Technology, 2018, pp.1-12. ⟨10.1007/s10772-018-09558-6⟩
International audience; This paper investigates the use of hidden Markov models (HMM) for Modern Standard Arabic speech synthesis. HMM-basedspeech synthesis systems require a description of each speech unit with a set of contextual features that spec
Externí odkaz:
https://explore.openaire.eu/search/publication?articleId=doi_dedup___::49fd6a113bf451397539e9c1ca08f216
https://hal.inria.fr/hal-01936963/file/IJST_Accepted_Version.pdf
https://hal.inria.fr/hal-01936963/file/IJST_Accepted_Version.pdf
Publikováno v:
9th International Conference on Speech Prosody
9th International Conference on Speech Prosody, Jun 2018, Poznań, Poland
9th International Conference on Speech Prosody, Jun 2018, Poznań, Poland
International audience; Duration modeling is a key task for every parametric speech synthesis system. Though such parametric systems have been adapted to many languages, no special attention was paid to explicitly handling Arabic speech characteristi
Externí odkaz:
https://explore.openaire.eu/search/publication?articleId=doi_dedup___::ebfbda24a4a53b83db8a41e285f2ce56
https://hal.inria.fr/hal-01889917/document
https://hal.inria.fr/hal-01889917/document
Publikováno v:
The 14th International Conference on Auditory-Visual Speech Processing
The 14th International Conference on Auditory-Visual Speech Processing, KTH, Aug 2017, Stockholm, Sweden
AVSP
The 14th International Conference on Auditory-Visual Speech Processing, KTH, Aug 2017, Stockholm, Sweden
AVSP
Proceedings on line: http://avsp2017.loria.fr/proceedings/; International audience; In the context of developing an expressive audiovisual speech synthesis system, the quality of the audiovisual corpus from which the 3D visual data will be extracted
Externí odkaz:
https://explore.openaire.eu/search/publication?articleId=doi_dedup___::273eb1c3fe9f5983c96242ea53ef99b2
https://hal.inria.fr/hal-01596614/document
https://hal.inria.fr/hal-01596614/document