Zobrazeno 1 - 10
of 28
pro vyhledávání: '"Yu Ting Yeung"'
Publikováno v:
Interspeech 2022.
Publikováno v:
Interspeech 2022.
Publikováno v:
ICASSP 2022 - 2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
This study propose a fully automated system for speech correction and accent reduction. Consider the application scenario that a recorded speech audio contains certain errors, e.g., inappropriate words, mispronunciations, that need to be corrected. T
Externí odkaz:
https://explore.openaire.eu/search/publication?articleId=doi_dedup___::55fe48cd42469df24ebd2ed1ff932b55
Publikováno v:
Interspeech 2021.
Autor:
Yu Ting Yeung, Nianzu Zheng, Disong Wang, Helen Meng, Liqun Deng, Xunying Liu, Yang Zhang, Xiao Chen
Publikováno v:
ICASSP
Sequence-to-sequence (seq2seq) learning has greatly improved text-to-speech (TTS) synthesis performance, but effective implementation on resource-restricted devices remains challenging as seq2seq models are usually computationally expensive and memor
Dysarthric speech detection (DSD) systems aim to detect characteristics of the neuromotor disorder from speech. Such systems are particularly susceptible to domain mismatch where the training and testing data come from the source and target domains r
Externí odkaz:
https://explore.openaire.eu/search/publication?articleId=doi_dedup___::c4c1f9fd40331e1f0ac2c2f21e3944ed
One-shot voice conversion (VC), which performs conversion across arbitrary speakers with only a single target-speaker utterance for reference, can be effectively achieved by speech representation disentanglement. Existing work generally ignores the c
Externí odkaz:
https://explore.openaire.eu/search/publication?articleId=doi_dedup___::7a0f25900987f706d0b515bf54af00d1
Autor:
Nianzu Zheng, Liqun Deng, Wenyong Huang, Yu Ting Yeung, Baohua Xu, Yuanyuan Guo, Yasheng Wang, Xiao Chen, Xin Jiang, Qun Liu
Mispronunciation detection and diagnosis (MDD) is a popular research focus in computer-aided pronunciation training (CAPT) systems. End-to-end (e2e) approaches are becoming dominant in MDD. However an e2e MDD model usually requires entire speech utte
Externí odkaz:
https://explore.openaire.eu/search/publication?articleId=doi_dedup___::9ae6803b10e9d4bcda4209575a1448bd
Publikováno v:
INTERSPEECH
Transformer has achieved competitive performance against state-of-the-art end-to-end models in automatic speech recognition (ASR), and requires significantly less training time than RNN-based models. The original Transformer, with encoder-decoder arc
Externí odkaz:
https://explore.openaire.eu/search/publication?articleId=doi_dedup___::0a5c52f16f3135a4f89fa72375f4d97b