Zobrazeno 1 - 10
of 11
pro vyhledávání: '"Min-Jae Hwang"'
Publikováno v:
2022 International Conference on Electronics, Information, and Communication (ICEIC).
Publikováno v:
2022 International Conference on Electronics, Information, and Communication (ICEIC).
Publikováno v:
Interspeech 2021.
Publikováno v:
Interspeech 2021.
Publikováno v:
ICASSP
This paper proposes voicing-aware conditional discriminators for Parallel WaveGAN-based waveform synthesis systems. In this framework, we adopt a projection-based conditioning method that can significantly improve the discriminator's performance. Fur
Publikováno v:
ICASSP
In this paper, we propose a text-to-speech (TTS)-driven data augmentation method for improving the quality of a non-autoregressive (AR) TTS system. Recently proposed non-AR models, such as FastSpeech 2, have successfully achieved fast speech synthesi
Externí odkaz:
https://explore.openaire.eu/search/publication?articleId=doi_dedup___::867cb3922f4ca8e8c0e17fa067f5f51b
http://arxiv.org/abs/2010.13421
http://arxiv.org/abs/2010.13421
Publikováno v:
ICASSP
In this paper, we propose an improved LPCNet vocoder using a linear prediction (LP)-structured mixture density network (MDN). The recently proposed LPCNet vocoder has successfully achieved high-quality and lightweight speech synthesis systems by comb
Externí odkaz:
https://explore.openaire.eu/search/publication?articleId=doi_dedup___::95fafdd73a792901939d26bd68afc12d
http://arxiv.org/abs/2001.11686
http://arxiv.org/abs/2001.11686
Publikováno v:
IEEE Transactions on Multimedia. 20:45-54
This paper proposes a blind digital audio water- marking algorithm that utilizes the quantization index modulation (QIM) and the singular value decomposition (SVD) of stereo audio signals. Conventional SVD-based blind audio watermarking algorithms la
Autor:
Min-Jae Hwang, Hong-Goo Kang
Publikováno v:
INTERSPEECH
In this paper, we propose a deep learning (DL)-based parameter enhancement method for a mixed excitation linear prediction (MELP) speech codec in noisy communication environment. Unlike conventional speech enhancement modules that are designed to obt
Publikováno v:
INTERSPEECH