Zobrazeno 1 - 10
of 60
pro vyhledávání: '"Yongwon Jeong"'
Publikováno v:
2018 International Conference on Electronics, Information, and Communication (ICEIC).
Convolutional neural networks (CNNs) have been shown to improve classification tasks such as automatic speech recognition (ASR). Furthermore, the CNN with very deep architecture lowered the word error rate (WER) in reverberant and noisy environments.
Autor:
Yongwon Jeong
Publikováno v:
Journal of Signal Processing Systems. 82:303-310
This paper presents the basis-based speaker adaptation method that includes approaches using principal component analysis (PCA) and two-dimensional PCA (2DPCA). The proposed method partitions the hidden Markov model (HMM) mean vectors of training mod
Publikováno v:
O-COCOSDA
The performance of automatic speech recognition (ASR) has been greatly improved by deep neural network (DNN) acoustic models. However, DNN-based systems still perform poorly in reverberant environments. Convolutional neural network (CNN) acoustic mod
Autor:
Yongwon Jeong
Publikováno v:
Speech Communication. 58:1-10
We present an adaptation of a hidden Markov model (HMM)-based automatic speech recognition system to the target speaker and noise environment. Given HMMs built from various speakers and noise conditions, we build tensorvoices that capture the interac
Autor:
Hyung Soon Kim, Yongwon Jeong
Publikováno v:
IEICE Transactions on Information and Systems. :2195-2199
Autor:
Yongwon Jeong
Publikováno v:
IEICE Transactions on Information and Systems. :2200-2204
Autor:
Yongwon Jeong
Publikováno v:
Speech Communication. 55:340-346
We present a unified framework for basis-based speaker adaptation techniques, which subsumes eigenvoice speaker adaptation using principal component analysis (PCA) and speaker adaptation using two-dimensional PCA (2DPCA). The basic idea is to partiti
Publikováno v:
IEICE Transactions on Information and Systems. :2152-2155
Autor:
Yongwon Jeong
Publikováno v:
IEICE Transactions on Information and Systems. :1402-1405
Autor:
Yongwon Jeong
Publikováno v:
IEEE Transactions on Audio, Speech, and Language Processing. 20:2352-2364
In this paper, we describe basis-based speaker adaptation techniques using the matrix representation of training models. Bases are obtained from training models by decomposition techniques for matrix-variate objects: two-dimensional principal compone