Zobrazeno 1 - 10
of 14
pro vyhledávání: '"Monisankha Pal"'
Meta-Learning With Latent Space Clustering in Generative Adversarial Network for Speaker Diarization
Autor:
Somer L. Bishop, Manoj Kumar, Catherine Lord, Shrikanth S. Narayanan, Tae Jin Park, Monisankha Pal, So Hyun Kim, Raghuveer Peri
Publikováno v:
IEEE/ACM Trans Audio Speech Lang Process
The performance of most speaker diarization systems with x-vector embeddings is both vulnerable to noisy environments and lacks domain robustness. Earlier work on speaker diarization using generative adversarial network (GAN) with an encoder network
Publikováno v:
International Journal of Speech Technology
International Journal of Speech Technology, 2021, ⟨10.1007/s10772-020-09785-w⟩
International Journal of Speech Technology, Springer Verlag, 2021, ⟨10.1007/s10772-020-09785-w⟩
International Journal of Speech Technology, 2021, ⟨10.1007/s10772-020-09785-w⟩
International Journal of Speech Technology, Springer Verlag, 2021, ⟨10.1007/s10772-020-09785-w⟩
International audience; In this paper, we introduce a frame selection strategy for improved detection of spoofed speech. A countermeasure (CM) system typically uses a Gaussian mixture model (GMM) based classifier for computing the log-likelihood scor
Externí odkaz:
https://explore.openaire.eu/search/publication?articleId=doi_dedup___::5924ce1335b5dfca20009d8538edc661
https://hal.science/hal-03008912
https://hal.science/hal-03008912
Autor:
Arindam Jati, Shrikanth S. Narayanan, Monisankha Pal, Chin-Cheng Hsu, Raghuveer Peri, Wael AbdAlmageed
Publikováno v:
ICASSP
Deep neural network based speaker recognition systems can easily be deceived by an adversary using minuscule imperceptible perturbations to the input speech samples. These adversarial attacks pose serious security threats to the speaker recognition s
Externí odkaz:
https://explore.openaire.eu/search/publication?articleId=doi_dedup___::20fcf9fe42fa8e72e770541d58635b84
Autor:
Chin-Cheng Hsu, Raghuveer Peri, Monisankha Pal, Wael AbdAlmageed, Shrikanth S. Narayanan, Arindam Jati
Publikováno v:
Computer Speech & Language. 68:101199
Robust speaker recognition, including in the presence of malicious attacks, is becoming increasingly important and essential, especially due to the proliferation of several smart speakers and personal agents that interact with an individual's voice c
Autor:
Goutam Saha, Monisankha Pal
Publikováno v:
IEEE/ACM Transactions on Audio, Speech, and Language Processing. 25:2071-2084
In this paper, we propose a new voice conversion (VC) method using i-vectors which consider low-dimensional representation of speech utterances. An attempt is made to restrict the i-vector variability in the intermediate computation of total variabil
Publikováno v:
IEEE Journal of Selected Topics in Signal Processing. 11:605-617
Recent advancements in voice conversion (VC) and speech synthesis research make speech-based biometric systems highly prone to spoofing attacks. This can provoke an increase in false acceptance rate in such systems and requires countermeasure to miti
Autor:
Somer L. Bishop, Shrikanth S. Narayanan, Manoj Kumar, Tae Jin Park, Raghuveer Peri, So Hyun Kim, Catherine Lord, Monisankha Pal
Publikováno v:
ICASSP
In this work, we propose deep latent space clustering for speaker diarization using generative adversarial network (GAN) backprojection with the help of an encoder network. The proposed diarization system is trained jointly with GAN loss, latent vari
Externí odkaz:
https://explore.openaire.eu/search/publication?articleId=doi_dedup___::920eb4dc19bedaf0f215a8bc51c0a6d4
http://arxiv.org/abs/1910.11398
http://arxiv.org/abs/1910.11398
Autor:
Raghuveer Peri, Monisankha Pal, Tae Jin Park, Panayiotis G. Georgiou, Naveen Kumar, Shrikanth S. Narayanan, Ruchir Travadi, Arindam Jati
Publikováno v:
INTERSPEECH
Autor:
Rimita Lahiri, Shrikanth S. Narayanan, Panayiotis G. Georgiou, Manoj Kumar, Monisankha Pal, Raghuveer Peri, Nikolaos Flemotomos, Tae Jin Park
Publikováno v:
INTERSPEECH
Publikováno v:
ICASSP
In this paper, we address the problem of speaker recognition in challenging acoustic conditions using a novel method to extract robust speaker-discriminative speech representations. We adopt a recently proposed unsupervised adversarial invariance arc
Externí odkaz:
https://explore.openaire.eu/search/publication?articleId=doi_dedup___::cb9576f216fc5c01054a563d945b0bb2