Unsupervised pretraining transfers well across languages
Autor: | Pierre-Emmanuel Mazaré, Emmanuel Dupoux, Morgane Riviere, Armand Joulin |
---|---|
Přispěvatelé: | Facebook AI Research [Paris] (FAIR), Facebook, Apprentissage machine et développement cognitif (CoML), Inria de Paris, Institut National de Recherche en Informatique et en Automatique (Inria)-Institut National de Recherche en Informatique et en Automatique (Inria)-Laboratoire de sciences cognitives et psycholinguistique (LSCP), Département d'Etudes Cognitives - ENS Paris (DEC), École normale supérieure - Paris (ENS Paris), Université Paris sciences et lettres (PSL)-Université Paris sciences et lettres (PSL)-École normale supérieure - Paris (ENS Paris), Université Paris sciences et lettres (PSL)-Université Paris sciences et lettres (PSL)-École des hautes études en sciences sociales (EHESS)-Centre National de la Recherche Scientifique (CNRS)-Département d'Etudes Cognitives - ENS Paris (DEC), Université Paris sciences et lettres (PSL)-Université Paris sciences et lettres (PSL)-École des hautes études en sciences sociales (EHESS)-Centre National de la Recherche Scientifique (CNRS), Laboratoire de sciences cognitives et psycholinguistique (LSCP), CIFAR program in Learning in Machines & Brains CIFAR LMB program, ANR-19-P3IA-0001,PRAIRIE,PaRis Artificial Intelligence Research InstitutE(2019), ANR-17-EURE-0017,FrontCog,Frontières en cognition(2017), ANR-10-IDEX-0001,PSL,Paris Sciences et Lettres(2010), École normale supérieure - Paris (ENS-PSL), Université Paris sciences et lettres (PSL)-Université Paris sciences et lettres (PSL)-École normale supérieure - Paris (ENS-PSL), Université Paris sciences et lettres (PSL)-Université Paris sciences et lettres (PSL)-École des hautes études en sciences sociales (EHESS)-Centre National de la Recherche Scientifique (CNRS)-Inria de Paris, Institut National de Recherche en Informatique et en Automatique (Inria)-Institut National de Recherche en Informatique et en Automatique (Inria) |
Jazyk: | angličtina |
Rok vydání: | 2020 |
Předmět: |
FOS: Computer and information sciences
Sound (cs.SD) Computer Science - Machine Learning Computer science InformationSystems_INFORMATIONSTORAGEANDRETRIEVAL 02 engineering and technology 010501 environmental sciences computer.software_genre 01 natural sciences [INFO.INFO-CL]Computer Science [cs]/Computation and Language [cs.CL] Computer Science - Sound Machine Learning (cs.LG) Transcription (linguistics) [INFO.INFO-LG]Computer Science [cs]/Machine Learning [cs.LG] Audio and Speech Processing (eess.AS) FOS: Electrical engineering electronic engineering information engineering 0202 electrical engineering electronic engineering information engineering 0105 earth and related environmental sciences Computer Science - Computation and Language business.industry Low resources 020206 networking & telecommunications ComputingMethodologies_PATTERNRECOGNITION Unsupervised pretraining [INFO.INFO-SD]Computer Science [cs]/Sound [cs.SD] Artificial intelligence business Computation and Language (cs.CL) computer Natural language processing Cross-lingual Electrical Engineering and Systems Science - Audio and Speech Processing |
Zdroj: | ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and Signal Processing ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and Signal Processing, May 2020, Barcelona / Virtual, Spain. pp.7414-7418, ⟨10.1109/ICASSP40776.2020.9054548⟩ ICASSP |
Popis: | Cross-lingual and multi-lingual training of Automatic Speech Recognition (ASR) has been extensively investigated in the supervised setting. This assumes the existence of a parallel corpus of speech and orthographic transcriptions. Recently, contrastive predictive coding (CPC) algorithms have been proposed to pretrain ASR systems with unlabelled data. In this work, we investigate whether unsupervised pretraining transfers well across languages. We show that a slight modification of the CPC pretraining extracts features that transfer well to other languages, being on par or even outperforming supervised pretraining. This shows the potential of unsupervised methods for languages with few linguistic resources. 6 pages. Accepted at ICASSP 2020. However the 2 pages of supplementary materials will appear only in the arxiv version |
Databáze: | OpenAIRE |
Externí odkaz: |