Zobrazeno 1 - 10
of 23
pro vyhledávání: '"Myle Ott"'
Publikováno v:
Proceedings of the International AAAI Conference on Web and Social Media. 7:390-399
User-generated comments in online social media have recently been gaining increasing attention as a viable source of general-purpose descriptive annotations for digital objects like photos or videos. Because users have different levels of expertise,
Autor:
Jing Xu, Stephen Roller, Eric Michael Smith, Jason Weston, Yinhan Liu, Emily Dinan, Y-Lan Boureau, Mary Williamson, Naman Goyal, Myle Ott, Da Ju
Publikováno v:
EACL
Building open-domain chatbots is a challenging area for machine learning research. While prior work has shown that scaling neural models in the number of parameters and the size of the data they are trained on gives improved results, we highlight oth
Publikováno v:
Proceedings of the 6th Workshop on Representation Learning for NLP (RepL4NLP-2021).
Recent work has demonstrated the effectiveness of cross-lingual language model pretraining for cross-lingual understanding. In this study, we present the results of two larger multilingual masked language models, with 3.5B and 10.7B parameters. Our t
Publikováno v:
EACL
In this work, we study how the finetuning stage in the pretrain-finetune framework changes the behavior of a pretrained neural language generator. We focus on the transformer encoder-decoder model for the open-domain dialogue response generation task
Publikováno v:
EMNLP (Findings)
The state of the art on many NLP tasks is currently achieved by large pre-trained language models, which require a considerable amount of computation. We explore a setting where many different predictions are made on a single piece of text. In that c
Externí odkaz:
https://explore.openaire.eu/search/publication?articleId=doi_dedup___::b7e9855dc87dcb31c324e0d75b180c8a
Publikováno v:
ClinicalNLP@EMNLP
A large array of pretrained models are available to the biomedical NLP (BioNLP) community. Finding the best model for a particular task can be difficult and time-consuming. For many applications in the biomedical and clinical domains, it is crucial t
Autor:
Edouard Grave, Vishrav Chaudhary, Guillaume Wenzek, Luke Zettlemoyer, Kartikay Khandelwal, Veselin Stoyanov, Naman Goyal, Myle Ott, Alexis Conneau, Francisco Guzmán
Publikováno v:
ACL
This paper shows that pretraining multilingual language models at scale leads to significant performance gains for a wide range of cross-lingual transfer tasks. We train a Transformer-based masked language model on one hundred languages, using more t
Autor:
Fabio Petroni, Fabrizio Silvestri, Sebastian Riedel, Myle Ott, Vassilis Plachouras, Luca Massarelli, Tim Rocktäschel, Aleksandra Piktus
Publikováno v:
EMNLP (Findings)
Recent progress in pre-trained language models led to systems that are able to generate text of an increasingly high quality. While several works have investigated the fluency and grammatical correctness of such models, it is still unclear to which e
Autor:
Jiatao Gu, Jiajun Shen, Junxian He, Michael Auli, Peng-Jen Chen, Matthew Le, Marc'Aurelio Ranzato, Myle Ott
Publikováno v:
EACL
While we live in an increasingly interconnected world, different places still exhibit strikingly different cultures and many events we experience in our every day life pertain only to the specific place we live in. As a result, people often talk abou
Externí odkaz:
https://explore.openaire.eu/search/publication?articleId=doi_dedup___::1ce4165096d248e3f51629421c920fcc
http://arxiv.org/abs/1909.13151
http://arxiv.org/abs/1909.13151
Publikováno v:
ACL
Back-translation is a widely used data augmentation technique which leverages target monolingual data. However, its effectiveness has been challenged since automatic metrics such as BLEU only show significant improvements for test examples where the
Externí odkaz:
https://explore.openaire.eu/search/publication?articleId=doi_dedup___::5bb95df2b53bdcc07bf8c55b97d2a183
http://arxiv.org/abs/1908.05204
http://arxiv.org/abs/1908.05204