Zobrazeno 1 - 10
of 27 548
pro vyhledávání: '"Vocab"'
Autor:
Gräf, Maike
Publikováno v:
Praxis Sprache. 2024, Issue 4, p232-234. 3p.
Autor:
Michael Fisher
All teachers know that a robust vocabulary gives students the communication skills they need to do well on tests and shine in the classroom--and the best way to ensure successful vocabulary instruction is to embrace new and engaging strategies that d
Autor:
Färber, Michael, Popovic, Nicholas
In this paper, we propose Vocab-Expander at https://vocab-expander.com, an online tool that enables end-users (e.g., technology scouts) to create and expand a vocabulary of their domain of interest. It utilizes an ensemble of state-of-the-art word em
Externí odkaz:
http://arxiv.org/abs/2308.03519
Autor:
Sumit Singh, Uma Shanker Tiwary
Publikováno v:
IEEE Access, Vol 12, Pp 22707-22717 (2024)
Named entities are random, like emerging entities and complex entities. Most of the large language model’s tokenizers have fixed vocab; hence, they tokenize out-of-vocab (OOV) words into multiple sub-words during tokenization. During fine-tuning fo
Externí odkaz:
https://doaj.org/article/6efdec15324f4c7ba423951d94cfa011
Autor:
Kang, Wang-Cheng, Cheng, Derek Zhiyuan, Chen, Ting, Yi, Xinyang, Lin, Dong, Hong, Lichan, Chi, Ed H.
Recommender system models often represent various sparse features like users, items, and categorical features via embeddings. A standard approach is to map each unique feature value to an embedding vector. The size of the produced embedding table gro
Externí odkaz:
http://arxiv.org/abs/2002.08530
Akademický článek
Tento výsledek nelze pro nepřihlášené uživatele zobrazit.
K zobrazení výsledku je třeba se přihlásit.
K zobrazení výsledku je třeba se přihlásit.
Autor:
Zhu, Wei, Cheung, Daniel
In this work, we represent CMV-BERT, which improves the pretraining of a language model via two ingredients: (a) contrastive learning, which is well studied in the area of computer vision; (b) multiple vocabularies, one of which is fine-grained and t
Externí odkaz:
http://arxiv.org/abs/2012.14763
Autor:
Zhu, Wei
Despite the development of pre-trained language models (PLMs) significantly raise the performances of various Chinese natural language processing (NLP) tasks, the vocabulary for these Chinese PLMs remain to be the one provided by Google Chinese Bert
Externí odkaz:
http://arxiv.org/abs/2011.08539