Zobrazeno 1 - 10
of 14 776
pro vyhledávání: '"Tsvetkov, AS"'
Autor:
Park, Chan Young, Li, Shuyue Stella, Jung, Hayoung, Volkova, Svitlana, Mitra, Tanushree, Jurgens, David, Tsvetkov, Yulia
This study introduces ValueScope, a framework leveraging language models to quantify social norms and values within online communities, grounded in social science perspectives on normative structures. We employ ValueScope to dissect and analyze lingu
Externí odkaz:
http://arxiv.org/abs/2407.02472
Autor:
Ahia, Orevaoghene, Aremu, Anuoluwapo, Abagyan, Diana, Gonen, Hila, Adelani, David Ifeoluwa, Abolade, Daud, Smith, Noah A., Tsvetkov, Yulia
Yor\`ub\'a an African language with roughly 47 million speakers encompasses a continuum with several dialects. Recent efforts to develop NLP technologies for African languages have focused on their standard dialects, resulting in disparities for dial
Externí odkaz:
http://arxiv.org/abs/2406.19564
Autor:
Zhang, Yizhuo, Wang, Heng, Feng, Shangbin, Tan, Zhaoxuan, Han, Xiaochuang, He, Tianxing, Tsvetkov, Yulia
Large language models (LLMs) demonstrate great potential for problems with implicit graphical structures, while recent works seek to enhance the graph reasoning capabilities of LLMs through specialized instruction tuning. The resulting 'graph LLMs' a
Externí odkaz:
http://arxiv.org/abs/2406.15992
Autor:
Feng, Shangbin, Sorensen, Taylor, Liu, Yuhan, Fisher, Jillian, Park, Chan Young, Choi, Yejin, Tsvetkov, Yulia
While existing alignment paradigms have been integral in developing large language models (LLMs), LLMs often learn an averaged human preference and struggle to model diverse preferences across cultures, demographics, and communities. We propose Modul
Externí odkaz:
http://arxiv.org/abs/2406.15951
Autor:
Feng, Shangbin, Shi, Weijia, Wang, Yike, Ding, Wenxuan, Ahia, Orevaoghene, Li, Shuyue Stella, Balachandran, Vidhisha, Sitaram, Sunayana, Tsvetkov, Yulia
Multilingual LLMs often have knowledge disparities across languages, with larger gaps in under-resourced languages. Teaching LLMs to abstain in the face of knowledge gaps is thus a promising strategy to mitigate hallucinations in multilingual setting
Externí odkaz:
http://arxiv.org/abs/2406.15948
Autor:
Li, Shuyue Stella, Balachandran, Vidhisha, Feng, Shangbin, Ilgen, Jonathan, Pierson, Emma, Koh, Pang Wei, Tsvetkov, Yulia
In high-stakes domains like clinical reasoning, AI assistants powered by large language models (LLMs) are yet to be reliable and safe. We identify a key obstacle towards reliability: existing LLMs are trained to answer any question, even with incompl
Externí odkaz:
http://arxiv.org/abs/2406.00922
Autor:
Ahuja, Kabir, Balachandran, Vidhisha, Panwar, Madhur, He, Tianxing, Smith, Noah A., Goyal, Navin, Tsvetkov, Yulia
Transformers trained on natural language data have been shown to learn its hierarchical structure and generalize to sentences with unseen syntactic structures without explicitly encoding any structural bias. In this work, we investigate sources of in
Externí odkaz:
http://arxiv.org/abs/2404.16367
Autor:
Chiu, Yu Ying, Jiang, Liwei, Antoniak, Maria, Park, Chan Young, Li, Shuyue Stella, Bhatia, Mehar, Ravi, Sahithya, Tsvetkov, Yulia, Shwartz, Vered, Choi, Yejin
Frontier large language models (LLMs) are developed by researchers and practitioners with skewed cultural backgrounds and on datasets with skewed sources. However, LLMs' (lack of) multicultural knowledge cannot be effectively assessed with current me
Externí odkaz:
http://arxiv.org/abs/2404.06664
Autor:
Faisal, Fahim, Ahia, Orevaoghene, Srivastava, Aarohi, Ahuja, Kabir, Chiang, David, Tsvetkov, Yulia, Anastasopoulos, Antonios
Language technologies should be judged on their usefulness in real-world use cases. An often overlooked aspect in natural language processing (NLP) research and evaluation is language variation in the form of non-standard dialects or language varieti
Externí odkaz:
http://arxiv.org/abs/2403.11009
Autor:
Kassem, Aly M., Mahmoud, Omar, Mireshghallah, Niloofar, Kim, Hyunwoo, Tsvetkov, Yulia, Choi, Yejin, Saad, Sherif, Rana, Santu
In this paper, we introduce a black-box prompt optimization method that uses an attacker LLM agent to uncover higher levels of memorization in a victim agent, compared to what is revealed by prompting the target model with the training data directly,
Externí odkaz:
http://arxiv.org/abs/2403.04801