Zobrazeno 1 - 10
of 1 381
pro vyhledávání: '"Vassilieva, A."'
Autor:
Tao, Tianhua, Li, Junbo, Tan, Bowen, Wang, Hongyi, Marshall, William, Kanakiya, Bhargav M, Hestness, Joel, Vassilieva, Natalia, Shen, Zhiqiang, Xing, Eric P., Liu, Zhengzhong
Large Language Models (LLMs) specializing in code generation (which are also often referred to as code LLMs), e.g., StarCoder and Code Llama, play increasingly critical roles in various software development scenarios. It is also crucial for code LLMs
Externí odkaz:
http://arxiv.org/abs/2411.04156
Autor:
Gosal, Gurpreet, Xu, Yishi, Ramakrishnan, Gokul, Joshi, Rituraj, Sheinin, Avraham, Zhiming, Chen, Mishra, Biswajit, Vassilieva, Natalia, Hestness, Joel, Sengupta, Neha, Sahu, Sunil Kumar, Jia, Bokang, Pandit, Onkar, Katipomu, Satheesh, Kamboj, Samta, Ghosh, Samujjwal, Pal, Rahul, Mullah, Parvez, Doraiswamy, Soundar, Chami, Mohamed El Karim, Nakov, Preslav
We present an efficient method for adapting a monolingual Large Language Model (LLM) to another language, addressing challenges of catastrophic forgetting and tokenizer limitations. We focus this study on adapting Llama 2 to Arabic. Our two-stage app
Externí odkaz:
http://arxiv.org/abs/2407.12869
Publikováno v:
S\'eminaire Lotharingien de Combinatoire 91B (2024) (Proceedings of the 36th FPSAC) Article #86, 12 pp
In our previous works we introduced a $q$-deformation of the generating functions for enriched $P$-partitions. We call the evaluation of this generating functions on labelled chains, the $q$-fundamental quasisymmetric functions. These functions inter
Externí odkaz:
http://arxiv.org/abs/2406.01166
Autor:
Christophe, Clément, Kanithi, Praveen K, Munjal, Prateek, Raha, Tathagata, Hayat, Nasir, Rajan, Ronnie, Al-Mahrooqi, Ahmed, Gupta, Avani, Salman, Muhammad Umar, Gosal, Gurpreet, Kanakiya, Bhargav, Chen, Charles, Vassilieva, Natalia, Amor, Boulbaba Ben, Pimentel, Marco AF, Khan, Shadab
This study presents a comprehensive analysis and comparison of two predominant fine-tuning methodologies - full-parameter fine-tuning and parameter-efficient tuning - within the context of medical Large Language Models (LLMs). We developed and refine
Externí odkaz:
http://arxiv.org/abs/2404.14779
Autor:
Dey, Nolan, Soboleva, Daria, Al-Khateeb, Faisal, Yang, Bowen, Pathria, Ribhu, Khachane, Hemant, Muhammad, Shaheer, Zhiming, Chen, Myers, Robert, Steeves, Jacob Robert, Vassilieva, Natalia, Tom, Marvin, Hestness, Joel
We introduce the Bittensor Language Model, called "BTLM-3B-8K", a new state-of-the-art 3 billion parameter open-source language model. BTLM-3B-8K was trained on 627B tokens from the SlimPajama dataset with a mixture of 2,048 and 8,192 context lengths
Externí odkaz:
http://arxiv.org/abs/2309.11568
Autor:
Shen, Zhiqiang, Tao, Tianhua, Ma, Liqun, Neiswanger, Willie, Liu, Zhengzhong, Wang, Hongyi, Tan, Bowen, Hestness, Joel, Vassilieva, Natalia, Soboleva, Daria, Xing, Eric
This paper aims to understand the impacts of various data combinations (e.g., web text, Wikipedia, GitHub, books) on the pretraining of large language models using SlimPajama. SlimPajama is a rigorously deduplicated, multi-source dataset, which has b
Externí odkaz:
http://arxiv.org/abs/2309.10818
We construct a new family $\left( \eta_{\alpha}^{\left( q\right) }\right) _{\alpha\in\operatorname*{Comp}}$ of quasisymmetric functions for each element $q$ of the base ring. We call them the "enriched $q$-monomial quasisymmetric functions". When $r:
Externí odkaz:
http://arxiv.org/abs/2309.01118
Autor:
Sengupta, Neha, Sahu, Sunil Kumar, Jia, Bokang, Katipomu, Satheesh, Li, Haonan, Koto, Fajri, Marshall, William, Gosal, Gurpreet, Liu, Cynthia, Chen, Zhiming, Afzal, Osama Mohammed, Kamboj, Samta, Pandit, Onkar, Pal, Rahul, Pradhan, Lalit, Mujahid, Zain Muhammad, Baali, Massa, Han, Xudong, Bsharat, Sondos Mahmoud, Aji, Alham Fikri, Shen, Zhiqiang, Liu, Zhengzhong, Vassilieva, Natalia, Hestness, Joel, Hock, Andy, Feldman, Andrew, Lee, Jonathan, Jackson, Andrew, Ren, Hector Xuguang, Nakov, Preslav, Baldwin, Timothy, Xing, Eric
We introduce Jais and Jais-chat, new state-of-the-art Arabic-centric foundation and instruction-tuned open generative large language models (LLMs). The models are based on the GPT-3 decoder-only architecture and are pretrained on a mixture of Arabic
Externí odkaz:
http://arxiv.org/abs/2308.16149
Publikováno v:
S\'eminaire Lotharingien de Combinatoire 89B (2023), Proceedings of the 35th Conference on Formal Power Series 2023, Article #46
Building up on our previous works regarding $q$-deformed $P$-partitions, we introduce a new family of subalgebras for the ring of quasisymmetric functions. Each of these subalgebras admits as a basis a $q$-analogue to Gessel's fundamental quasisymmet
Externí odkaz:
http://arxiv.org/abs/2301.00309
Autor:
Alexandra Vassilieva, Markus Harboe Olsen, Jane Skjøth‐Rasmussen, Kirsten Møller, Martin Kryspin Sørensen
Publikováno v:
Physiological Reports, Vol 12, Iss 20, Pp n/a-n/a (2024)
Abstract Hyperlactatemia is common during tumor craniotomy, but the underlying pathophysiology is unclear. This study measured simultaneous arterial and jugular‐bulb lactate concentrations in patients undergoing brain tumor craniotomy to investigat
Externí odkaz:
https://doaj.org/article/736d1c1ccb89471689ca451386cafaed