Zobrazeno 1 - 10
of 237
pro vyhledávání: '"Shen, Yilin"'
Autor:
Gao, Shangqian, Lin, Chi-Heng, Hua, Ting, Zheng, Tang, Shen, Yilin, Jin, Hongxia, Hsu, Yen-Chang
Large Language Models (LLMs) have achieved remarkable success in various natural language processing tasks, including language modeling, understanding, and generation. However, the increased memory and computational costs associated with these models
Externí odkaz:
http://arxiv.org/abs/2410.11988
Autor:
Lin, Chi-Heng, Gao, Shangqian, Smith, James Seale, Patel, Abhishek, Tuli, Shikhar, Shen, Yilin, Jin, Hongxia, Hsu, Yen-Chang
Large Language Models (LLMs) have reshaped the landscape of artificial intelligence by demonstrating exceptional performance across various tasks. However, substantial computational requirements make their deployment challenging on devices with limit
Externí odkaz:
http://arxiv.org/abs/2408.09632
Traditional language models operate autoregressively, i.e., they predict one token at a time. Rapid explosion in model sizes has resulted in high inference times. In this work, we propose DynaMo, a suite of multi-token prediction language models that
Externí odkaz:
http://arxiv.org/abs/2405.00888
Publikováno v:
Proceedings of 24th INTERSPEECH Conference (INTERSPEECH 2023), Dublin, Ireland
State-of-the-art spoken language understanding (SLU) models have shown tremendous success in benchmark SLU datasets, yet they still fail in many practical scenario due to the lack of model compositionality when trained on limited training data. In th
Externí odkaz:
http://arxiv.org/abs/2312.15815
Open World Compositional Zero-Shot Learning (OW-CZSL) is known to be an extremely challenging task, which aims to recognize unseen compositions formed from seen attributes and objects without any prior assumption of the output space. In order to achi
Externí odkaz:
http://arxiv.org/abs/2312.02191
Vision Transformers (ViTs) have emerged as powerful backbones in computer vision, outperforming many traditional CNNs. However, their computational overhead, largely attributed to the self-attention mechanism, makes deployment on resource-constrained
Externí odkaz:
http://arxiv.org/abs/2312.01026
Recent work has demonstrated a remarkable ability to customize text-to-image diffusion models to multiple, fine-grained concepts in a sequential (i.e., continual) manner while only providing a few example images for each concept. This setting is know
Externí odkaz:
http://arxiv.org/abs/2311.18763
Autor:
Shen, Yilin
The Variational Quantum Eigensolver (VQE) is a quantum algorithm used to find the ground state energy of a given Hamiltonian. The key component of VQE is the ansatz, which is a trial wavefunction that the algorithm uses to approximate the ground stat
Externí odkaz:
http://arxiv.org/abs/2310.02511
Autor:
Srinivasa, Rakshith Sharma, Cho, Jaejin, Yang, Chouchang, Saidutta, Yashas Malur, Lee, Ching-Hua, Shen, Yilin, Jin, Hongxia
This paper considers contrastive training for cross-modal 0-shot transfer wherein a pre-trained model in one modality is used for representation learning in another domain using pairwise data. The learnt models in the latter domain can then be used f
Externí odkaz:
http://arxiv.org/abs/2309.14580
Large Language Models (LLMs) are progressively being utilized as machine learning services and interface tools for various applications. However, the security implications of LLMs, particularly in relation to adversarial and Trojan attacks, remain in
Externí odkaz:
http://arxiv.org/abs/2306.06815