Zobrazeno 1 - 10
of 13 599
pro vyhledávání: '"Chen, Qing‐An"'
Autor:
Bai, Jinbin, Ye, Tian, Chow, Wei, Song, Enxin, Chen, Qing-Guo, Li, Xiangtai, Dong, Zhen, Zhu, Lei, Yan, Shuicheng
Diffusion models, such as Stable Diffusion, have made significant strides in visual generation, yet their paradigm remains fundamentally different from autoregressive language models, complicating the development of unified language-vision models. Re
Externí odkaz:
http://arxiv.org/abs/2410.08261
Autor:
Zhu, Li-Fang, Koermann, Fritz, Chen, Qing, Selleby, Malin, Neugebauer, Joerg, Grabowski, and Blazej
Melting properties are critical for designing novel materials, especially for discovering high-performance, high-melting refractory materials. Experimental measurements of these properties are extremely challenging due to their high melting temperatu
Externí odkaz:
http://arxiv.org/abs/2408.08654
Autor:
Chen, Sijia, Wang, Yibo, Wu, Yi-Feng, Chen, Qing-Guo, Xu, Zhao, Luo, Weihua, Zhang, Kaifu, Zhang, Lijun
Tool-augmented large language models (LLMs) leverage tools, often in the form of APIs, to enhance their reasoning capabilities on complex tasks, thus taking on the role of intelligent agents interacting with the real world. The recently introduced To
Externí odkaz:
http://arxiv.org/abs/2406.07115
Autor:
Zhang, Yi-Kai, Lu, Shiyin, Li, Yang, Ma, Yanqing, Chen, Qing-Guo, Xu, Zhao, Luo, Weihua, Zhang, Kaifu, Zhan, De-Chuan, Ye, Han-Jia
Multimodal large language models (MLLMs), initiated with a trained LLM, first align images with text and then fine-tune on multimodal mixed inputs. However, the MLLM catastrophically forgets the text-only instructions, which do not include images and
Externí odkaz:
http://arxiv.org/abs/2406.03496
Autor:
Sun, Hai-Long, Zhou, Da-Wei, Li, Yang, Lu, Shiyin, Yi, Chao, Chen, Qing-Guo, Xu, Zhao, Luo, Weihua, Zhang, Kaifu, Zhan, De-Chuan, Ye, Han-Jia
The rapid development of Multimodal Large Language Models (MLLMs) like GPT-4V has marked a significant step towards artificial general intelligence. Existing methods mainly focus on aligning vision encoders with LLMs through supervised fine-tuning (S
Externí odkaz:
http://arxiv.org/abs/2406.02539
Current Multimodal Large Language Models (MLLMs) typically integrate a pre-trained LLM with another pre-trained vision transformer through a connector, such as an MLP, endowing the LLM with visual capabilities. However, the misalignment between two e
Externí odkaz:
http://arxiv.org/abs/2405.20797
The superradiant phase transition (SRPT) is forbidden in the standard isotropic Dicke model due to the so-called no-go theorem induced by A-square term. In the framework of the Dicke model, we demonstrate that SRPTs can occur at both zero and finite
Externí odkaz:
http://arxiv.org/abs/2405.19776
Autor:
Ma, Yilin, Huang, Jiajia, Gao, Shengyong, Li, iangyu, Yi, Zhibin, Xiao, Diwen, Chan, Cheuk Kai Kevin, Pan, Ding, Chen, Qing
Rechargeable Zn batteries offer safe, inexpensive energy storage, but when deeply discharged to compete with lithium-ion batteries, they are plagued by parasitic reactions at the Zn anodes. We apply super-concentrated alkaline electrolytes to suppres
Externí odkaz:
http://arxiv.org/abs/2405.07675
The recent introduction of prompt tuning based on pre-trained vision-language models has dramatically improved the performance of multi-label image classification. However, some existing strategies that have been explored still have drawbacks, i.e.,
Externí odkaz:
http://arxiv.org/abs/2405.06926
The non-Hermitian one-photon and two-photon quantum Rabi models (QRMs) with imaginary couplings are respectively solved through the Bogoliubov operators approach. Transcendental functions responsible for exact solutions are derived, whose zeros produ
Externí odkaz:
http://arxiv.org/abs/2402.09749