Zobrazeno 1 - 10
of 1 184
pro vyhledávání: '"WANG, Zhendong"'
Autor:
Wang, Zhendong, Li, Zhaoshuo, Mandlekar, Ajay, Xu, Zhenjia, Fan, Jiaojiao, Narang, Yashraj, Fan, Linxi, Zhu, Yuke, Balaji, Yogesh, Zhou, Mingyuan, Liu, Ming-Yu, Zeng, Yu
Diffusion models, praised for their success in generative tasks, are increasingly being applied to robotics, demonstrating exceptional performance in behavior cloning. However, their slow generation process stemming from iterative denoising steps pos
Externí odkaz:
http://arxiv.org/abs/2410.21257
Score identity Distillation (SiD) is a data-free method that has achieved state-of-the-art performance in image generation by leveraging only a pretrained diffusion model, without requiring any training data. However, the ultimate performance of SiD
Externí odkaz:
http://arxiv.org/abs/2410.14919
Aligning large language models with human preferences has emerged as a critical focus in language modeling research. Yet, integrating preference learning into Text-to-Image (T2I) generative models is still relatively uncharted territory. The Diffusio
Externí odkaz:
http://arxiv.org/abs/2406.06382
Diffusion-based text-to-image generation models trained on extensive text-image pairs have shown the capacity to generate photorealistic images consistent with textual descriptions. However, a significant limitation of these models is their slow samp
Externí odkaz:
http://arxiv.org/abs/2406.01561
Traditional language model alignment methods, such as Direct Preference Optimization (DPO), are limited by their dependence on static, pre-collected paired preference data, which hampers their adaptability and practical applicability. To overcome thi
Externí odkaz:
http://arxiv.org/abs/2405.20830
Offline reinforcement learning (RL) leverages pre-collected datasets to train optimal policies. Diffusion Q-Learning (DQL), introducing diffusion models as a powerful and expressive policy class, significantly boosts the performance of offline RL. Ho
Externí odkaz:
http://arxiv.org/abs/2405.19690
We introduce Score identity Distillation (SiD), an innovative data-free method that distills the generative capabilities of pretrained diffusion models into a single-step generator. SiD not only facilitates an exponentially fast reduction in Fr\'eche
Externí odkaz:
http://arxiv.org/abs/2404.04057
Autor:
Chen, Xuxi, Wang, Zhendong, Sow, Daouda, Yang, Junjie, Chen, Tianlong, Liang, Yingbin, Zhou, Mingyuan, Wang, Zhangyang
In the rapidly advancing arena of large language models (LLMs), a key challenge is to enhance their capabilities amid a looming shortage of high-quality training data. Our study starts from an empirical strategy for the light continual training of LL
Externí odkaz:
http://arxiv.org/abs/2402.14270
In the field of large language models (LLMs), aligning models with the diverse preferences of users is a critical challenge. Direct Preference Optimization (DPO) has played a key role in this area. It works by using pairs of preferences derived from
Externí odkaz:
http://arxiv.org/abs/2402.10958
Autor:
Bai, Chen, Shao, Zeman, Zhang, Guoxiang, Liang, Di, Yang, Jie, Zhang, Zhuorui, Guo, Yujian, Zhong, Chengzhang, Qiu, Yiqiao, Wang, Zhendong, Guan, Yichen, Zheng, Xiaoyin, Wang, Tao, Lu, Cheng
Realistic video simulation has shown significant potential across diverse applications, from virtual reality to film production. This is particularly true for scenarios where capturing videos in real-world settings is either impractical or expensive.
Externí odkaz:
http://arxiv.org/abs/2401.17509