Zobrazeno 1 - 1
of 1
pro vyhledávání: '"Emmersberger, Michael"'
Autor:
Daxberger, Erik, Weers, Floris, Zhang, Bowen, Gunter, Tom, Pang, Ruoming, Eichner, Marcin, Emmersberger, Michael, Yang, Yinfei, Toshev, Alexander, Du, Xianzhi
Sparse Mixture-of-Experts models (MoEs) have recently gained popularity due to their ability to decouple model size from inference efficiency by only activating a small subset of the model parameters for any given input token. As such, sparse MoEs ha
Externí odkaz:
http://arxiv.org/abs/2309.04354