Zobrazeno 1 - 3
of 3
pro vyhledávání: '"Aguilera, Aurora Cobo"'
Language models (LM) have grown with non-stop in the last decade, from sequence-to-sequence architectures to the state-of-the-art and utter attention-based Transformers. In this work, we demonstrate how the inclusion of deep generative models within
Externí odkaz:
http://arxiv.org/abs/2108.10764
Deep learning requires regularization mechanisms to reduce overfitting and improve generalization. We address this problem by a new regularization method based on distributional robust optimization. The key idea is to modify the contribution from eac
Externí odkaz:
http://arxiv.org/abs/2006.02734
Publikováno v:
In Neural Networks April 2023 161:565-574