Popis: |
WATCH NOW 🔴📺👉UFC 282 Live It's almost time to watch UFC 282 live stream online this weekend, but we hope you haven't been expecting the card from this past summer. At least when it comes to the main event. Wolf, Thomas; Debut, Lysandre; Sanh, Victor; Chaumond, Julien; Delangue, Clement; Moi, Anthony; Cistac, Perric; Ma, Clara; Jernite, Yacine; Plu, Julien; Xu, Canwen; Le Scao, Teven; Gugger, Sylvain; Drame, Mariama; Lhoest, Quentin; Rush, Alexander M. PyTorch 2.0 stack support We are very excited by the newly announced PyTorch 2.0 stack. You can enable torch.compile on any of our models, and get support with the Trainer (and in all our PyTorch examples) by using the torchdynamo training argument. For instance, just add --torchdynamo inductor when launching those examples from the command line. This API is still experimental and may be subject to changes as the PyTorch 2.0 stack matures. Note that to get the best performance, we recommend: using an Ampere GPU (or more recent) sticking to fixed shaped for now (so use --pad_to_max_length in our examples) Repurpose torchdynamo training args towards torch._dynamo by @sgugger in #20498 Audio Spectrogram Transformer The Audio Spectrogram Transformer model was proposed in AST: Audio Spectrogram Transformer by Yuan Gong, Yu-An Chung, James Glass. The Audio Spectrogram Transformer applies a Vision Transformer to audio, by turning audio into an image (spectrogram). The model obtains state-of-the-art results for audio classification. Add Audio Spectogram Transformer by @NielsRogge in #19981 Jukebox The Jukebox model was proposed in Jukebox: A generative model for music by Prafulla Dhariwal, Heewoo Jun, Christine Payne, Jong Wook Kim, Alec Radford, Ilya Sutskever. It introduces a generative music model which can produce minute long samples that can be conditionned on an artist, genres and lyrics. Add Jukebox model (replaces #16875) by @ArthurZucker in #17826 Switch Transformers The SwitchTransformers model was proposed in Switch Transformers: Scaling to Trillion Parameter Models with Simple and Efficient Sparsity by William Fedus, Barret Zoph, Noam Shazeer. It is the first MoE model supported in transformers, with the largest checkpoint currently available currently containing 1T parameters. Add Switch transformers by @younesbelkada and @ArthurZucker in #19323 RocBert The RoCBert model was proposed in RoCBert: Robust Chinese Bert with Multimodal Contrastive Pretraining by HuiSu, WeiweiShi, XiaoyuShen, XiaoZhou, TuoJi, JiaruiFang, JieZhou. It's a pretrained Chinese language model that is robust under various forms of adversarial attacks. Add RocBert by @sww9370 in #20013 CLIPSeg The CLIPSeg model was proposed in Image Segmentation Using Text and Image Prompts |