| AltCLIP: Altering the Language Encoder in CLIP for Extended Language Capabilities | Nov 12, 2022 | Contrastive LearningCross-Modal Retrieval | CodeCode Available | 4 | 5 |
| MAD-X: An Adapter-Based Framework for Multi-Task Cross-Lingual Transfer | Apr 30, 2020 | Cross-Lingual Transfernamed-entity-recognition | CodeCode Available | 2 | 5 |
| MASSIVE: A 1M-Example Multilingual Natural Language Understanding Dataset with 51 Typologically-Diverse Languages | Apr 18, 2022 | intent-classificationIntent Classification | CodeCode Available | 2 | 5 |
| Zero-Shot Tokenizer Transfer | May 13, 2024 | XLM-R | CodeCode Available | 2 | 5 |
| X^2-VLM: All-In-One Pre-trained Model For Vision-Language Tasks | Nov 22, 2022 | AllCross-Modal Retrieval | CodeCode Available | 2 | 5 |
| DeBERTaV3: Improving DeBERTa using ELECTRA-Style Pre-Training with Gradient-Disentangled Embedding Sharing | Nov 18, 2021 | Language ModelingLanguage Modelling | CodeCode Available | 2 | 5 |
| AdapterHub: A Framework for Adapting Transformers | Jul 15, 2020 | XLM-R | CodeCode Available | 2 | 5 |
| DUMB: A Benchmark for Smart Evaluation of Dutch Models | May 22, 2023 | XLM-R | CodeCode Available | 1 | 5 |
| ARBERT & MARBERT: Deep Bidirectional Transformers for Arabic | Dec 27, 2020 | DiversityXLM-R | CodeCode Available | 1 | 5 |
| Code-Mixing on Sesame Street: Dawn of the Adversarial Polyglots | Mar 17, 2021 | Cross-Lingual TransferXLM-R | CodeCode Available | 1 | 5 |