SOTAVerified

Cross-Lingual Transfer

Cross-lingual transfer refers to transfer learning using data and models available for one language for which ample such resources are available (e.g., English) to solve tasks in another, commonly more low-resource, language.

Papers

Showing 150 of 782 papers

TitleStatusHype
T-FREE: Subword Tokenizer-Free Generative LLMs via Sparse Representations for Memory-Efficient EmbeddingsCode2
MIND Your Language: A Multilingual Dataset for Cross-lingual News RecommendationCode2
The Belebele Benchmark: a Parallel Reading Comprehension Dataset in 122 Language VariantsCode2
Crosslingual Generalization through Multitask FinetuningCode2
Model and Data Transfer for Cross-Lingual Sequence Labelling in Zero-Resource SettingsCode2
Overcoming Catastrophic Forgetting in Zero-Shot Cross-Lingual GenerationCode2
mGPT: Few-Shot Learners Go MultilingualCode2
MAD-X: An Adapter-Based Framework for Multi-Task Cross-Lingual TransferCode2
Middle-Layer Representation Alignment for Cross-Lingual Transfer in Fine-Tuned LLMsCode1
Bridging the Gap: Enhancing LLM Performance for Low-Resource African Languages with New Benchmarks, Fine-Tuning, and Cultural AdjustmentsCode1
Multilingual Large Language Models: A Systematic SurveyCode1
From One to Many: Expanding the Scope of Toxicity Mitigation in Language ModelsCode1
IRCoder: Intermediate Representations Make Language Models Robust Multilingual Code GeneratorsCode1
AdaMergeX: Cross-Lingual Transfer with Large Language Models via Adaptive Adapter MergingCode1
ColBERT-XM: A Modular Multi-Vector Representation Model for Zero-Shot Multilingual Information RetrievalCode1
Investigating Cultural Alignment of Large Language ModelsCode1
LEIA: Facilitating Cross-lingual Knowledge Transfer in Language Models with Entity-based Data AugmentationCode1
UltraLink: An Open-Source Knowledge-Enhanced Multilingual Supervised Fine-tuning DatasetCode1
Turning English-centric LLMs Into Polyglots: How Much Multilinguality Is Needed?Code1
TaCo: Enhancing Cross-Lingual Transfer for Low-Resource Languages in LLMs through Translation-Assisted Chain-of-Thought ProcessesCode1
Lost in Translation, Found in Spans: Identifying Claims in Multilingual Social MediaCode1
CLARA: Multilingual Contrastive Learning for Audio Representation AcquisitionCode1
SIB-200: A Simple, Inclusive, and Big Evaluation Dataset for Topic Classification in 200+ Languages and DialectsCode1
mCLIP: Multilingual CLIP via Cross-lingual TransferCode1
Allophant: Cross-lingual Phoneme Recognition with Articulatory AttributesCode1
Zambezi Voice: A Multilingual Speech Corpus for Zambian LanguagesCode1
Revisiting non-English Text Simplification: A Unified Multilingual BenchmarkCode1
mPLM-Sim: Better Cross-Lingual Similarity and Transfer in Multilingual Pretrained Language ModelsCode1
MasakhaPOS: Part-of-Speech Tagging for Typologically Diverse African LanguagesCode1
ReadMe++: Benchmarking Multilingual Language Models for Multi-Domain Readability AssessmentCode1
Machine-Created Universal Language for Cross-lingual TransferCode1
ScandEval: A Benchmark for Scandinavian Natural Language ProcessingCode1
Investigating the Translation Performance of a Large Multilingual Language Model: the Case of BLOOMCode1
Frustratingly Easy Label Projection for Cross-lingual TransferCode1
ALIGN-MLM: Word Embedding Alignment is Crucial for Multilingual Pre-trainingCode1
GreenPLM: Cross-Lingual Transfer of Monolingual Pre-Trained Language Models at Almost No CostCode1
Prompt-Tuning Can Be Much Better Than Fine-Tuning on Cross-lingual Understanding With Multilingual Language ModelsCode1
On the Calibration of Massively Multilingual Language ModelsCode1
An Empirical Study on Cross-X Transfer for Legal Judgment PredictionCode1
CONCRETE: Improving Cross-lingual Fact-checking with Cross-lingual RetrievalCode1
Cross-View Language Modeling: Towards Unified Cross-Lingual Cross-Modal Pre-trainingCode1
The Importance of Being Parameters: An Intra-Distillation Method for Serious GainsCode1
The Geometry of Multilingual Language Model RepresentationsCode1
Pre-training Data Quality and Quantity for a Low-Resource Language: New Corpus and BERT Models for MalteseCode1
Enhancing Cross-lingual Transfer by Manifold MixupCode1
Towards Making the Most of Cross-Lingual Transfer for Zero-Shot Neural Machine TranslationCode1
Make the Best of Cross-lingual Transfer: Evidence from POS Tagging with over 100 LanguagesCode1
IndicXNLI: Evaluating Multilingual Inference for Indian LanguagesCode1
Adapting Pre-trained Language Models to African Languages via Multilingual Adaptive Fine-TuningCode1
Few-Shot Cross-lingual Transfer for Coarse-grained De-identification of Code-Mixed Clinical TextsCode1
Show:102550
← PrevPage 1 of 16Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1PaLM 2 (few-shot)Accuracy94.4Unverified
2mT0-13BAccuracy84.45Unverified
3RoBERTa Large (translate test)Accuracy76.05Unverified
4BLOOMZAccuracy75.5Unverified
5MAD-X BaseAccuracy60.94Unverified
6mGPTAccuracy55.5Unverified