| Plausible May Not Be Faithful: Probing Object Hallucination in Vision-Language Pre-training | Oct 14, 2022 | HallucinationImage Augmentation | CodeCode Available | 0 |
| Mixture of Attention Heads: Selecting Attention Heads Per Token | Oct 11, 2022 | Computational EfficiencyLanguage Modeling | CodeCode Available | 1 |
| MAP: Multimodal Uncertainty-Aware Vision-Language Pre-training Model | Oct 11, 2022 | Contrastive LearningImage-text matching | CodeCode Available | 1 |
| Revisiting and Advancing Chinese Natural Language Understanding with Accelerated Heterogeneous Knowledge Pre-training | Oct 11, 2022 | GPUKnowledge Graphs | —Unverified | 0 |
| The Effectiveness of Masked Language Modeling and Adapters for Factual Knowledge Injection | Oct 3, 2022 | Language ModelingLanguage Modelling | CodeCode Available | 0 |
| KUL@SMM4H’22: Template Augmented Adaptive Pre-training for Tweet Classification | Oct 1, 2022 | Data AugmentationLanguage Modeling | —Unverified | 0 |
| A Closer Look at Parameter Contributions When Training Neural Language and Translation Models | Oct 1, 2022 | Causal Language ModelingLanguage Modeling | —Unverified | 0 |
| Taking Actions Separately: A Bidirectionally-Adaptive Transfer Learning Method for Low-Resource Neural Machine Translation | Oct 1, 2022 | Generative Adversarial NetworkLanguage Modeling | —Unverified | 0 |
| Towards Making the Most of Pre-trained Translation Model for Quality Estimation | Oct 1, 2022 | DenoisingLanguage Modeling | —Unverified | 0 |
| Bidirectional Language Models Are Also Few-shot Learners | Sep 29, 2022 | DenoisingLanguage Modeling | —Unverified | 0 |