SOTAVerified

TriviaQA

Papers

Showing 51100 of 124 papers

TitleStatusHype
Deep Understanding based Multi-Document Machine Reading Comprehension0
Dense Sparse Retrieval: Using Sparse Language Models for Inference Efficient Dense Retrieval0
DYNAMAX: Dynamic computing for Transformers and Mamba based architectures0
Efficient Transformer Knowledge Distillation: A Performance Review0
Evaluating Neural Model Robustness for Machine Comprehension0
FiE: Building a Global Probability Space by Leveraging Early Fusion in Encoder for Open-Domain Question Answering0
Fine-Grained Self-Endorsement Improves Factuality and Reasoning0
FIT-RAG: Black-Box RAG with Factual Information and Token Reduction0
GNN-encoder: Learning a Dual-encoder Architecture via Graph Neural Networks for Dense Passage Retrieval0
Grow-and-Clip: Informative-yet-Concise Evidence Distillation for Answer Explanation0
HASH-RAG: Bridging Deep Hashing with Retriever for Efficient, Fine Retrieval and Augmented Generation0
Improving Generated and Retrieved Knowledge Combination Through Zero-shot Generation0
Improving Unsupervised Question Answering via Summarization-Informed Question Generation0
Integrated Triaging for Fast Reading Comprehension0
Keyword-based Query Comprehending via Multiple Optimized-Demand Augmentation0
Machine Reading Comprehension with Enhanced Linguistic Verifiers0
MEMEN: Multi-layer Embedding with Memory Networks for Machine Comprehension0
MemoReader: Large-Scale Reading Comprehension through Neural Memory Controller0
Mitigating LLM Hallucinations via Conformal Abstention0
Multi-Mention Learning for Reading Comprehension with Neural Cascades0
Noisy Pair Corrector for Dense Retrieval0
On the Influence of Masking Policies in Intermediate Pre-training0
Pretrained Encyclopedia: Weakly Supervised Knowledge-Pretrained Language Model0
Quick Dense Retrievers Consume KALE: Post Training Kullback Leibler Alignment of Embeddings for Asymmetrical dual encoders0
Ranking and Sampling in Open-Domain Question Answering0
RECONSIDER: Improved Re-Ranking using Span-Focused Cross-Attention for Open Domain Question Answering0
Relation-Guided Pre-Training for Open-Domain Question Answering0
Researchy Questions: A Dataset of Multi-Perspective, Decompositional Questions for LLM Web Agents0
Rethinking Data Synthesis: A Teacher Model Training Recipe with Interpretation0
Self-Training Large Language Models for Tool-Use Without Demonstrations0
SFR-RAG: Towards Contextually Faithful LLMs0
ShED-HD: A Shannon Entropy Distribution Framework for Lightweight Hallucination Detection on Edge Devices0
Simple and Effective Semi-Supervised Question Answering0
SKILL: Structured Knowledge Infusion for Large Language Models0
Smarnet: Teaching Machines to Read and Comprehend Like Human0
Sorted LLaMA: Unlocking the Potential of Intermediate Layers of Large Language Models for Dynamic Inference0
Speculative RAG: Enhancing Retrieval Augmented Generation through Drafting0
Studying Strategically: Learning to Mask for Closed-book QA0
The Generative AI Paradox on Evaluation: What It Can Solve, It May Not Evaluate0
Tradeoffs in Sentence Selection Techniques for Open-Domain Question Answering0
UnitedQA: A Hybrid Approach for Open Domain Question Answering0
Vision-centric Token Compression in Large Language Model0
When to Read Documents or QA History: On Unified and Selective Open-domain QA0
Semantic Caching of Contextual Summaries for Efficient Question-Answering with Language Models0
Layer Importance and Hallucination Analysis in Large Language Models via Enhanced Activation Variance-Sparsity0
LACIE: Listener-Aware Finetuning for Confidence Calibration in Large Language ModelsCode0
KV Prediction for Improved Time to First TokenCode0
RE-RAG: Improving Open-Domain QA Performance and Interpretability with Relevance Estimator in Retrieval-Augmented GenerationCode0
Just Ask for Calibration: Strategies for Eliciting Calibrated Confidence Scores from Language Models Fine-Tuned with Human FeedbackCode0
Judging the Judges: Evaluating Alignment and Vulnerabilities in LLMs-as-JudgesCode0
Show:102550
← PrevPage 2 of 3Next →

No leaderboard results yet.