SOTAVerified

Question Answering

Question answering can be segmented into domain-specific tasks like community question answering and knowledge-base question answering. Popular benchmark datasets for evaluation question answering systems include SQuAD, HotPotQA, bAbI, TriviaQA, WikiQA, and many others. Models for question answering are typically evaluated on metrics like EM and F1. Some recent top performing models are T5 and XLNet.

( Image credit: SQuAD )

Papers

Showing 85518575 of 10817 papers

TitleStatusHype
Exploring the BERT Cross-Lingual Transferability: a Case Study in Reading Comprehension0
Exploring the Effectiveness of Convolutional Neural Networks for Answer Selection in End-to-End Question Answering0
Exploring the Effectiveness of GPT Models in Test-Taking: A Case Study of the Driver's License Knowledge Test0
Exploring the Effectiveness of Object-Centric Representations in Visual Question Answering: Comparative Insights with Foundation Models0
Exploring the Feasibility of ChatGPT for Event Extraction0
Exploring the Frontier of Vision-Language Models: A Survey of Current Methodologies and Future Directions0
Exploring the Ideal Depth of Neural Network when Predicting Question Deletion on Community Question Answering0
Exploring the Impact of Occupational Personas on Domain-Specific QA0
Exploring the Impact of Table-to-Text Methods on Augmenting LLM-based Question Answering with Domain Hybrid Data0
Exploring the In-context Learning Ability of Large Language Model for Biomedical Concept Linking0
Modeling Uncertainty and Using Post-fusion as Fallback Improves Retrieval Augmented Generation with LLMs0
Exploring The Landscape of Distributional Robustness for Question Answering Models0
Do Large Language Models have Shared Weaknesses in Medical Question Answering?0
Exploring the Role of Knowledge Graph-Based RAG in Japanese Medical Question Answering with Small-Scale LLMs0
Exploring The Spatial Reasoning Ability of Neural Models in Human IQ Tests0
Exploring the Viability of Synthetic Query Generation for Relevance Prediction0
How Additional Knowledge can Improve Natural Language Commonsense Question Answering?0
Exploring Weaknesses of VQA Models through Attribution Driven Insights0
EXPR at SemEval-2018 Task 9: A Combined Approach for Hypernym Discovery0
Extended Named Entity Recognition API and Its Applications in Language Education0
Extending Neural Question Answering with Linguistic Input Features0
Extending the Frontier of ChatGPT: Code Generation and Debugging0
EKTVQA: Generalized use of External Knowledge to empower Scene Text in Text-VQA0
Extracting Context-Rich Entailment Rules from Wikipedia Revision History0
Extracting Definitions and Hypernym Relations relying on Syntactic Dependencies and Support Vector Machines0
Show:102550
← PrevPage 343 of 433Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1IE-Net (ensemble)EM90.94Unverified
2FPNet (ensemble)EM90.87Unverified
3IE-NetV2 (ensemble)EM90.86Unverified
4SA-Net on Albert (ensemble)EM90.72Unverified
5SA-Net-V2 (ensemble)EM90.68Unverified
6FPNet (ensemble)EM90.6Unverified
7Retro-Reader (ensemble)EM90.58Unverified
8EntitySpanFocusV2 (ensemble)EM90.52Unverified
9TransNets + SFVerifier + SFEnsembler (ensemble)EM90.49Unverified
10EntitySpanFocus+AT (ensemble)EM90.45Unverified