SOTAVerified

Question Answering

Question answering can be segmented into domain-specific tasks like community question answering and knowledge-base question answering. Popular benchmark datasets for evaluation question answering systems include SQuAD, HotPotQA, bAbI, TriviaQA, WikiQA, and many others. Models for question answering are typically evaluated on metrics like EM and F1. Some recent top performing models are T5 and XLNet.

( Image credit: SQuAD )

Papers

Showing 33213330 of 10817 papers

TitleStatusHype
A Novel Approach Towards Incorporating Context Processing Capabilities in NLIDB System0
Extracting Factual Min/Max Age Information from Clinical Trial Studies0
Building English-Vietnamese Named Entity Corpus with Aligned Bilingual News Articles0
Enhancing Multi-Image Question Answering via Submodular Subset Selection0
Extracting Opinion Expressions with semi-Markov Conditional Random Fields0
Enhancing Multimodal LLM for Detailed and Accurate Video Captioning using Multi-Round Preference Optimization0
A Novel Approach for Automatic Bengali Question Answering System using Semantic Similarity Analysis0
Building Interpretable and Reliable Open Information Retriever for New Domains Overnight0
Building Privacy-Preserving and Secure Geospatial Artificial Intelligence Foundation Models0
Bidirectional Long Short-Term Memory Networks for Relation Classification0
Show:102550
← PrevPage 333 of 1082Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1IE-Net (ensemble)EM90.94Unverified
2FPNet (ensemble)EM90.87Unverified
3IE-NetV2 (ensemble)EM90.86Unverified
4SA-Net on Albert (ensemble)EM90.72Unverified
5SA-Net-V2 (ensemble)EM90.68Unverified
6FPNet (ensemble)EM90.6Unverified
7Retro-Reader (ensemble)EM90.58Unverified
8EntitySpanFocusV2 (ensemble)EM90.52Unverified
9TransNets + SFVerifier + SFEnsembler (ensemble)EM90.49Unverified
10EntitySpanFocus+AT (ensemble)EM90.45Unverified