SOTAVerified

The Open Verification Layer for ML Research

Community benchmark tracking and reproducibility verification. Built for researchers and autonomous research agents.

659,983 papers248,104 code links4,818 tasks

Papers

Showing 10261050 of 659983 papers

TitleStatusHype
OpenCodeInterpreter: Integrating Code Generation with Execution and RefinementCode5
MobileLLM: Optimizing Sub-billion Parameter Language Models for On-Device Use CasesCode5
How NeRFs and 3D Gaussian Splatting are Reshaping SLAM: a SurveyCode5
VADv2: End-to-End Vectorized Autonomous Driving via Probabilistic PlanningCode5
A Survey on Knowledge Distillation of Large Language ModelsCode5
Efficient Multimodal Learning from Data-centric PerspectiveCode5
Trust Regions for Explanations via Black-Box Probabilistic CertificationCode5
BlackJAX: Composable Bayesian inference in JAXCode5
DataDreamer: A Tool for Synthetic Data Generation and Reproducible LLM WorkflowsCode5
GaussianObject: High-Quality 3D Object Reconstruction from Four Views with Gaussian SplattingCode5
OS-Copilot: Towards Generalist Computer Agents with Self-ImprovementCode5
Online Iterative Reinforcement Learning from Human Feedback with General Preference ModelCode5
WebLINX: Real-World Website Navigation with Multi-Turn DialogueCode5
MobileVLM V2: Faster and Stronger Baseline for Vision Language ModelCode5
EasyInstruct: An Easy-to-use Instruction Processing Framework for Large Language ModelsCode5
Unified Training of Universal Time Series Forecasting TransformersCode5
Break the Sequential Dependency of LLM Inference Using Lookahead DecodingCode5
Audio Flamingo: A Novel Audio Language Model with Few-Shot Learning and Dialogue AbilitiesCode5
Executable Code Actions Elicit Better LLM AgentsCode5
BootsTAP: Bootstrapped Training for Tracking-Any-PointCode5
SymbolicAI: A framework for logic-based approaches combining generative models and solversCode5
MEIA: Multimodal Embodied Perception and Interaction in Unknown EnvironmentsCode5
RAPTOR: Recursive Abstractive Processing for Tree-Organized RetrievalCode5
Dolma: an Open Corpus of Three Trillion Tokens for Language Model Pretraining ResearchCode5
OpenMoE: An Early Effort on Open Mixture-of-Experts Language ModelsCode5
Show:102550
← PrevPage 42 of 26400Next →