SOTAVerified

ARC

Papers

Showing 125 of 554 papers

TitleStatusHype
Fast Text-to-Audio Generation with Adversarial Post-TrainingCode7
DataDecide: How to Predict Best Pretraining Data with Small ExperimentsCode3
ARC Prize 2024: Technical ReportCode3
The Surprising Effectiveness of Test-Time Training for Few-Shot LearningCode3
Digitizing Touch with an Artificial Multimodal FingertipCode3
CAX: Cellular Automata Accelerated in JAXCode3
Monte Carlo Tree Search Boosts Reasoning via Iterative Preference LearningCode3
Addressing the Abstraction and Reasoning Corpus via Procedural Example GenerationCode3
Neural networks for abstraction and reasoning: Towards broad generalization in machinesCode3
ST-MoE: Designing Stable and Transferable Sparse Expert ModelsCode3
Finetuned Language Models Are Zero-Shot LearnersCode3
A Parallelizable Lattice Rescoring Strategy with Neural Language ModelsCode3
The Jumping Reasoning Curve? Tracking the Evolution of Reasoning Performance in GPT-[n] and o-[n] Models on Multimodal PuzzlesCode2
Searching Latent Program SpacesCode2
Language Models are Hidden Reasoners: Unlocking Latent Reasoning Capabilities via Self-RewardingCode2
Combining Induction and Transduction for Abstract ReasoningCode2
ARCLE: The Abstraction and Reasoning Corpus Learning Environment for Reinforcement LearningCode2
Scaling Laws with Vocabulary: Larger Models Deserve Larger VocabulariesCode2
Flow of Reasoning:Training LLMs for Divergent Problem Solving with Minimal ExamplesCode2
Yuan 2.0-M32: Mixture of Experts with Attention RouterCode2
Con Instruction: Universal Jailbreaking of Multimodal Large Language Models via Non-Textual ModalitiesCode1
HELM: Hyperbolic Large Language Models via Mixture-of-Curvature ExpertsCode1
PHT-CAD: Efficient CAD Parametric Primitive Analysis with Progressive Hierarchical TuningCode1
DRACO: Differentiable Reconstruction for Arbitrary CBCT OrbitsCode1
IterGen: Iterative Semantic-aware Structured LLM Generation with BacktrackingCode1
Show:102550
← PrevPage 1 of 23Next →

No leaderboard results yet.