SOTAVerified

Vision-Language-Action

Papers

Showing 76100 of 157 papers

TitleStatusHype
ReBot: Scaling Robot Learning with Real-to-Sim-to-Real Robotic Video Synthesis0
Refined Policy Distillation: From VLA Generalists to RL Experts0
ReVLA: Reverting Visual Domain Limitation of Robotic Foundation Models0
RLRC: Reinforcement Learning-based Recovery for Compressed Vision-Language-Action Models0
RoboCerebra: A Large-scale Benchmark for Long-horizon Robotic Manipulation Evaluation0
RoboMamba: Efficient Vision-Language-Action Model for Robotic Reasoning and Manipulation0
RoboMIND: Benchmark on Multi-embodiment Intelligence Normative Data for Robot Manipulation0
RoboMonkey: Scaling Test-Time Sampling and Verification for Vision-Language-Action Models0
Robotic Control via Embodied Chain-of-Thought Reasoning0
Robotic Policy Learning via Human-assisted Action Preference Optimization0
ROSA: Harnessing Robot States for Vision-Language and Action Alignment0
RT-cache: Efficient Robot Trajectory Retrieval System0
Run-time Observation Interventions Make Vision-Language-Action Models More Visually Robust0
SAFE: Multitask Failure Detection for Vision-Language-Action Models0
SafeVLA: Towards Safety Alignment of Vision-Language-Action Model via Constrained Learning0
SARA-RT: Scaling up Robotics Transformers with Self-Adaptive Robust Attention0
SOLAMI: Social Vision-Language-Action Modeling for Immersive Interaction with 3D Autonomous Characters0
Survey on Vision-Language-Action Models0
Towards a Generalizable Bimanual Foundation Policy via Flow-based Video Prediction0
Towards Natural Language-Driven Assembly Using Foundation Models0
A Taxonomy for Evaluating Generalist Robot Policies0
TraceVLA: Visual Trace Prompting Enhances Spatial-Temporal Awareness for Generalist Robotic Policies0
TrackVLA: Embodied Visual Tracking in the Wild0
Unified Vision-Language-Action Model0
Uni-NaVid: A Video-based Vision-Language-Action Model for Unifying Embodied Navigation Tasks0
Show:102550
← PrevPage 4 of 7Next →

No leaderboard results yet.