OpenVLA: An Open-Source Vision-Language-Action Model
Paper
• 2406.09246
• Published
• 43
CogACT: A Foundational Vision-Language-Action Model for Synergizing
Cognition and Action in Robotic Manipulation
Paper
• 2411.19650
• Published
Octo: An Open-Source Generalist Robot Policy
Paper
• 2405.12213
• Published
• 29
Diffusion-VLA: Scaling Robot Foundation Models via Unified Diffusion and
Autoregression
Paper
• 2412.03293
• Published
robotics-diffusion-transformer/rdt-1b
Robotics
• Updated
• 314
• 100
OmniManip: Towards General Robotic Manipulation via Object-Centric
Interaction Primitives as Spatial Constraints
Paper
• 2501.03841
• Published
• 56
Beyond Sight: Finetuning Generalist Robot Policies with Heterogeneous
Sensors via Language Grounding
Paper
• 2501.04693
• Published
• 3
Spatial Forcing: Implicit Spatial Representation Alignment for
Vision-language-action Model
Paper
• 2510.12276
• Published
• 147
Robot Learning: A Tutorial
Paper
• 2510.12403
• Published
• 124
RoboOmni: Proactive Robot Manipulation in Omni-modal Context
Paper
• 2510.23763
• Published
• 56
RynnVLA-002: A Unified Vision-Language-Action and World Model
Paper
• 2511.17502
• Published
• 28
MobileVLA-R1: Reinforcing Vision-Language-Action for Mobile Robots
Paper
• 2511.17889
• Published
• 5
MIND-V: Hierarchical Video Generation for Long-Horizon Robotic Manipulation with RL-based Physical Alignment
Paper
• 2512.06628
• Published
• 13
GigaBrain-0.5M*: a VLA That Learns From World Model-Based Reinforcement Learning
Paper
• 2602.12099
• Published
• 57