EmbodiedOneVision: Interleaved Vision-Text-Action Pretraining for General Robot Control Paper • 2508.21112 • Published Aug 28, 2025 • 77
F1: A Vision-Language-Action Model Bridging Understanding and Generation to Actions Paper • 2509.06951 • Published Sep 8, 2025 • 32
FastUMI-100K: Advancing Data-driven Robotic Manipulation with a Large-scale UMI-style Dataset Paper • 2510.08022 • Published Oct 9, 2025
Are We Ready for RL in Text-to-3D Generation? A Progressive Investigation Paper • 2512.10949 • Published Dec 11, 2025 • 45
Fast-UMI: A Scalable and Hardware-Independent Universal Manipulation Interface Paper • 2409.19499 • Published Sep 29, 2024
SpatialVLA: Exploring Spatial Representations for Visual-Language-Action Model Paper • 2501.15830 • Published Jan 27, 2025 • 13
FreeGaussian: Annotation-free Controllable 3D Gaussian Splats with Flow Derivatives Paper • 2410.22070 • Published Oct 29, 2024
Hume: Introducing System-2 Thinking in Visual-Language-Action Model Paper • 2505.21432 • Published May 27, 2025 • 4