QUAR-VLA: Vision-Language-Action Model for Quadruped Robots Paper • 2312.14457 • Published Dec 22, 2023 • 1
PiTe: Pixel-Temporal Alignment for Large Video-Language Model Paper • 2409.07239 • Published Sep 11, 2024 • 15
OpenHelix: A Short Survey, Empirical Analysis, and Open-Source Dual-System VLA Model for Robotic Manipulation Paper • 2505.03912 • Published May 6 • 9
SSR: Enhancing Depth Perception in Vision-Language Models via Rationale-Guided Spatial Reasoning Paper • 2505.12448 • Published May 18 • 10
Accelerating Vision-Language-Action Model Integrated with Action Chunking via Parallel Decoding Paper • 2503.02310 • Published Mar 4 • 1
CEED-VLA: Consistency Vision-Language-Action Model with Early-Exit Decoding Paper • 2506.13725 • Published Jun 16 • 1
QUART-Online: Latency-Free Large Multimodal Language Model for Quadruped Robot Learning Paper • 2412.15576 • Published Dec 20, 2024
ReconVLA: Reconstructive Vision-Language-Action Model as Effective Robot Perceiver Paper • 2508.10333 • Published Aug 14 • 1
VLA-Adapter: An Effective Paradigm for Tiny-Scale Vision-Language-Action Model Paper • 2509.09372 • Published Sep 11 • 236
Towards a Unified Understanding of Robot Manipulation: A Comprehensive Survey Paper • 2510.10903 • Published 27 days ago
Spatial Forcing: Implicit Spatial Representation Alignment for Vision-language-action Model Paper • 2510.12276 • Published 25 days ago • 142
VLA^2: Empowering Vision-Language-Action Models with an Agentic Framework for Unseen Concept Manipulation Paper • 2510.14902 • Published 23 days ago • 13
VLA^2: Empowering Vision-Language-Action Models with an Agentic Framework for Unseen Concept Manipulation Paper • 2510.14902 • Published 23 days ago • 13 • 2
VLA^2: Empowering Vision-Language-Action Models with an Agentic Framework for Unseen Concept Manipulation Paper • 2510.14902 • Published 23 days ago • 13
VLA-RFT: Vision-Language-Action Reinforcement Fine-tuning with Verified Rewards in World Simulators Paper • 2510.00406 • Published Oct 1 • 64
VLA-Adapter: An Effective Paradigm for Tiny-Scale Vision-Language-Action Model Paper • 2509.09372 • Published Sep 11 • 236
view post Post 4133 FalconMamba 7B - a new model from TII (Technology Innovation Institute) is out !- Blogpost: https://huggingface.co/blog/falconmamba- Link to collection: tiiuae/falconmamba-7b-66b9a580324dd1598b0f6d4a- Link to playground: tiiuae/falcon-mamba-playground 🔥 12 12 + Reply