[RSS 2025] Learning to Act Anywhere with Task-centric Latent Actions
-
Updated
Jun 20, 2025 - Python
[RSS 2025] Learning to Act Anywhere with Task-centric Latent Actions
🔥 SpatialVLA: a spatial-enhanced vision-language-action model that is trained on 1.1 Million real robot episodes. Accepted at RSS 2025.
[ICLR'25] LLaRA: Supercharging Robot Learning Data for Vision-Language Policy
[CVPR 2025, Spotlight] SimLingo (CarLLava): Vision-Only Closed-Loop Autonomous Driving with Language-Action Alignment
Official implementation for "FutureSightDrive: Thinking Visually with Spatio-Temporal CoT for Autonomous Driving"
Emma-X: An Embodied Multimodal Action Model with Grounded Chain of Thought and Look-ahead Spatial Reasoning
📚这个仓库是在arxiv上收集的有关VLN,VLA, SLAM,Gaussian Splatting,非线性优化等相关论文。每天都会自动更新!issue区域是最新10篇论文
🦾 A Dual-System VLA with System2 Thinking
Pipeline which takes raw VLA data and yields science-ready data products, i.e. star formation rates
Very Large Array Image Download and Time Lapse Builder
Add a description, image, and links to the vla topic page so that developers can more easily learn about it.
To associate your repository with the vla topic, visit your repo's landing page and select "manage topics."