StarVLA: A Lego-like Codebase for Vision-Language-Action Model Developing
-
Updated
Apr 9, 2026 - Python
StarVLA: A Lego-like Codebase for Vision-Language-Action Model Developing
[RSS 2025] Learning to Act Anywhere with Task-centric Latent Actions
Dexbotic: Open-Source Vision-Language-Action Toolbox
🚀🚀🚀A collection of some awesome public projects about Large Language Model(LLM), Vision Language Model(VLM), Vision Language Action(VLA), AI Generated Content(AIGC), the related Datasets and Applications.
InternRobotics' open platform for building generalized navigation foundation models.
[NeurIPS 2025 spotlight] Official implementation for "FutureSightDrive: Thinking Visually with Spatio-Temporal CoT for Autonomous Driving"
🔥 SpatialVLA: a spatial-enhanced vision-language-action model that is trained on 1.1 Million real robot episodes. Accepted at RSS 2025.
Unified Codebase for Advanced World Models.
DeepThinkVLA: Enhancing Reasoning Capability of Vision-Language-Action Models
[CVPR 2025, Spotlight] SimLingo (CarLLava): Vision-Only Closed-Loop Autonomous Driving with Language-Action Alignment
🌐 Vision-Language-Action Models for Autonomous Driving: Past, Present, and Future
EO: Open-source Unified Embodied Foundation Model Series
[CVPR 2026] WAM-Flow: Parallel Coarse-to-Fine Motion Planning via Discrete Flow Matching for Autonomous Driving
CLI for Physical AI Skill Models
Add a description, image, and links to the vla topic page so that developers can more easily learn about it.
To associate your repository with the vla topic, visit your repo's landing page and select "manage topics."