SmolVLA: A Vision-Language-Action Model for Affordable and Efficient Robotics
Paper
β’
2506.01844
β’
Published
β’
66
We release large pre-training datasets to accelerate open LLM development. Part of the Hugging Face Science team (hf.co/science)