SmolVLA: A Vision-Language-Action Model for Affordable and Efficient Robotics Paper • 2506.01844 • Published 4 days ago • 74
view article Article Tiny Agents in Python: a MCP-powered agent in ~70 lines of code By celinah and 3 others • 15 days ago • 122
view changelog Changelog Xet is now the default storage option for new users and organizations 15 days ago • 58
view article Article Cohere on Hugging Face Inference Providers 🔥 By burtenshaw and 6 others • Apr 16 • 126
view article Article Introducing Three New Serverless Inference Providers: Hyperbolic, Nebius AI Studio, and Novita 🔥 By reach-vb and 6 others • Feb 18 • 99
view article Article π0 and π0-FAST: Vision-Language-Action Models for General Robot Control By danaaubakirova and 3 others • Feb 4 • 158
Meta Llama 3 Collection This collection hosts the transformers and original repos of the Meta Llama 3 and Llama Guard 2 releases • 5 items • Updated Dec 6, 2024 • 776
Idefics2 🐶 Collection Idefics2-8B is a foundation vision-language model. In this collection, you will find the models, datasets and demo related to its creation. • 11 items • Updated May 6, 2024 • 91
MM1: Methods, Analysis & Insights from Multimodal LLM Pre-training Paper • 2403.09611 • Published Mar 14, 2024 • 128
Masked Audio Generation using a Single Non-Autoregressive Transformer Paper • 2401.04577 • Published Jan 9, 2024 • 44
MAGNeT Collection Masked Audio Generation using a Single Non-Autoregressive Transformer • 9 items • Updated Apr 4, 2024 • 40
QuIP: 2-Bit Quantization of Large Language Models With Guarantees Paper • 2307.13304 • Published Jul 25, 2023 • 2