Open Source
Explore the latest AI open-source projects from GitHub and HuggingFace.
Explore the latest AI open-source projects from GitHub and HuggingFace.
## Introduction LeRobot is Hugging Face's open-source robotics framework that aims to make AI for robotics accessible through end-to-end learning. With over 22,000 GitHub stars and a thriving community of contributors, LeRobot has become one of the most influential projects in democratizing robot learning. Its recent paper acceptance at ICLR 2026 further validates its technical contributions to the field. The project emerged from a simple but powerful insight: the robotics community lacks the shared infrastructure that has accelerated progress in NLP and computer vision. LeRobot addresses this gap by providing standardized models, datasets, and tools that lower the barrier to entry for robot learning research and application development. Hugging Face's acquisition of Pollen Robotics further signals the company's commitment to making physical AI as accessible as language AI. ## Architecture and Design LeRobot is built on a modular architecture centered around three pillars: standardized datasets, pre-trained policies, and a hardware-agnostic robot interface. The LeRobotDataset format is the project's data backbone. It provides a standardized schema for storing robot trajectories including observations (images, proprioception), actions, rewards, and metadata. Datasets are hosted on the Hugging Face Hub, enabling the same discoverability and version control workflow that has proven effective for NLP and vision datasets. Version 0.5.0 introduced a faster dataset loading pipeline that significantly reduces training startup time. The policy library implements state-of-the-art robot learning algorithms in a consistent API. Supported policies include ACT (Action Chunking with Transformers), Diffusion Policy, TDMPC, VQ-BeT, and the recently added autoregressive vision-language-action (VLA) models. Each policy can be trained on any LeRobotDataset and deployed to any supported robot through the unified interface. | Specification | Detail | |---------------|--------| | Framework | PyTorch (Python 3.12+, Transformers v5) | | Policies | ACT, Diffusion, TDMPC, VQ-BeT, VLA | | Dataset Format | LeRobotDataset (Hugging Face Hub) | | Simulation | MuJoCo, Gymnasium, Hub-loadable envs | | Robots | SO-100, Koch v1.1, Moss v1, Aloha, first humanoid | | License | Apache 2.0 | | Stars | 22,200+ | | Contributors | 200+ PRs in v0.5.0 alone | ## Key Capabilities LeRobot v0.5.0, the latest release, represents the project's most ambitious expansion to date with over 200 merged pull requests and 50+ new contributors. **Hardware-Agnostic Robot Interface**: LeRobot provides a Python-native control interface that standardizes interaction across diverse robot platforms. Whether controlling a low-cost SO-100 arm, a Koch v1.1 bimanual setup, or the project's first supported humanoid robot, the same high-level API applies. This abstraction allows policies trained in simulation or on one robot to be evaluated on different hardware with minimal code changes. **Pre-Trained Policy Zoo**: The Hugging Face Hub hosts a growing collection of pre-trained robot policies that can be downloaded and deployed directly. These models span manipulation tasks (pick-and-place, insertion, folding) and have been validated for sim-to-real transfer. The policy zoo follows the same model card convention as language and vision models, providing standardized documentation of training data, evaluation results, and usage instructions. **Hub-Loadable Simulation Environments**: Version 0.5.0 introduced the ability to load simulation environments directly from the Hugging Face Hub. This means researchers can share not just datasets and models, but complete reproducible training setups including environment configurations, reward functions, and evaluation protocols. **DIY Robot Ecosystem**: LeRobot has cultivated a vibrant community of builders who construct affordable robots using 3D-printed parts and commodity actuators. Detailed build guides, CAD files, and bill-of-materials listings enable anyone with a 3D printer to assemble a functional robot arm for under $300. YouTube tutorials and an active Discord community provide support throughout the build process. **Autoregressive VLA Models**: The v0.5.0 release reintroduced autoregressive vision-language-action models, which generate robot actions conditioned on both visual observations and natural language task descriptions. This capability enables instruction-following robots that can interpret open-ended commands rather than requiring task-specific training. ## Developer Integration LeRobot provides a streamlined workflow from data collection to deployment: ```python from lerobot import LeRobotDataset, Policy # Load a pre-trained policy from the Hub policy = Policy.from_pretrained("lerobot/act-so100-pick") # Run inference on a robot from lerobot.robot import Robot robot = Robot("so100") robot.connect() observation = robot.get_observation() action = policy.predict(observation) robot.execute(action) ``` For training custom policies, LeRobot provides data collection tools that record demonstrations through teleoperation, convert them to the LeRobotDataset format, and upload them to the Hub. Training scripts support distributed training, wandb logging, and checkpoint management out of the box. ## Limitations LeRobot's policy implementations, while covering major algorithms, may lag behind the very latest research publications by several months. The hardware-agnostic interface introduces a small abstraction overhead that may matter for time-critical control loops. The DIY robot ecosystem, while affordable, produces hardware with lower precision and repeatability than commercial research platforms. Dataset quality on the Hub varies significantly as community contributions are not uniformly curated. Real-time control support for high-frequency servoing applications remains an active development area. ## Who Should Use This LeRobot is ideal for robotics researchers who want to benchmark policies across standardized tasks and datasets, educators teaching robot learning courses with affordable hardware, hobbyists and makers building their first AI-controlled robot, and teams developing vision-language-action models who need a diverse collection of robot manipulation data. The Hugging Face ecosystem integration makes it particularly appealing for practitioners already familiar with the Hub workflow for NLP and vision tasks.