Open Source
Explore the latest AI open-source projects from GitHub and HuggingFace.
Explore the latest AI open-source projects from GitHub and HuggingFace.
## Introduction DimOS (Dimensional OS) is an agentic operating system for generalist robotics that enables developers to control humanoids, quadrupeds, drones, and manipulators using Python and natural language. With 1,600+ GitHub stars and 230+ forks, DimOS represents a new approach to robotics software where AI agents directly control physical hardware through natural language commands and structured skill execution. Currently in pre-release beta, the project has been trending on GitHub as interest in embodied AI surges. Traditional robotics software stacks rely on ROS (Robot Operating System) with complex message passing, launch files, and steep learning curves. DimOS reimagines this by making AI agents the primary control interface, allowing developers to command robots with instructions like "go find the kitchen" rather than writing low-level motion planning code. ## Architecture and Design DimOS is organized around four core capability areas: | Capability | Description | |-----------|------------| | Navigation and Mapping | SLAM, dynamic obstacle avoidance, route planning, autonomous exploration | | Perception | Object detection, 3D projections, vision language models, audio processing | | Agentive Control | Natural language commands with MCP (Model Context Protocol) support | | Spatial Memory | Spatio-temporal RAG, dynamic memory, object localization and permanence | **Navigation and Mapping** provides full autonomous navigation capabilities including Simultaneous Localization and Mapping (SLAM), dynamic obstacle avoidance, route planning, and autonomous exploration. These can run through native DimOS implementations or through ROS integration for teams with existing ROS infrastructure. **Perception** combines traditional computer vision (object detectors, 3D point cloud projections) with modern vision language models for semantic understanding of environments. Audio processing enables voice command interfaces for hands-free robot control. **Agentive Control** is the defining feature. Rather than programming robot behaviors imperatively, developers describe goals in natural language. The AI agent layer translates these into executable plans, calling appropriate navigation, perception, and manipulation skills. MCP (Model Context Protocol) support means the robot's skills are exposed as standardized tools that any compatible LLM can invoke. **Spatial Memory** implements spatio-temporal RAG (Retrieval-Augmented Generation) for robots, enabling them to remember where objects were seen, track changes over time, and maintain object permanence even when items leave the visual field. ## Supported Hardware DimOS supports a wide range of robotic platforms: - **Quadrupeds**: Unitree Go2 Pro/Air, Unitree B1 - **Humanoids**: Unitree G1 - **Manipulators**: XArm, AgileX Piper - **Drones**: MAVLink-compatible, DJI Mavic - **Sensors**: Force/torque sensors, depth cameras ## Key Capabilities **No ROS Required**: Basic operation works without ROS installation, dramatically lowering the barrier to entry for developers coming from web or AI backgrounds. **MCP Server**: Exposes robot skills as MCP tools, enabling any MCP-compatible LLM to control the robot through standardized tool calling protocols. **Simulation Support**: MuJoCo integration allows development and testing without physical hardware, enabling faster iteration cycles. **Modular Architecture**: Standardized message passing with multi-language support (C++, Lua, TypeScript via LCM interop) allows mixing components written in different languages. ## Limitations DimOS is in pre-release beta, meaning APIs are subject to breaking changes and production deployment is not yet recommended. Hardware support, while growing, is currently concentrated on Unitree platforms with varying levels of maturity for other manufacturers. The natural language control layer adds latency compared to direct programmatic control, making it unsuitable for time-critical reactive behaviors. Documentation is still catching up with the codebase, and some advanced features lack comprehensive examples. The project requires significant compute resources on the robot or a connected workstation for running the LLM inference that powers agent control. ## Who Should Use This DimOS is ideal for robotics researchers exploring embodied AI and language-grounded robot control. AI developers interested in extending their agent frameworks to physical hardware will find the MCP integration particularly compelling. Teams prototyping new robot applications can use the natural language interface for rapid experimentation before optimizing with traditional control methods. Academic labs working on spatial reasoning, navigation, and human-robot interaction benefit from the integrated perception and spatial memory systems. Early-stage robotics startups looking for a modern alternative to ROS can evaluate DimOS as a foundation for their products.