Open Source
Explore the latest AI open-source projects from GitHub and HuggingFace.
Explore the latest AI open-source projects from GitHub and HuggingFace.
MLC-LLM is a universal LLM deployment engine powered by ML compilation via Apache TVM. It compiles models once and runs them natively across NVIDIA, AMD, Apple, and Intel GPUs as well as mobile platforms including iOS and Android, with WebGPU support for browser-based inference. The unified MLCEngine provides an OpenAI-compatible REST API, Python, JavaScript, and mobile bindings from the same compiled artifact, enabling developers to deploy quantized LLMs from cloud to edge without platform-specific rewrites.
ollama
The simplest way to run LLMs locally with 165K+ GitHub stars. One-command deployment, 100+ models, REST API, and multi-platform support.
ggml-org
Pure C/C++ LLM inference engine supporting CPUs, Apple Silicon, CUDA, and Vulkan
sgl-project
High-performance LLM and multimodal model serving framework with RadixAttention and structured generation.
mlc-ai
Run LLMs entirely in the browser with WebGPU acceleration and OpenAI API compatibility — no server needed