Open Source
Explore the latest AI open-source projects from GitHub and HuggingFace.
Explore the latest AI open-source projects from GitHub and HuggingFace.
Sirchmunk is an open-source agentic search engine from Alibaba's ModelScope team that eliminates the need for vector databases, pre-indexing, and ETL pipelines entirely. Released in January 2026 and reaching v0.0.5 by March 2026, it enables raw data to become searchable intelligence in real-time without any embedding infrastructure. ## Why Sirchmunk Matters Traditional RAG (Retrieval-Augmented Generation) systems require expensive setup: vector databases, embedding models, chunking strategies, and batch re-indexing pipelines. Sirchmunk challenges this entire paradigm by operating directly on raw data files. There are no vectors, no pre-processing steps, and no stale indexes. Data is always current because the system reads source files at query time, using intelligent search algorithms to find relevant information. ## Key Features ### Embedding-Free Architecture Sirchmunk bypasses vector approximation entirely. Instead of converting documents into embeddings and performing approximate nearest neighbor search, it uses a combination of keyword extraction, context-window sampling, and LLM-powered evidence evaluation. This means zero information loss from vector quantization and zero infrastructure overhead from maintaining embedding indexes. ### Dual Search Modes The engine offers two primary search modes. FAST mode uses a greedy keyword cascade with only 2 LLM calls, returning results in 2-5 seconds. DEEP mode employs comprehensive Monte Carlo evidence sampling for thorough analysis, taking 10-30 seconds but providing higher accuracy. A FILENAME_ONLY mode is also available for quick file discovery without content analysis. ### Self-Evolving Knowledge Base Rather than treating data as static snapshots, Sirchmunk processes information as a continuous stream. Its Knowledge Clusters use SHA256 deterministic IDs for stability, and the system triggers LLM inference only when necessary through token-efficient reasoning. This means the knowledge base improves and adapts without manual re-indexing. ### MCP Integration Sirchmunk supports the Model Context Protocol (MCP), enabling direct integration with Claude Desktop and Cursor IDE. Developers can add Sirchmunk as an MCP tool to give their AI assistants instant search capability over local document collections without setting up any vector infrastructure. ## Technical Architecture | Component | Technology | |-----------|------------| | Backend | FastAPI 0.100+, Python 3.10+ | | Frontend | Next.js 14, TailwindCSS 3.4 | | Database | DuckDB (OLAP) | | Search Engine | ripgrep-all | | LLM Integration | OpenAI-compatible APIs | | Text Extraction | Kreuzberg library | | Protocol | Model Context Protocol (MCP) | ## How It Compares to Traditional RAG | Dimension | Traditional RAG | Sirchmunk | |-----------|-----------------|----------| | Setup Cost | High infrastructure overhead | Zero infrastructure required | | Data Freshness | Batch re-indexing (stale) | Dynamic, real-time updates | | Scalability | Linear cost growth | Low RAM/CPU consumption | | Accuracy | Vector approximation | Deterministic and contextual | | Workflow | Complex ETL pipelines | Drop-and-search deployment | ## Installation and Usage Getting started requires just two commands: `pip install sirchmunk` followed by `sirchmunk init`. The system can be deployed via Docker with pre-built images from Alibaba Cloud Container Registry, or run locally with the built-in web UI served alongside the API. ## Limitations - Still in early development (v0.0.5) with evolving APIs - Star count is relatively low (381) compared to established RAG frameworks - DEEP mode search latency of 10-30 seconds may be too slow for real-time applications - Requires OpenAI-compatible LLM API access for intelligent search features ## Conclusion Sirchmunk represents a compelling alternative to the increasingly complex RAG infrastructure stack. By proving that effective document search can work without embeddings or vector databases, it challenges assumptions that have driven significant engineering investment across the industry. For teams tired of maintaining vector database infrastructure, Sirchmunk offers a refreshingly simple approach: point it at your files and search.