Open Source
Explore the latest AI open-source projects from GitHub and HuggingFace.
Explore the latest AI open-source projects from GitHub and HuggingFace.
GLM-V is an open-source multimodal vision-language model family from Z.ai (the team behind the GLM series) covering GLM-4.6V, GLM-4.5V, and GLM-4.1V-Thinking. These models achieve versatile multimodal reasoning via scalable reinforcement learning, supporting text, images, video, documents, GUI agents, and 3D spatial tasks. GLM-4.6V features a 128K context window and native multimodal tool use, accepting images directly as tool parameters.
hacksider
Real-time AI face swap and one-click video deepfake with only a single image
harry0703
AI-powered short video generator that automates scripting, footage sourcing, subtitles, and composition — supporting 10+ LLM providers and batch production.
microsoft
Microsoft's official 1-bit LLM inference framework achieving 1.37x-6.17x speedup and up to 82% energy reduction, enabling 100B parameter models to run on consumer CPUs.
bytedance
ByteDance's open-source multimodal AI agent stack for GUI automation with vision-language model integration