AI & Machine Learning

Top 350+ AI GitHub Projects 2026: The Complete Open Source Landscape

Discover the ultimate collection of 350+ top AI GitHub projects for 2026. From Claude Code to MLX and RAGFlow, we map the 13 core domains of the AI open-source ecosystem.

Top 350+ AI GitHub Projects 2026: The Complete Open Source Landscape

Introduction: The Golden Age of the AI Open Source Ecosystem

In 2026, the AI open-source ecosystem has reached a level of maturity that was once unimaginable. The days of relying solely on closed-source APIs are fading as the community delivers tools that match or exceed proprietary performance. From Claude Code surpassing 113K stars to Dify hitting 138K and LangFlow soaring to 147K, the numbers reflect a global movement toward decentralized, controllable intelligence.

This comprehensive guide serves as your definitive map for the 2026 AI landscape, compiling over 350 top AI-related GitHub projects across 13 core domains. Whether you are an AI engineer building autonomous agents, a data scientist fine-tuning the latest LLMs, or a developer integrating multimodal capabilities into your applications, this list provides the full technical blueprint of our era’s open-source revolution.

💡 Why bookmark this list?

  • Covers the latest Agent development paradigms and the Model Context Protocol (MCP) ecosystem.
  • Includes Apple Silicon-optimized MLX tools that unlock the full potential of Mac hardware.
  • Organizes a complete production toolchain, from research and memory systems to infrastructure.

Why is the AI Open Source Ecosystem Booming in 2026?

Direct answer: The boom is driven by three decisive shifts: the transition to Agentic workflows where AI executes tasks independently, the localization of compute through frameworks like MLX and Ollama, and the standardization of protocols like MCP that enable seamless interoperability between different AI systems.

The shift from centralized cloud-only models to hybrid and local deployments has made open-source projects the backbone of modern enterprise AI. Developers now prioritize data sovereignty and system latency, leading to the massive adoption of local inference engines and self-hosted RAG (Retrieval-Augmented Generation) pipelines.


What are the Core Categories of the 2026 AI GitHub Projects?

Direct answer: The 2026 AI landscape is organized into 13 specialized domains, including AI Agents, LLM Frameworks, Audio Processing, Video Generation, and RAG Infrastructure. This specialization ensures that developers can find production-grade tools for every layer of the AI stack, from raw hardware acceleration to high-level agent orchestration.

Ecosystem Overview

CategoryProject CountRepresentative Projects
AI Agents & Coding Tools100+Claude Code, OpenHands, Cursor
LLM Frameworks & Inference60+Ollama, vLLM, llama.cpp
Audio, Voice & Music30+GPT-SoVITS, ChatTTS
Video Generation & Editing25+OpenCut, Remotion
Memory Systems & RAG20+RAGFlow, LightRAG
Vision & Multimodal35+ComfyUI, InternVL
Training & Fine-tuning40+LlamaFactory, Unsloth
Tools & Infrastructure50+Dify, LangFlow
macOS Specialized20+MLX, QuickRecorder

Which AI Agent and Coding Tools are Redefining Productivity?

Direct answer: AI Agents have evolved into autonomous engineers capable of managing entire repositories. Tools like Claude Code and OpenHands can now navigate files, execute CLI commands, and perform surgical bug fixes with 95%+ success rates on common issues, effectively acting as 24/7 technical team members.

Mainstream Agent Frameworks

ProjectStarsDescription
anthropics/claude-code113KClaude Code - Agentic coding tool for terminal
anthropics/skills116KAgent Skills public repository
openai/codex75KLightweight terminal coding agent
openai/symphony15KIsolated autonomous execution
NousResearch/hermes-agent75KAgent that grows with you
FoundationAgents/MetaGPT67KMulti-agent framework for roles
FoundationAgents/OpenManus56KOpen-source Manus alternative
microsoft/autogen57KAgentic AI programming framework
langchain-ai/langchain133KAgent engineering platform
langchain-ai/langgraph29KGraphical language agent building
crewAIInc/crewAI49KRole-playing orchestration
bytedance/deer-flow61KLong-term SuperAgent framework
OpenHands/OpenHands71KAI-powered development
browser-use/browser-use88KAI Agent web access
agentscope-ai/agentscope24KBuilding trustworthy agents
stitionai/devika19KAgentic software engineer
Aider-AI/aider43KTerminal AI pair programming
Pythagora-io/gpt-pilot34KFirst real AI developer
cline/cline60KIDE autonomous coding agent
cursor/cursor33KAI code editor
SWE-agent/SWE-agent19KFix GitHub Issues automatically
RooCodeInc/Roo-Code23KEditor AI Agent
openinterpreter/open-interpreter63KNatural language computer interface
antonosika/gpt-engineer55KCLI code generation platform
gptme/gptme4KYour terminal agent
QwenLM/qwen-code23KOpen-source terminal AI agent
myshell-ai/AIlice1.4KFully autonomous agent
Fosowl/agenticSeek26KLocal Manus AI
kortix-ai/suna20KAutonomous company OS
moonshotai/kimi-cli8KKimi Code CLI
google-gemini/gemini-cli101KOpen-source terminal AI Agent

Claude Code Ecosystem

ProjectStarsDescription
claude-code-best/claude-code16KRunnable Claude Code
shareAI-lab/learn-claude-code53KNano Claude Code agent
steipete/CodexBar11KAgent usage statistics
musistudio/claude-code-router32KClaude Code infrastructure
doriandarko/claude-engineer11KClaude 3.5 Sonnet CLI
cranot/claude-code-guide2.6KFull CLI guide
wasabeef/claude-code-cookbook1KSetup and recipes

What are the Leading LLM Frameworks and Inference Engines?

Direct answer: Ollama continues to dominate for local accessibility, while vLLM and llama.cpp lead in high-throughput enterprise environments. For the Apple ecosystem, MLX has become the industry standard, providing hardware-native acceleration that outperforms standard ports on Mac Silicon.

Primary Inference Frameworks

ProjectStarsDescription
ollama/ollama169KOne-click LLM local deployment
vllm-project/vllm76KHigh-throughput inference engine
sgl-project/sglang26KHigh-performance serving
ggml-org/llama.cpp103KC/C++ LLM inference
huggingface/transformers159KSOTA ML models
huggingface/trl18KRL training for transformers
karpathy/llm.c30KRaw C/CUDA training
google/gemma.cpp7KGemma lightweight C++
unslothai/unsloth61KTraining Web UI
kvcache-ai/ktransformers17KHeterogeneous LLM inference
Tiiny-AI/PowerInfer9KFast local serving
NVIDIA/TensorRT-LLM13KNVIDIA LLM inference

Apple Silicon MLX Ecosystem

ProjectStarsDescription
ml-explore/mlx25KApple Silicon array framework
Blaizzy/mlx-vlm4KMac Vision-Language models
Blaizzy/mlx-audio7KApple MLX TTS/STT/STS
ml-explore/mlx-lm5KMLX running LLMs
jundot/omlx10KApple Silicon inference server
walter-grace/mac-code786Free Claude Code for Mac

How are Audio, Vision, and Multimodal Projects Advancing?

Direct answer: 2026 has seen a breakthrough in studio-quality generation from minimal data. Tools like ComfyUI and InternVL lead the multimodal visual space, while GPT-SoVITS and ChatTTS have democratized high-fidelity voice cloning with as little as 60 seconds of audio.

Audio, Voice & Music

ProjectStarsDescription
OpenBMB/VoxCPM12KNo-tokenizer TTS
RVC-Boss/GPT-SoVITS57K1-min voice data TTS
2noise/ChatTTS39KGenerative speech model
FunAudioLLM/CosyVoice21KMultilingual voice generation
index-tts/index-tts20KIndustrial grade TTS
SYSTRAN/faster-whisper22KFaster Whisper transcription
ace-step/ACE-Step-1.59KLocal music generation
Anjok07/ultimatevocalremovergui24KAI Vocal removal

Video Generation & Vision

ProjectStarsDescription
OpenCut-app/OpenCut48KOpen CapCut alternative
remotion-dev/remotion43KProgrammatic React video
Lightricks/LTX-26KA/V generation model
OpenGVLab/InternVL10KOpen GPT-4o alternative
OpenBMB/MiniCPM-o24KGemini 2.5 Flash level MLLM
Comfy-Org/ComfyUI109KDiffusion model GUI
PaddlePaddle/PaddleOCR76KMulti-format OCR toolkit
opendatalab/MinerU60KPDF to Markdown extraction

What Tools Support Memory, Training, and Infrastructure?

Direct answer: Modern RAG (Retrieval-Augmented Generation) has shifted toward GraphRAG for deeper semantic links. Infrastructure projects like Dify and LangFlow provide the visual “operating system” for agent deployment, while LlamaFactory remains the standard for unified fine-tuning across hundreds of models.

Memory Systems & RAG

ProjectStarsDescription
MemPalace/mempalace44KHighest rated AI memory
mem0ai/mem053KUniversal memory layer
infiniflow/ragflow78KEnterprise RAG engine
HKUDS/LightRAG33KFast and simple RAG
stanford-oval/storm28KLLM knowledge curation
chroma-core/chroma27KAI data infrastructure

Training & Infrastructure

ProjectStarsDescription
hiyouga/LlamaFactory70KUnified fine-tuning (ACL 2024)
unslothai/unsloth61KFast local model training
langgenius/dify138KProduction Agentic workflows
langflow-ai/langflow147KAI-driven Agent builder
astral-sh/uv83KFast Python package manager
ray-project/ray42KAI compute engine

Frequently Asked Questions (FAQ)

Why should I choose open-source AI over proprietary APIs like GPT-4?

Open-source AI in 2026 provides superior data privacy, lower long-term costs, and complete control over the model’s behavior. Advanced open-source models have now closed the performance gap with proprietary models for most industrial and enterprise applications.

Can I run these 2026 AI models on a standard laptop?

Yes, thanks to optimized inference engines like Ollama and MLX, many 7B to 30B parameter models can run smoothly on a modern laptop with 16GB+ RAM. Apple Silicon Macs are particularly effective due to their unified memory architecture.

What is the Model Context Protocol (MCP)?

MCP is an open standard that allows AI agents to securely connect to diverse data sources and tools without needing custom integrations for every single software. It has become the universal “plug-in” system for the 2026 Agent ecosystem.

Is fine-tuning still necessary if I use RAG?

RAG is best for providing real-time, accurate data, while fine-tuning is used to teach a model a specific style, format, or specialized domain-specific reasoning. Most production systems in 2026 use a hybrid approach of both.

How do I stay updated with the fast-moving GitHub ecosystem?

Starring repositories is a good start, but following “Awesome” curated lists and using AI-integrated code search tools like Bloop or SeaGOAT allows you to navigate the thousands of active projects efficiently.


Further Reading


Data sources: GitHub Star statistics, open-source community contributions, and curated collections by xiaotianfotos. Data current as of April 2026; star counts are approximations.

TAG
CATEGORIES