New Open Source AI Projects on GitHub and Hugging Face - April 2026

Matthew Diakonov··6 min read

New Open Source AI Projects on GitHub and Hugging Face - April 2026

Every month brings hundreds of new AI repositories and model uploads across GitHub and Hugging Face. Most disappear into noise. This post filters for the projects that actually gained traction in April 2026: the ones developers starred, downloaded, and started building with.

Top New GitHub Repositories - April 2026

| Repository | Stars (first 2 weeks) | Language | What It Does | |---|---|---|---| | google/adk-python | 8,200+ | Python | Google Agent Development Kit for building multi-agent systems | | meta-llama/llama-stack | 6,400+ | Python | Unified deployment stack for Llama 4 family models | | openai/codex-cli | 5,800+ | TypeScript | Terminal-native coding agent with sandboxed execution | | block/goose | 4,900+ | Rust | Local-first AI agent framework with MCP support | | huggingface/smolagents | 4,100+ | Python | Lightweight agent library with tool-use and code execution | | microsoft/markitdown | 3,600+ | Python | Convert any document format to clean Markdown for LLM input | | deepseek-ai/DeepSeek-V3 | 3,200+ | Python | MoE inference code for DeepSeek V3 (671B total, 37B active) | | qwen-ai/qwen3-coder | 2,800+ | Python | Code-specialized Qwen 3 with 128K context and tool calling | | mistralai/codestral-2 | 2,400+ | Python | Apache 2.0 licensed code generation model and inference code | | unsloth/unsloth | 2,100+ (monthly gain) | Python | 2x faster fine-tuning with 70% less memory, now supports Llama 4 |

April 2026: Where New AI Projects Get PublishedGitHubCode, frameworks,tools, CLIsHugging FaceModels, datasets,Spaces demosBoth PlatformsModel + inferencecode togetherAgent Frameworks (Goose, ADK, Codex CLI)Models (Qwen 3, Codestral 2, Llama 4)Developer Workflow: Clone repo + Download model = Working AI tool

Top New Hugging Face Models and Spaces - April 2026

The Hugging Face Hub saw several high-impact model uploads in the first half of April 2026. These are filtered by download velocity and community engagement, not just raw size.

| Model / Space | Downloads (first week) | Type | Key Differentiator | |---|---|---|---| | meta-llama/Llama-4-Scout-17B | 1,200,000+ | MoE LLM | 17B active params, runs on single 48GB GPU | | meta-llama/Llama-4-Maverick-17B | 890,000+ | MoE LLM | Optimized for creative and conversational tasks | | Qwen/Qwen3-72B | 640,000+ | Dense LLM | First open model to beat GPT-4o on MMLU-Pro | | Qwen/Qwen3-Coder-32B | 420,000+ | Code LLM | 128K context, native tool calling | | mistralai/Codestral-2-22B | 380,000+ | Code LLM | Apache 2.0 license, unrestricted commercial use | | google/gemma-3-9b | 310,000+ | Dense LLM | Newly opened for commercial use | | deepseek-ai/DeepSeek-V3-Base | 280,000+ | MoE LLM | 671B total params, 37B active, research reference | | unsloth/Llama-4-Scout-GGUF | 250,000+ | Quantized | 4-bit GGUF for llama.cpp, day-one availability | | HuggingFaceTB/SmolVLM2-2.2B | 180,000+ | Vision LLM | Tiny multimodal model for edge deployment | | black-forest-labs/FLUX.1-Kontext | 160,000+ | Image gen | In-context image editing, text rendering |

Notable Patterns in April 2026

MoE Architecture Goes Mainstream

The biggest shift visible across both platforms: mixture-of-experts is no longer experimental. Llama 4 Scout, DeepSeek V3, and several Qwen variants all use MoE to deliver large-model performance at smaller inference costs. For developers, this means you can now run "70B-class" intelligence on hardware that previously topped out at 13B dense models.

GitHub Stars Alone Do Not Tell the Story

Several April releases got massive star counts from hype but limited actual usage. The projects in the table above were selected based on a combination of:

  • Star velocity (first 14 days)
  • Fork count and PR activity (people building on it, not just bookmarking)
  • Issues opened (people actually running it and hitting real problems)
  • Downstream integrations (other projects depending on it)

Hugging Face Spaces as Launch Strategy

More teams are using Hugging Face Spaces as interactive demos alongside their GitHub code releases. This pattern, where the model weights live on the Hub, the inference code lives on GitHub, and a Spaces demo provides instant try-before-you-clone access, became the default launch strategy in April.

How to Find New Projects Early

If you want to catch these before they hit trending lists:

  1. GitHub Trending filtered by language and date range gives you raw signal but high noise
  2. Hugging Face Daily Papers surfaces arxiv papers with linked model weights
  3. GitHub Release Radar emails you about new releases from repos you watch
  4. Hugging Face Collections curated by domain experts (search "april 2026 releases")
  5. r/LocalLLaMA and r/MachineLearning usually surface notable drops within hours

Practical tip

If you are evaluating whether a new project is worth adopting, check the issues tab first. A repo with 5,000 stars and 3 issues is likely hype; one with 2,000 stars and 200 issues means people are actually using it in real workflows.

What Matters Most for Builders

If you are deciding what to try from this month's releases:

  • Need a local coding assistant? Codestral 2 (Apache 2.0) or Qwen3-Coder-32B give you strong code completion without API costs
  • Building an agent? Google ADK or Goose provide the most mature MCP integration
  • Want multimodal on edge hardware? SmolVLM2-2.2B runs on devices with as little as 4GB RAM
  • Fine-tuning on custom data? Unsloth's Llama 4 support cuts training time and memory usage in half

The velocity of open source AI releases shows no sign of slowing. What has changed is the quality floor: projects released in April 2026 ship with quantized weights, working inference code, and interactive demos from day one. The era of "here is a paper and maybe we will release weights eventually" is effectively over.

Related Posts