Open Source AI Projects and Tools Announcements: April 11-12, 2026

Matthew Diakonov··12 min read

Open Source AI Projects and Tools Announcements: April 11-12, 2026

April 11 and 12 produced a concentrated burst of open source AI announcements. MiniMax officially launched its self-evolving M2.7 agent model. OpenClaw announced Active Memory and native Codex support. OpenAI released major Codex CLI upgrades covering voice, MCP, and remote workflows. Google's Gemma 4 GGUF quantizations received a critical compatibility fix from the community. And Archon emerged as a new contender in the coding harness space. Here is every notable announcement from those two days.

Announcement Overview

| Project | Date | Type | License | Announcement | |---|---|---|---|---| | MiniMax M2.7 | April 12 | Model launch | Open weights | 230B MoE with self-evolving training, 56% SWE-Pro | | OpenClaw 2026.4.10 | April 11 | Platform release | Open source | Active Memory plugin, native Codex provider | | Codex CLI | April 11 | Tool release | Open source | Realtime voice v2, MCP upgrades, remote workflows | | Gemma 4 GGUF | April 11 | Compatibility fix | Apache 2.0 | Chat template correction for llama.cpp and Ollama | | claude-mem v11.0 | April 11 | Milestone | Open source | Sync CLI launch, crossed 46K GitHub stars | | Archon | April 11 | Framework launch | Open source | YAML deterministic pipelines, 14K+ stars |

MiniMax M2.7 Launch

MiniMax announced M2.7 on April 12, a 230-billion parameter sparse mixture-of-experts model that participated in its own training. The model helped design reinforcement learning experiments during development, a capability MiniMax calls "self-evolving training."

Performance Benchmarks

| Benchmark | M2.7 | Category | |---|---|---| | SWE-Pro | 56.22% | Software engineering | | VIBE-Pro | 55.6% | End-to-end project delivery | | Terminal Bench 2 | 57.0% | Complex engineering |

The MoE architecture activates only a subset of parameters during each forward pass, keeping inference costs lower than a comparable dense model. MiniMax reports a 30% internal evaluation improvement from the self-evolving training approach.

# Pull MiniMax M2.7 from Ollama (requires multi-GPU)
ollama pull minimax-m2.7

# Download from Hugging Face
huggingface-cli download MiniMaxAI/MiniMax-M2.7 \
  --local-dir ./models/minimax-m2.7

# Serve via vLLM with tensor parallelism
python -m vllm.entrypoints.openai.api_server \
  --model MiniMaxAI/MiniMax-M2.7 \
  --tensor-parallel-size 4

Hardware requirements

M2.7 requires a multi-GPU setup for local inference even with sparse activation. For single-GPU users, Ollama and OpenRouter hosted endpoints provide the fastest path to evaluation.

OpenClaw 2026.4.10 Announcement

OpenClaw shipped version 2026.4.10 on April 11, announcing 17 new features and over 20 fixes. The two headline announcements: Active Memory and native Codex integration.

Active Memory Plugin

The Active Memory plugin introduces a dedicated memory sub-agent that runs before every response. Rather than requiring users to manually invoke memory commands, the plugin automatically retrieves relevant preferences, prior context, and session details.

Configuration modes:

  • Message mode injects recent context with lower token overhead
  • Full context mode performs deep retrieval across all sessions
  • Verbose inspection via /verbose for debugging retrieval decisions
  • Transcript persistence for auditing what the memory agent retrieves

Native Codex Provider

Version 2026.4.10 adds a dedicated Codex provider path with its own authentication, thread management, and context compaction. Models accessed via codex/gpt-* now route through Codex-specific APIs, while openai/gpt-* continues using the original provider.

Additional announcements in 2026.4.10 include local MLX voice support for macOS Talk Mode, SSRF hardening, launchd stability fixes, and Teams integration with pins, reactions, and read receipts.

# Update to OpenClaw 2026.4.10
npm install -g openclaw@2026.4.10

# Enable Active Memory
openclaw config set plugins.active-memory.enabled true

# Configure Codex provider
openclaw config set providers.codex.auth oauth

Codex CLI: Voice, MCP, and Remote Workflow Announcements

OpenAI released a major Codex CLI update on April 11 covering three areas: realtime voice, enhanced MCP support, and remote workflow capabilities.

Voice mode now defaults to the v2 WebRTC transport. You can configure transport backends, select voices, and use native terminal media support for hands-free coding.

MCP support gained resource reads, tool-call metadata, custom-server tool search, server-driven elicitations, file-parameter uploads, and improved plugin cache reliability. Teams running MCP servers alongside Codex see significantly smoother integration.

Remote workflows added egress websocket transport, remote --cd forwarding, and sandbox-aware filesystem APIs. The experimental codex exec-server subcommand enables orchestrating Codex from external tools.

The codex-core crate was also decomposed: MCP, tools, config, model management, auth, feedback, and protocol each became separate crates, reducing compile times.

# Install latest Codex CLI
npm install -g @openai/codex@latest

# Start voice session
codex --voice

# Connect to remote workspace
codex --remote wss://your-server.example.com

Gemma 4 GGUF Compatibility Fix

Google released Gemma 4 on April 2 under Apache 2.0. On April 11, the community announced corrected GGUF quantization files with Google's latest chat template and llama.cpp compatibility fixes.

Quantization Options

| Variant | Size | VRAM Required | Target Hardware | |---|---|---|---| | Gemma 4 E2B | ~1.5GB | 2GB | Phones, Raspberry Pi, edge | | Gemma 4 E4B | ~3GB | 4GB | Laptops, lightweight tasks | | Gemma 4 26B-A4B (Q4_K_M) | ~15GB | 16GB | Development workstations | | Gemma 4 31B Dense (Q4_K_M) | ~18GB | 20GB | High-capability local inference |

The fix addresses a chat template mismatch that caused malformed output formatting in llama.cpp and Ollama. Anyone who downloaded Gemma 4 GGUF files before April 11 should re-download to get corrected versions.

Gemma 4 benchmarks show dramatic improvements over Gemma 3: AIME 2026 math went from 20.8% to 89.2%, LiveCodeBench coding from 29.1% to 80.0%, and GPQA science from 42.4% to 84.3%.

# Pull corrected Gemma 4 via Ollama
ollama pull gemma4:26b

# Download GGUF from Hugging Face
huggingface-cli download unsloth/gemma-4-26B-A4B-it-GGUF \
  --include "gemma-4-26B-A4B-it-Q4_K_M.gguf" \
  --local-dir ./models/gemma4

Archon: Coding Harness Builder Announcement

Archon announced itself as the first dedicated coding harness builder on April 11, quickly reaching 14,000+ GitHub stars. The framework uses YAML-based deterministic pipelines for defining coding agent workflows.

Rather than emergent agent coordination, Archon lets developers define explicit flow control through YAML configuration files. This deterministic approach reduces the unpredictability that plagues multi-step agent orchestration and makes debugging reproducible.

claude-mem v11.0: Sync CLI and 46K Stars

The claude-mem persistent memory plugin for Claude Code crossed 46,100 GitHub stars. The v11.0 release (shipped April 4, with adoption ramping through April 11-12) introduced a new claude-mem-sync CLI:

  • push / pull / sync / status commands for memory transfer
  • Bidirectional sync of observations and session summaries via SSH/SCP
  • Automatic deduplication to prevent memory bloat across machines

Five lifecycle hooks (SessionStart, UserPromptSubmit, PostToolUse, Stop, SessionEnd) capture agent activity without manual intervention. Local SQLite storage with Chroma vector search handles retrieval.

# Install claude-mem
claude plugins install claude-mem

# Sync between machines
claude-mem-sync push --target user@remote:~/.claude-mem
claude-mem-sync pull --source user@remote:~/.claude-mem

How the Announcements Connect

April 11-12 Announcements: Ecosystem MapMEMORYclaude-mem v11.046K stars, sync CLIOpenClaw Active MemoryAuto retrieval sub-agentHARNESSOpenClaw 2026.4.10Codex native, MLX voiceArchonYAML pipelinesCodex CLIVoice, MCP, remoteAGENTSClaude CodeCodexGemini CLIGooseINFERENCEllama.cpp / GGUFvLLM / SGLangOllamaMODELSMiniMax M2.7230B MoE, self-evolvingGemma 4 (fixed)Apache 2.0, GGUF fixGLM-5.1 (MIT)Community quantsPROTOCOLMCP (Linux Foundation AAIF governance)10,000+ servers, 97M installs, vendor-neutral protocolTeal = announced April 11-12 | Gray = existing ecosystem

Memory systems feed harnesses, which orchestrate agents. Agents use the inference layer to run models. MCP connects everything as the vendor-neutral protocol. Every teal-highlighted component received a formal announcement on April 11 or 12.

Key Themes Across the Announcements

Three patterns emerge from the April 11-12 announcements:

Memory is becoming automatic. Both OpenClaw Active Memory and claude-mem v11.0 move toward agents that remember context without explicit user commands. This shifts the burden from "tell the agent what to remember" to "tell the agent what to forget."

Voice is entering coding workflows. Codex CLI's v2 WebRTC voice and OpenClaw's MLX voice for macOS signal that hands-free coding interaction is moving from experiment to supported feature. Both announcements position voice as a primary input method, not a novelty.

Self-improving training has arrived in open source. MiniMax M2.7's self-evolving training, where the model participates in designing its own RL experiments, is the first open source example of this approach at scale. The deployed model is frozen at inference time, but the training methodology could reshape how open source models are developed.

Things to Watch After April 12

  1. M2.7 community quantizations. GGUF and other quantized formats from Unsloth and community teams will determine whether M2.7 reaches single-GPU users.
  2. OpenClaw Active Memory usage patterns. Full context mode increases token usage and latency significantly. Best practices around mode selection will emerge from early adopter feedback.
  3. Codex voice reliability. The v2 WebRTC path is new. Real-world testing will surface platform-specific issues around latency and transcription quality.
  4. Gemma 4 fine-tuning wave. With GGUF fixes landed and Unsloth support live, domain-specific fine-tunes for coding, medical, and legal use cases are next.
  5. Memory format standardization. claude-mem and OpenClaw Active Memory point toward the same future but use different formats. Convergence would benefit the entire ecosystem.

Quick Start Guide

# Try MiniMax M2.7 (multi-GPU required)
ollama pull minimax-m2.7
ollama run minimax-m2.7 "Explain your MoE architecture"

# Try Gemma 4 26B locally (16GB+ VRAM)
ollama pull gemma4:26b
ollama run gemma4:26b "Summarize the April 11 GGUF changes"

# Try OpenClaw with Active Memory (any machine)
npm install -g openclaw@2026.4.10
openclaw config set plugins.active-memory.enabled true

# Try Codex CLI with voice (any machine)
npm install -g @openai/codex@latest
codex --voice

# Try claude-mem for Claude Code (any machine)
claude plugins install claude-mem

Fastest path

For a quick test of the April 11-12 announcements, update Codex CLI. The MCP improvements and voice mode require no model downloads, and the update completes in under a minute. For local model testing, Gemma 4 26B via Ollama offers the best capability-to-hardware ratio.

Wrapping Up

April 11-12, 2026 delivered a concentrated set of announcements across every layer of the open source AI stack. MiniMax launched the first self-evolving open model at 230B MoE scale. OpenClaw announced automatic memory retrieval and native Codex integration. Codex CLI shipped voice, MCP, and remote workflow capabilities. The community fixed Gemma 4's GGUF deployment issues. And Archon emerged as a purpose-built coding harness framework. The common thread: the open source AI ecosystem is shifting from shipping models to shipping complete, memory-aware development environments.

Fazm is an open source macOS AI agent that works with MCP extensions and the open source models covered in this roundup. Open source on GitHub.

Related Posts