A new LLM or foundation model has been released or announced almost every week of May 2026. The interesting question is not the list. It is how fast you can put one to work.
Every roundup of new models in May 2026 looks the same: names, parameter counts, benchmark numbers, links to the model card. None of that tells you whether the new model is good enough to drive your real work tomorrow morning. The only thing that answers that is running the model inside a real agent loop, on a real task, for ten minutes. This guide is about doing exactly that, using one specific Fazm feature most people miss.
Direct answer, verified 2026-05-15
There is no canonical "released in May 2026" list for foundation models. The two surfaces that hold the ground truth are:
- Each lab's own changelog or blog: anthropic.com/news, openai.com/blog, ai.meta.com/blog, deepmind.google.
- Open-weights releases sorted by date: huggingface.co/models?sort=created.
Once you have a candidate model, the actual question is how fast you can drive your existing workflow with it. With Fazm's custom API endpoint, the answer is on the order of an hour: paste the URL into a new chat window, fork an existing conversation as the test prompt, and read both transcripts side by side.
The thesis: a model release is not a workflow upgrade until it survives an agent loop
Every May 2026 roundup leans on the same shape. A new model ships, the post lists its parameter count, its context window, its score on a few benchmarks, the price per million tokens. That is enough information to file the model in your head. It is not enough information to switch to it.
The reason is structural. Benchmarks measure single-turn answer quality on a fixed set of prompts. The work most people actually do with these models is multi-turn, tool-calling, and stateful: a coding session that runs for two hours and edits twenty files, a research task that opens ten tabs and pulls quotes back, a workflow that drives a spreadsheet and a calendar in the same conversation. None of the public benchmarks measure that, and the failure modes in that regime are not the failure modes a benchmark would catch. A model can be best-in-class on MMLU and still fall apart on turn nine of an agent loop, because the loss function it was trained against was not the one your work cares about.
The honest test is to put the new model behind your existing agent harness, give it a real task you care about, and read the transcript. The cost of doing that should be measured in minutes, not days. Where most setups fall down is the wiring: stitching a brand-new endpoint into the harness you already use is a project, not an afternoon. The point of the rest of this page is that it does not have to be.
The anchor: Fazm has a custom API endpoint per chat window
This is the feature most pages about new model releases never mention, because most pages are not written by people who ship a model client. Fazm wraps Claude Code via ACP and Codex via codex-acp inside a native macOS UI. The agent loop on the other side of the wrapper is the real one. What the wrapper adds, beyond the persistent sessions and one-click forking, is that the backend URL is a per-chat setting.
Concretely: open a new chat, swap the endpoint to any Anthropic-compatible URL, and that window's agent is now driven by whatever model lives at that endpoint. A corporate proxy. GitHub Copilot's Anthropic-shaped gateway. An OpenRouter model. Your own LiteLLM instance running a fresh May 2026 release behind a translation shim. The chat window does not need a new build, a new plugin, or even a restart.
That single property is what makes the "ten minutes of real agent traffic" test cheap. You do not have to wait for someone to publish an integration for the new model. You wire it in yourself, the same evening it lands, and the rest of your workflow keeps working unchanged.
“A model release is information. A model running inside your existing agent loop is leverage. The job of a desktop client is to keep the gap between those two states as small as possible.”
What "swap the backend per chat window" actually means
Claude (Anthropic)
Native: bring your Claude Pro or Max account, the agent loop is real Claude Code via ACP. Default for most users.
Codex (OpenAI)
Bundled via codex-acp as a swappable backend. Pick it per chat window, the rest of the UI does not change.
Custom API endpoint
Any Anthropic-compatible URL: a corporate proxy, GitHub Copilot's gateway, OpenRouter, your own LiteLLM instance. The chat window does not need to know which model is on the other side.
New May 2026 model
If the lab ships an Anthropic-shaped endpoint, point Fazm at it directly. If they only ship OpenAI-shaped, a tiny reverse proxy translates and the same chat window now runs the new model.
A worked test: pointing a new May 2026 model at a real task
Suppose a new open-weights instruct model lands today on Hugging Face. The provider exposes an OpenAI-compatible HTTP endpoint, because that is the most common shape this year. You want to know whether it can replace your current backend for a coding loop you actually care about. Here is the flow from zero to a side-by-side transcript.
The whole loop above is bounded by how long it takes to bring up the shim. The Fazm side, the part that actually used to be painful (sessions, forking, pointing at a new endpoint), is three menu clicks. That is the point. The wrapper is no longer the bottleneck on whether you adopt a new model.
If the new model survives ten turns of your real task, you promote it: that becomes the default endpoint for new windows, the old chats keep running on the old backend until you decide to fork them too. If it does not survive, you close the window. Nothing else in your setup changed. That asymmetry is what keeps the cost of trying things low.
What to actually look for in the transcript
Most reviews of a new model in May 2026 grade the first response. That is the wrong frame. The first response of any half-decent model in 2026 is fine. The texture you are looking for shows up later, and it has three signatures.
First, on-topic drift. Around turn six or seven, does the model still remember the constraint you set in turn one, or has it quietly started solving a different problem? This is where context handling shows up. A model that compacts silently will lose your constraint and confidently produce something off-spec.
Second, tool-call shape. When the model decides to call a tool, are the arguments well-formed against the schema, or is it hallucinating field names? A model that gets the schema wrong on turn three will keep getting it wrong, because it does not know it was wrong. This is the cheapest way to spot a model that was not trained on enough real tool-use data.
Third, recovery. When a tool returns an error or an unexpected shape, does the model adjust, or does it loop on the same broken call? Recovery is the single hardest behavior to pre-train, and it is the behavior that determines whether an agent loop terminates successfully or burns through your token budget chasing its tail.
Why the wrapper matters more than usual right now
May 2026 has had an unusual cadence. The release pace from both the major labs and the open-weights community is high enough that any given week, something on your shortlist is already obsolete. The cost of switching has to drop in proportion, or you stop switching at all and you settle for whatever you started the month on.
That is the uncopyable reason to use a wrapper that holds sessions, forks chats in one click, and treats the model endpoint as configuration rather than a hardcoded build target. The wrapper is the substrate that makes "try the new model on a real task today" a normal Tuesday move instead of a weekend project. It is the thing that turns the stream of May 2026 announcements from a spectator sport into something you act on.
menu clicks to point a window at a new endpoint
turns of real agent traffic to know if a model holds up
click to fork the chat and run an A/B
The honest caveats
This pattern only works as well as the wire format on the other side. If a new model is shipped behind an entirely bespoke API shape with no published shim, you will spend a few hours writing the shim before the rest of the flow applies. The number of new models in that situation in May 2026 is small but not zero.
The other caveat is rate limits. A free or low-tier endpoint for a fresh model often will not let you do the ten-turn test without backing off. That is fine; it just means the test takes an evening rather than an hour. The wrapper part of the equation does not change.
Finally: the test is not whether the new model is the best available in absolute terms. It is whether it is better, on your real workflow, than the one you were already using. Most switches do not happen because that comparison was never made. The whole point of dropping the cost of running it is that you will, finally, make it.
Want a hand wiring a brand-new May 2026 model into your existing agent loop?
Book a 20-minute call. We will look at the model, your task, and what the cleanest endpoint shim is for your stack.
Frequently asked questions
What counts as a 'large language model released in May 2026' for the purposes of this guide?
Anything a model lab or open-source group has either shipped weights for or made available behind an API during May 2026. That includes new base or instruct models from the major labs, open-weights releases on Hugging Face, fine-tunes of older bases that are useful enough to land on the Trending tab, and provider-side announcements that promise rollout this month. It does not include rumours or leaks. The line is whether you can, today, point an HTTP client at an endpoint and get a response back. If you can, it counts.
Why does it matter how a new model performs as a desktop agent backend, and not just on benchmarks?
Benchmarks measure single-turn answer quality. Agent loops measure something different: whether the model can hold a long, growing conversation, follow multi-step plans, call tools without hallucinating their schemas, and recover when a tool returns an error. A model that wins MMLU can still fail the second turn of an agent loop, because the failure mode is structural, not knowledge-based. You only learn this by running the model inside a real harness. Reading a leaderboard is not a substitute for running ten minutes of agent traffic through it.
What does Fazm specifically let you do with a freshly-released LLM that other clients do not?
Fazm wraps Claude Code (via ACP) and Codex (codex-acp) in a native macOS UI, and the model behind that wrapper is configurable per chat window. The key feature is custom API endpoint support: any Anthropic-compatible gateway, corporate proxy, or third-party router. So when a new model lands in May 2026 and someone exposes it behind an Anthropic-shaped endpoint (or you put a thin shim in front of it), you can open a fresh chat window in Fazm, paste the endpoint URL, and that window's agent loop is now driven by the new model. The persistent session, one-click fork, and accessibility-API-driven actions all keep working. The model is the only thing that changed.
Where do the model providers actually announce a release on the day it ships?
There is no single feed. The most reliable surfaces are: the lab's own blog (anthropic.com/news, openai.com/blog, ai.meta.com/blog, deepmind.google), the model card on Hugging Face for open releases, GitHub release tags for tooling around the model, and the official X account of the team. Aggregators like Hacker News and the r/LocalLLaMA subreddit tend to surface things within hours. Press coverage typically lags by a day or two and is filtered through one writer's read. If you want to be early, watch the lab's own blog and the Hugging Face profile that lab uses, not the press.
How do I tell, in under an hour, whether a new May 2026 model is good enough to switch my workflow to?
Set up an honest A/B. Open Fazm. Open one chat window pointed at your current model. Open a second chat window pointed at the new endpoint. Use the one-click fork to give both windows the same starting context, then send the same multi-turn task to each: a real bug from your work, a real document you need rewritten, a real refactor across three files. Read both transcripts side by side. The thing you are looking for is not which one wrote a prettier first paragraph. It is which one was still on-topic and using tools correctly by turn ten. That is the test that maps to whether your workflow will actually be better tomorrow.
What about model releases that are 'announced' in May 2026 but not yet shipped?
Treat them as marketing until the endpoint exists. The thing that matters for any working developer or operator is whether the model is reachable from your code today. An announcement with a 'coming this summer' tag is information about future capacity planning, not something you can build on this week. The only useful response to a pure announcement is to subscribe to the relevant changelog so you know the day the endpoint actually opens.
Do I need to wait for someone to publish a Claude Code or ACP integration for a brand-new model?
No, and that is the point of the custom API endpoint feature. ACP just needs an Anthropic-compatible HTTP shape on the other side. If the model lab provides that natively, you point Fazm at it directly. If they expose only an OpenAI-compatible shape, a small reverse proxy (LiteLLM, OpenRouter, your own twenty-line script) translates the request and response. The Fazm chat window does not know or care which model is on the other side, as long as the wire format matches. So 'wait for an integration' is rarely the right answer in May 2026; the integration is a shim you can write yourself in an evening.
The companion pieces that go with this one.
Related guides
AI tech developments, May 11 to 12, 2026
A mid-May 2026 snapshot of model releases and project activity, the kind of roundup this guide tells you how to act on.
Open-source LLM releases, May 2026
Specifically the open-weights side of the same month, and how those releases fit into a desktop agent workflow.
Hugging Face or GitHub for new AI projects, May 13 2026
Companion piece on how to vet brand-new projects by their release cadence, with a worked example.
Comments (••)
Leave a comment to see what others are saying.Public and anonymous. No signup.