Parameter Update: 2026-08

"peaceclaude" edition

Parameter Update: 2026-08

Missed last week's post because I was in Paris for HackEurope, so this week's busier than usual. And then there's all the DoW drama which honestly deserves a post on its own.

DoW Drama

For the past couple weeks, the US Department of War and Anthropic have been going through a very loud discussion around the usage of Claude within the department. While the DoW insisted on "any lawful use” of frontier models inside defense/intel contexts, Anthropic insisted on two explicit carve-outs (mass domestic surveillance + fully autonomous weapons), claiming that law/policy can lag behind capabilities or might be changed, but these two applications should never be permissable.

Despite Anthropic going all-out in a blog post, claiming to be under attack from Chinese labs distilling their models (which.. so what? and also - why does Claude insist on being a DeepSeek model when talked to in Chinese if you're not doing the same thing?), talks soon broke down.

Immediately after, the Pentagon/DoW went ahead and forced a phase-out of Claude within the department. Furthermore, in a step that is hard to read as anything but punative, it designated Anthropic a “supply chain risk”, which, to my understanding, forces any company doing business with the DoW to also phase out usage of Anthropic products.

While Anthropic has announced plans to challenge the designation in court, the process will take a while, and leaves Anthropic in a precarious position. Adding insult to injury, hours after the designation was made public, and hours before US strikes on Iran, OpenAI announced they struck a deal with the DoW to step in as Anthropic's replacement.

OpenAI claims the deal it has made with the DoW includes the same redlines Anthropic insisted on, and is actually more enforcable than Anthropic's old contract, because it relies on cloud-only deployments with cleared OpenAI personnel "in the loop". However, they also directly shared portions of the contract's language and, while I am not a lawyer, it feels like the wording leaves huge, gaping holes that pretty much exactly match Anthropic's claims about the contract they rejected.

No idea how this whole thing will play out, and there must be more to this story than what's circulating on twitter, but at this point, the whole thing feels like an extremely stupid own goal for both the DoW and OpenAI.

Deepmind

Nano Banana 2

While Nano Banana Pro has been by far the best image generation model for a bit now, it was always pretty slow and expensive. Nano Banana 2, then, is positioned as a faster, cheaper alternative that produces similar quality images that, ideally, pop a bit more than the original Nano Banana (which tended to output realistic, but sometimes boring, pictures). In my testing, it's still clearly worse than Nano Banana Pro at most tasks, and the "more lively" images end up looking cheaper than I would like. That being said, it is better at character consistency and supports Search Grounding, which feels like an underdiscussed addition. Overall, it feels like a very good baseline model that excel at 95% of requests, with Nano Banana Pro handling the remainder.

Gemini 3.1 Pro

After OpenAI's and Anthropic's model launches over the last few weeks, I've been waiting on Google to pull back up. Two weeks ago, we got an updated Gemini 3 Deep Think, which was a bit of a teaser for this drop - Gemini 3.1 Pro.

It pulls way ahead in ARC-AGI2 (33% -> 77% from Gemini 3 to 3.1!), and is basically SOTA in every benchmark but SWE-Bench. That being said, benchmarks are worth less and less these days, so let's see if they actually fixed the hallucination issues that effectively killed Gemini 3 for agentic coding the moment Opus 4.5 became available.

Lyria 3

Here's something I didn't have on my bingo card (but maybe should have) - Google has dropped a new music model. And it's good!

I haven't played around with it enough to compare it to Suno (their V5 model is excellent), but given that this will be available for free to all Gemini users, they might have a hard time competing.

Anthropic

Opus 3 Retirement

In contrast to the GPT-4o drama OpenAI has been facing over the last couple.. months?, Anthropic seems to be trying their best to ensure Opus 3 gets the retirement it deserves. For one, the model will continue being available for paid users for the time being. Secondly, they have started a separate access program for researchers that want to keep playing with it. Finally, and this one is a bit weird, they gave it a substack?

For the next couple months, Opus 3 will get to share it's thoughts with the world. I love these kinds of experiments, so I am delighted to see this, though I am curious about exactly how much of these blogs Opus is writing on it's volition (vs. just being prompted to do so).

OpenAI

gpt-realtime-1.5

Very minor update to the realtime audio model, but most use cases should run a bit smoother now. What perplexes me is that we haven't seen any sort of agentic/long-running audio-native tasks/deployments that I am aware of, nor have we seen the models getting much better since GPT-4o Advanced Voice Mode dropped. I wonder why that is? Could just be latency constraints?

WebSockets in Responses API

Small protocol update with potentially large consequences - the Responses API supports WebSockets now. In practice, that means much faster tool calls (up to 40%, apparently) for long-running agents at the cost of a slight delay during initial agent start.

Block: Workforce Reduction

There's a new data point in the ongoing debate of when exactly all software engineers will lose their jobs - Block (ex Square, owning CashApp/Tidal/...) has fired over 40% of their employees, with a focus on cutting developers. CEO Jack Dorsey (who also is the original founder of twitter, btw) has promised the company will provide a very generous benefits package for people affected (20 weeks salary + 1 week per year of tenure, equity vested through the end of may, 6 months of health care, corporate devices, and $5,000), so this is a pretty sharp contrast to the Twitter cuts Elon made (that many people are comparing this to). There is also an argument that this is mostly a reckoning from the COVID hiring spree many tech companies went on, so maybe there is limited generalization.

However you spin it, though, 4000 employees lost their jobs, and Block's stock is up >20%.