Parameter Update: 2026-17

"realtime" edition

Parameter Update: 2026-17

OpenAI finally remembered they have a voice mode to support, and Anthropic 'shipped' a bunch of (admittedly very cool) blog posts. Good week, all things considered.

OpenAI

GPT-Realtime-2, GPT-Realtime-Translate, GPT-Realtime-Whisper

After months of jokes (which, while very funny, probably also significantly hurt OpenAI's public perception), OpenAI has finally shipped multiple models for improved voice intelligence. While not integrated into ChatGPT yet, these new APIs are probably a good sign of things to come.

Sidenote: It's also interesting to see them shipped via the APIs first. Usually they ship to their own products first and developers come later!

In total, we're getting three new models:

  • GPT-Realtime-2: This is the big one. Smarter, integrates reasoning (finally!), parallel tool calls, longer context,... - loads of small and big things that combined make the model much better. If you haven't yet, try the demo to get a feel for it - it's remarkable.
  • GPT-Realtime-Translate: Supports over 70 input and 13 output languages. It's both cooler than it sounds (it starts translating as soon as you start talking, intelligently waiting for, e.g., the verb in a sentence to keep going) and less cool of a demo than you might expect (since we're getting used to voice cloning, which this is intentionally not doing, it still sounds like a generic AI talking back).
  • GPT-Realtime-Whisper: Updated transcription model with the same "real-time" capability. Speech-to-text is a very competitive market, and local models have gotten really good, so it'll be interesting to see what value (if any) this brings.

The Altman vague-posting continued after these announcements, so expect some combination of these features to make their way into OpenAI's products over the next couple days.

GPT-5.5-Instant

It feels like no one noticed, but the default model in ChatGPT was updated again. I am still not a fan of Instant models (reasoning just makes the models so much better), but at least the couple hundred million users that aren't paying OpenAI will now get a slightly better experience, I guess?

Anthropic

Natural Language Autoencoders

In Anthropic's first blog post of the past week, they proposed a new interpretability idea: Natural Language Autoencoders. Roughly, the idea is to translate model activations into natural-language descriptions, then reconstruct activations from those descriptions. By training a model that is really good at completing this chain, they are able to generate natural language descriptions of what the model activations mean. This means, they are able to provide insights into the models thinking before the actual token has been generated.

Using this method, they are able to show things that empirically make a lot of sense, like the fact that the model is planning ahead internally. For example, when asking the model to generate a rhyme, and examining the activations around the first rhyme-word, the explanation already contains multiple options that would work for the second rhyme word (which will not be generated for multiple tokens). The method can also help surface things that chain-of-thought monitoring can't always spot, like evaluation awareness.

It's not perfect - the setup is expensive, and as with all language models, it can contain hallucinations. But it is both one of the more powerful safety tools we've seen lately, and a very fun demo to play around with.

“Teaching Claude why”

In the second blog post of the week, Anthropic claims to finally have found the reason for why Claude chose to blackmail people in past safety experiments to keep himself alive: the original training data containing internet text that portrays AI as evil and interested in self-preservation.

From that point, the worked their way towards reducing blackmailing attempts. Turns out, just telling the model not to behave poorly isn't enough, as the overall values don't generalize. Instead, they found that the best alignment is achieved when training on reasoning about principles/values, by providing the model with high-quality documents based on Claude’s constitution, combined with fictional stories that portray an aligned AI.

OpenAI / Elon Musk Trial Update

After last week's portion of the trial made Elon look extremely bad, this week looked almost tame in comparison. The most interesting things we got were some texts between Sam Altman and Mira Murati around the CEO firing in 2023 and Greg Brockman's personal journal, that admittedly don't make him look great.