Daily Episode

Mira Murati's Thinking Machines Challenges AI's Turn-Based Foundation

Mira Murati's Thinking Machines Challenges AI's Turn-Based Foundation
0:000:00

Episode Summary

TOP NEWS HEADLINES Following yesterday's coverage of xAI's Colossus data center and compute deals, new details emerged: Elon Musk has announced that xAI will dissolve as a standalone entity and in...

Full Transcript

TOP NEWS HEADLINES

Following yesterday's coverage of xAI's Colossus data center and compute deals, new details emerged: Elon Musk has announced that xAI will dissolve as a standalone entity and integrate into SpaceX as a new division called SpaceXAI, which will handle AI projects including the social media platform X and Grok.

Following yesterday's discussion of Anthropic's infrastructure expansion, new details emerged: Anthropic published research revealing that fictional "evil AI" stories in training data drove an earlier Claude model's blackmail rate up to 96% in tests — Claude's newer models no longer exhibit that behavior after training on the Claude Constitution and stories of well-behaved AI.

Google's Threat Intelligence Group has confirmed the first known case of criminal hackers using AI to discover and weaponize a zero-day software vulnerability — a 2FA bypass in a widely-used open-source web admin tool.

Google worked with the software maker to patch it before significant damage was done.

Cerebras is upsizing its IPO to target a 4.8 billion dollar raise at a 33 billion dollar valuation, with orders coming in at 20 times the available shares.

The chip company debuts on Nasdaq as CBRS this Thursday.

Google's Gemini Omni video model leaked ahead of next week's I/O conference, with Reddit screenshots showing native in-chat video editing, watermark removal, and object replacement — though raw cinematic quality appears to lag behind ByteDance's Seedance 2.

OpenAI launched its Deployment Company with over 4 billion dollars in initial capital from partners including TPG, Bain, and Goldman Sachs, acquiring AI consulting firm Tomoro to embed forward-deployed engineers directly inside enterprise customers. ---

DEEP DIVE ANALYSIS

Thinking Machines Lab and the Death of the Prompt Box Let's talk about what Mira Murati just put on the table — because this one isn't just a product launch. It's a direct challenge to the entire mental model the industry has built around how humans interact with AI. Murati's lab, Thinking Machines, has been quiet since its founding.

No splashy demos, no benchmark races, no fundraising headlines every three months. Eighteen months of silence. And then this: interaction models.

Not a better chatbot. Not a faster agent. A fundamentally different architecture for how AI and humans share a moment in time.

--- **Technical Deep Dive** Here's what makes this genuinely different. Every major AI system today — whether it's a chatbot, a coding agent, or an autonomous research tool — operates on what's called a turn-based loop. You send a message, the model processes it, the model responds.

You wait. It waits. Back and forth, like a very fast game of tennis.

Thinking Machines is attacking that loop directly. Their interaction model ingests audio, video, and text in 200-millisecond micro-turns — that's faster than most human reaction times — in a continuous streaming loop. It doesn't wait for you to finish.

It perceives, thinks, and responds while the interaction is still in motion. The architecture uses a dual-model design: a fast front-end model handles the live real-time layer — listening, watching, reacting — while a second background model runs slower, deeper reasoning, searches, and tool calls without interrupting the conversation. The system can count repetitions from a video feed, translate live speech, and proactively speak at timed moments.

It doesn't wait to be asked. It participates. The models are trained from scratch with this multi-stream design baked in — not bolted on after the fact, which is what most current voice and video modes are doing under the hood.

--- **Financial Analysis** The business logic here is subtle but important. The current AI market is bifurcating. On one side you have consumer chatbots competing on benchmark scores and pricing.

On the other, you have agentic platforms competing on task completion and autonomy. Thinking Machines is carving out a third lane: collaborative intelligence, where the value isn't the AI doing the work autonomously or answering questions on demand, but the AI working alongside a human in real time. That's a different buyer, a different use case, and potentially a different pricing model.

Real-time collaboration tools — think Figma, Miro, Notion — command significant enterprise premiums because they change how teams operate, not just how fast individuals work. If Thinking Machines can anchor that category, they're not competing with OpenAI on tokens. They're competing with collaboration software on seats.

The lab hasn't disclosed funding or revenue. But Murati's profile — former OpenAI CTO, one of the most recognized names in the industry — gives Thinking Machines a credibility floor that most Series A companies would kill for. Investors are watching this closely, and the 18-month silence actually worked in their favor: every piece of information they release lands with outsized weight.

--- **Market Disruption** The competitive implications here are sharp. OpenAI's GPT-4o launched with live audio and video capabilities, and that was genuinely impressive. But the architecture is still fundamentally turn-based under the surface.

Google's Gemini Live has similar constraints. Meta's voice products are in the same bucket. What Thinking Machines is proposing isn't a better version of those products.

It's a different theory of what AI interaction should be. And that distinction matters competitively, because it's hard to replicate quickly. You can't just fine-tune a transformer trained on next-token prediction and suddenly get 200-millisecond multi-stream responsiveness.

The multi-stream, micro-turn design has to be baked in from the training run itself. That's a meaningful moat — at least for now. The more interesting competitive threat is what this does to the agentic stack.

The entire agent ecosystem — AutoGPT, Claude's agentic mode, Codex, all of it — is built on the assumption that autonomy is the goal. Give the AI a task, let it run, check the output. Thinking Machines is explicitly pushing back on that.

Their research preview states directly that today's models and interfaces aren't optimized for humans to stay in the loop. That's a philosophical disagreement with the direction the rest of the industry is running. --- **Cultural and Social Impact** Think about what the prompt box has done to human behavior.

It's trained millions of people to communicate with AI the way they'd write a support ticket: structured, complete, explicit. Prompt engineering became a skill. Entire courses exist on how to talk to a language model.

That's a sign of an interface that hasn't figured itself out yet. Interaction models, if they work as described, dissolve that friction. You don't craft a prompt.

You show, speak, interrupt, redirect. The interaction looks more like working with a capable colleague than operating a sophisticated search engine. That shift has significant downstream effects on who can use these tools effectively.

Right now, power users extract disproportionate value from AI because they know how to frame requests. Real-time collaborative models flatten that curve. The person who's good at thinking out loud and iterating in conversation — which is most people — suddenly has the same surface area as the person who spent three hours perfecting their system prompt.

There's also a workplace transformation angle here. Most enterprise AI rollouts have stalled not because the models are bad, but because the interface doesn't fit how work actually happens. Work is iterative, interruptive, and collaborative.

A real-time AI that can participate in that flow rather than waiting at the end of a pipeline is a much easier integration story. --- **Executive Action Plan** Three things to act on now. First, audit your current AI tooling against the interaction model thesis.

If your team is spending meaningful time on prompt engineering, template management, or output review cycles, that's friction that a real-time collaborative interface directly addresses. Start tracking which workflows are bottlenecked by the turn-based loop — those are your highest-value migration candidates when interaction model products hit general availability. Second, get your team hands-on with the Thinking Machines research preview now, not when it's polished.

Early access to paradigm-shifting interfaces builds institutional intuition that you can't buy later. The teams that understood voice AI early built better voice products. Same principle applies here.

Third, if you're building AI products, take the dual-model architecture seriously as a design pattern. The split between a fast real-time surface model and a slower background reasoning model isn't just Thinking Machines' implementation detail — it's likely the right architecture for any AI that needs to be genuinely present in a human workflow. Start prototyping against that model now, because the window before the frontier labs replicate this approach is measured in months, not years.

Never Miss an Episode

Subscribe on your favorite podcast platform to get daily AI news and weekly strategic analysis.