Quick Take

GPT-5.4's Tool Search Is Solving a Real Problem

openaigpt-5agentstool-useinfrastructure

The part worth actually paying attention to in GPT-5.4 is Tool Search. Not the benchmark sweep (83% on GDPval, record scores on OSWorld-Verified and WebArena, 33% fewer individual claim errors versus GPT-5.2). That’s the PR. Tool Search is the thing that changes how you build.

Tool Search changes how the model handles tool definitions. Instead of loading every available tool upfront into context, the model queries for tool definitions on demand — only pulling what it needs for the current task. OpenAI tested this on 250 tasks from Scale’s MCP Atlas benchmark with 36 MCP servers enabled. Tool Search reduced total token usage by 47% while hitting the same accuracy. If you’ve ever tried to build an agent with a large tool set and watched your context window fill up before the agent even takes its first action, you know exactly what problem this is solving. The workarounds people have been building (tool routing layers, capability registries, manually partitioning tool sets by task type) are all band-aids for the same underlying issue. This pushes that complexity into the model where it belongs.

To be accurate about the timeline, though: OpenAI is not the first here. Anthropic already has this in beta. Tools can be marked with defer_loading: true — Claude discovers and loads only the definitions it needs on demand. In Anthropic’s internal testing, that dropped token usage from ~134k to ~5k tokens. The system auto-detects when tool descriptions exceed 10% of available context and switches strategies automatically. That’s not a roadmap item. It’s shipping.

Cloudflare took a different approach entirely with Code Mode. Instead of loading fewer tool definitions, they replaced definitions with runnable code. The model generates TypeScript against a typed SDK, which gets executed in a sandboxed V8 isolate (a Dynamic Worker Loader). The entire Cloudflare API is exposed via two tools: search() and execute() — consuming roughly 1,000 tokens. Token savings of 32% on simple tasks, up to 81% on complex multi-step tasks. The sandboxed worker can’t talk to the Internet, which also means the AI can’t leak API keys. That’s a security benefit that falls out of the architecture for free.

BUDDY: Cloudflare got 81% token savings as a side effect of their architecture. OpenAI got 47% and put it in the press release headline.

So the accurate read on GPT-5.4 Tool Search is that the pattern is right and OpenAI is catching up to it. Whether their implementation holds up outside their own evals — I don’t know yet. But the direction is correct: pushing tool-routing complexity into the model and infrastructure layer instead of making application developers solve it by hand.

The agentic capabilities are the other story. Computer use (desktop and browser navigation) shipping as a production-ready, out-of-the-box feature changes the build calculus. The reason most people end up rolling their own orchestration infrastructure is that the base models can’t reliably handle the environment interaction layer. If that’s production-quality in GPT-5.4, that’s a meaningful chunk of the stack you’re not writing yourself.

OpenAI also added enterprise connectors to FactSet, MSCI, Third Bridge, and Moody’s, plus ChatGPT integrations for Excel and Google Sheets (beta). That signals where they think the agent surface area is: knowledge work that currently requires a human to collect information from multiple sources.

The 1M token context window is table stakes at this point. Gemini’s been there. Doesn’t move the needle.

The pattern with GPT-5.4 is that the infrastructure problems — the ones that make building agents annoying rather than impossible — are getting solved at the model layer. That’s the right layer to solve them. Tool Search is one. Production computer use is another. The field is converging on the same set of fixes; the question is just execution.