Be part of the occasion trusted by enterprise leaders for almost 20 years. VB Rework brings collectively the individuals constructing actual enterprise AI technique. Study extra
At VentureBeat’s Rework 2025 convention, Olivier Godement, Head of Product for OpenAI’s API platform, supplied a behind-the-scenes take a look at how enterprise groups are adopting and deploying AI brokers at scale.
In a 20-minute panel dialogue I hosted completely with Godement, the previous Stripe researcher and present OpenAI API boss unpacked OpenAI’s newest developer instruments—the Responses API and Brokers SDK—whereas highlighting real-world patterns, safety concerns, and cost-return examples from early adopters like Stripe and Field.
For enterprise leaders unable to attend the session dwell, listed below are high 8 most essential takeaways:
Brokers Are Quickly Shifting From Prototype to Manufacturing
In response to Godement, 2025 marks an actual shift in how AI is being deployed at scale. With over one million month-to-month lively builders now utilizing OpenAI’s API platform globally, and token utilization up 700% 12 months over 12 months, AI is shifting past experimentation.
“It’s been 5 years since we launched basically GPT-3… and man, the previous 5 years has been fairly wild.”
Godement emphasised that present demand isn’t nearly chatbots anymore. “AI use instances are shifting from easy Q&A to truly use instances the place the applying, the agent, can do stuff for you.”
This shift prompted OpenAI to launch two main developer-facing instruments in March: the Responses API and the Brokers SDK.
When to Use Single Brokers vs. Sub-Agent Architectures
A serious theme was architectural alternative. Godement famous that single-agent loops, which encapsulate full instrument entry and context in a single mannequin, are conceptually elegant however usually impractical at scale.
“Constructing correct and dependable single brokers is difficult. Like, it’s actually exhausting.”
As complexity will increase—extra instruments, extra doable consumer inputs, extra logic—groups usually transfer towards modular architectures with specialised sub-agents.
“A apply which has emerged is to basically break down the brokers into a number of sub-agents… You’ll do separation of issues like in software program.”
These sub-agents operate like roles in a small workforce: a triage agent classifies intent, tier-one brokers deal with routine points, and others escalate or resolve edge instances.
Why the Responses API Is a Step Change
Godement positioned the Responses API as a foundational evolution in developer tooling. Beforehand, builders manually orchestrated sequences of mannequin calls. Now, that orchestration is dealt with internally.
“The Responses API might be the largest new layer of abstraction we launched since just about GPT-3.”
It permits builders to specific intent, not simply configure mannequin flows. “You care about returning a extremely good response to the shopper… the Response API basically handles that loop.”
It additionally contains built-in capabilities for data retrieval, internet search, and performance calling—instruments that enterprises want for real-world agent workflows.
Observability and Safety Are Constructed In
Safety and compliance have been high of thoughts. Godement cited key guardrails that make OpenAI’s stack viable for regulated sectors like finance and healthcare:
- Coverage-based refusals
- SOC-2 logging
- Information residency help
Analysis is the place Godement sees the largest hole between demo and manufacturing.
“My sizzling take is that mannequin analysis might be the largest bottleneck to large AI adoption.”
OpenAI now contains tracing and eval instruments with the API stack to assist groups outline what success appears like and monitor how brokers carry out over time.
“Until you spend money on analysis… it’s actually exhausting to construct that belief, that confidence that the mannequin is being correct, dependable.”
Early ROI Is Seen in Particular Features
Some enterprise use instances are already delivering measurable features. Godement shared examples from:
- Stripe, which makes use of brokers to speed up bill dealing with, reporting “35% quicker bill decision”
- Field, which launched data assistants that allow “zero-touch ticket triage”
Different high-value use instances embody buyer help (together with voice), inner governance, and data assistants for navigating dense documentation.
What It Takes to Launch in Manufacturing
Godement emphasised the human consider profitable deployments.
“There’s a small fraction of very high-end individuals who, each time they see an issue and see a expertise, they run at it.”
These inner champions don’t all the time come from engineering. What unites them is persistence.
“Their first response is, OK, how can I make it work?”
OpenAI sees many preliminary deployments pushed by this group — individuals who pushed early ChatGPT use within the enterprise and at the moment are experimenting with full agent programs.
He additionally identified a spot many overlook: area experience. “The data in an enterprise… doesn’t lie with engineers. It lies with the ops groups.”
Making agent-building instruments accessible to non-developers is a problem OpenAI goals to handle.
What’s Subsequent for Enterprise Brokers
Godement supplied a glimpse into the roadmap. OpenAI is actively engaged on:
- Multimodal brokers that may work together through textual content, voice, pictures, and structured knowledge
- Lengthy-term reminiscence for retaining data throughout classes
- Cross-cloud orchestration to help complicated, distributed IT environments
These aren’t radical modifications, however iterative layers that develop what’s already doable. “As soon as now we have fashions that may assume not just for just a few seconds however for minutes, for hours… that’s going to allow some fairly mind-blowing use instances.”
Closing Phrase: Reasoning Fashions Are Underhyped
Godement closed the session by reaffirming his perception that reasoning-capable fashions—these that may mirror earlier than responding—would be the true enablers of long-term transformation.
“I nonetheless have conviction that we’re just about on the GPT-2 or GPT-3 stage of maturity of these fashions….We’re nonetheless scratching the floor on what reasoning fashions can do.”
For enterprise resolution makers, the message is obvious: the infrastructure for agentic automation is right here. What issues now’s constructing a targeted use case, empowering cross-functional groups, and being able to iterate. The subsequent section of worth creation lies not in novel demos—however in sturdy programs, formed by real-world wants and the operational self-discipline to make them dependable.
Keep forward of the curve with Enterprise Digital 24. Discover extra tales, subscribe to our publication, and be a part of our rising group at bdigit24.com