← Back to OpenAI updates ← Terug naar OpenAI-updates
OpenAI ARTICLE ARTIKEL 21 May 2025 21 mei 2025

New tools and features in the Responses API New tools and features in the Responses API

Loading… Loading…

Article details Artikelgegevens
AI maker AI-maker OpenAI Type Type Article Artikel Published Gepubliceerd 21 May 2025 21 mei 2025 Updates Updates Videos Video's View original article Bekijk origineel artikel
Why it matters Waarom dit telt

Quick editorial signal Snelle redactionele duiding

7 min
Impact Impact

Worth checking before choosing or changing a subscription. Handig om te checken voordat je een abonnement kiest of wijzigt.

Audience Voor wie Developers Developers
Level Niveau Expert Expert
  • Track this as a OpenAI update, not just a standalone headline. Bekijk dit als OpenAI-update, niet alleen als losse headline.
  • Check plan details before changing subscriptions or advising a team. Controleer plandetails voordat je abonnementen wijzigt of een team adviseert.
  • Likely worth revisiting after people have used the release in practice. Waarschijnlijk de moeite waard om opnieuw te bekijken zodra mensen het in praktijk gebruiken.
model apps video pricing

Today, we’re adding new built-in tools to the Responses API—our core API primitive for building agentic applications. This includes support for all remote Model Context Protocol (MCP) servers⁠(opens in a new window), as well as tools like image generation⁠(opens in a new window), Code Interpreter⁠(opens in a new window), and improvements to file search⁠(opens in a new window). These tools are available across our GPT‑4o series, GPT‑4.1 series, and OpenAI o-series reasoning models. o3 and o4-mini can now call tools and functions directly within their chain-of-thought in the Responses API, producing answers that are more contextually rich and relevant. Using o3 and o4-mini with the Responses API preserves reasoning tokens across requests and tool calls, improving model intelligence and reducing the cost and latency for developers.

We’re also introducing new features in the Responses API that improve reliability, visibility, and privacy for enterprises and developers. These include background mode⁠(opens in a new window) to handle long-running tasks asynchronously and more reliably, support for reasoning summaries⁠(opens in a new window), and support for encrypted reasoning items⁠(opens in a new window).

Since releasing the Responses API in March 2025 with tools like web search, file search, and computer use, hundreds of thousands of developers have used the API to process trillions of tokens across our models. Customers have used the API to build a variety of agentic applications, including Zencoder⁠(opens in a new window)’s coding agent, Revi⁠(opens in a new window)’s market intelligence agent for private equity and investment banking, and MagicSchool AI⁠(opens in a new window)'s education assistant—all of which use web search to pull relevant, up-to-date information into their app. Now developers can build agents that are even more useful and reliable with access to the new tools and features released today.

New remote MCP server support

We’re adding support for remote MCP servers⁠(opens in a new window) in the Responses API, building on the release of MCP support in the Agents SDK⁠(opens in a new window). MCP is an open protocol that standardizes how applications provide context to LLMs. By supporting MCP servers in the Responses API, developers will be able to connect our models to tools hosted on any MCP server with just a few lines of code. Here are some examples showing how developers can use remote MCP servers with the Responses API today:

Shopify Twilio Stripe DeepWiki (Devin)

Python

1response = client.responses.create(2 model="gpt-4.1",3 tools=[{4 "type": "mcp",5 "server_label": "shopify",6 "server_url": "https://pitchskin.com/api/mcp",7 }],8 input="Add the Blemish Toner Pads to my cart"9)

The Blemish Toner Pads have been added to your cart! You can proceed to checkout here:

Popular remote MCP servers include Cloudflare⁠(opens in a new window), HubSpot⁠(opens in a new window), Intercom⁠(opens in a new window), PayPal⁠(opens in a new window), Plaid⁠(opens in a new window), Shopify⁠(opens in a new window), Stripe⁠(opens in a new window), Square⁠(opens in a new window), Twilio⁠(opens in a new window), Zapier⁠(opens in a new window), and more. We expect the ecosystem of remote MCP servers to grow quickly in the coming months, making it easier for developers to build powerful agents that can connect to the tools and data sources their users already rely on. In order to best support the ecosystem and contribute to this developing standard, OpenAI has also joined the steering committee for MCP.

To learn how to spin up your own remote MCP server, check out this guide from Cloudflare⁠(opens in a new window). To learn how to use the MCP tool in the Responses API, check out this guide⁠(opens in a new window) in our API Cookbook.

Updates to image generation, Code Interpreter, and file search

With built-in tools in the Responses API, developers can easily create more capable agents with just a single API call. By calling multiple tools while reasoning, models now achieve significantly higher tool calling performance on industry-standard benchmarks like Humanity’s Last Exam (source). Today, we’re adding new tools including:

* Image generation: In addition to using the Images API⁠(opens in a new window), developers can now access our latest image generation model—gpt-image-1—as a tool within the Responses API. This tool supports real-time streaming—allowing developers to see previews of the image as it’s being generated—and multi-turn edits—allowing developers to prompt the model to granularly refine these images step-by-step. Learn more⁠(opens in a new window).

* Code Interpreter:Developers can now use the Code Interpreter⁠(opens in a new window) tool within the Responses API. This tool is useful for data analysis, solving complex math and coding problems, and helping the models deeply understand and manipulate images (e.g., thinking with images). The ability for models like o3 and o4-mini to use the Code Interpreter tool within their chain-of-thought has resulted in improved performance across several benchmarks including Humanity’s Last Exam (source). Learn more⁠(opens in a new window).

* File search:Developers can now access the file search⁠(opens in a new window) tool in our reasoning models. File search enables developers to pull relevant chunks of their documents into the model’s context based on the user query. We’re also introducing updates to the file search tool that allow developers to perform searches across multiple vector stores and support attribute filtering with arrays. Learn more⁠(opens in a new window).

New features in the Responses API

In addition to the new tools, we’re also adding support for new features in the Responses API, including:

* Background mode:As seen in agentic products like Codex, deep research, and Operator, reasoning models can take several minutes to solve complex problems. Developers can now use background mode to build similar experiences on models like o3 without worrying about timeouts or other connectivity issues—background mode kicks off these tasks asynchronously. Developers can either poll these objects to check for completion, or start streaming events whenever their application needs to catch up on the latest state. Learn more⁠(opens in a new window).

1response = client.responses.create(2 model="o3",3 input="Write me an extremely long story.",4 reasoning={ "effort": "high" },5 background=True6)

1response = client.responses.create(2 model="o3",3 input="Write me an extremely long story.",4 reasoning={ "effort": "high" },5 background=True6)

1response = client.responses.create(2 model="o4-mini",3 tools=[4 {5 "type": "code_interpreter",6 "container": {"type": "auto"}7 }8 ],9 instructions=(10 "You are a personal math tutor. "11 "When asked a math question, run code to answer the question."12 ),13 input="I need to solve the equation 3x + 11 = 14. Can you help me?",14 reasoning={"summary": "auto"}15)

Python

1response = client.responses.create(2 model="o3",3 input="Implement a simple web server in Rust from scratch.",4 store=False,5 include=["reasoning.encrypted_content"]6)

Pricing and availability

All of these tools and features are now available in the Responses API, supported in our GPT‑4o series, GPT‑4.1 series, and our OpenAI o-series reasoning models (o1, o3, o3‑mini, and o4-mini). Image generation is only supported on o3 of our reasoning model series.

Pricing for existing tools remains the same. Image generation costs $5.00/1M text input tokens, $10.00 / 1M image input tokens, and $40.00 / 1M image output tokens, with 75% off cached input tokens. Code Interpreter costs $0.03 per container. File search costs $0.10/GB of vector storage per day and $2.50/1k tool calls. There is no additional cost to call the remote MCP server tool—you are simply billed for output tokens from the API. Learn more about pricing⁠(opens in a new window) in our docs.

We’re excited to see what you build!

All of these tools and features are now available in the Responses API, supported in our GPT‑4o series, GPT‑4.1 series, and our OpenAI o-series reasoning models (o1, o3, o3‑mini, and o4-mini). Image generation is only supported on o3 of our reasoning model series.

Pricing for existing tools remains the same. Image generation costs $5.00/1M text input tokens, $10.00 / 1M image input tokens, and $40.00 / 1M image output tokens, with 75% off cached input tokens. Code Interpreter costs $0.03 per container. File search costs $0.10/GB of vector storage per day and $2.50/1k tool calls. There is no additional cost to call the remote MCP server tool—you are simply billed for output tokens from the API. Learn more about pricing⁠(opens in a new window) in our docs.

Author

OpenAI

* API Platform

Author

OpenAI

Help shape what we cover next Help bepalen wat we hierna volgen

Anonymous feedback, no frontend account needed. Anonieme feedback, zonder front-end account.

Watch related videos Bekijk gerelateerde video's

Open videos → Open video's →

More from OpenAI Meer van OpenAI

All updates Alle updates

Gemini komt eraan