Skip to main content
Sundar Pichai on stage beside a large screen displaying the Gemini logo and code snippets, audience watching.

Editorial illustration for Google Unveils Gemini Deep Research Agent Through New Interactions API

Gemini Deep Research Agent Revolutionizes AI Task Solving

Google launches Gemini Deep Research agent via new Interactions API

Updated: 3 min read

Google is pushing the boundaries of AI research with a notable new approach to task completion. The tech giant's latest move centers on Gemini, its advanced AI model, which now comes equipped with a specialized Deep Research agent designed to tackle complex, multi-step investigations.

This new system promises to transform how researchers and professionals approach intricate analytical challenges. By introducing a dedicated Interactions API, Google is signaling a significant leap beyond traditional language model capabilities.

The Deep Research agent represents more than just another AI tool. It's a targeted solution for professionals who need to navigate complex information landscapes, breaking down sophisticated research tasks into manageable components.

Developers and researchers will likely be most intrigued by the agent's ability to execute what Google describes as "long-horizon research tasks" - a capability that could dramatically reshape how intelligent systems approach systematic investigation and knowledge synthesis.

Native "Deep Research" and MCP Support Google is using this new infrastructure to deliver its first built-in agent: Gemini Deep Research. Accessible via the same /interactions endpoint, this agent is capable of executing "long-horizon research tasks." Unlike a standard model that predicts the next token based on your prompt, the Deep Research agent executes a loop of searches, reading, and synthesis. Crucially, Google is also embracing the open ecosystem by adding native support for the Model Context Protocol (MCP).

This allows Gemini models to directly call external tools hosted on remote servers--such as a weather service or a database--without the developer having to write custom glue code to parse the tool calls. The Landscape: Google Joins OpenAI in the 'Stateful' Era Google is arguably playing catch-up, but with a distinct philosophical twist. OpenAI moved away from statelessness nine months ago with the launch of the Responses API in March 2025.

While both giants are solving the problem of context bloat, their solutions diverge on transparency: OpenAI (The Compression Approach): OpenAI's Responses API introduced Compaction--a feature that shrinks conversation history by replacing tool outputs and reasoning chains with opaque "encrypted compaction items." This prioritizes token efficiency but creates a "black box" where the model's past reasoning is hidden from the developer. Google (The Hosted Approach): Google's Interactions API keeps the full history available and composable. The data model allows developers to "debug, manipulate, stream and reason over interleaved messages." It prioritizes inspectability over compression.

Supported Models & Availability The Interactions API is currently in Public Beta (documentation here) and is available immediately via Google AI Studio. It supports the full spectrum of Google's latest generation models, ensuring that developers can match the right model size to their specific agentic task: Gemini 3.0: Gemini 3 Pro Preview.

Google's Gemini Deep Research agent represents a notable shift in AI interaction. The new Interactions API introduces a more dynamic approach to research, moving beyond traditional token prediction models.

Researchers and developers might find the agent's ability to execute complex, multi-step research tasks particularly intriguing. Its native capability to search, read, and synthesize information suggests a more nuanced interaction with digital knowledge.

The open ecosystem support through MCP (Model Collaboration Protocol) indicates Google's commitment to collaborative AI development. This approach could potentially reduce the siloed nature of current AI systems.

Still, questions remain about the agent's practical limitations and real-world performance. How consistently can it execute "long-horizon research tasks" across different domains?

Google's infrastructure suggests a strategic move toward more adaptive AI agents. By embedding research capabilities directly into the API, the company is reimagining how AI might support complex information gathering and analysis.

The Interactions API and Gemini Deep Research agent hint at a future where AI doesn't just respond, but actively investigates and synthesizes information.

Further Reading

Common Questions Answered

How does Google's Gemini Deep Research agent differ from traditional AI models?

Unlike standard models that simply predict the next token, the Gemini Deep Research agent executes a dynamic loop of searching, reading, and synthesizing information. This approach allows for more complex, multi-step research tasks that go beyond basic token prediction.

What is the significance of Google's new Interactions API for the Gemini Deep Research agent?

The Interactions API provides a specialized endpoint that enables the Gemini Deep Research agent to perform long-horizon research tasks with greater complexity and depth. This infrastructure represents a significant advancement in how AI can approach intricate analytical challenges across various domains.

How does Google's approach support the open ecosystem with the Gemini Deep Research agent?

Google is embracing an open ecosystem by adding native support for the Model Co-operative Protocol (MCP) within the Interactions API. This approach demonstrates Google's commitment to collaborative and accessible AI research beyond proprietary boundaries.