Amazon Bedrock AgentCore now supports stateful MCP with user input, LLM sampling, and progress streaming
Amazon has introduced stateful MCP client capabilities on Bedrock AgentCore Runtime, enabling agents to pause mid-execution for user input, request LLM-generated content, and stream real-time progress updates. The update transforms one-way tool execution into bidirectional conversations between MCP servers and clients, supporting interactive workflows previously impossible with stateless implementations.
Amazon Bedrock AgentCore Now Supports Stateful MCP with User Input and LLM Sampling
Amazon has introduced stateful MCP (Model Context Protocol) client capabilities on Bedrock AgentCore Runtime, enabling interactive, multi-turn agent workflows that require mid-execution pauses for user input and real-time feedback.
The update addresses a core limitation of previous stateless MCP implementations: agents couldn't pause execution to request clarification from users, generate content via LLM, or provide progress updates during long-running operations. Stateful mode removes this constraint by maintaining session continuity and enabling three new client capabilities.
The Three New Client Capabilities
Elicitation allows servers to pause execution and request structured user input mid-workflow. Servers send an elicitation/create request with a message and optional JSON schema describing expected response format. The client renders an input interface, and users can accept (providing data), decline, or cancel. Two modes are supported: form mode for direct data collection and URL mode for sensitive interactions like OAuth flows or payment processing.
Sampling enables servers to request LLM-generated completions from the client via sampling/createMessage. This allows tool logic on the server to use language model capabilities without maintaining its own model credentials. Practical uses include generating personalized summaries, creating natural-language explanations of structured data, or producing recommendations based on conversation context.
Progress notifications allow servers to report incremental progress during long-running operations using ctx.report_progress(progress, total). Clients can display updates as progress bars or status indicators, keeping users informed instead of watching blank screens during multi-step operations like cross-source data searches.
Technical Implementation
Enabling stateful mode requires a single configuration flag: setting stateless_http=False in server startup. When activated, AgentCore Runtime provisions a dedicated microVM for each user session with isolated CPU, memory, and filesystem. Sessions persist for up to 8 hours or until 15 minutes of inactivity (configurable via idleRuntimeSessionTimeout).
The protocol maintains continuity through an Mcp-Session-Id header. The server returns this identifier during initialization handshake; clients include it in every subsequent request to route back to the same session. All three capabilities are opt-in at the client level—clients declare supported capabilities during initialization, and servers must only use advertised capabilities.
Stateful mode contrasts with the previous stateless implementation, where each HTTP request was independent with no shared context between calls. Stateless mode remains suitable for simple tool serving, while stateful mode supports interactive, multi-turn workflows.
What This Means
This release completes MCP's bidirectional protocol implementation on AWS's managed agent platform. Teams can now build more sophisticated agents that interact naturally with users and leverage LLM capabilities without creating separate model connections. The session-based architecture with configurable timeouts provides flexibility for both short interactions and extended operations. For organizations already standardized on MCP, this reduces friction in moving agents from development to Bedrock production.
Related Articles
Amazon Nova Multimodal Embeddings adds audio search capabilities to Bedrock
Amazon Nova Multimodal Embeddings, announced October 28, 2025, now supports audio content for semantic search alongside text, images, and video. The model offers four embedding dimension options (3,072, 1,024, 384, 256) and uses Matryoshka Representation Learning to balance accuracy with storage efficiency.
ChatGPT now integrates Tubi TV app for searching 300,000+ movies and shows
OpenAI's ChatGPT has integrated Tubi TV, the ad-supported streaming service, allowing users to search Tubi's catalog of over 300,000 movies and TV episodes directly through the AI. Tubi becomes the first streaming service to integrate with ChatGPT's app ecosystem, available across web, desktop, and mobile platforms.
Google Gemini adds interactive visualization generation with real-time parameter adjustment
Google Gemini can now generate interactive visualizations directly within the chat interface, allowing users to tweak variables, rotate 3D models, and explore data in real time. The feature activates through phrases like "show me" or "help me visualize" when using the Gemini Pro model. This follows Anthropic's Claude launch of similar interactive diagram capabilities in mid-March.
Anthropic exits Claude Cowork research preview with enterprise features, launches Claude Managed Agents beta
Anthropic has promoted Claude Cowork from research preview to general availability, adding six enterprise features including role-based access controls, group spend limits, and usage analytics. The company simultaneously launched Claude Managed Agents in public beta—a composable API suite for building and deploying cloud-hosted agents without custom infrastructure work.
Comments
Loading...