Company
Anthropic
Title
Building Production Agentic Systems with Platform-Level LLMOps Features
Industry
Tech
Year
2025
Summary (short)
Anthropic's presentation at the AI Engineer conference outlined their platform evolution for building high-performance agentic systems, using Claude Code as the primary example. The company identified three core challenges in production LLM deployments: harnessing model capabilities through API features, managing context windows effectively, and providing secure computational infrastructure for autonomous agent operation. Their solution involved developing platform-level features including extended thinking modes, tool use APIs, Model Context Protocol (MCP) for standardized external system integration, memory management for selective context retrieval, context editing capabilities, and secure code execution environments with container orchestration. The combination of memory tools and context editing demonstrated a 39% performance improvement on internal benchmarks, while their infrastructure solutions enabled Claude Code to run autonomously on web and mobile platforms with session persistence and secure sandboxing.
## Overview This case study presents Anthropic's platform strategy for supporting production deployment of agentic LLM systems, delivered as a conference talk by Caitlyn, who leads the Claude developer platform team. The presentation uses Claude Code—an agentic coding product—as the primary illustrative example throughout, though the insights apply broadly to any production agentic system built on Claude. The talk addresses developers building agents integrated against LLM APIs and focuses on what Anthropic describes as "raising the ceiling of intelligence"—helping developers extract maximum performance from their models in production environments. Anthropic's framework for maximizing model performance in production revolves around three interconnected pillars: harnessing model capabilities through well-designed API features, managing the context window to ensure optimal information density, and providing infrastructure that allows Claude to operate autonomously and securely. This represents a thoughtful platform-level approach to LLMOps that goes beyond simply exposing model endpoints and instead considers the full operational lifecycle of agentic systems. ## Harnessing Model Capabilities Through API Design The first pillar focuses on how Anthropic exposes Claude's trained capabilities through their API as customizable features. This represents an interesting LLMOps philosophy: as the research team trains Claude to improve at various tasks, the platform team must provide corresponding API primitives that allow developers to access those capabilities effectively. One key example is **extended thinking**. Claude's performance on complex tasks scales with the amount of reasoning time allocated, so Anthropic exposed this as a controllable API feature. Developers can decide whether to have Claude think longer for complex problems or provide quick answers for simpler queries. This is implemented with a token budget mechanism, allowing developers to specify how many tokens Claude should "spend" on reasoning. For Claude Code specifically, this becomes crucial because the agent must balance between debugging complex systems (requiring extended thinking) and providing quick responses to straightforward queries. This represents a pragmatic approach to LLMOps where computational cost and response latency can be balanced against task complexity. The second capability exposed is **tool use**. Claude has been trained to reliably call tools, and Anthropic exposes this through both built-in tools (like web search) and custom tool definitions. Developers define tools with a name, description, and input schema, and Claude learns when to invoke them with appropriate arguments. For Claude Code, this is foundational—the agent continuously calls tools to read files, search codebases, write to files, and rerun tests. The reliability of tool calling becomes critical in production because unreliable tool invocation would cascade into poor agent performance and potentially dangerous operations (like writing to incorrect files). From an LLMOps perspective, this API design philosophy is notable because it acknowledges that model capabilities alone are insufficient—those capabilities must be surfaced through well-designed interfaces that allow production systems to control and leverage them effectively. However, it's worth noting that the presentation doesn't deeply discuss failure modes, error handling, or monitoring for these tool calls in production, which would be critical operational concerns. ## Context Window Management The second pillar addresses what is arguably one of the most challenging aspects of production LLM systems: context management. Anthropic identifies that "getting the right context at the right time in the window is one of the most important things that you can do to maximize performance." For agentic coding systems like Claude Code, this becomes particularly complex because the context might include technical designs, entire codebases, instructions, tool call histories, and more. The challenge is ensuring the optimal subset of available information is present in the context window at any given moment. Anthropic introduced three complementary mechanisms for context management: **Model Context Protocol (MCP)** was introduced a year before this presentation and has gained community adoption as a standardized way for agents to interact with external systems. For Claude Code, this enables integration with systems like GitHub or Sentry, providing access to information and tools beyond what's explicitly in the agent's context window. From an LLMOps perspective, MCP represents an important standardization effort—rather than each developer building custom integrations, a common protocol enables interoperability and reduces integration overhead. However, the presentation doesn't detail how MCP handles authentication, rate limiting, or error scenarios in production deployments, which would be important operational considerations. **Memory tools** complement MCP by helping Claude decide what context to store outside the window and when to retrieve it. Anthropic's initial implementation uses a client-side file system, giving developers control over their data while allowing Claude to intelligently store information for later retrieval. For Claude Code, this could include codebase patterns or git workflow preferences. Critically, Claude learns when to pull this stored context back into the window only when relevant. This addresses a fundamental LLMOps challenge: context windows are finite and expensive, so efficient utilization directly impacts both performance and operational costs. **Context editing** provides the inverse capability—removing irrelevant information from the context window. Anthropic's first implementation focuses on clearing old tool results, which can consume significant window space and may not remain relevant for future reasoning. For Claude Code, which calls hundreds of tools during a session, this becomes essential for maintaining a clean, relevant context. The combination of memory and context editing yielded a 39% performance improvement on Anthropic's internal benchmarks, demonstrating the significant operational impact of context management. Anthropic is expanding these capabilities by providing larger context windows (up to one million tokens for some models) while simultaneously teaching Claude to understand its own context utilization—essentially making the model aware of how much "room" it has left. This meta-awareness allows Claude to adapt its behavior based on available context space. From a balanced LLMOps perspective, these context management tools represent sophisticated solutions to real production challenges. However, several operational questions remain unaddressed: How do developers debug context management decisions? How are context management failures surfaced and monitored? What happens when memory retrieval fails or returns stale information? These observability and debugging considerations are crucial for production deployments but aren't extensively covered in this platform-focused presentation. ## Agent Infrastructure and Autonomous Operation The third pillar represents Anthropic's most ambitious vision: "give Claude a computer and just let it do its thing." This philosophy emerges from the observation that with access to writing and executing code, Claude can accomplish virtually anything. However, this requires substantial infrastructure to execute safely and reliably in production. The motivating use case was launching Claude Code on web and mobile platforms. When Claude Code runs locally, it uses the user's machine as its computer, but web and mobile deployments required solving several hard infrastructure problems: - **Secure execution environments**: Claude needs to write and run code that hasn't been explicitly approved by users, requiring robust sandboxing - **Container orchestration at scale**: Supporting many concurrent sessions with proper resource isolation - **Session persistence**: Users start sessions, walk away, and expect to return to completed work, requiring persistent execution environments Anthropic's solution involved developing a **code execution tool** exposed through their API. This tool allows Claude to write and run code in secure sandboxed environments hosted on Anthropic's servers, abstracting away container management and security concerns from developers. For Claude Code specifically, this enables scenarios like "make an animation more sparkly" where Claude needs to write, execute, and iterate on code autonomously. Building on this foundation, Anthropic introduced **agent skills**—folders of scripts, instructions, and resources that Claude can access and execute within its sandbox environment. Claude determines when to use skills based on user requests and skill descriptions. Skills can combine with MCP tools, where MCP provides access to external systems and context while skills provide the expertise to use those resources effectively. For Claude Code, a web design skill might ensure landing pages follow specific design systems and patterns, with Claude recognizing when to apply this expertise. From an LLMOps perspective, this infrastructure approach is impressive but raises important operational considerations. The presentation emphasizes that Anthropic handles orchestration, security, and sandboxing, but doesn't detail monitoring, logging, resource limits, cost controls, or failure recovery mechanisms. In production deployments, developers would need visibility into what code is being executed, resource consumption, execution failures, and security events. The abstraction of complexity is valuable, but operational transparency remains essential. The session persistence challenge is particularly interesting from an LLMOps standpoint. Users can start intensive operations and disconnect, expecting the agent to continue working. This requires not just persistent containers but also state management, progress tracking, and potentially notification systems to alert users when work completes. The presentation doesn't explore these operational aspects in depth. ## Platform Evolution and Future Direction Anthropic articulates three ongoing evolution vectors for their platform: First, continued **capability exposure**—as Claude improves existing capabilities and gains new ones, Anthropic will evolve their API to make these accessible to developers. This represents a commitment to keeping pace with model research through platform features. Second, enhanced **context management**—expanding the tools for deciding what to pull into context, what to store for later, and what to remove. This recognizes that context management will remain a critical operational challenge as models and applications grow more sophisticated. Third, deeper investment in **agent infrastructure**—continuing work on orchestration, secure environments, and sandboxing. Anthropic identifies these as the biggest barriers to the "just give Claude a computer" vision, and plans to address them at the platform level rather than expecting developers to solve these problems individually. ## Critical Assessment This case study presents a sophisticated platform-level approach to LLMOps for agentic systems, but several considerations merit attention: **Strengths**: Anthropic demonstrates thoughtful API design that exposes model capabilities through developer-friendly primitives. The context management tools address real production challenges, and the quantified 39% improvement from memory and context editing provides concrete evidence of impact. The infrastructure for secure code execution represents significant engineering investment that would be prohibitively expensive for most developers to replicate. **Limitations and Questions**: The presentation is understandably focused on capabilities and features, but leaves many operational questions unanswered. How do developers monitor agent behavior in production? What observability tools exist for debugging context management decisions or tool call failures? How are costs controlled when agents operate autonomously? What safeguards prevent runaway execution or resource consumption? How are errors and exceptions surfaced and handled? Additionally, the case study heavily features Claude Code as the exemplar, which is Anthropic's own product. While this provides concrete examples, it would be valuable to understand how third-party developers experience these platform features in diverse production environments. The 39% performance improvement is based on Anthropic's internal benchmarks, which may not generalize to all use cases. The "give Claude a computer" vision is compelling but raises security and control questions that aren't fully addressed. While sandboxing is mentioned, the presentation doesn't detail what level of access Claude has, what guardrails exist, or how developers can restrict capabilities when needed. **Operational Maturity**: The platform features described represent significant LLMOps sophistication, particularly around context management and secure execution. However, the presentation focuses more on enabling capabilities than on the operational disciplines (monitoring, logging, alerting, cost management, testing, evaluation) that typically define mature LLMOps practices. These may exist but aren't the focus of this developer-facing talk. **Vendor Lock-in Considerations**: The platform features described, while powerful, create dependencies on Anthropic's infrastructure. Developers using code execution, skills, memory, and context editing are tightly coupled to Anthropic's platform, which differs from the more portable approach of using standard LLM APIs. This isn't inherently problematic but represents a tradeoff between capability and portability that production teams should consider. Overall, this case study illustrates how LLM providers are evolving beyond simple inference APIs toward comprehensive platforms for production agentic systems, taking on infrastructure complexity to enable developer productivity. The approach demonstrates technical sophistication and addresses real operational challenges, though many questions about observability, debugging, and operational governance remain for teams deploying these systems in production.

Start deploying reproducible AI workflows today

Enterprise-grade MLOps platform trusted by thousands of companies in production.