Company
Anthropic
Title
Building Production-Ready Agentic Systems with the Claude Developer Platform
Industry
Tech
Year
2025
Summary (short)
Anthropic's Claude Developer Platform team discusses their evolution from a simple API to a comprehensive platform for building autonomous AI agents in production. The conversation covers their philosophy of "unhobbling" models by reducing scaffolding and giving Claude more autonomous decision-making capabilities through tools like web search, code execution, and context management. They introduce the Claude Code SDK as a general-purpose agentic harness that handles the tool-calling loop automatically, making it easier for developers to prototype and deploy agents. The platform addresses key production challenges including prompt caching, context window management, observability for long-running tasks, and agentic memory, with a roadmap focused on higher-order abstractions and self-improving systems.
## Overview This case study presents insights from Anthropic's Claude Relations lead (Alex), Product Management lead (Brad), and Engineering lead (Katelyn) for the Claude Developer Platform. The discussion focuses on how Anthropic has evolved their production infrastructure from a simple API into a comprehensive platform designed to support autonomous agentic workflows at scale. The conversation reveals Anthropic's philosophy on building production LLM systems: reducing scaffolding and constraints on models while providing the right tools and infrastructure to enable autonomous decision-making. ## Platform Evolution and Philosophy The Claude Developer Platform represents a significant evolution from what was originally called the "Anthropic API." The platform now encompasses APIs, SDKs, comprehensive documentation, and console experiences—everything developers need to build production applications on Claude. Importantly, Anthropic's internal products like Claude Code are built directly on this public platform, which demonstrates confidence in the platform's production readiness and ensures that internal and external users benefit from the same infrastructure improvements. The team emphasizes a key philosophical shift in how to approach LLM systems in production: moving away from heavy scaffolding and predefined workflows toward giving models more autonomy. This reflects their observation that as models become more capable, the guardrails and constraints that developers previously built around them actually become liabilities rather than assets. Brad notes that they've encountered customers who upgraded to newer models but reported only marginal improvements—until investigation revealed that their scaffolding was constraining the model's ability to demonstrate its enhanced capabilities. ## Defining Agents and Autonomy The team provides a clear definition of what Anthropic considers an "agent" in the context of production systems. While acknowledging that "agent" has become somewhat of a buzzword with varying definitions, they emphasize that true agency involves the model having autonomy to choose which tools to call, handle results, and determine next steps. This contrasts with workflow systems where developers predefine the execution path. The distinction is important for production deployments: predefined workflows can be useful and reliable, but they inherently limit how much benefit you can extract from model improvements. When you build with agentic patterns that allow the model to make autonomous decisions, each new model release can deliver compounding improvements without requiring changes to your application code. This represents a fundamentally different approach to building production systems compared to traditional software engineering. ## The Scaffolding Problem A central theme throughout the discussion is what the team calls "unhobbling" the model. The intuition is that current generation models contain significantly more intelligence than developers have been able to unlock, but excessive scaffolding prevents that intelligence from being expressed. The team argues that many frameworks and orchestration tools have become too heavy and too opinionated, getting in the way of what the model is trying to accomplish. This has led to an interesting discourse in the field, with some practitioners arguing that agents can be as simple as "just a while loop." Anthropic's position is nuanced: they acknowledge that many agent systems have become overly complex, but they also recognize that the right kind of infrastructure support is valuable. Their goal is to provide lightweight, opinionated tooling that helps developers get the most out of the model without imposing heavy frameworks that constrain model behavior. ## The Claude Code SDK as General-Purpose Agentic Harness A significant practical development is the Claude Code SDK, which was originally built for coding applications but has evolved into a general-purpose agentic harness. The SDK provides an out-of-the-box solution for running the agentic loop—automating tool calling, handling results, and managing the interaction cycle. This is particularly valuable because it eliminates the need for every developer to implement their own version of prompt caching management, tool call handling, and loop logic. The team emphasizes that when they removed all the coding-specific scaffolding from Claude Code to "unhobble" the model, they discovered there wasn't much coding-specific left—just a generic agentic loop with access to a file system, Linux command-line tools, and the ability to write and execute code. These are sufficiently generic capabilities that the SDK works well for a wide variety of use cases beyond coding. For production deployments, developers can use the SDK's runtime wherever they need it. However, the team is also working on higher-order abstractions that will make it even easier to deploy agentic systems at scale while maintaining the observability and control that enterprises require. ## Production Tools and Features The platform includes several features specifically designed to address production challenges: **Web Search and Web Fetch**: Server-side tools that enable Claude to autonomously search the web and fetch content from specific URLs. The implementation is deliberately minimal—the team provides the tools to the model with minimal prompting, and Claude autonomously decides when and how to use them. For example, when conducting research, Claude will perform searches, evaluate results, decide which links are most promising, and fetch detailed content from those links—all without requiring explicit orchestration from the developer. This demonstrates the "unhobbling" philosophy in practice: give the model the right tools and trust its reasoning capabilities. **Prompt Caching**: An optimization that helps reduce costs and latency for repeated API calls with similar context. This is particularly important for agentic workflows that may involve many tool calls over extended periods. **Batch API**: A separate API endpoint for handling batch processing workloads, which is important for production use cases that need to process large volumes of requests efficiently. **Code Execution**: Provides Claude with a VM where it can write and execute code, see results, and iterate. This enables capabilities like data analysis with charts and graphs, image manipulation, and other computational tasks. The team views this as an early step toward giving Claude a persistent computer environment. **Context Management Tools**: Given that Claude supports 200K tokens by default (with 1M tokens available in beta for Sonnet), managing context effectively is crucial for production systems. The platform includes features to help: - **Automatic tool call removal**: The model can remove older tool calls from the context that are no longer needed. This is based on the observation that decluttering the context actually helps the model focus better, similar to how a human works better with a clean desk. The system includes guardrails to prevent removing recent or critical tool calls, and uses "tombstoning"—leaving notes about what was removed so the model maintains awareness of its history. - **Agentic Memory**: A tool that enables Claude to take notes during task execution and review those notes later. This addresses a limitation where models currently perform similarly each time they run a task, unlike humans who improve with practice. With memory capabilities, Claude can learn that certain websites are more reliable, specific search strategies work better, or particular databases should be prioritized—and apply those learnings in subsequent runs. ## Production Considerations: Business Value and Use Case Selection Beyond the technical capabilities, the team emphasizes the importance of clearly defining business value when deploying agents in production. They've observed that the most successful customer implementations are those where there's clear articulation of expected outcomes: how many engineering hours will be saved, how much manual work will be eliminated, what specific business process will be improved. This clarity helps in properly scoping the agent project and measuring success. This pragmatic focus on business outcomes is important context when evaluating the platform's capabilities. While the technical features are impressive, Anthropic is clearly thinking about how enterprises will actually adopt and justify agentic systems in production environments. ## Observability and Control A critical challenge for production agentic systems is observability, especially as tasks become longer-running and more autonomous. The team acknowledges this as one of the most common concerns from users: when you give an agent autonomy to work in the background, how do you ensure it's doing the right thing? How do you audit its behavior? How do you tune prompts or adjust tool-calling strategies based on what you observe? Anthropic is prioritizing observability as a key platform capability. Their position is that if they're going to encourage giving models more autonomy, they need to provide the infrastructure for developers to audit, monitor, and tune their systems effectively. This is particularly important for the longer-running tasks that the platform is increasingly enabling. ## Future Direction: Higher-Order Abstractions and Self-Improvement The roadmap centers on two complementary themes: higher-order abstractions that make it simpler to get the best outcomes from Claude, and observability that enables continuous improvement. Katelyn describes a vision of a "flywheel" where these capabilities combine with features like memory to create systems that don't just deliver good outcomes, but self-improving outcomes that get better over time. This represents an interesting production paradigm: rather than building systems that perform consistently at a fixed level, the goal is to build systems that learn and improve through use. This requires careful infrastructure support—not just for the learning mechanisms themselves, but for observing and validating that the improvements are genuine and aligned with desired outcomes. Brad's excitement about "giving Claude a computer" points to another direction: moving from isolated task execution to persistent environments where Claude can organize files, set up tools the way it prefers, and maintain state across sessions. The code execution feature is positioned as just the "baby step" toward this vision. This raises important questions about security, isolation, and control that production deployments will need to address. ## Critical Assessment While this case study provides valuable insights into Anthropic's production platform, it's important to note several limitations and considerations: **Source Limitations**: This is a promotional conversation featuring Anthropic employees discussing their own platform. The claims about performance, customer success, and model capabilities should be understood in that context. There are no specific metrics, customer names (beyond vague references), or quantitative results provided. **Observability Gap**: While the team acknowledges observability as critical and promises future capabilities, the current state seems relatively limited. For enterprises considering production deployment of autonomous agents, the lack of detailed discussion about current observability tools is notable. **Scaffolding Debate**: The "unhobbling" philosophy is presented quite confidently, but the team also acknowledges that there's active debate in the field about whether heavy frameworks are necessary. Their position that scaffolding becomes a liability as models improve is reasonable but somewhat unproven at scale. Different use cases may require different levels of control and constraint. **Memory and State Management**: While the agentic memory feature sounds promising, leaving developers to manage where memory is stored could introduce complexity and consistency challenges in production deployments. The division of responsibility between platform and developer isn't fully clear. **Scale and Enterprise Readiness**: When asked whether the Claude Code SDK is ready for enterprise deployment, Katelyn's response is somewhat equivocal—it can be used if you can deploy the runtime, but they're working on higher-order abstractions for scale. This suggests the current state may require significant engineering effort for production use. **Model Dependency**: The entire philosophy depends heavily on continued model improvements. If model capabilities plateau, the benefit of the "unhobbling" approach versus more structured workflows becomes less clear. ## Technical Contributions to LLMOps Despite these caveats, the case study does illuminate several valuable contributions to LLMOps practice: **Agentic Loop Abstraction**: The Claude Code SDK's approach of providing a reusable agentic harness addresses a real pain point where every team implements their own version of tool-calling loops and context management. **Context Management Strategies**: The automatic tool call removal with tombstoning and the agentic memory tool represent concrete approaches to managing long-running agent contexts in production. **Platform Consistency**: Building internal products on the same public platform is a strong signal of production readiness and ensures alignment between internal and external use cases. **Tool Design Philosophy**: The minimalist approach to tools like web search—providing capabilities with minimal prompting and trusting model reasoning—offers an alternative to heavily prompted and constrained tool use that may be worth exploring. The discussion provides a useful window into how a leading AI company is thinking about production infrastructure for agentic systems, even if specific implementation details and empirical validation would strengthen the case study considerably.

Start deploying reproducible AI workflows today

Enterprise-grade MLOps platform trusted by thousands of companies in production.