ZenML

AI-Powered Account Planning Assistant for Sales Teams

AWS Sales 2025
View original source

AWS Sales developed an AI-powered account planning draft assistant to streamline their annual account planning process, which previously took up to 40 hours per customer. Using Amazon Bedrock and a comprehensive RAG architecture, the solution helps sales teams generate high-quality account plans by synthesizing data from multiple internal and external sources. The system has successfully reduced planning time significantly while maintaining quality, allowing sales teams to focus more on customer engagement.

Industry

Tech

Technologies

Overview

AWS Sales developed an internal generative AI-powered assistant to streamline the process of creating account plans (APs), which are comprehensive strategy documents that sales personnel draft for established AWS customers. Prior to this solution, account managers spent up to 40 hours per customer creating these documents, leading to significant organizational overhead. The solution, launched in October 2024, builds on an existing internal tool called Field Advisor and leverages Amazon Bedrock to help sales teams create comprehensive and insightful APs in less time.

It’s worth noting that this case study comes from AWS itself, describing an internal tool built using their own cloud services. While this provides interesting insights into how AWS uses its own products, readers should be aware that it also serves as a showcase for AWS services. The reported benefits and time savings come from internal testimonials rather than independent verification.

Business Problem and Use Cases

The account plans draft assistant addresses four primary use cases:

The first is account plan draft generation, where Amazon Bedrock is used to make internal and external data sources available for generating draft content for key sections of APs. This enables sales teams to quickly create initial drafts for sections such as customer overviews, industry analysis, and business priorities, which previously required hours of research across the internet and disparate internal AWS tools.

The second is data synthesis, where the assistant pulls relevant information from multiple sources including CRM systems, financial reports, news articles, and previous APs to provide a holistic view of customers. This multi-source aggregation is a classic use case for RAG-based systems.

Third, built-in quality checks help ensure that APs meet internal standards for comprehensiveness, accuracy, and strategic alignment. This quality assurance capability is an important aspect of production LLM systems, though the case study doesn’t elaborate on how these checks are implemented.

Finally, the system offers customization capabilities, allowing account managers to upload proprietary documents to match their unique customer knowledge and strategic approach while still benefiting from AI-generated drafts.

Technical Architecture

The solution employs a sophisticated multi-component architecture built on AWS services. When a user initiates the workflow in the internal CRM system, it triggers the account plan draft assistant through a pre-signed URL. The assistant then orchestrates a multi-source data collection process, performing web searches while also pulling account metadata from OpenSearch, Amazon DynamoDB, and Amazon S3 storage.

Core Components

Amazon Bedrock serves as the foundation, providing programmatic API access to high-performing foundation models along with vector search capabilities and metadata filtering using Amazon Bedrock Knowledge Bases. The knowledge bases are populated using sales-enablement materials, historic APs, and other relevant documents curated by AWS Glue jobs. This RAG architecture enables the system to semantically search and use metadata filtering to retrieve relevant context from diverse sources.

AWS Lambda functions support two distinct use cases in the architecture. The async resolver Lambda function interfaces with the front-end client CRM and orchestrates async job IDs for the client to poll. This layer also handles input validations, user request throttling, and cache management. Worker Lambda functions perform the actual content generation, working concurrently to generate different sections of APs by using publicly available data, internal data, and curated data in the Bedrock knowledge bases. These functions invoke various LLMs using Amazon Bedrock and store the final content in the AP’s DynamoDB database corresponding to each async job ID.

Amazon DynamoDB maintains the state of each user request by tracking async job IDs, tracks throttling quota (both global request count and per-user request count), and acts as a cache. This stateful tracking is essential for the asynchronous processing model the system employs.

AWS Glue jobs curate and transform data from various internal and external data sources. These jobs push data to internal data sources (APs, internal tooling team S3 buckets, and other internal services) and to Bedrock Knowledge Bases, facilitating high-quality output through RAG.

Amazon SQS enables decoupling between the management plane and data plane. This decoupling is crucial in allowing the data plane worker functions to concurrently process different sections of the APs and ensure that APs can be generated within specified timeframes.

The system also includes a custom web frontend built using a ReactJS-based micro-frontend architecture, which integrates directly into the CRM system for a seamless user experience. This allows account managers to access the assistant without leaving their familiar work environment.

Data Management and RAG Implementation

The solution uses Amazon Bedrock’s out-of-the-box knowledge base management capabilities. Through its RAG architecture, the system semantically searches and uses metadata filtering to retrieve relevant context from diverse sources including internal sales enablement materials, historic APs, SEC filings, news articles, executive engagements, and data from CRM systems.

The connectors built into Amazon Bedrock handle data ingestion from Amazon S3, relational database management systems (RDBMS), and third-party APIs. The knowledge base capabilities enable filtering and prioritization of source documents when generating responses, resulting in what AWS describes as higher quality and more relevant content in generated AP sections.

Security and Compliance Considerations

The case study emphasizes that security and compliance are paramount when dealing with customer data. AWS IAM Identity Center is used for enterprise single sign-on so that only authorized users can access the account plans draft assistant. Through Field Advisor, various internal authorization mechanisms help ensure that users generating APs only access data they already have permissions for. This role-based access control is a critical aspect of any production LLM system handling sensitive business information.

Operational Workflow

The notification chain is another notable operational aspect. When content generation is complete, Amazon SQS and an internal notifications service API gateway deliver updates using Slack direct messaging and store searchable records in OpenSearch for future reference. This integration with existing communication tools (Slack) demonstrates thoughtful consideration of user workflow and adoption.

Results and User Feedback

Since its launch, thousands of sales teams have reportedly used the assistant to draft sections of their APs. According to testimonials included in the case study, one enterprise account manager reported saving at least 15 hours on their latest enterprise account plan, allowing more time for actual customer engagement rather than research and writing. A mid-market account manager noted that the tool helps rapidly generate baseline plans that can then be prioritized and customized, which is particularly valuable when managing multiple accounts.

These testimonials should be taken with some caution as they are internal AWS employees discussing an internal AWS tool built with AWS services. However, the reported time savings from 40 hours down to significantly less represents a meaningful productivity improvement if accurate.

Future Development Plans

AWS outlines an ambitious roadmap toward creating what they call a “zero-touch account planner” that could generate a full AP for a customer, incorporating best practices observed across customers to provide sales teams with best-in-class strategies. Planned enhancements include deeper integration with purpose-built planning tools including automatic generation of value maps and stakeholder maps, enhanced personalization tailored by industry, account size, and individual user preferences, improved collaboration features for multiple team members to work together on refining AI-generated plans, and expanded recommendation capabilities to provide “what next?” ideas to sales teams.

Key LLMOps Patterns

Several notable LLMOps patterns emerge from this case study. The asynchronous processing model with job tracking addresses the latency challenges inherent in LLM-based content generation, allowing users to initiate requests and receive notifications when processing completes. The concurrent generation of different AP sections by separate worker Lambda functions demonstrates effective parallelization for complex multi-part document generation.

The RAG architecture with metadata filtering shows a mature approach to grounding LLM outputs in relevant, authoritative sources. The throttling and quota management at both global and per-user levels indicates consideration for resource management and cost control in a production environment.

The micro-frontend integration approach demonstrates how AI capabilities can be embedded into existing tools rather than requiring users to adopt entirely new workflows, which is often critical for enterprise adoption.

Overall, this case study provides a detailed look at how a large technology organization has deployed LLM-based capabilities for a specific internal use case, with particular attention to the architectural decisions around asynchronous processing, data management, and security controls that are essential for production LLM systems.

More Like This

Building a Search Engine for AI Agents: Infrastructure, Product Development, and Production Deployment

Exa.ai 2025

Exa.ai has built the first search engine specifically designed for AI agents rather than human users, addressing the fundamental problem that existing search engines like Google are optimized for consumer clicks and keyword-based queries rather than semantic understanding and agent workflows. The company trained its own models, built its own index, and invested heavily in compute infrastructure (including purchasing their own GPU cluster) to enable meaning-based search that returns raw, primary data sources rather than listicles or summaries. Their solution includes both an API for developers building AI applications and an agentic search tool called Websites that can find and enrich complex, multi-criteria queries. The results include serving hundreds of millions of queries across use cases like sales intelligence, recruiting, market research, and research paper discovery, with 95% inbound growth and expanding from 7 to 28+ employees within a year.

question_answering data_analysis chatbot +44

Scaling AI Product Development with Rigorous Evaluation and Observability

Notion 2025

Notion AI, serving over 100 million users with multiple AI features including meeting notes, enterprise search, and deep research tools, demonstrates how rigorous evaluation and observability practices are essential for scaling AI product development. The company uses Brain Trust as their evaluation platform to manage the complexity of supporting multilingual workspaces, rapid model switching, and maintaining product polish while building at the speed of AI industry innovation. Their approach emphasizes that 90% of AI development time should be spent on evaluation and observability rather than prompting, with specialized data specialists creating targeted datasets and custom LLM-as-a-judge scoring functions to ensure consistent quality across their diverse AI product suite.

document_processing content_moderation question_answering +52

Multi-Agent AI System for Investment Thesis Validation Using Devil's Advocate

Linqalpha 2026

LinqAlpha, a Boston-based AI platform serving over 170 institutional investors, developed Devil's Advocate, an AI agent that systematically pressure-tests investment theses by identifying blind spots and generating evidence-based counterarguments. The system addresses the challenge of confirmation bias in investment research by automating the manual process of challenging investment ideas, which traditionally required time-consuming cross-referencing of expert calls, broker reports, and filings. Using a multi-agent architecture powered by Claude Sonnet 3.7 and 4.0 on Amazon Bedrock, integrated with Amazon Textract, Amazon OpenSearch Service, Amazon RDS, and Amazon S3, the solution decomposes investment theses into assumptions, retrieves counterevidence from uploaded documents, and generates structured, citation-linked rebuttals. The system enables investors to conduct rigorous due diligence at 5-10 times the speed of traditional reviews while maintaining auditability and compliance requirements critical to institutional finance.

document_processing question_answering structured_output +33