Company
PGA Tour
Title
AI-Powered Content Generation and Shot Commentary System for Live Golf Tournament Coverage
Industry
Media & Entertainment
Year
2025
Summary (short)
The PGA Tour faced the challenge of engaging fans with golf content across multiple tournaments running nearly every week of the year, generating meaningful content from 31,000+ shots per tournament across 156 players, and maintaining relevance during non-tournament days. They implemented an agentic AI system using AWS Bedrock that generates up to 800 articles per week across eight different content types (betting profiles, tournament previews, player recaps, round recaps, purse breakdowns, etc.) and a real-time shot commentary system that provides contextual narration for live tournament play. The solution achieved 95% cost reduction (generating articles at $0.25 each), enabled content publication within 5-10 minutes of live events, resulted in billions of annual page views for AI-generated content, and became their highest-engaged content on non-tournament days while maintaining brand voice and factual accuracy through multi-agent validation workflows.
## Overview The PGA Tour case study represents a comprehensive implementation of LLMs in production for automated content generation at scale in the sports media domain. As described by David Provann (VP of Digital Architecture) and Murali Bakht (AWS Solutions Architect), the PGA Tour operates golf tournaments globally nearly every week of the year across four different tour levels, generating petabytes of data from military-grade radar systems, 14K cameras, walking scorers, and other tracking technologies that capture approximately 31,000 shots per tournament from 156 players. The organization faced the dual challenge of making golf comprehensible and engaging to fans while managing the operational complexity of content creation across multiple formats, audiences, and time-sensitive publication schedules. The implementation focuses on two primary AI systems: an agentic content generation platform that produces approximately 800 articles per week across eight content types, and a shot commentary system that provides real-time contextual narration during live tournament play. Both systems exemplify production LLMOps at scale, with careful attention to validation, cost optimization, operational monitoring, and brand consistency. ## Business Context and Technical Foundation The PGA Tour's digital strategy operates as a layered platform with their website serving as the comprehensive base layer containing exhaustive data going back to 1864 for scoring and 2012 for detailed shot data. Their mobile apps see 7X higher consumption than web but with focused, repeated short-duration sessions. The organization has built substantial infrastructure on AWS including a data lake containing all scoring, ball trajectory, and player data in highly structured format, plus a media lake storing thousands of hours of golf coverage across seven simultaneous live video streams per event. The unique operational challenges include the 150+ acre playable areas with no fixed boundaries, variable scoring formats (stroke play, match play, Stableford), and what the team describes as living in "edge cases" - unusual situations like players hitting balls onto clubhouses or hole-in-ones that damage the cup mid-round. This operational reality informed their production-first thinking for AI implementations, requiring systems that could handle inconsistent, non-uniform data while maintaining absolute accuracy since being second in sports reporting means being last. ## Product and Strategic Approach The PGA Tour took a product-first approach to AI implementation starting in 2023, applying their existing product framework that evaluates features based on three criteria: value for fans, value for stakeholders, and value for brand. They explicitly avoided doing "AI projects" in favor of "projects that utilize AI when it makes sense," representing a mature perspective on technology adoption. Their initial experimentation included chatbots (which David noted he doesn't favor due to brand exposure risks), followed by pivoting to practical applications like shot commentary and automated content generation. A critical strategic decision was their willingness to walk away from ideas that didn't work. David emphasized the "bravery to walk away" rather than continuing to invest in marginal improvements. Image generation was specifically called out as an area where they've had limited success due to player rights, image rights, and IP complications, demonstrating pragmatic boundaries for their AI adoption despite potential value. ## Shot Commentary System Launched at The Players Championship in 2025, the shot commentary system represents real-time AI content generation integrated with their Torcast 3D data visualization platform. The system generates two layers of content: factual descriptions (distance, position relative to pin) and contextual analysis (probability of making the putt, impact on leaderboard position, FedEx Cup implications). The team held themselves to a high standard of providing "commentary, not narration," meaning the system must add meaningful context rather than merely describing observable facts. Importantly, the system includes logic to recognize when context isn't meaningful - for example, a first drive on the first hole may warrant only a basic factual statement. This represents sophisticated prompt engineering and decision logic that avoids the common AI pitfall of always trying to generate maximum content. The commentary must be varied, engaging, and non-repetitive while maintaining factual accuracy, requiring integration with live scoring data, historical player statistics, and real-time tournament standings. ## Agentic Content Generation Architecture The content generation system uses AWS Bedrock with a multi-agent architecture orchestrated through AWS's Agent Core runtime. The workflow involves five distinct agent roles: **Research Agent**: Retrieves both structured data from PGA Tour APIs (player statistics, recent performance, tournament data) and unstructured data from PDF media guides containing player backgrounds, education, career history, and other biographical information. This dual data sourcing approach combines the precision of structured databases with the richness of narrative documents. **Data Agent**: Specifically handles querying and extracting information from both data sources, making it available to downstream agents. This separation of concerns allows specialized optimization for different data types. **Writer Agent**: Receives a work order containing all research data and generates the actual article content based on specific format requirements, target audience specifications, and content type. The writer operates under explicit brand and style guidelines provided through prompting. **Editor Agent**: Reviews generated content for adherence to PGA Tour style guidelines and brand standards. If content doesn't meet standards, the editor sends the request back to the writer for revision in an iterative loop. Only when the editor validates the content does it proceed to the validation stage. The editor agent also coordinates with an image selection process. **Validator Agent**: Performs fact-checking by extracting factual claims from the generated content and comparing them against authoritative data from the PGA Tour APIs. For example, if content claims "Scottie Scheffler scored a birdie on the 18th hole," the validator retrieves the actual scoring data to confirm accuracy. Failed validation sends content back to the editor for correction. This multi-agent workflow represents a sophisticated production LLMOps pattern that mirrors human editorial processes with clear separation of responsibilities, quality gates, and feedback loops. ## Technical Infrastructure and Implementation The system is implemented on AWS infrastructure with several key components optimizing for cost, reliability, and scalability: **Queue-Based Request Management**: Incoming content generation requests are stored in DynamoDB and placed into an SQS queue. This queue prevents hitting token usage limits when generating hundreds of articles simultaneously and provides natural rate limiting and retry capabilities. The pattern enables graceful handling of the 800+ articles generated per week without overwhelming downstream services. **AWS Agent Core Runtime**: The agent orchestration runs in AWS Agent Core, which provides significant cost advantages. During periods when agents are waiting for external API responses (from PGA Tour data APIs, image repositories, or Bedrock LLM calls), Agent Core does not charge for compute time. Given the I/O-bound nature of the multi-agent workflow with numerous external calls, this architectural choice significantly reduces operational costs. **Lambda-Based Invocation**: A Lambda function acts as the Agent Core invoker, pulling messages from SQS and initiating the agent workflow. This serverless pattern aligns with the event-driven nature of content generation requests. **S3 for Content Storage**: Generated content is written to S3 buckets, which trigger downstream Lambda functions that feed into the content ingest workflow for publication to various channels (website, mobile app, social media). **Bedrock LLM Integration**: The system uses AWS Bedrock for LLM inference, with multiple model calls throughout the agent workflow. They specifically adopted the Nova model for image selection tasks, achieving 75% cost reduction compared to previous approaches due to Nova's price-performance advantages. **Observability**: The team implemented comprehensive monitoring using AWS CloudWatch dashboards with custom metrics tracking agent latency, success rates, and operational health. Importantly, they extended their existing operational support patterns to these AI systems rather than creating separate support processes, maintaining consistent operational practices. ## Content Types and Formats The system generates eight distinct content types serving different engagement patterns: - **Betting Profiles**: Generated Monday mornings for all 156 players, completed by 9:30 AM for rapid SEO indexing. These drive the highest engagement on non-tournament days (Monday-Wednesday) when traffic is typically low, targeting fantasy and gambling users who consume statistical content to inform decisions. - **Tournament Previews**: Pre-tournament articles setting context for upcoming events. - **Tournament Recaps**: Post-tournament summaries of outcomes and key moments. - **Player Recaps**: Individual player performance summaries for each of 156 players, recognizing that fans exist for players at all leaderboard positions, not just leaders. - **Round Recaps**: After each of the four rounds (Thursday through Sunday), articles covering round-specific performance. - **Purse Breakdowns**: Prize money distribution by finish position - noted as generating vertical traffic spikes as fans immediately search for earnings information post-tournament. - **Points Breakdowns**: FedEx Cup points implications based on tournament finishes. - **Betting Profile Summaries**: Aggregated analysis across player profiles. Content is generated in multiple formats: long-form for website and mobile app, single-paragraph for social media, and short-form for push notifications. This format flexibility requires sophisticated prompting that adjusts length, style, and information density based on channel requirements. ## Validation and Quality Assurance The validation approach represents mature LLMOps practice addressing the core challenge of AI reliability in production: **Fact Extraction and Verification**: The system extracts factual claims from generated content and validates them against authoritative structured data sources. This programmatic fact-checking addresses the hallucination problem inherent to LLMs. **Multi-Stage Review**: The editor agent provides a first quality gate checking style and brand compliance before content proceeds to fact validation, creating multiple opportunities to catch issues. **LLM-as-Judge**: While not explicitly detailed, the editor agent likely employs LLM-as-judge patterns where an LLM evaluates content against guidelines, a common technique for style and brand consistency checks. **Automated vs. Human Review**: New features initially run with human oversight to establish confidence in quality. Once validation scoring meets thresholds, the team "flips the auto button on" to enable fully automated publication. This graduated automation approach manages risk while scaling. **Source Data Verification**: The validator compares output against original input data to ensure information fidelity throughout the generation process. The presentation notes that these validation techniques vary by content type - "some of it's simple rejects, some of it's LLM as a judge, some of it's take the original data and check it's in the output" - indicating a pragmatic, fit-for-purpose approach rather than one-size-fits-all validation. ## Performance and Cost Metrics The system demonstrates impressive production metrics: **Cost Efficiency**: Articles are generated at $0.25 each, representing a 95% cost reduction compared to previous approaches. This dramatic cost reduction is attributed to automation replacing human writing for repetitive statistical content, though the presentation acknowledges this doesn't replace human journalists for all content types. **Speed to Market**: Content is published 5-10 minutes after events conclude (e.g., round completion at 5:00 PM, articles live by 5:05-5:10 PM). This speed advantage makes PGA Tour the fastest source for golf tournament content, driving SEO advantages and capturing search traffic before other outlets publish. **Scale**: Currently generating 140-180 articles per week with plans to reach 800 articles per week by end of 2025. This represents scaling from prototype to full production deployment. **Engagement**: AI-generated content receives billions of page views annually. Betting profiles specifically are the highest-engaged content on non-tournament days, demonstrating clear business value beyond cost savings. **Image Selection Cost Reduction**: Adoption of Amazon Nova models for image appropriateness checking achieved 75% cost reduction for this specific task, showing ongoing optimization as new models become available. ## Operational Considerations Several operational aspects demonstrate production maturity: **Support Model**: The team maintained their existing support footprint without expanding headcount despite adding AI systems. This was achieved by integrating AI system monitoring into existing CloudWatch dashboards and operational workflows, treating AI systems as standard production components rather than special cases. **Edge Case Handling**: Given golf's inherent variability and the team's self-described existence in "edge cases," the system was designed with production resilience from the start. The queue-based architecture, retry mechanisms, and validation loops all contribute to handling unexpected scenarios. **Brand Protection**: Generated articles are clearly labeled as AI-generated at the bottom, maintaining transparency with audiences. This addresses both ethical considerations and manages expectations around content origin. **Continuous Improvement**: The validation scoring and automated/manual toggle approach enables continuous quality improvement. As confidence grows, more content types transition to fully automated publication. ## Development Practice Evolution David discussed how AI tools are influencing their development practices and hiring, representing a forward-looking LLMOps consideration: **AI-Assisted Development**: The team is evaluating tools like Amazon Q Developer (Quiro), Claude Code, and Cursor for development acceleration. Initial findings suggest these tools work significantly better with senior developers who can critically evaluate generated code, recognize pattern issues (like secrets in plain text), and provide more sophisticated prompting. **Quality vs. Appearance**: A key insight is that AI-generated code may "look great" to junior developers (proper formatting, camel case, etc.) but contain subtle issues visible only to experienced practitioners. AI agents "like to succeed" and "want you to be happy with the result," which can mask problems. **Hiring Implications**: The effectiveness of AI coding tools with senior talent is influencing hiring strategy, potentially shifting from traditional junior-mid-senior team structures toward more senior-heavy teams that can effectively leverage AI assistance. **Operational AI Applications**: Beyond content generation, the team is exploring AI for development processes, operational support, content management and planning, AWS account analysis, and spending anomaly detection - applying AI to internal processes rather than only customer-facing features. ## Strategic Lessons and Limitations The case study offers several strategic insights about production LLMs: **Selective Application**: The "bravery to walk away" philosophy and focus on using AI "where it makes sense" rather than everywhere represents mature technology adoption. Not every problem benefits from AI, and forcing fit reduces effectiveness. **Image Generation Challenges**: The explicit acknowledgement that image generation hasn't worked due to player rights and IP issues demonstrates domain-specific constraints that technical capabilities alone can't overcome. Production LLMOps must account for legal, rights, and regulatory realities. **Brand Voice Maintenance**: The focus on commentary quality, style adherence, and brand consistency throughout the agent workflow shows that generating content at scale requires more than just LLM calls - it requires sophisticated orchestration to maintain organizational voice. **Production-First Design**: The emphasis on planning for production operations from the beginning, including monitoring, support integration, and validation, reflects lessons learned from sports technology operations where failure is highly visible and immediate. **Value-Based Prioritization**: The three-criteria framework (fan value, stakeholder value, brand value) provides clear guidance for AI investment decisions and prevents technology-driven rather than value-driven projects. ## Content Strategy and SEO The content generation system serves a deliberate engagement strategy addressing different audience needs and temporal patterns: **Non-Tournament Day Engagement**: Betting profiles and statistical content maintain audience engagement during Monday-Wednesday periods when live tournament content isn't available, smoothing traffic patterns and maintaining SEO presence. **Long-Tail Player Coverage**: Generating content for all 156 players regardless of leaderboard position serves global fan bases with diverse player interests, recognizing that engagement isn't limited to tournament leaders. **Speed Advantage**: The 5-10 minute publication window after events creates first-mover SEO advantages, positioning PGA Tour as the authoritative first source for golf content in search results. **Format Diversity**: Multi-format content (long-form website, social media paragraphs, push notification snippets) from a single generation workflow maximizes content ROI and serves different consumption contexts. **Vertical Search Spikes**: Content like purse breakdowns targets predictable high-volume search patterns (fans searching earnings immediately post-tournament), capturing traffic with timely, targeted content. The presentation notes that PGA Tour leads SEO for their sport, with users coming to PGA Tour first for scoring over other platforms - an unusual achievement they attribute partly to their content strategy powered by AI-generated articles. ## Technical Debt and Future Direction Looking forward, the team is focusing on operational improvements across their development, support, and content planning processes rather than expanding AI into every possible domain. This measured approach to AI expansion reflects recognition that not all improvements come from adding more AI, but from optimizing existing implementations and applying AI thoughtfully to internal processes that benefit from automation. The plan to scale from 140-180 articles to 800 articles per week by end of 2025 suggests confidence in the architecture's scalability and the validation approach's effectiveness. The cost metrics (95% reduction, $0.25 per article) provide clear economic justification for this scaling. Overall, the PGA Tour implementation represents a mature production LLMOps case study with clear business value, thoughtful architecture, comprehensive validation, operational discipline, and strategic focus on problems where AI provides measurable advantage. The multi-agent workflow pattern, queue-based request management, cost-optimized runtime selection, and graduated automation approach offer replicable patterns for other organizations implementing production LLM systems at scale.

Start deploying reproducible AI workflows today

Enterprise-grade MLOps platform trusted by thousands of companies in production.