Company
Meta
Title
Scaling Meta AI's Feed Deep Dive from Launch to Product-Market Fit
Industry
Media & Entertainment
Year
2025
Summary (short)
Meta launched Feed Deep Dive as an AI-powered feature on Facebook in April 2024 to address information-seeking and context enrichment needs when users encounter posts they want to learn more about. The challenge was scaling from launch to product-market fit while maintaining high-quality responses at Meta scale, dealing with LLM hallucinations and refusals, and providing more value than users would get from simply scrolling Facebook Feed. Meta's solution involved evolving from traditional orchestration to agentic models with planning, tool calling, and reflection capabilities; implementing auto-judges for online quality evaluation; using smart caching strategies focused on high-traffic posts; and leveraging ML-based user cohort targeting to show the feature to users who derived the most value. The results included achieving product-market fit through improved quality and engagement, with the team now moving toward monetization and expanded use cases.
## Overview Meta's Feed Deep Dive represents a comprehensive LLMOps case study demonstrating how a major technology company scaled an AI-powered feature from initial launch to product-market fit on one of the world's largest social media platforms. Launched in April 2024 on Facebook, Feed Deep Dive addresses a specific user need: information-seeking and context enrichment when encountering posts that spark curiosity. The feature needed to compete with Facebook Feed itself—one of the most successful and engaging products in social media—meaning every interaction had to provide more value than users would get from simply continuing to scroll. The presentation details Meta's journey from launch through the first year of operation, initially focusing on basic engagement across languages and regions, then pivoting in the second year to quality improvements through agentic models and reasoning while simultaneously improving latency and reducing capacity costs. This case study is particularly valuable because it addresses the real-world challenges of operating LLMs at Meta scale, where the stakes are high and user expectations are shaped by highly optimized competing features within the same ecosystem. ## Core Technical Challenges Meta identified three fundamental challenges in building product-market fit for Feed Deep Dive. First, the system needed to provide consistently high-quality experiences despite well-known LLM issues like hallucinations and refusals. Second, the solution had to operate at Meta scale while continuously improving latency. Third, and perhaps most uniquely, Feed Deep Dive faced an internal competition problem: because it's a feature built on top of Facebook Feed, every time a user engages with it they're being pulled away from an already highly engaging experience, creating an unusually high bar for value delivery. ## Context Understanding and Enrichment The presentation provides detailed examples of how context understanding evolved as a core capability. In one case study, Meta showed a meme from a TV show with minimal inherent context—just a few dialogue lines that could be interpreted literally (leading to questions about "rebooking flights"). The initial approach was too literal, but by leveraging content sentiment analysis from user comments, entity extraction to identify actors and show titles, and external web searches, the system could triangulate the actual meaning. This transformed the experience from answering literal questions about rebooking to understanding that this was a critical scene from a show and providing rich context about why it mattered to the story. Another compelling example involved an ambiguous post showing a rooster at what appeared to be a farm. Frame-by-frame analysis from a visual language model only identified "a rooster in a grassy field," leading to generic prompts about rooster breeding and behavior. However, the post was actually about an equestrian training event. Meta's solution involved extracting mentioned people, searching for their profiles (discovering they were equestrian trainers), examining the creator's previous posts (revealing they operated an equestrian training farm), identifying recent events, and applying step-by-step reasoning to connect these pieces. The result was accurately identifying the post's true subject and generating relevant prompts about equestrian lessons and event details—something even humans would struggle to determine without conducting similar research. ## Evolution from Traditional Orchestration to Agentic Models While the context enrichment examples demonstrated clear value, Meta recognized fundamental limitations in their heuristic-based approach. They were hitting token window limits, encountering needle-in-haystack problems where LLMs couldn't identify the most critical context, making numerous API calls that went unused, spending capacity on summarization to manage context windows, and critically, accumulating latency that caused users to abandon the experience and return to scrolling Feed. Meta's response was transitioning to agentic approaches with three key features: planning, tool calling, and reflection. Instead of predetermined context gathering, the agentic model would first generate a plan for which tools (APIs) to call. In the equestrian example, this meant initially calling APIs to understand mentioned people and multimodal content. Through reflection, when these didn't match expectations, the agent would make additional tool calls to retrieve similar posts from the author. After generating a proposed response, the system would reflect on quality before showing it to users. This transition represents a critical LLMOps evolution from static orchestration to dynamic, reasoning-based approaches. However, Meta is transparent about the trade-off: agentic approaches introduce additional latency. The decision to accept this trade-off reflects their findings about user behavior and product-market fit, which we'll explore further. ## Auto-Judges and Quality Evaluation Meta implemented auto-judges as a critical component for handling refusals—situations where the LLM equivocates or doesn't provide a direct answer. The presentation shows an actual user complaint where a suggested prompt led to a non-answer, breaking user trust and engagement habits. Meta identified three primary refusal causes: insufficient context, infrastructure failures during context generation, and integrity/safety reasons. The traditional offline evaluation approach involved sampling production traffic, using auto-judges to evaluate refusals, comparing with human raters, and fine-tuning models based on discrepancies. However, this cycle took weeks—too slow given that specific prompts and responses are typically only relevant for a few days. While this improved long-term model quality, it failed to address immediate issues. Meta's solution was implementing online auto-judges running continuously at scale. Leveraging the power law distribution where 5-10% of posts generate about 50% of traffic, they focused on ensuring refusal-free workflows for top posts. Critically, they didn't feed refusal detection back to the prompt generation model but rather to the prompt ranking model, which decides which prompts to show which users at what time. This allowed real-time downranking of prompts likely to cause refusals while promoting prompts likely to generate quality responses. This approach demonstrates sophisticated LLMOps thinking: rather than only improving the generative model, Meta optimized the system orchestrating when and how to deploy that model. It's a pragmatic recognition that in production systems, you can achieve quality improvements through multiple intervention points. ## Scaling Through Smart Caching Beyond quality, Meta needed to address capacity costs and latency at scale. Their primary strategy was smart caching, again leveraging the fact that 5-10% of posts drive 50% of impressions and 30% of clicks. By ensuring excellent, fast experiences for these high-traffic posts, they could optimize for the majority of user interactions while dramatically reducing recomputation costs. However, caching introduces the classic trade-off between latency/capacity and freshness. Meta's solution was topic-aware TTL (time-to-live) tuning. Sports and news content received very short TTLs due to fast-moving information environments, while how-tos and recipes—extremely popular on Feed and Feed Deep Dive—could tolerate much longer TTLs. The system detects topic and intent to set appropriate TTLs dynamically. Meta went further with adaptive caching based on online hit rate monitoring. For topics with shorter TTLs but high cache hit rates, they could accelerate eviction to favor freshness, accepting the latency cost because their data showed deeply engaged users were willing to wait longer for high-quality, fresh answers. This represents sophisticated production optimization where technical decisions are driven by observed user behavior patterns rather than theoretical optimization. ## Retention and Product-Market Fit Meta's approach to building product-market fit involved four key steps: defining the product funnel, identifying and understanding key cohorts, identifying key content types, and identifying jobs the product does well for users. For the product funnel, they defined metrics including click-through rates, deep sessions, weekly active usage, and L7ness/L28ness (engagement over 7 and 28-day windows), establishing thresholds for when the product was delivering value. This allowed pattern identification across the remaining dimensions. Cohort identification involved analyzing Feed usage patterns, Feed Deep Dive usage patterns, and monetizability potential. Using deep neural networks for targeting, they could increase feature prevalence for users deriving value while rolling back or reducing exposure for users not getting value, allowing those users to continue engaging with Feed, Reels, and other Facebook offerings. This is a critical LLMOps insight often missed in case studies: successful production deployment isn't about showing AI features to everyone—it's about intelligent targeting to maximize value and minimize negative impacts. The implementation leveraged U2U (user-to-user similarity) and U2M (user-to-media affinity) models operating in real-time. Feature engineering happened continuously with constant A/B testing, and importantly, they incorporated LLM-extracted features into their deep neural networks. This represents a fascinating LLMOps pattern where LLMs aren't just the end-user-facing component but also contribute to the ML infrastructure determining when and how to deploy LLM features. ## Key Learnings and Critical Assessment Meta identified three major learnings from this journey. First, "context is king," but more specifically, they evolved from believing "more context is good" to understanding that "right context is even better." Finding the right context, rather than maximizing context, became the key insight—one that agentic models help address through selective tool calling and reflection. Second, they confirmed the general principle that "speed wins," but with an important caveat: deeply engaged users are willing to accept higher latency for high-quality answers. This nuance is critical for LLMOps practitioners. The conventional wisdom about minimizing latency isn't wrong, but it's incomplete. Different user cohorts have different preferences, and systems sophisticated enough to identify deeply engaged users can make different trade-offs for different populations. Third, the transition to agentic models provided the right balance between capacity, latency, and quality. This isn't presented as a silver bullet—the latency cost is acknowledged—but as the appropriate tool for their specific context. From a critical perspective, several aspects deserve consideration. The presentation comes directly from Meta and naturally emphasizes successes while being relatively light on failure modes beyond the examples shown. The claim about achieving product-market fit should be assessed carefully—Meta is moving toward monetization, which will be the real test. The current metric discussion focuses on engagement and quality rather than business outcomes. The reliance on the power law distribution (5-10% of posts generating 50% of traffic) is presented as an advantage for optimization, but it's worth considering whether this creates a "rich get richer" dynamic where popular content receives better AI treatment, potentially reinforcing existing biases. The presentation doesn't address potential equity concerns in how the feature is deployed across different content types or creators. The auto-judge approach feeding into prompt ranking rather than model improvement is pragmatic and clever, but it's essentially a workaround for slow model improvement cycles. While the online approach is valuable, the underlying challenge of model improvement latency remains unresolved in the presentation. ## Future Directions Meta is exploring new use cases and moving toward monetization, focusing on intents where monetization is possible, contextual ad integrations, and social assistance features. The latter includes social graph understanding and privacy-preserving social AI to extend Feed Deep Dive into social use cases beyond pure information-seeking. The monetization focus is particularly interesting from an LLMOps perspective. Many organizations are still trying to prove basic utility; Meta is moving to prove business value through advertising integration. This will require additional infrastructure for understanding commercial intent, maintaining separation between organic AI responses and paid placements, and likely new evaluation frameworks for measuring not just engagement but business outcomes. ## Production Scale Insights This case study is particularly valuable for its honest treatment of production scale challenges. Many LLMOps discussions focus on model quality in isolation; Meta's presentation integrates quality, latency, capacity costs, user targeting, caching strategies, and evaluation infrastructure into a holistic view of what it takes to operate LLMs at scale. The multi-modal nature of the implementation—processing video transcripts, image OCR, user comments, creator history, and external web searches—demonstrates the complexity of real-world LLM applications. These aren't simple question-answering systems but sophisticated orchestrations of multiple AI capabilities, data sources, and reasoning steps. The year-long journey from launch to product-market fit provides a realistic timeline for mature organizations deploying LLM features. The initial focus on geographic and language expansion, followed by the pivot to quality and efficiency improvements, reflects a staged approach that many organizations might benefit from understanding as they plan their own LLM deployments. Overall, Meta's Feed Deep Dive case study represents a mature, sophisticated LLMOps implementation operating at genuine scale with careful attention to the full stack of challenges: quality, latency, capacity, user targeting, evaluation, and business value. While the presentation naturally emphasizes successes and comes from the implementing organization, it provides valuable insights into the real-world complexity of production LLM systems and the engineering trade-offs required to achieve product-market fit in a highly competitive environment.

Start deploying reproducible AI workflows today

Enterprise-grade MLOps platform trusted by thousands of companies in production.