Company
Thumbtack
Title
Building and Implementing a Company-Wide GenAI Strategy
Industry
Tech
Year
2024
Summary (short)
Thumbtack developed and implemented a comprehensive generative AI strategy focusing on three key areas: enhancing their consumer product with LLMs for improved search and data analysis, transforming internal operations through AI-powered business processes, and boosting employee productivity. They established new infrastructure and policies for secure LLM deployment, demonstrated value through early wins in policy violation detection, and successfully drove company-wide adoption through executive sponsorship and careful expectation management.
## Overview Thumbtack is a consumer-facing technology company that operates a marketplace connecting homeowners with service professionals across the United States. The company has an established machine learning practice that handles problems like search, ranking, customer and professional recommendations, marketing optimization, and fraud detection. This case study documents how Thumbtack developed and implemented a comprehensive generative AI strategy starting in summer 2023, with the article published in late 2024 reflecting on approximately 18 months of experience. The case study provides valuable insights into how a mid-sized technology company approached the challenge of integrating LLMs into production systems while balancing exploration with execution on existing product roadmaps. Rather than focusing on a single technical implementation, this case study is notable for its strategic and organizational perspective on LLMOps adoption at scale. ## Strategic Framework Thumbtack structured their generative AI strategy around three core questions that shaped their investment priorities: The first pillar focused on leveraging gen AI to build better consumer products. The company had already begun adopting LLMs for search understanding, using them to better interpret customer queries in natural language and match customers with appropriate professionals. The strategic goal was to explore both augmenting and potentially replacing existing ML techniques with LLMs where applicable, while also exploring new capabilities around extracting product insights from structured and unstructured data. The second pillar addressed how gen AI could transform internal business operations. This included identifying bottlenecks to adoption and creating proof points through specific use cases. The company recognized that business processes often involve manual review and human intervention points that could be augmented or automated with generative AI. The third pillar targeted employee productivity improvements, with the expectation that gen AI would eventually become embedded in tools and workflows across engineering, sales, operations, and support functions. ## ML Infrastructure and Production Architecture A critical aspect of Thumbtack's LLMOps approach was their investment in flexible infrastructure that could support both internally-hosted open-source models and external API-based services. The company had established a dedicated ML infrastructure team in 2022, which positioned them well to extend their capabilities for LLM workloads. The ML infrastructure team began supporting deployment of open-source models like LLaMA 2 for internal teams in 2023. Recognizing the rapidly evolving landscape, they deliberately created optionality in their architecture to avoid lock-in to any single approach. This meant building capabilities to work with self-hosted models while also creating secure integration paths to external vendors like OpenAI. A significant technical investment was the creation of a new layer on top of their existing unified inference framework. This layer was designed to scrub personally identifiable information (PII) before data could be sent to external gen AI APIs, addressing privacy and compliance concerns that are essential for production deployments. This PII-aware proxy approach enabled teams to explore external LLM APIs without violating the company's privacy policy or exposing sensitive customer data. The ML infrastructure team also created shared compute clusters specifically for open-source LLMs, allowing data scientists to experiment with different models and use cases without needing to provision their own infrastructure. This democratization of access was a deliberate strategy to accelerate exploration and reduce barriers to adoption. ## Production Use Cases and Human-in-the-Loop Patterns One of the more interesting LLMOps patterns described in the case study is the approach to policy violation detection. Thumbtack has policies that govern acceptable behavior on their platform, and violations are typically detected through a combination of rules and machine learning models that flag potential issues for human review. The company identified this as an ideal early use case for gen AI augmentation because the existing human-in-the-loop workflow provided a natural safety net against hallucinations and false positives. Since humans were already reviewing flagged items, the gen AI system could be deployed to improve efficiency without the full burden of needing to be perfectly accurate. This pattern of selecting initial use cases where human oversight already exists represents a pragmatic approach to managing LLM reliability concerns in production. The case study notes that this approach resulted in "significant efficiency gains," though specific metrics are not provided. The key insight here is that augmenting existing human review processes with gen AI is a lower-risk path to productization compared to fully automated decision-making systems. ## Unstructured Data Analysis Another production use case highlighted involves using gen AI for exploratory analysis of unstructured data. The example given is mining customer reviews for insights about professional-customer interactions, such as understanding pricing patterns for specific services in particular geographic areas. Previously, extracting such insights required significant data engineering and natural language processing work. With LLMs, data scientists could interactively explore unstructured data sources and then productize the analytical processes that proved valuable. This represents a shift from batch-oriented NLP pipelines to more interactive, exploratory workflows enabled by LLM capabilities. ## Developer Productivity and Tooling Thumbtack implemented a pilot program for GitHub Copilot as part of their employee productivity pillar. The approach was deliberately structured as a measured experiment, with the goal of tracking both adoption rates and productivity signals. The company used this as a proof point to build momentum for broader adoption of gen AI-powered tools across the organization. Beyond coding assistance, the company invested in AI capabilities for business operational tooling, including real-time AI agent assist for customer support, automated quality assurance, and conversational AI trained on product and policy documentation. The goal was to expand support access to more customers and professionals with 24/7 availability. ## Governance and Policy Framework A notable aspect of Thumbtack's LLMOps approach was the emphasis on creating new policies and governance frameworks specifically for AI/ML usage. The company worked cross-functionally with IT, Platform, and Legal teams to update their AI/ML usage policies to track and approve usage across different use cases. This governance layer is essential for enterprise LLM adoption, as it provides the organizational structure needed to manage risks around data privacy, model behavior, and compliance. The case study emphasizes that without these updated policies and infrastructure frameworks, exploration and productization of gen AI would have been significantly constrained. ## Organizational Change Management The case study provides valuable insights into the organizational aspects of LLMOps adoption. Executive sponsorship is highlighted as a critical success factor, with sponsors at the executive level pushing the team to think holistically about generative AI strategy and to consider transforming business processes beyond just product use cases. The adoption strategy included explicit mentions of gen AI in company-level planning memos, ensuring that teams were actively encouraged to explore these technologies during planning exercises. This top-down signaling combined with bottom-up proof points created organizational momentum for broader adoption. Communication of learnings and success stories was treated as an important ongoing activity, helping to ground expectations while building excitement. The company held periodic check-ins with leadership to track progress and adjust investment levels as needed. ## Lessons Learned and Honest Reflections The case study is refreshingly candid about the challenges faced during the adoption journey. In 2023, the team encountered a range of attitudes: some stakeholders underestimated how transformative gen AI would be, while others overestimated what could be achieved or focused too heavily on risks like hallucination. The 18-month planning horizon was a deliberate choice to balance forward-looking investment with the uncertainty inherent in a rapidly evolving technology space. This allowed the company to make meaningful commitments without overcommitting based on assumptions that might not hold. The emphasis on early wins in relatively straightforward problems that were less sensitive to hallucination issues (like augmenting human review processes) helped build credibility and momentum before tackling more challenging fully-automated use cases. ## Critical Assessment While the case study provides a comprehensive view of Thumbtack's strategic approach to gen AI adoption, it should be noted that specific quantitative results are largely absent. The article mentions "significant efficiency gains" and "immediate value" but does not provide concrete metrics that would allow readers to assess the magnitude of impact. The case study reads partly as a strategic framework document and partly as a retrospective, which is valuable for understanding the organizational and governance aspects of LLMOps but provides less technical depth on the actual production implementations. Readers looking for detailed technical architectures or specific model performance characteristics will find this case study more focused on the meta-level concerns of driving enterprise adoption. That said, the emphasis on creating optionality in infrastructure, the pragmatic approach to selecting initial use cases with existing human oversight, and the attention to governance frameworks represent practical patterns that are applicable to other organizations navigating similar adoption journeys.

Start deploying reproducible AI workflows today

Enterprise-grade MLOps platform trusted by thousands of companies in production.