Adobe's Information Architect Jessica Talisman discusses how to build and maintain taxonomies for AI and search systems. The case study explores the challenges and best practices in creating taxonomies that bridge the gap between human understanding and machine processing, covering everything from metadata extraction to ontology development. The approach emphasizes the importance of human curation in AI systems and demonstrates how well-structured taxonomies can significantly improve search relevance, content categorization, and business operations.
This case study entry is based on a source URL that returned a 404 error, meaning the actual content was not accessible at the time of analysis. The following notes are constructed based solely on the URL title and metadata available, and should be treated with appropriate caution as the underlying details cannot be verified.
The URL suggests this was an episode from the “How AI Is Built” podcast (Season 2, Episode 8) featuring discussion about building taxonomies and data models to remove ambiguity from AI and search systems. The episode appears to be associated with Adobe, a major technology company known for creative software, document management, and enterprise solutions.
Based on the episode title, the discussion likely centered on how structured knowledge organization—specifically taxonomies and data models—can improve the performance and reliability of AI systems, particularly in search applications. This is a critical topic in LLMOps because:
The Challenge of Ambiguity in AI and Search: Large Language Models and search systems often struggle with ambiguous terms, concepts that have multiple meanings, and content that lacks proper categorization. Without proper taxonomies, AI systems may return irrelevant results, misinterpret user intent, or fail to make meaningful connections between related content.
Role of Taxonomies in Production AI: Taxonomies provide a structured vocabulary and hierarchical organization of concepts that help AI systems understand relationships between entities. In production environments, well-designed taxonomies can:
Data Models for AI Applications: Data models define how information is structured, stored, and related within a system. For AI applications in production, robust data models are essential for:
Adobe, as a company with extensive experience in content management, digital experience platforms, and creative tools, would have significant expertise in taxonomy and data model development. Their products like Adobe Experience Manager, Adobe Sensei (their AI and machine learning framework), and various creative cloud applications all require sophisticated knowledge organization systems.
Adobe’s potential use cases for taxonomies in AI include:
While the specific details of Adobe’s implementation cannot be verified from the available source, the general topic of taxonomies and data models intersects with several key LLMOps concerns:
Knowledge Management: Organizations deploying LLMs in production must carefully consider how to structure their knowledge bases. Taxonomies help create organized, queryable knowledge repositories that can be used to augment LLM responses with accurate, contextual information.
Reducing Ambiguity: One of the persistent challenges in LLM deployment is handling ambiguous queries or content. Taxonomies can provide disambiguation by establishing clear definitions and relationships between concepts, helping the AI system understand context more accurately.
Scalability and Maintenance: Production AI systems require data structures that can scale and be maintained over time. Well-designed taxonomies and data models facilitate updates, additions, and modifications without breaking existing functionality.
Evaluation and Quality: Structured data models enable better evaluation of AI system performance by providing clear categories and relationships against which outputs can be measured.
It is important to emphasize that without access to the actual content of the podcast episode, this analysis is necessarily speculative. The specific implementation details, results achieved, challenges encountered, and lessons learned by Adobe cannot be accurately reported. The URL’s 404 status means that any claims about specific technical approaches, metrics, or outcomes would be unfounded.
Readers seeking detailed information about Adobe’s approach to taxonomies and data models in AI systems should consult official Adobe documentation, engineering blogs, or attempt to access the podcast content through alternative means.
Based on industry knowledge, organizations implementing taxonomies for AI systems typically consider:
These practices would likely be relevant to any enterprise-scale implementation, including Adobe’s, though the specific details of their approach remain unverified due to the inaccessible source material.
Snorkel developed a specialized benchmark dataset for evaluating AI agents in insurance underwriting, leveraging their expert network of Chartered Property and Casualty Underwriters (CPCUs). The benchmark simulates an AI copilot that assists junior underwriters by reasoning over proprietary knowledge, using multiple tools including databases and underwriting guidelines, and engaging in multi-turn conversations. The evaluation revealed significant performance variations across frontier models (single digits to ~80% accuracy), with notable error modes including tool use failures (36% of conversations) and hallucinations from pretrained domain knowledge, particularly from OpenAI models which hallucinated non-existent insurance products 15-45% of the time.
Articul8 developed a generative AI platform to address enterprise challenges in manufacturing and supply chain management, particularly for a European automotive manufacturer. The platform combines public AI models with domain-specific intelligence and proprietary data to create a comprehensive knowledge graph from vast amounts of unstructured data. The solution reduced incident response time from 90 seconds to 30 seconds (3x improvement) and enabled automated root cause analysis for manufacturing defects, helping experts disseminate daily incidents and optimize production processes that previously required manual analysis by experienced engineers.
Notion AI, serving over 100 million users with multiple AI features including meeting notes, enterprise search, and deep research tools, demonstrates how rigorous evaluation and observability practices are essential for scaling AI product development. The company uses Brain Trust as their evaluation platform to manage the complexity of supporting multilingual workspaces, rapid model switching, and maintaining product polish while building at the speed of AI industry innovation. Their approach emphasizes that 90% of AI development time should be spent on evaluation and observability rather than prompting, with specialized data specialists creating targeted datasets and custom LLM-as-a-judge scoring functions to ensure consistent quality across their diverse AI product suite.