Tech
Palo Alto Networks
Company
Palo Alto Networks
Title
Automated Log Classification System for Device Security Infrastructure
Industry
Tech
Year
2026
Summary (short)
Palo Alto Networks' Device Security team faced challenges with reactively processing over 200 million daily service and application log entries, resulting in delayed response times to critical production issues. In partnership with AWS Generative AI Innovation Center, they developed an automated log classification pipeline powered by Amazon Bedrock using Anthropic's Claude Haiku model and Amazon Titan Text Embeddings. The solution achieved 95% precision in detecting production issues while reducing incident response times by 83%, transforming reactive log monitoring into proactive issue detection through intelligent caching, context-aware classification, and dynamic few-shot learning.
## Overview Palo Alto Networks' Device Security team, which provides Cloud-Delivered Security Services (CDSS) using machine learning and automated discovery to enforce Zero Trust principles, faced a significant operational challenge in managing their infrastructure logs. The team needed to detect early warning signs of potential production issues from a massive volume of over 200 million daily service and application log entries. The reactive nature of their existing log processing approach resulted in delayed response times to critical issues, leaving the team at risk for potential service degradation and multi-week outages that could impact customers. To address this operational challenge, Palo Alto Networks partnered with the AWS Generative AI Innovation Center (GenAIIC) to develop an automated log classification pipeline powered by Amazon Bedrock. The solution represents a comprehensive production implementation of LLM-based log analysis that balances scale, accuracy, and cost-efficiency while integrating seamlessly with existing infrastructure. ## Production Architecture and LLMOps Implementation The solution architecture demonstrates sophisticated LLMOps practices through a three-stage processing pipeline that handles the massive daily log volume. The system integrates with Palo Alto Networks' existing FluentD and Kafka pipeline for data ingestion, processes logs through multiple stages of analysis, and outputs results to Amazon S3 and Amazon Redshift for further analysis and reporting. The three-stage processing pipeline represents the core LLMOps innovation. **Stage 1** implements smart caching and deduplication using Amazon Aurora for exact matching and Amazon Titan Text Embeddings for semantic matching. This stage addresses the fundamental scalability challenge of LLM-based systems by intelligently identifying duplicate log entries for the same code event. Rather than invoking an LLM to classify every log individually, the system employs a cascading approach: exact matching first, then overlap similarity, and finally semantic similarity only if no earlier match is found. This design achieved over 99% cache hit rates during testing, meaning that more than 99% of the 200 million daily logs corresponded to duplicate events despite containing different timestamps, log levels, and phrasing variations. **Stage 2** handles context retrieval for the remaining less than 1% of truly unique logs. This stage uses Amazon Titan Text Embeddings to perform vector similarity search against Palo Alto Networks' labeled historical dataset. Rather than using static few-shot examples in prompts, the system dynamically retrieves the most contextually relevant historical examples for each unique log. This approach ensures that each log receives appropriate guidance for classification based on similar past cases, improving accuracy for Palo Alto Networks' internal systems and evolving log patterns that traditional rule-based systems struggle to handle. **Stage 3** performs the actual classification using Amazon Bedrock with Anthropic's Claude Haiku model. Unique logs are processed alongside their dynamically selected historical examples to produce structured predictions including severity classification (Priority 1/P1 for critical issues, Priority 2/P2 for high priority, and Priority 3/P3 for low priority) along with detailed reasoning for each decision. The model uses carefully crafted prompts that combine domain expertise with the dynamically selected examples, helping subject matter experts (SMEs) quickly prioritize responses and take preventive action before potential outages occur. ## Prompt Engineering and Model Selection The prompt engineering approach demonstrates production-grade practices for log classification. The system prompt establishes clear task definition, severity definitions tied to operational impact, provides example classifications with reasoning, and structures the target log input. The prompt explicitly defines P1 as requiring immediate action for system-wide outages and repeated application crashes, P2 as warranting attention during business hours for performance issues and partial service disruption, and P3 as addressable when resources are available for minor bugs and intermittent issues. The choice of Anthropic's Claude Haiku model represents a deliberate LLMOps decision balancing cost, performance, and latency requirements. While the case study doesn't explicitly discuss model selection criteria, Haiku is positioned as a faster, more cost-effective model suitable for high-volume classification tasks. The structured output requirements (severity classification plus reasoning) align well with Haiku's capabilities for instruction-following and structured generation. ## Caching Strategy and Cost Optimization The intelligent caching architecture represents one of the most significant LLMOps innovations in this implementation. By reducing LLM invocations by over 99%, the system makes AI processing economically viable at enterprise scale. The multi-layered approach (exact matching, overlap similarity, semantic similarity) progressively narrows down logs that require full LLM processing, with each layer providing increasingly sophisticated but more expensive matching. The caching system uses Amazon Aurora as the storage layer, enabling subsecond responses for cached results. This architecture choice supports the real-time processing requirements while maintaining the ability to quickly lookup historical classifications. The cache becomes increasingly valuable over time as more logs are classified and stored, creating a compounding effect where operational use enhances system performance and reduces costs. ## Dynamic Few-Shot Learning and Continuous Improvement The dynamic few-shot retrieval system represents a sophisticated approach to handling evolving log patterns without requiring system modifications. Rather than hardcoding example logs in prompts, the system maintains a growing dataset of labeled examples that SMEs validate over time. When processing a unique log, vector similarity search identifies the most relevant historical examples to include as few-shot demonstrations in the prompt. This approach provides several LLMOps advantages. The system automatically improves as SMEs validate classifications and label new examples, creating a continuous learning cycle. Each validated classification becomes part of the retrieval dataset, improving accuracy for similar future logs while increasing cache hit rates. The system accommodates new log categories and patterns without code changes—when performance needs improvement for novel log types, SMEs simply label additional examples and the dynamic retrieval automatically incorporates this knowledge. The vector similarity search uses Amazon Titan Text Embeddings, which generates semantic representations of log entries that capture meaning beyond keyword matching. This enables the system to identify relevant examples even when logs use different phrasing or terminology to describe similar issues. ## Integration with Existing Infrastructure The solution demonstrates practical LLMOps implementation by integrating seamlessly with Palo Alto Networks' existing data infrastructure. The system receives logs from their established FluentD and Kafka pipeline, processes them through the classification stages, and outputs results back into their data ecosystem through Amazon S3 and Amazon Redshift. This integration approach minimizes disruption to existing workflows while adding AI-powered capabilities. The SME review interface enables subject matter experts to validate classifications, which serves dual purposes: providing operational oversight for critical decisions and generating labeled data for continuous system improvement. This human-in-the-loop approach balances automation with appropriate human validation, particularly important for P1 severity classifications that require immediate action. ## Production Results and Operational Impact The system achieved 95% precision and 90% recall specifically for P1 severity logs, demonstrating strong performance on the most critical classification task. These metrics indicate that when the system flags a log as P1, it's correct 95% of the time, and it successfully identifies 9 out of 10 truly critical issues. The remaining critical issues can be caught by existing monitoring systems, providing defense in depth. The 83% reduction in debugging time represents significant operational improvement, allowing SMEs to focus on strategic improvements rather than routine log analysis. The system enables proactive issue detection, identifying potential problems before they impact customers and preventing the multi-week outages that previously disrupted service. The over 99% cache hit rate confirms the effectiveness of the intelligent caching strategy in handling the 200 million daily log volume cost-effectively. ## LLMOps Lessons and Design Principles Several important LLMOps principles emerge from this implementation. The continuous learning design creates compounding value as each SME validation improves future performance. The intelligent caching architecture makes AI processing economically viable at production scale by reducing expensive LLM operations to a small fraction of total volume. The adaptive system handles evolving requirements without code changes through dynamic few-shot learning. The explainable classifications with detailed reasoning drive operational confidence, transforming AI outputs from predictions into actionable intelligence. ## Balanced Assessment While the case study presents impressive results, several considerations warrant attention. The 95% precision and 90% recall figures specifically apply to P1 severity logs, and the case study doesn't provide detailed performance metrics for P2 and P3 classifications. The over 99% cache hit rate depends on log patterns remaining relatively stable—significant changes in application behavior or new service deployments might temporarily reduce cache effectiveness until sufficient new examples are classified and cached. The dynamic few-shot learning approach requires ongoing SME involvement to label examples and validate classifications. While this creates a continuous improvement cycle, it also represents an operational cost that organizations must account for. The system's effectiveness depends heavily on the quality and representativeness of labeled examples in the retrieval dataset. The choice of Amazon Titan Text Embeddings and Claude Haiku represents vendor lock-in to AWS and Anthropic services. While this integration provides operational benefits, organizations should consider portability and vendor dependency in their architectural decisions. The case study doesn't discuss fallback mechanisms if the LLM service experiences issues or latency spikes. The semantic similarity matching in the caching layer and the vector similarity search for few-shot example retrieval both depend on embedding quality. The effectiveness of Amazon Titan Text Embeddings for Palo Alto Networks' specific log formats and technical terminology may vary, though the results suggest good performance for their use case. ## Conclusion Palo Alto Networks' automated log classification system demonstrates mature LLMOps practices for deploying LLMs in production at scale. The architecture balances multiple competing concerns—handling massive volume, maintaining accuracy, controlling costs, and enabling continuous improvement—through intelligent caching, dynamic few-shot learning, and thoughtful integration with existing infrastructure. The system transforms reactive log monitoring into proactive issue detection while delivering measurable operational improvements in precision, recall, response time, and cost efficiency. For organizations considering similar implementations, this case study provides a practical blueprint showing how careful architectural design can make LLM-based systems economically viable for high-volume operational use cases.

Start deploying reproducible AI workflows today

Enterprise-grade MLOps platform trusted by thousands of companies in production.