ZenML

Unlocking Patient Population Insights Using Smart Subgroups and LLMs

Aetion 2025
View original source

Aetion developed a system to help healthcare researchers discover patterns in patient populations using natural language queries. The solution combines unsupervised machine learning for patient clustering with Amazon Bedrock and Claude 3 LLMs to enable natural language interaction with the data. This allows users unfamiliar with real-world healthcare data to quickly discover patterns and generate hypotheses, reducing analysis time from days to minutes while maintaining scientific rigor.

Industry

Healthcare

Technologies

Overview

Aetion is a healthcare technology company that provides real-world evidence software to biopharma companies, payors, and regulatory agencies. Their platform transforms real-world patient data into actionable evidence for evaluating the safety, effectiveness, and value of medical interventions. The company serves top pharmaceutical companies, regulatory agencies including the FDA and EMA, and operates across the US, Canada, Europe, and Japan.

The case study describes how Aetion developed the Smart Subgroups Interpreter, a generative AI feature within their Discover product that allows users to interact with patient population clusters using natural language. This represents a practical application of LLMs in a production healthcare analytics environment, where the goal is to democratize access to complex patient data insights.

Problem Statement

Real-world patient data contains a wealth of information about patient characteristics, outcomes, and treatment effectiveness. However, researchers traditionally need deep familiarity with complex datasets and must express queries in structured formats to extract meaningful insights. This creates barriers for users who may not have extensive technical expertise with the underlying data structures, and many valuable patterns and trends can remain undiscovered as a result.

Aetion’s challenge was to make patient population analysis more accessible by enabling natural language interaction with clustered patient data, reducing the time and expertise required to generate hypotheses and evidence from real-world data.

Technical Architecture and LLM Integration

The Smart Subgroups Interpreter combines unsupervised machine learning with generative AI in a multi-step pipeline deployed on AWS infrastructure.

Data Processing Pipeline

The workflow begins with users defining a patient population using the Aetion Measure Library (AML), which standardizes variable definitions using scientifically validated algorithms. The Aetion Evidence Platform (AEP) then computes over 1,000 AML features for each patient across categories including diagnoses, therapies, and procedures. These features are stored in Amazon S3 with AWS KMS encryption.

The Smart Subgroups component trains a topic model using these patient features to determine optimal cluster numbers and assign patients to clusters. A trained classification model identifies the most distinctive features within each cluster, and their prevalences are used to describe cluster characteristics. This component is deployed as a Kubernetes job that runs on demand.

LLM Selection and Integration

Aetion selected Amazon Bedrock as their LLM platform due to its model selection from multiple providers, security posture, extensibility, and ease of use. Within Amazon Bedrock, they chose Anthropic’s Claude 3 Haiku specifically for its speed and what they describe as a “satisfactory intelligence level” for their use case. This choice reflects a pragmatic production consideration: balancing model capability with latency requirements for interactive user experiences.

The solution architecture includes:

Prompt Engineering Approach

The system uses what Aetion describes as “composite prompt engineering techniques” to generate responses. This involves versioned prompt templates that combine multiple information sources:

This approach represents a form of structured data augmentation in prompts, where domain-specific scientific knowledge and computed statistics are systematically injected into the LLM context to enable accurate, domain-relevant responses.

User Interaction Patterns

The system supports several types of natural language queries:

The LLM responses cite relevant subgroup statistics, grounding the natural language output in the underlying data.

Production Deployment Considerations

Several LLMOps practices are evident in this implementation:

Infrastructure Choices

The use of Kubernetes for deployment enables scalability and portability across environments. Running the clustering component as on-demand Kubernetes jobs suggests an event-driven architecture that scales resources based on user requests rather than maintaining constant compute capacity.

Versioned Prompt Templates

The explicit mention of “versioned prompt templates” indicates that Aetion treats prompts as managed artifacts, likely enabling iteration, testing, and rollback capabilities. This is a mature LLMOps practice that recognizes prompts as critical components of the production system.

Security Architecture

Healthcare data requires stringent security controls. The architecture incorporates:

Integration with Existing Platform

The LLM capability is embedded within Aetion’s broader Evidence Platform rather than deployed as a standalone application. This integration pattern, referred to as “AetionAI,” represents an approach where generative AI enhances existing workflows rather than replacing them. Users can move from AI-powered exploration in Discover to rigorous causal inference analysis in Aetion Substantiate.

Outcomes and Claims

Aetion reports that the Smart Subgroups Interpreter enables users unfamiliar with real-world data to discover patterns using natural language. The claimed benefit is reducing the time to generate decision-grade evidence from days to minutes, without requiring support staff.

It’s worth noting that these outcomes are self-reported by Aetion and AWS in what is essentially a promotional blog post. The claims about time savings (from days to minutes) should be understood as marketing assertions rather than independently verified metrics. Additionally, the term “decision-grade evidence” is Aetion’s own terminology and may carry specific meaning within their product framework that differs from clinical or regulatory definitions.

Technical Observations

The architecture demonstrates several interesting patterns for healthcare LLM applications:

The combination of unsupervised learning (clustering) with LLMs for interpretation represents a hybrid approach where traditional ML handles the analytical heavy lifting while the LLM serves as an interface layer. This division of labor may provide more reliable results than asking an LLM to perform the clustering analysis directly.

The feature expansion technique—where coded medical features are enriched with detailed descriptions before being sent to the LLM—addresses the challenge of domain-specific terminology and provides the model with sufficient context to generate medically meaningful responses.

The choice of Claude 3 Haiku over more capable models in the Claude family suggests optimization for latency in an interactive use case. This is a practical production trade-off that many organizations face when deploying LLMs in user-facing applications.

Limitations and Considerations

The case study does not address several important LLMOps concerns:

These gaps may simply reflect the scope of the blog post rather than actual limitations of the implementation, but they represent areas that production healthcare AI systems typically need to address carefully.

More Like This

Building a Microservices-Based Multi-Agent Platform for Financial Advisors

Prudential 2025

Prudential Financial, in partnership with AWS GenAI Innovation Center, built a scalable multi-agent platform to support 100,000+ financial advisors across insurance and financial services. The system addresses fragmented workflows where advisors previously had to navigate dozens of disconnected IT systems for client engagement, underwriting, product information, and servicing. The solution features an orchestration agent that routes requests to specialized sub-agents (quick quote, forms, product, illustration, book of business) while maintaining context and enforcing governance. The platform-based microservices architecture reduced time-to-value from 6-8 weeks to 3-4 weeks for new agent deployments, enabled cross-business reusability, and provided standardized frameworks for authentication, LLM gateway access, knowledge management, and observability while handling the complexity of scaling multi-agent systems in a regulated financial services environment.

healthcare fraud_detection customer_support +48

AI-Powered Clinical Outcome Assessment Review Using Generative AI

Clario 2025

Clario, a clinical trials endpoint data provider, developed an AI-powered solution to automate the analysis of Clinical Outcome Assessment (COA) interviews in clinical trials for psychosis, anxiety, and mood disorders. The traditional approach of manually reviewing audio-video recordings was time-consuming, logistically complex, and introduced variability that could compromise trial reliability. Using Amazon Bedrock and other AWS services, Clario built a system that performs speaker diarization, multi-lingual transcription, semantic search, and agentic AI-powered quality review to evaluate interviews against standardized criteria. The solution demonstrates potential for reducing manual review effort by over 90%, providing 100% data coverage versus subset sampling, and decreasing review turnaround time from weeks to hours, while maintaining regulatory compliance and improving data quality for submissions.

healthcare regulatory_compliance high_stakes_application +28

AI-Powered Social Intelligence for Life Sciences

Indegene 2025

Indegene developed an AI-powered social intelligence solution to help pharmaceutical companies extract insights from digital healthcare conversations on social media. The solution addresses the challenge that 52% of healthcare professionals now prefer receiving medical content through social channels, while the life sciences industry struggles with analyzing complex medical discussions at scale. Using Amazon Bedrock, SageMaker, and other AWS services, the platform provides healthcare-focused analytics including HCP identification, sentiment analysis, brand monitoring, and adverse event detection. The layered architecture delivers measurable improvements in time-to-insight generation and operational cost savings while maintaining regulatory compliance.

healthcare content_moderation classification +39