Healthcare
National Healthcare Group
Company
National Healthcare Group
Title
Implementing LLMs for Patient Education and Healthcare Communication
Industry
Healthcare
Year
2024
Summary (short)
National Healthcare Group addressed the challenge of inconsistent and time-consuming patient education by implementing LLM-powered chatbots integrated into their existing healthcare apps and messaging platforms. The solution provides 24/7 multilingual patient education, focusing on conditions like eczema and medical test preparation, while ensuring privacy and accuracy. The implementation emphasizes integration with existing platforms rather than creating new standalone solutions, combined with careful monitoring and refinement of responses.
## Overview This case study, presented at the NLP Summit 2024 by a medical doctor and health services researcher working as a clinical advisor for the AI Lab at National Healthcare Group in Singapore, explores the practical application of large language models for patient education in healthcare settings. The presenter brings a unique perspective as a practicing clinician rather than a data scientist, focusing on real-world implementation challenges and operational considerations for deploying LLMs in patient-facing healthcare applications. The core thesis is that patient education represents the ideal entry point for LLMs in healthcare because it sidesteps many of the more challenging barriers present in clinical applications—namely the need for clinician buy-in, concerns about patient confidentiality, and the higher stakes of clinical decision-making. The speaker argues that patient education materials are abundant, easily accessible, non-confidential, and don't require direct physician approval, making them a lower-risk proving ground for healthcare LLM implementations. ## Problem Statement and Motivation Traditional patient education faces several operational challenges that LLMs can potentially address. Time constraints during clinical consultations mean doctors often resort to printing materials that patients frequently discard. Information inconsistency across different providers—doctors, dieticians, nurses—can confuse patients and lead to suboptimal health decisions. Medical jargon creates literacy barriers for patients without higher education. Additionally, maintaining educational materials in multiple languages (the presenter showed an example with English, Malay, Chinese, and Tamil versions of diabetes materials in Singapore) is extremely resource-intensive, traditionally requiring two-way translation validation processes. The speaker emphasizes the business case for better patient education: it leads to improved health outcomes through better treatment adherence, reduced healthcare costs through fewer unnecessary hospital visits, patient empowerment in shared decision-making, and enhanced patient satisfaction (which helps reduce complaints and medical-legal issues). ## Technical Approach and Architecture The implementation strategy focuses heavily on integration rather than building standalone systems. A key operational principle articulated is "do not reinvent the wheel"—the recommendation is to integrate LLM capabilities into existing healthcare apps and portals rather than creating new platforms that patients won't adopt. For National Healthcare Group, this means embedding LLM functionality within their existing health app and using WhatsApp as a communication channel rather than developing a proprietary messaging solution. The speaker mentions several LLM options, including OpenAI's ChatGPT, Claude, and open-source models like Mistral 7B that can be deployed on-premise if organizations have sufficient computational resources. The choice of on-premise versus cloud deployment has implications for data privacy and processing speed, though the speaker notes that open-source models require significant processing power to perform adequately. For translation workflows, LLMs have transformed what was previously a labor-intensive process. The traditional gold standard of two-way translation (English to Malay, then Malay back to English to verify accuracy) can now be automated and extended to multiple language chains (English to Malay to Chinese to Tamil and back) for validation. This represents a significant operational efficiency gain for organizations maintaining multilingual patient materials. ## Production Deployment Considerations The presentation outlines a comprehensive operational framework for deploying LLMs in patient education: **Training and Adoption**: Staff need training on how to use the systems, and providers need guidance on explaining the chatbot to patients. This human-centric approach to deployment acknowledges that technology adoption requires organizational change management, not just technical implementation. **Monitoring and Metrics**: The speaker emphasizes the importance of tracking whether systems are actually being used and measuring their effectiveness. In an environment where resources are scrutinized ("is the money well spent"), demonstrating value through usage metrics and outcomes is essential for continued investment. **Iterative Refinement**: The implementation is framed as an ongoing process, not a one-time deployment. Collecting patient feedback, analyzing top FAQs generated from user interactions, and continuously refining responses are all part of the operational model. This feedback loop allows the system to identify gaps in existing educational materials—if patients are repeatedly asking questions not covered by current resources, that signals a content gap to address. **Human-in-the-Loop Validation**: Currently, responses from the LLM are manually reviewed to ensure accuracy. This is particularly important given the stakes of health information—even educational materials need to be correct. The speaker acknowledges this creates operational overhead but considers it necessary at the current stage of LLM maturity in healthcare. ## Data Privacy and Safety Considerations A significant portion of the operational framework addresses data privacy, which is particularly sensitive in healthcare. The key insight is that patient education chatbots should be designed to provide information rather than collect it. The system should discourage or prevent patients from inputting sensitive health information such as diagnoses (e.g., "I have HIV"), lab results, or medical records. This is an interesting architectural decision that shapes the entire system design—rather than building a personalized health assistant that requires patient data, the goal is an educational resource that maintains user privacy by design. This approach reduces regulatory burden and liability while still providing value. ## Accuracy and Liability The speaker addresses the challenge of ensuring LLM-generated health information is accurate and up-to-date. Healthcare information has liability implications, so chatbots include disclaimers directing users to consult medical professionals for personalized advice. The manual review process for LLM outputs represents the current approach to quality assurance, though this may evolve as confidence in LLM accuracy increases. An interesting observation from the presentation is that studies have shown ChatGPT can provide more empathetic responses than burned-out human clinicians. However, empathy must be balanced with appropriate messaging—for terminal illness queries, the system should provide hope without creating false expectations, acknowledging that outcomes vary and new treatments emerge. ## Current Projects Two concrete projects are mentioned as being in development or seeking funding: The **eczema patient education project** is being developed in collaboration with clinicians from the National Skin Centre. It will integrate with messaging platforms (WhatsApp, Telegram, Facebook, or LINE depending on the country) to provide information about eczema causes, avoidance triggers, symptom monitoring, and when to seek medical care. The **lung function test preparation project** focuses on procedural guidance rather than condition education. It addresses the complex instructions patients receive before diagnostic tests—such as medication restrictions before spirometry or the multi-step bowel preparation regimen for colonoscopy. These instructions are often provided on paper and can be confusing; an LLM-based system can provide interactive, clarifying explanations. ## Future Directions The presentation outlines several forward-looking capabilities that would extend the current text-based approach: **Multimodal Integration**: Incorporating voice (speech-to-text and text-to-speech) would improve accessibility, particularly for visually impaired users who could ask questions verbally and receive spoken responses. **Personalized Health Coaching**: Moving beyond passive information provision to active, actionable guidance based on lifestyle and habits. This would require more sophisticated personalization while maintaining privacy boundaries. **Real-time Health Monitoring Integration**: Connecting with wearable devices (Fitbit, smartwatches) to provide contextual advice based on activity levels, heart rate, and other biometric data. This could enable proactive interventions like alerts when step counts drop or potential fall detection. ## Balanced Assessment It's worth noting that this presentation describes projects in various stages of development rather than fully deployed production systems with measured outcomes. The eczema project is described as having recently applied for a grant, suggesting it's still in early stages. The benefits articulated (improved outcomes, reduced costs, enhanced satisfaction) are projected or anticipated rather than measured results from live deployments. The speaker acknowledges significant challenges that remain unresolved, including ensuring accuracy without excessive manual review overhead, managing liability, and maintaining the human touch in AI-mediated patient communication. The recommendation to start with patient education rather than clinical applications is pragmatic but also reflects the current limitations of LLM reliability for higher-stakes medical use cases. The emphasis on integration with existing platforms and iterative refinement represents sound operational thinking, recognizing that successful healthcare AI deployment requires organizational change management alongside technical implementation. The privacy-by-design approach—building systems that provide information without collecting sensitive patient data—is a thoughtful architectural decision that simplifies compliance and reduces risk.

Start deploying reproducible AI workflows today

Enterprise-grade MLOps platform trusted by thousands of companies in production.