Education
UC Santa Barbara
Company
UC Santa Barbara
Title
Scaling AI-Powered Student Support Chatbots Across Campus
Industry
Education
Year
2025
Summary (short)
UC Santa Barbara implemented an AI-powered chatbot platform called "Story" (powered by Gravity's Ivy and Ocelot services) to address challenges in student support after COVID-19, particularly helping students navigate campus services and reducing staff workload. Starting with a pilot of five departments in 2022, UCSB scaled to 19 chatbot instances across diverse student services over two and a half years. The implementation resulted in nearly 40,000 conversations, with 30% occurring outside business hours, significantly reducing phone and email volume to departments while enabling staff to focus on more complex student inquiries. The university took a phased cohort approach, training departments in groups over 10-week periods, with student testers providing crucial feedback on language and expectations before launch.
## Overview UC Santa Barbara's implementation of AI-powered chatbots represents a comprehensive case study in scaling LLM-based conversational AI across a large higher education institution. The university, which serves approximately 25,000 undergraduate students across over 200 majors, partnered with Gravity (using their Ivy and Ocelot platforms) to deploy a chatbot named "Story" (after Storke Tower on campus). The initiative began in 2022 as part of a work group focused on student return to campus after COVID-19, addressing the challenge of helping two entering classes of students who had never physically navigated the campus. Over two and a half years, the deployment grew from five pilot departments to 19 chatbot instances, generating nearly 40,000 conversations. ## Problem Context and Motivation The university faced several interconnected challenges that made an AI chatbot solution attractive. With student services distributed across many different buildings on a large campus, students frequently were sent from one building to another trying to find answers. Staff members reported being overloaded with questions, many of which were already answered on dense, information-rich websites that students struggled to navigate. The university operated with limited staffing capacity but wanted to ensure that students with complex, nuanced questions could still reach the right person for in-depth support. Additionally, students had evolving expectations around 24/7 access to information, with questions arising at 2 AM or on Saturday mornings—times when traditional office hours couldn't accommodate them. Email fatigue was also identified as a significant issue, with students receiving dozens of emails per week from campus. ## Technical Architecture and Platform Capabilities The Gravity platform leverages a sophisticated technical architecture that combines multiple AI and operational capabilities. At its core, the system uses Retrieval-Augmented Generation (RAG) to crawl and index university websites every 24 hours, automatically updating the knowledge base when website content changes. This automatic web crawling was identified as a critical feature that reduced the manual effort required from departments, as some alternative solutions would have required staff to manually build the entire knowledge base. The platform implements a confidence scoring system that only provides answers when it achieves above 85% confidence in the response accuracy. When confidence is below this threshold, the system can be configured to either provide "low confidence responses" with links to relevant resources or immediately escalate to human support. The Student Health department at UCSB, for example, chose to disable low confidence responses entirely and route uncertain queries directly to staff with an intake form for follow-up. The AI system uses advanced natural language processing to handle variations in how students phrase questions, including slang terminology and informal language. According to the platform provider, the system achieves approximately 98% accuracy in understanding different terminologies and question formulations. The system employs generative AI as a fallback mechanism—when the primary knowledge base doesn't contain a high-confidence answer, it queries large language models using context from the website to attempt to reformulate and answer the question, always subject to confidence thresholds and guardrails. ## Security, Privacy, and Compliance Security and privacy were critical considerations throughout the deployment. The platform is SOC 2 compliant and implements comprehensive personally identifiable information (PII) scrubbing to ensure that no sensitive student data is transmitted to external large language models. The system operates within what the university described as a "closed AI" or "AI within a safety net" model—the chatbot can only provide information that has been explicitly trained into it from approved websites or custom responses, preventing it from "going rogue" and generating unauthorized content. The platform includes threat detection and monitoring capabilities, with configurable thresholds for threat words or concerning situations. When potential issues are detected, staff receive instant notifications. All chat transcripts are logged and auditable, allowing the university to trace any answer back to its source—either a specific webpage or a custom response created by staff. This auditability was highlighted as particularly important for addressing concerns about chatbot accuracy, as staff could always identify why a particular answer was provided and correct it for future interactions. ## Implementation Methodology and Scaling Strategy UCSB took a deliberately phased approach to implementation, which proved crucial to their success. The initial pilot included five strategically selected departments representing diverse student populations and service types: Orientation Programs (serving new students and parents), Registrar's Office (core academic enrollment questions), Student Health (high phone call volume), Educational Opportunity Program (serving first-generation and income-eligible students), and a Well-being website that aggregated information from multiple departments. The selection criteria for pilot departments included having websites that were relatively up-to-date or could be updated quickly, and department readiness to commit to the training and ongoing maintenance process. Trisha Rasone, the Assistant Dean for Student Life who served as project manager, emphasized being honest with departments that maintaining the chatbot would require work—particularly in the early "infant" stages—with approximately two one-hour sessions per week recommended initially for reviewing and training the bot. After the successful pilot, UCSB adopted a cohort model for scaling, bringing on groups of four to five departments together and guiding them through a structured 10-week implementation process aligned with the university's quarter system. This cohort approach allowed departments to learn from each other, test each other's chatbots, and share best practices. The university found this peer learning environment more effective than one-off implementations. Rasone served as a centralized project manager interfacing between departments and Gravity, while each department committed one to two staff members to complete training and be responsible for their specific chatbot instance. ## Knowledge Management and the "Shared Brain" Concept A particularly sophisticated aspect of the implementation is the "shared brain" architecture. Each department builds its own library of questions and answers, but can selectively share content with other departments' chatbot instances. This one-to-many model means that if the Financial Aid office creates comprehensive responses about aid eligibility, the Admissions chatbot can inherit those answers without recreating them. However, departments retain control over what they share—for example, "When do we open?" might be relevant only to specific offices. The shared brain concept proved especially important as the deployment scaled. With 19 departments now represented, students could ask questions from any entry point and potentially receive answers drawing from multiple departmental knowledge bases. The university discovered challenges with terminology—the same word might mean different things to different departments (e.g., "counselor" in academic advising vs. psychological counseling). To address this, UCSB developed a standardized list of shared terminology and requires each department to create custom responses defining what specific terms mean in their context, ensuring the bot provides department-appropriate answers. The platform also comes with content packs—approximately 20 years of curated higher education data that institutions can activate to give their chatbots a more comprehensive baseline knowledge from day one. ## Student Testing and Language Optimization One of the most important implementation lessons involved student testing. Initially, departments wanted to test chatbots internally with colleagues before launch. However, the university quickly established a formal requirement for student testers to interact with bots before deployment. This revealed critical insights: students use very different language than administrators and staff, often employing informal terms or slang. Students also had different expectations about what departments should know—for example, the Registrar's Office received many questions about billing and financial aid because students associated registration blocks with billing issues, even though these topics weren't in the Registrar's domain. Through student testing, departments learned to add more comprehensive redirections to other services, train the bot on variant phrasings, and set realistic expectations about the bot's scope. The university found that this upfront investment in understanding student language patterns significantly improved accuracy and satisfaction once bots went live. ## User Interface and Pathway Design The chatbot interface includes a welcome message with prominent "pathway buttons" that provide quick access to the top seven to eight most common questions for each department. When clicked, these buttons can lead to immediate answers, additional nested buttons for more detailed topics, or links to relevant website sections. Departments identified these pathway topics by brainstorming "the top 10 basic questions we get asked where the information is on our front page"—queries that consume staff time but are straightforward to answer. Analytics showed that pathway buttons are heavily utilized across bots, providing efficient self-service for common inquiries. However, students can also type free-form questions, and the system will process them using the full NLP and RAG capabilities. This dual approach accommodates both students who know roughly what they're looking for and benefit from guided options, and those who prefer conversational interaction. ## Multi-Channel Support: Live Chat and SMS The platform supports multiple communication channels beyond the web-based chatbot. Departments can optionally enable live chat, allowing students to seamlessly transition from chatbot to human agent when needed. Importantly, this feature is entirely optional—it was highlighted that departments hesitant about staffing live chat could launch with just the automated chatbot and add live chat later. This flexibility was crucial for gaining buy-in from resource-constrained departments. SMS messaging emerged as a particularly powerful channel. UCSB initially piloted SMS with a targeted campaign for Advising Day, where 75 students who hadn't registered received text messages directing them to the orientation website. Within 48 hours over a weekend, approximately 30% signed up and attended—a dramatically higher response rate than email campaigns had achieved. This success led to plans for broader SMS deployment, including wellness check-ins from the Dean of Students where students can text back to the chatbot for two-way conversations. The platform supports both one-way SMS broadcasts and two-way "bot-backed campaigns" where students can respond and receive automated answers through decision trees, reducing the need for staff to manually respond to thousands of texts. The SMS functionality addresses the email fatigue problem while meeting students in their preferred communication channel. ## Analytics, Monitoring, and Continuous Improvement The platform provides comprehensive analytics that inform both operational improvements and strategic decisions. Key analytics include pathway button usage tracking, cluster analysis showing the most frequently asked questions about specific topics and how they're phrased, and comparative reports showing how a department's top 100 questions and answers compare to similar departments at other institutions using the Gravity platform. Staff regularly review actual chat transcripts to understand how students are asking questions and identify gaps in the knowledge base. The university established a practice of allocating calendar time each week to review random chatbot conversations, building this into the ongoing maintenance workflow. Departments also learned to update welcome message banners with timely information (e.g., "Registration opens today" or "Orientation reservations are now live") based on predictable seasonal question patterns. The analytics revealed significant patterns in usage timing, with 30% of conversations occurring outside traditional business hours and on weekends. Some departments like Letters and Science Advising and Student Health saw even higher percentages, confirming that the chatbot was filling a genuine gap in service availability. ## Organizational Structure and Governance While the chatbot platform is housed within the Student Affairs division, UCSB deliberately positioned it as a campus-wide resource rather than a divisional tool. Rasone emphasized that the university doesn't have a dedicated full-time chatbot team—she serves as project manager as one component of her broader role, and each participating department has designated staff who manage their specific instance. The university established a quarterly meeting cadence where all department chatbot administrators convene to share experiences, discuss challenges, and learn about new features. These administrators serve as ambassadors and advocates for the chatbot within their broader campus networks. The IT department was involved in the contract approval process and initial training to understand the technical architecture, but their ongoing operational role is limited primarily to adding chatbot widgets to websites—the actual content management and maintenance is handled by student services staff. This organizational model proved crucial for demonstrating that an AI chatbot platform could be successfully managed without extensive technical expertise. Rasone explicitly noted that she comes from a student services background rather than IT, and that the vendor provides sufficient technical support and backend AI capabilities that departments can focus on content and student experience rather than technical infrastructure. ## Impact and Outcomes After two and a half years of operation across 19 departments, the implementation has generated nearly 40,000 conversations. Specific departments reported measurable impacts: Orientation Programs saw decreased phone and email traffic during their critical eight-week registration windows, with remaining inquiries being more nuanced questions requiring human expertise. Student Health reported similar reductions in phone volume, particularly for routine questions like "How do I make an appointment?" that were previously consuming staff time despite being clearly documented on the website. The 30% of usage occurring outside business hours represents thousands of student interactions that would otherwise have been delayed until offices reopened, potentially creating frustration and workflow bottlenecks. Staff members reported being able to focus more attention on complex, individualized student needs rather than repetitive basic inquiries. The chatbot also generated unexpected value in identifying website content issues. When departments saw questions the bot couldn't answer well, they realized their websites were missing important information or had outdated content. In one case, the bot pulled incorrect information that turned out to be from another department's website that referenced the first department—enabling targeted corrections that wouldn't have been discovered otherwise. The Orientation department used chatbot analytics to inform a complete website redesign, prioritizing visibility for content that generated the most questions. ## Technical Roadmap and Future Capabilities The vendor discussed ongoing development efforts focused on reducing the manual maintenance burden through increased automation. Current pilot programs are exploring automated content refreshing where the system uses RAG and AI to identify low-confidence responses, research better answers from available sources, and present them to staff for approval rather than requiring staff to manually create all custom responses. This "human-centered AI" approach maintains staff control while dramatically reducing the time investment. Other capabilities on the roadmap include predictive analytics to identify content gaps and process inefficiencies, AI-powered campaign creation where staff can describe a desired outreach campaign and the system generates appropriate content and targeting, and deeper integrations with Student Information Systems (SIS) and Customer Relationship Management (CRM) platforms. The system already supports API-based integrations with campus systems and custom webhooks for department-specific workflows like financial aid calculations, but these capabilities are positioned as growth opportunities rather than requirements for initial deployment. ## Balanced Assessment and Considerations While the case study presents a largely positive narrative, several important caveats and challenges merit consideration. The university acknowledged that chatbot accuracy concerns are valid—any AI system will sometimes provide incorrect or incomplete answers. However, they frame this in context: students currently receive varying quality answers when asking questions at different service desks around campus, with no systematic way to track or correct misinformation. The chatbot's auditability and traceability actually provide more quality control than many existing support channels. The "closed AI" approach using RAG and confidence thresholds reduces but doesn't eliminate the risk of hallucinations or inappropriate responses. The 85% confidence threshold and PII scrubbing represent operational guardrails, but departments still need active monitoring and maintenance. The recommendation of two hours per week initially, scaling down over time, represents a real ongoing cost that resource-constrained institutions must consider. The cohort implementation model and the emphasis on student testing add time and complexity to deployment compared to simply turning on a chatbot for all departments simultaneously. However, the university's experience suggests this measured approach yields higher quality implementations and better departmental buy-in. The shared brain concept, while powerful, introduces coordination challenges around terminology and content ownership. The need for departments to define common terms specifically for their context shows that knowledge management complexity grows with scale. The case study comes from a webinar hosted by the vendor and Internet2, which naturally emphasizes positive outcomes. Independent validation of the claimed 98% accuracy in terminology understanding and the specific usage statistics would strengthen confidence in the results. The relatively limited discussion of student satisfaction metrics (as opposed to usage volume) leaves some questions about the quality of the student experience beyond simply measuring conversation counts. The SMS success story, while compelling, was based on a small pilot of 75 students. Scaling to campus-wide SMS campaigns may encounter different response patterns, and the concern about "text fatigue" mirrors the email fatigue problem the university was trying to solve—suggesting that careful governance will be required as SMS usage grows. Overall, UC Santa Barbara's implementation represents a thoughtful, pragmatic approach to deploying LLM-based conversational AI in a complex institutional environment. The emphasis on phased rollout, student-centered testing, organizational change management, and realistic expectations about ongoing maintenance provides valuable lessons for other institutions considering similar implementations. The technical architecture balancing automated web crawling with human oversight, and combining knowledge base retrieval with generative AI fallbacks, demonstrates a mature approach to LLMOps in the higher education context.

Start deploying reproducible AI workflows today

Enterprise-grade MLOps platform trusted by thousands of companies in production.