Claribel — USDA GenAI Chatbot

USDA Farm Loan Programs — Conversational AI Interface Design

Role UX Architect
Timeline May – Aug 2025
Methods Stakeholder Interviews, Competitive Analysis, Iterative Prototyping, Developer Handoff
Tools Figma
Figma mockups of the USDA Claribel chatbot interface

The Problem

USDA Farm Loan Programs staff navigate 2,000–3,000 pages of policy handbooks to answer questions from farmers applying for loans. New customer service representatives relied on Ctrl+F or asking a more experienced colleague. Seasoned loan officials needed deeper cross-referencing across multiple policy areas. Leadership wanted to deploy a GenAI chatbot that could surface policy answers with citations to source material — but the team had no UI design and no UX perspective on how to make AI-generated responses trustworthy and actionable in a government context.

The Approach

I started by interviewing the product owner to identify the three core user types: customer service reps looking for quick answers while on the phone with a farmer, subject matter expert loan officials doing deep policy research across multiple handbooks, and executives who needed a mobile-optimized view to look up information while in meetings. Each persona had fundamentally different needs — the CSR needs a short, actionable answer fast; the SME needs comprehensive source material and cross-references; the executive needs the same information delivered cleanly on a phone screen.

Two screens showing a chat in progress with citations and conversation history
The core interface: two screens showing a chat in progress with citations and conversation history.

The stakeholder's three priorities were clear: access to ground truth source documents was non-negotiable, the system needed to help users construct accurate prompts, and there had to be a structured feedback mechanism so the team could improve the AI's responses over time.

I designed a personality system that adapted the chatbot's behavior to the user's context. A "Customer Service Coach" mode delivers short answers with suggestions for phone or email follow-up. A "Technical Research Assistant" mode provides longer, more detailed responses with additional references and diagrams. Users could also filter which policy documents the AI draws from — critical because the AI model was getting confused between loan making and loan servicing questions when the full corpus was included.

Settings screen showing how to enable the Technical Research Assistant mode and filter which policy documents the AI references
Users can enable the Technical Research Assistant mode and filter which policy documents the AI references — solving a prompt accuracy problem identified during stakeholder interviews.

For the citation experience, I designed a reference system where numbered citations in the response link to a split-screen document viewer showing the highlighted source passage alongside the AI's answer. This was the product owner's highest priority — staff needed to verify AI responses against the actual handbook language and cite those sources in official correspondence like adverse decision letters.

Split-screen showing chat on bottom and source document on top with highlighted passage
Citations link to a reference viewer showing the exact source passage — the most critical trust-building feature for government use.

I also designed the feedback mechanism with structured categories (factually correct/incorrect, contextually relevant/irrelevant, easy/difficult to understand, practical/impractical) plus an optional free-text field, giving the development team actionable data to improve the model.

Feedback form with thumbs up/down and structured quality dimension categories
Structured feedback captures specific quality dimensions rather than just thumbs up/down, giving the team actionable improvement data.

Since the development team didn't have Figma access, I delivered fully dimensioned specs — redlined screenshots with spacing, font sizes, weights, and corner radii annotated — so engineers could implement without ambiguity.

Redlined Figma screen showing dimension annotations, font specs, and spacing callouts
Developer handoff specs with full dimensions, created as annotated screenshots because the engineering team didn't have Figma licenses.

Reflection

Designing for AI-generated content requires different trust patterns than traditional interfaces. In a government context, the AI's answer isn't the product — the verifiable source citation is. Everything in the UI had to reinforce that the chatbot is a tool for finding information faster, not a replacement for policy expertise. The personality and filtering systems were my solution to a prompt engineering problem that was showing up as a UX problem — the AI was giving confused answers not because the model was bad, but because users had no way to scope their questions.