Problem
Internal teams needed a Q&A assistant that could support technical research and non-technical marketing workflows from the same knowledge base.
GenAI Final Project
Final project for UC Berkeley DATASCI 267. I designed and evaluated a retrieval-augmented generation system that answers the same query differently for engineering and marketing users, while staying grounded in source context.
Problem
Internal teams needed a Q&A assistant that could support technical research and non-technical marketing workflows from the same knowledge base.
Approach
Built a LangChain RAG pipeline with Qdrant retrieval, Cohere reranking, and persona-specific prompts to control depth, tone, and answer length.
Outcome
Improved quality over baseline runs with strong groundedness/relevance scores and clearer role-specific responses across validation examples.
I evaluated outputs with persona-aware criteria: groundedness and relevance for both personas, plus accuracy for engineering responses and coherence for marketing responses. I also compared generated answers against gold responses with semantic similarity scoring.
The strongest setup combined MMR retrieval, reranking, and separate prompts. One practical takeaway was that similarity metrics can overrate answers when tone or technical depth diverges from user expectations, even when semantic content overlaps.