Agentic AI chatbot that routes queries across curriculum, events, and places; augments with Pinecone RAG; powered by GPT-4o-mini via LangChain
This full-stack web app implements an agentic workflow for Duke University FAQs. A router agent dispatches queries to specialized agents (curriculum, events, locations). For curriculum-heavy questions, a Pinecone RAG index augments answers with scraped, offline Duke content. The backend is Flask on EC2; the frontend (React + Vite) connects via a WebSocket path through AWS Lambda.
Central decision-maker with a system prompt that selects the right agent(s) and synthesizes cross-domain answers.
Curriculum, Events, and Locations agents inherit from a shared BaseAgent and wrap Duke API tools via LangChain.
Offline scraping → paragraph-based chunking with sliding windows → embeddings → Pinecone namespaces with metadata for traceability.
Typed service layer for curriculum, events, and places; subject codes sourced to data/metadata/subjects.json
.
Flask on EC2; WebSocket path via AWS Lambda to the backend. GPT-4o-mini as the LLM through LangChain tools.
Automatic metrics on 20 benchmark questions (saved to llm_eval_results.csv
): ROUGE-L, BERTScore-F1, BLEU.
A 5-user study rated answers for helpfulness/accuracy, avg 4.12/5.