Resume

Adaptive RAG Chatbot

An intelligent retrieval system that dynamically adapts query processing to reduce hallucination and improve response quality in LLMs.

Adaptive RAG Chatbot

Problem

Large Language Models often generate confident but incorrect responses when relevant context is missing. Traditional RAG systems rely on static retrieval pipelines, which treat all queries equally — leading to unnecessary latency for simple queries and insufficient context for complex ones.

Solution

Designed a query routing mechanism using LangGraph to classify queries and dynamically select retrieval strategies
Implemented hybrid retrieval combining vector search (Qdrant) with contextual filtering for improved relevance
Added a re-ranking layer to prioritize high-signal documents and remove noisy results
Integrated fallback mechanisms (FAISS) and memory (MongoDB) for fault-tolerant conversational continuity

Tech Stack

LangGraphQdrantFAISSMongoDBFastAPIGPT-4o

Architecture

Query → Intent Classification → Adaptive Routing → Hybrid Retrieval (Qdrant/FAISS + Filters) → Re-ranking → Context Construction → LLM → Response

Challenges

Balancing latency and accuracy was critical — deeper retrieval improves answer quality but increases response time. This was solved using adaptive routing to trigger complex pipelines only when needed. Ensuring factual grounding required strict context injection and prompt design to minimize hallucination.

What I’d Improve Next

  • • Introduce learning-to-rank models for advanced re-ranking
  • • Add feedback-driven reinforcement loop for retrieval optimization
  • • Optimize ANN indexing (HNSW tuning) for large-scale datasets