I built Knowledge Repo RAG Chatbot to learn how Retrieval-Augmented Generation (RAG) applications work end-to-end. It’s a chatbot that answers questions using resources from my own knowledge repository.

This is a Node & React project has two flavors:

  • Local stack: Qdrant (vector DB), Nomic embeddings (via vLLM), and a LLaMA model with GPT4All.
  • Cloud stack: OpenAI embeddings, Qdrant Cloud, and GPT completions via OpenAI API.

Key features include:

  • Data ingestion pipeline
  • Real-time chat with token streaming
  • Chat memory, history, and automatic title generation
  • Automated build and deployment

Planned improvements:

  • Pipeline to auto-update the vector store when the repo changes
  • Account system, authentication, rate-limit