An AI-powered Retrieval-Augmented Generation (RAG) application that allows users to ask questions directly from official QUEST (Quaid-e-Awam University of Engineering, Science and Technology) documents.
The app uses FAISS for vector search, HuggingFace embeddings, and Groq LLMs, wrapped in a modern Streamlit UI with a clean light-mode AI chat interface.
- 🔍 Semantic search over official QUEST documents
- 📚 FAISS vector database for fast retrieval
- 🤖 Groq-powered LLM responses
- 🧠 BAAI bge-base-en-v1.5 embeddings
- 💬 Modern AI chat-style UI (light mode, vibrant design)
- 🧾 Source-aware context injection
- ⚡ Fast, accurate, and production-ready RAG pipeline
-
Documents Ingestion
QUEST prospectus and official PDFs are chunked and embedded -
Vector Storage
Embeddings are stored in a FAISS index -
Query Processing
User question is embedded using the same model -
Retrieval
FAISS returns the most relevant document chunks -
Generation
Groq LLM generates answers strictly from retrieved context
| Component | Technology |
|---|---|
| Frontend | Streamlit |
| Vector Store | FAISS |
| Embeddings | HuggingFace BAAI/bge-base-en-v1.5 |
| LLM Provider | Groq |
| Frameworks | LangChain |
| Language | Python |
.
├── app.py # Main Streamlit application
├── faiss.index # FAISS vector index
├── metadata.pkl # Document metadata & mappings
├── README.md # Project documentation
├── RAG.ipynb # Jupyter Notebook for RAG pipeline
└── requirements.txt # Python dependencies
git clone https://github.com/your-username/quest-rag-assistant.git
cd quest-rag-assistantpython -m venv venv
source venv/bin/activate # On Windows: venv\Scripts\activatepip install -r requirements.txtCreate the file .streamlit/secrets.toml:
GROQ_API_KEY = "your_groq_api_key_here"streamlit run app.pyOpen in your browser:
http://localhost:8501
- What is the admission criteria for BE programs?
- What is the fee structure for undergraduate students?
- Which departments are offered at QUEST?
- What are the eligibility requirements?
- ✅ The assistant ONLY uses retrieved document context
- ✅ If information is missing, it responds with:
Not found in documents. - ✅ No hallucinated answers are generated
- Light-mode focused design
- Vibrant modern AI interface
- Chat-style question and answer flow
- Clean typography and spacing
- Streaming token-by-token responses
- Conversation memory
- Clickable citations per source
- Hybrid retrieval (BM25 + FAISS)
- Docker deployment
- Multi-PDF support
- FAISS for fast similarity search
- HuggingFace for open-source embeddings
- Groq for high-speed LLM inference
- Streamlit for rapid UI development
This project is intended for educational and research purposes.
⭐ If you find this project useful, please consider giving it a star!