| title | Rag Mini Wikipedia Demo |
|---|---|
| emoji | 💬 |
| colorFrom | yellow |
| colorTo | purple |
| sdk | gradio |
| sdk_version | 4.36.1 |
| app_file | app.py |
| pinned | false |
| license | apache-2.0 |
This project demonstrates a Retrieval-Augmented Generation (RAG) application for improved question answering using Large Language Models (LLMs). RAG overcomes LLM limitations by enabling access to specific information not included in their training data.
- Uses
rag-mini-wikipediadataset for factual information retrieval. - Employs
all-MiniLM-L6-v2for sentence encoding andFAISSfor efficient similarity search. - Leverages
meta-llama/Llama-2-7b-chat-hfmodel for response generation. - Built with Gradio 4.38.1 (a user interface library for machine learning).
- More informative responses by incorporating external knowledge.
- Ideal for applications like smart Q&A chatbots in corporate knowledge bases.
- Ensure you have the required libraries installed (refer to the project's requirements).
- Open a terminal and navigate to the project directory.
- Login to your Hugging Face account with appropriate token:
huggingface-cli login- Run the application using the following command:
python app.pyThis will launch the Gradio interface where you can interact with the RAG model.
Research notebook in the 'Research' folder explores chunking, prompt development, and future directions. Read it !