title | emoji | colorFrom | colorTo | sdk | sdk_version | app_file | pinned | license |
---|---|---|---|---|---|---|---|---|
Rag Mini Wikipedia Demo |
💬 |
yellow |
purple |
gradio |
4.36.1 |
app.py |
false |
apache-2.0 |
This project demonstrates a Retrieval-Augmented Generation (RAG) application for improved question answering using Large Language Models (LLMs). RAG overcomes LLM limitations by enabling access to specific information not included in their training data.
- Uses
rag-mini-wikipedia
dataset for factual information retrieval. - Employs
all-MiniLM-L6-v2
for sentence encoding andFAISS
for efficient similarity search. - Leverages
meta-llama/Llama-2-7b-chat-hf
model for response generation. - Built with Gradio 4.38.1 (a user interface library for machine learning).
- More informative responses by incorporating external knowledge.
- Ideal for applications like smart Q&A chatbots in corporate knowledge bases.
- Ensure you have the required libraries installed (refer to the project's requirements).
- Open a terminal and navigate to the project directory.
- Login to your Hugging Face account with appropriate token:
huggingface-cli login
- Run the application using the following command:
python app.py
This will launch the Gradio interface where you can interact with the RAG model.
Research notebook in the 'Research' folder explores chunking, prompt development, and future directions. Read it !