Spaces:
Running
Running
title: Personal RAG | |
emoji: ⚡ | |
colorFrom: purple | |
colorTo: blue | |
sdk: docker | |
pinned: false | |
license: mit | |
# Ronald AI: A Personal RAG Chatbot | |
[](https://chainlit.io) | |
[](https://www.llamaindex.ai/) | |
[](https://huggingface.co/) | |
[](https://openrouter.ai/opengvlab/internvl3-14b:free) | |
--- | |
## Overview | |
**Ronald AI** is my intelligent and friendly personal assistant, designed to answer my questions based on a specific set of documents. Built with Python using the powerful `Chainlit` framework for the user interface and `LlamaIndex` for efficient data indexing and retrieval, this application provides me with a seamless conversational experience. | |
The core of Ronald AI is its ability to understand and respond to queries strictly within the context of the documents I provide, ensuring accurate and relevant answers. It leverages a state-of-the-art language model from `OpenRouter` and a high-quality embedding model from `Hugging Face` to comprehend and process my questions effectively. | |
## Features | |
- **RAG pipeline**: Retrieves and grounds answers on local documents. | |
- **LLM Integration**: Uses OpenRouter's OpenGVLab: InternVL3 14B Chat model for generation. | |
- **Document Indexing**: Auto-loads or rebuilds index from `data/`. | |
- **Real-time Chat UI**: Powered by Chainlit with streaming responses. | |
- **HuggingFace Embeddings**: Uses `BAAI/bge-small-en-v1.5` for vector search. | |
- **Simple storage**: Uses LlamaIndex's persistent `StorageContext`. | |