reworked document adding and most of the project

This commit is contained in:
Philipp Mock 2026-03-25 08:57:10 +01:00
parent 08610e0b64
commit 7d70169ee3
4 changed files with 205 additions and 58 deletions

203
README.md
View File

@ -1,90 +1,205 @@
# Local RAG Setup
Minimal RAG implementation with LangChain, FAISS, and support for either Ollama or OpenAI (API-key needed).
Minimal RAG implementation with LangChain, FAISS, and either **Ollama** (local) or **OpenAI** (API key). A web chat UI is included.
## Dependencies
---
- `langchain` - Core framework
- `langchain-community` - Loaders, vectorstores
- `langchain-ollama` - Ollama integration
- `langchain-openai` - OpenAI integration
- `langchain-text-splitters` - Text splitting
- `langchain-huggingface` - HuggingFace embeddings
- `faiss-cpu` - Vector search
- `sentence-transformers` - Embeddings
- `pypdf` - PDF loading
- `fastapi` - Web server
- `uvicorn` - ASGI server
## What you need (before you start)
## Installation
- **Python 3.10 or newer** ([python.org](https://www.python.org/downloads/))
- **Git** (optional, only if you clone the project)
- Either:
- **Ollama** installed and running ([ollama.com](https://ollama.com)), with at least one model pulled, **or**
- An **OpenAI API key** (if you use OpenAI in the chat)
---
## Install dependencies (step by step)
Install [Miniconda](https://docs.conda.io/en/latest/miniconda.html) or [Anaconda](https://www.anaconda.com/) if you do not have Conda yet.
All commands below assume your terminal is open **in the project folder** (the folder that contains `requirements.txt`).
```bash
conda create -n local_rag python=3.10 -y
conda activate local_rag
pip install --upgrade pip
pip install -r requirements.txt
```
## Setup
Use `conda activate local_rag` in every new terminal session before running `python` or `uvicorn` for this project.
### Ollama (optional)
### OpenAI (only if you use the OpenAI provider in the chat)
In the same terminal **before** starting the server:
```bash
export OPENAI_API_KEY="your-key-here"
```
On Windows (Command Prompt): `set OPENAI_API_KEY=your-key-here`
---
## Run Ollama (only if you use Ollama)
In a **separate** terminal:
```bash
ollama serve
ollama pull mistral
```
### OpenAI (optional)
Set the API key when using OpenAI:
In another terminal, pull a model once (example):
```bash
export OPENAI_API_KEY="your-key"
ollama pull gpt-oss:20b
```
## Add Documents
The model name must match what you configure in `server.py` (see [Configuration reference](#configuration-reference)).
**Option 1:** Add PDFs from a folder via script. Edit `DATA_ROOT` in [add_pdfs.py](add_pdfs.py) to point at your folder, then run:
---
## Build the vector store from a folder of PDFs
The project includes [add_pdfs.py](add_pdfs.py). It finds every **`.pdf`** file under a folder you choose (including subfolders), then chunks, embeds, and saves to FAISS.
**Two modes** (set in the script):
| Setting | Behavior |
|---------|----------|
| `CLEAR_VECTORSTORE_FIRST = True` | Deletes the existing vector store folder, then builds a **new** index from the PDFs under `DATA_ROOT`. Use this for a full rebuild. |
| `CLEAR_VECTORSTORE_FIRST = False` | Keeps the current index (if it exists) and **merges** chunks from the PDFs under `DATA_ROOT` into it. Use this to add another batch of PDFs without wiping what you already indexed. |
**Steps:**
1. Open [add_pdfs.py](add_pdfs.py) in a text editor.
2. Set **`DATA_ROOT`** to the folder that contains your PDFs (absolute path or path relative to how you run the script).
3. Set **`CLEAR_VECTORSTORE_FIRST`** to `True` (fresh index) or `False` (append to existing store).
4. Optionally set **`VECTORSTORE_PATH`** (default: `./vectorstore`). It must match **`VECTORSTORE_PATH`** in [server.py](server.py) so the chat loads the same index.
5. From the project folder, with `conda activate local_rag` (or your chosen env name):
```bash
python add_pdfs.py
```
The script clears the existing vector store and indexes all PDFs recursively. Supports `.pdf`, `.txt`, `.md`.
Indexing can take a long time for many large PDFs. When it finishes, you should see `Vector store saved to ...`.
**Option 2:** Use `local_rag.py` programmatically:
**Note:** This script only indexes **PDF** files. To add `.txt` or `.md` files, use the Python snippet below or call `add_documents` yourself.
---
## Add more documents later (alternative to add_pdfs)
You can also merge files by hand with a short script (any mix of supported types):
```python
from local_rag import LocalRAG
rag = LocalRAG()
rag.add_documents(["path/to/doc1.pdf", "path/to/doc2.txt"])
rag = LocalRAG(vectorstore_path="./vectorstore") # same path as server.py
rag.add_documents([
"path/to/new1.pdf",
"path/to/notes.txt",
])
```
## Chat GUI
`add_documents` merges new chunks into the existing FAISS store and saves it again—the same behavior as [add_pdfs.py](add_pdfs.py) with `CLEAR_VECTORSTORE_FIRST = False`.
Start the server:
---
## Swap or experiment with different vector stores
The vector index is stored on disk under the folder given by **`VECTORSTORE_PATH`** (default `./vectorstore`). That folder contains files such as `index.faiss` and `index.pkl`.
**To use a different index:**
1. Set **`VECTORSTORE_PATH`** in both [server.py](server.py) and any script you use to build the index (e.g. [add_pdfs.py](add_pdfs.py)) to the **same** path, e.g. `./vectorstore_experiment`.
2. Rebuild the index (run `add_pdfs.py` or `add_documents`) so that folder is created.
3. **Restart** the web server so it loads the new path at startup.
**Tips:**
- Keep multiple copies of the folder (e.g. `vectorstore_backup`, `vectorstore_papers_only`) and swap `VECTORSTORE_PATH` to switch between them.
- If you change **chunk size**, **embedding model**, or **FAISS** usage in code, treat the old index as incompatible: use a new `VECTORSTORE_PATH` or delete the old folder and rebuild.
---
## Run the chat web app
With the Conda environment activated (`conda activate local_rag`) and (if needed) `OPENAI_API_KEY` set:
```bash
uvicorn server:app --reload
```
Open [http://localhost:8000](http://localhost:8000). The chat UI provides:
Open [http://127.0.0.1:8000](http://127.0.0.1:8000) or [http://localhost:8000](http://localhost:8000).
- **Provider switch** Toggle between Ollama and OpenAI without restart (OpenAI requires `OPENAI_API_KEY`)
- **Conversation history** Multi-turn chat with context
- **Markdown** Assistant replies rendered as markdown (headings, code, lists, links)
- Use the **LLM provider** dropdown: **Ollama** or **OpenAI** (OpenAI only works if the server was started with a valid `OPENAI_API_KEY`).
- You need a **non-empty vector store** (see above) for answers to work.
Ensure the vector store is populated and at least one provider (Ollama or OpenAI) is configured.
---
## API
## API (short reference)
- `POST /api/chat` `{ "message": "...", "history": [...], "llm_provider": "ollama"|"openai" }`
- `GET /api/providers` `{ "ollama": true, "openai": true|false }`
- `GET /api/health` Health and vectorstore status
| Endpoint | Purpose |
|----------|---------|
| `POST /api/chat` | Body: `message`, optional `history`, optional `llm_provider` (`ollama` or `openai`) |
| `GET /api/providers` | Which providers are available (`openai` false if no API key at startup) |
| `GET /api/health` | Server and whether a vector store is loaded |
## How it works
---
1. **Load documents** PDFs or text via PyPDFLoader / TextLoader
2. **Chunk** RecursiveCharacterTextSplitter (2000 chars, 400 overlap)
3. **Embed** sentence-transformers/all-MiniLM-L6-v2
4. **Store** FAISS vector store (similarity search with scores)
5. **Query** Retrieve chunks, optionally rephrase with conversation history, generate answer with selected LLM
## How it works (high level)
1. **Load documents** PDFs via `PyPDFLoader`, text via `TextLoader`.
2. **Chunk** `RecursiveCharacterTextSplitter` (defaults in [local_rag.py](local_rag.py)).
3. **Embed** Hugging Face `sentence-transformers/all-MiniLM-L6-v2`.
4. **Store** FAISS; retrieval uses `similarity_search_with_score`.
5. **Query** Optional rephrase with chat history, retrieval, then answer from the LLM.
---
## Configuration reference (what to edit)
These are the main places to change behavior without restructuring the app.
### [server.py](server.py)
| What | Where |
|------|--------|
| Ollama model name | `OLLAMA_MODEL = "..."` |
| OpenAI model name | `OPENAI_MODEL = "..."` |
| Where the FAISS index is loaded from | `VECTORSTORE_PATH = "./vectorstore"` (must match your indexing script) |
### [local_rag.py](local_rag.py) `LocalRAG.__init__`
| What | Where (approx.) |
|------|------------------|
| Default vector store folder | Parameter `vectorstore_path="./vectorstore"` |
| Embedding model | `HuggingFaceEmbeddings(model_name="sentence-transformers/...")` |
| Chunk size and overlap | Module-level `CHUNK_SIZE` and `CHUNK_OVERLAP` (used by `RecursiveCharacterTextSplitter` when adding documents) |
| Default Ollama / OpenAI model strings | Parameters `ollama_model`, `openai_model`, `ollama_base_url` |
Changing the embedding model or chunk settings requires **rebuilding** the vector store (old index is not compatible).
### [local_rag.py](local_rag.py) `query_with_history`
| What | Where |
|------|--------|
| Default number of chunks retrieved (`k`) | Module-level `RETRIEVAL_K` (overrides: pass `k=` to `query` / `query_with_history`) |
| Extra text appended only to the **FAISS query** (biases retrieval, not the final answer phrasing) | `QUERY_ADDITIONAL_INSTRUCTIONS` (concatenated to the search query before embedding) |
| **Rephrase** prompt (standalone question when there is chat history) | String `rephrase_prompt = f"""..."""` inside `query_with_history` |
| **Answer** prompt opening instructions only | Module-level `ANSWER_PROMPT` (edit the role / style lines). The block from chat history through `Answer:` is built in `query_with_history` |
### [add_pdfs.py](add_pdfs.py)
| What | Where |
|------|--------|
| Folder to scan for PDFs | `DATA_ROOT = Path("...")` |
| Output vector store folder | `VECTORSTORE_PATH = "./vectorstore"` (keep in sync with `server.py`) |
| Wipe index vs merge | `CLEAR_VECTORSTORE_FIRST = True` (delete and rebuild) or `False` (append to existing index) |
---
## Dependencies (for developers)
See [requirements.txt](requirements.txt) for the full list (LangChain, FAISS, sentence-transformers, FastAPI, uvicorn, etc.).

View File

@ -5,13 +5,23 @@ from pathlib import Path
from local_rag import LocalRAG
# Folder to scan for PDFs (recursively).
DATA_ROOT = Path("/Users/Philipp/Desktop/workspace/python/gpt_publikationen/data_vs")
# Must match server.py so the chat loads the same index.
VECTORSTORE_PATH = "./vectorstore"
# If True: delete the existing vector store folder, then index all PDFs from scratch.
# If False: load the existing index (if any) and merge new chunks from these PDFs into it.
CLEAR_VECTORSTORE_FIRST = True
if __name__ == "__main__":
if Path(VECTORSTORE_PATH).exists():
if CLEAR_VECTORSTORE_FIRST and Path(VECTORSTORE_PATH).exists():
shutil.rmtree(VECTORSTORE_PATH)
print(f"Cleared existing vector store: {VECTORSTORE_PATH}")
elif not CLEAR_VECTORSTORE_FIRST:
print(f"Appending to existing vector store (if any): {VECTORSTORE_PATH}")
pdfs = sorted(p for p in DATA_ROOT.rglob("*") if p.suffix.lower() == ".pdf")
print(f"Found {len(pdfs)} PDF(s) under {DATA_ROOT}")

View File

@ -12,6 +12,23 @@ from langchain_ollama import ChatOllama
from langchain_openai import ChatOpenAI
from langchain_text_splitters import RecursiveCharacterTextSplitter
# Pipeline tuning: used when adding documents (rebuild index after changing).
CHUNK_SIZE = 2000
CHUNK_OVERLAP = 400
# Retrieval / query tuning
RETRIEVAL_K = 5
# Appended to the embedding query for FAISS (does not affect the answer-generation prompt).
QUERY_ADDITIONAL_INSTRUCTIONS = (
"Do not return a list of references but prioritize meaningful text from abstracts, results and discussion sections."
)
# Opening instructions for the final answer LLM call only. The rest (history, context, question)
# is assembled in query_with_history.
ANSWER_PROMPT = """You are an assistant for question-answering. Use the chat history (if any) and the retrieved context below to answer the current question.
If you don't know the answer, say so. Keep the conversation coherent."""
class LocalRAG:
def __init__(
@ -19,7 +36,7 @@ class LocalRAG:
vectorstore_path="./vectorstore",
llm_provider="ollama",
ollama_model="gpt-oss:20b",
openai_model="gpt-5.2",
openai_model="gpt-5-mini",
ollama_base_url="http://localhost:11434",
):
"""Initialize local RAG system. llm_provider: 'ollama' or 'openai'."""
@ -32,11 +49,12 @@ class LocalRAG:
model_name="sentence-transformers/all-MiniLM-L6-v2"
)
# Text splitter
# Text splitter (used when adding documents; tune CHUNK_SIZE / CHUNK_OVERLAP above)
self.text_splitter = RecursiveCharacterTextSplitter(
chunk_size=2000,
chunk_overlap=400
chunk_size=CHUNK_SIZE,
chunk_overlap=CHUNK_OVERLAP,
)
print(f"Text splitter: chunk_size={CHUNK_SIZE}, chunk_overlap={CHUNK_OVERLAP}")
# LLM (Ollama or OpenAI)
if llm_provider == "openai":
@ -186,15 +204,19 @@ class LocalRAG:
for doc, score in docs_with_scores
]
def query(self, question, k=8):
def query(self, question, k=None):
"""Query the RAG system (no conversation history). Returns dict with 'answer' and 'retrieved'."""
if k is None:
k = RETRIEVAL_K
return self.query_with_history(question, chat_history=[], k=k)
def query_with_history(self, question, chat_history=None, k=8):
def query_with_history(self, question, chat_history=None, k=None):
"""Query the RAG with conversation history: rephrase question using history for retrieval,
then answer with full conversation + retrieved context in the prompt.
Returns dict with 'answer' and 'retrieved' (list of chunks with content, source, page).
"""
if k is None:
k = RETRIEVAL_K
if self.vectorstore is None:
return {
"answer": "Error: No documents loaded. Please add documents first.",
@ -203,9 +225,6 @@ class LocalRAG:
history_str = self._format_history(chat_history)
search_query = question
rag_query_instruction = (
"Do not return a list of references but prioritize meaningful text from abstracts, results and discussion sections."
)
print(f"[RAG] User question: {question!r}")
@ -224,7 +243,7 @@ Standalone question:"""
search_query = (rephrase_response.content if hasattr(rephrase_response, "content") else str(rephrase_response)).strip() or question
print(f"[RAG] Standalone search query (rephrased): {search_query!r}")
retrieval_query = f"{search_query}\n\n{rag_query_instruction}"
retrieval_query = f"{search_query}\n\n{QUERY_ADDITIONAL_INSTRUCTIONS}"
print(f"[RAG] Search query: {search_query!r}")
print(f"[RAG] Retrieval query sent to vector store: {retrieval_query!r}")
@ -239,8 +258,9 @@ Standalone question:"""
# 3) Answer using conversation history + retrieved context
history_block = f"Chat history:\n{history_str}\n\n" if history_str else ""
answer_prompt = f"""You are an assistant for question-answering. Use the chat history (if any) and the retrieved context below to answer the current question.
If you don't know the answer, say so. Keep the conversation coherent.
answer_prompt = (
ANSWER_PROMPT
+ f"""
{history_block}Relevant context from documents:
@ -249,6 +269,8 @@ Standalone question:"""
Current question: {question}
Answer:"""
)
print(f"[RAG] Composed answer prompt:\n{answer_prompt}")
response = self.llm.invoke(answer_prompt)
answer = response.content if hasattr(response, "content") else str(response)

View File

@ -12,7 +12,7 @@ from pydantic import BaseModel
from local_rag import LocalRAG
OLLAMA_MODEL = "gpt-oss:20b"
OPENAI_MODEL = "gpt-5.2"
OPENAI_MODEL = "gpt-5-mini"
VECTORSTORE_PATH = "./vectorstore"
# Dual RAG instances for on-the-fly provider switching