
Chat with your own docs using local LLMs — no cloud, no API keys
Every time you paste a contract, a client note, or a financial doc into ChatGPT to ask a question, that content hits OpenAI's servers. For most people that's fine. For a lot of workflows it isn't. I wanted something that worked like that — drop in a document, ask questions — but where nothing left my machine. I looked at the existing options and none of them felt right. Too heavy, too opinionated, or they still phoned home somewhere. So I built VaultMind. What it is VaultMind is a local RAG system. You drop in PDFs, Word docs, CSVs, markdown files, or paste URLs, and then chat with them using open-source LLMs running through Ollama. The entire stack — inference, vector storage, embeddings, frontend — runs on your computer. git clone https://github.com/airblackbox/VaultMind cd VaultMind bash start.sh That's it. start.sh pulls the embedding model and default LLM, starts the FastAPI backend, and opens the UI. First run takes a few minutes while Ollama downloads the models. After that it's
Continue reading on Dev.to Python
Opens in a new tab


