Don't want to share your private data? You can run your own local LLM to query your own documents in a couple steps with a RAG!
- Get a local LLM
 - Get Ollama
 - Follow the instructions to install.
 - Open a Terminal and download a model (e.g. llama3.1).
 - And serve the model locally.
 
ollama pull llama3.1
ollama serve
- Get the interface: We will use LangChain to handle the vector store and Steamlit to give you a fancy front-end to interact with the LLM and data upload.
 
git clone https://github.com/Sydney-Informatics-Hub/LLM-local-RAG/
cd LLM-local-RAG
conda create -n localrag python=3.11 pip
conda activate localrag
pip install langchain streamlit streamlit_chat chromadb fastembed pypdf langchain_community
- Start chatting!
 
streamlit run app.py
Upload your docs and start asking questions!
