Skip to content

Repository containing code for setting up RAG on your machine. Implemented OpenAI as well as HuggingFace llms and embedding models

License

Notifications You must be signed in to change notification settings

Open-Source-Coding-Enthusiasts/chat-with-documents-quickstart

Repository files navigation

Deployment options

Docker - Go to docker dir and then to cpu-only or gpu and run `docker compose up -d`
Kubernetes - Not yet supported (under construction)

Just remember to add necessary api keys for example if you want to use OpenAI models, provide OPENAI api key in docker-compose file
Usage
  1. Check if everything is connected correctly in Resources tab image

  2. Upload PDF document image

  3. Create new chat image

  4. Start Chatting, in this case using smallest possible llm (zephyr3b) image

More advanced configuration

Local LLM models inference is done using Ollama. So any model listed in here is supported.


Adding additional LLMs

Downloading Open Source models is done in a Resources Tab

image


Enabling prompt monitoring

Connecting monitoring service (Langfuse server) is done in Resources Tab

image

By default Langfuse URL will point to url defined as a MONITORING_SERVER_URL, so if you are deploying this app with docker compose, there is no need to change anything here.

You can of cource provide URL pointing to Langfuse cloud as well.

Setting up monitoring server
  1. Go to app_url:3000 and create account image

  2. Log in and create new project image

  3. Create a pair of API keys image

  4. Pass those keys into corresponding places in Resources Tab

About

Repository containing code for setting up RAG on your machine. Implemented OpenAI as well as HuggingFace llms and embedding models

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published