Skip to content

A simple application that helps to talk with raw texts using RAG (Retrieval Augmented Generation).

License

Notifications You must be signed in to change notification settings

shriamrut/talk-to-text

Repository files navigation

Talk to text

A simple application that helps to talk with raw texts using RAG (Retrieval Augmented Generation).

How to get started?

Have docker pre-installed in the host / machine that you are planning to run this app, and configured as well.

Note: For faster performance, good to use machines having GPUs.

  1. First build the image of the app using docker
docker build -t talk-to-text:latest .
  1. Then do docker compose, so that all the other services like vector database and nosql database comes up, that are used by the application to faciltate RAG.
docker compose up -d

Then go to url to access the fast API api docs to trigger the APIs, and get started.

http://localhost:8080/docs
Basic steps involved
  1. Upload a raw text using the texts API.
  2. Create a conversation using the textId obtained in the text API.
  3. Then post the query that you want to get from the text to the conversation.

TODOs

  1. Use Ollama instead of Hugging Face
  2. Use fastAPI's router interface for multiple routings
  3. Provide API way to customize the hugging face model used. Same goes for prompt template.
  4. Provide API way to customize text chunking, and vectorization.
  5. Add UI / UX to have a Chat UI.
  6. Use multiple models in same conversations. (Query can be directed to a specific model, based on '@' annotation)

About

A simple application that helps to talk with raw texts using RAG (Retrieval Augmented Generation).

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published