-
Notifications
You must be signed in to change notification settings - Fork 7.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Training on own data? #532
Comments
vote this! Same question. ✋ |
LangChain allows you to connect a language model to other sources of data. you can also take a look https://github.com/imartinez/privateGPT |
Its known as "Chat with PDF" or "Talk to PDF/Book/Document", look it up |
I never intended to "train" on own data, but it was more about letting the GPT access a file repository to take into consideration when asking it questions. Chat with a datalake is what I wanted to achieve. |
Well, I think you can do this by performing a semantic search over your text data (embeddings) and feed the relevant ones to chat models and get your answers. |
Semantic search is one option agreed. But it sounds better than how it actually performs. Is there a way to fine-tune (domain adaptation) the gpt4all model using my local enterprise data, such that gpt4all "knows" about the local data as it does the open data (from wikipedia etc) |
Tried that with Came here to test out GPT4All to see if this is any better |
yes, this is a duplicate. please open a new, updated issue if this is still relevant to you. |
Is there a way to feed GPT4all own data so that it can be trained on the information? I would like to be able to feed it my emails, my PDF files and a bunch of other data that I have, and use GPT4all's chat to trawl through this data and spit out information for me. Is this something that is being worked on, or currently possible?
The text was updated successfully, but these errors were encountered: