Replies: 2 comments 10 replies
-
Long story short, yes. Most documents would overflow the context just by themselves, so "traditional" approaches are out. Typically, this is done with vector storage which is way out of scope of llama.cpp. Check out Private GPT: https://github.com/imartinez/privateGPT |
Beta Was this translation helpful? Give feedback.
-
What you're looking for is RAG (retrieval-augmented generation). The PrivateGPT mentioned before has examples, there is also https://python.langchain.com/docs/modules/data_connection/ |
Beta Was this translation helpful? Give feedback.
-
One of the biggest use case of LLMs especially for businesses is chatting with PDF and Docs privately.
Would it be difficult to add this as feature in llama.cpp?
Beta Was this translation helpful? Give feedback.
All reactions