Made a RAG-centric, Open-Source UI based on llama.cpp - With Advanced Source Citations & Referencing: Pinpointing Page-Numbers, Incorporating Extracted Images, Text-highlighting & Document-Readers alongside Local LLM-generated Responses #7928
Replies: 2 comments 5 replies
-
Thanks for sharing - looks like a cool project!
Absolutely, please open a PR Curious to know more about your experience using |
Beta Was this translation helpful? Give feedback.
-
That's an excellent suggestion and I will dig into it. Thanks so much once again!
…________________________________
From: Georgi Gerganov ***@***.***>
Sent: Friday, June 14, 2024 2:43:15 AM
To: ggerganov/llama.cpp ***@***.***>
Cc: Abheek Gulati ***@***.***>; Author ***@***.***>
Subject: Re: [ggerganov/llama.cpp] Made a RAG-centric, Open-Source UI based on llama.cpp - With Advanced Source Citations & Referencing: Pinpointing Page-Numbers, Incorporating Extracted Images, Text-highlighting & Document-Readers alongside Local LLM-generated...
I also humbly request you for feedback and suggestions on LARS: if there's anything specific I should focus on, any llama.cpp feature I should integrate and look into, any of the myriad of other amazing projects that you feel will benefit LARS and I should look into, I'm all ears and very eager to hear!
From user perspective I cannot provide much feedback as I am mainly interested in how 3rd party projects use llama.cpp at the developer level.
One thing that might be interesting to explore in that regard is to try to use llama.cpp also for computing the embeddings. Currently, if I understand correctly the code, LARS uses a Hugging Face engine to compute the embeddings:
https://github.com/abgulati/LARS/blob/c11dd07d9b8125efc3a4935716626434d5b94b92/web_app/app.py#L2060-L2064
Instead, it might be possible to utilize llama.cpp-based implementation. From a quick look at LangChain docs, there is this:
https://python.langchain.com/v0.1/docs/integrations/text_embedding/llamacpp/
I'm interested in that because I want to know what are the current limitations when it comes to computing embeddings with llama.cpp and how to improve it. Both in terms of performance and model support
—
Reply to this email directly, view it on GitHub<#7928 (reply in thread)>, or unsubscribe<https://github.com/notifications/unsubscribe-auth/AEZB3X3SQY7CEIEVHIRK7Y3ZHK3LHAVCNFSM6AAAAABJJINWASVHI2DSMVQWIX3LMV43SRDJONRXK43TNFXW4Q3PNVWWK3TUHM4TONZSHA3DC>.
You are receiving this because you authored the thread.Message ID: ***@***.***>
|
Beta Was this translation helpful? Give feedback.
-
Recently Open-Sourced my Citation-Centric Local-LLM Application: RAG with your LLM of choice, with your documents, on your machine.
Please read-on for more details, or check out my GitHub repo (attained 200 stars in under a week!) for the complete scoop!
Introducing LARS: The LLM & Advanced Referencing Solution! There are many desktop applications for running LLMs locally, but LARS aims to be the ultimate open-source RAG-centric LLM application.
Towards this end, LARS takes the concept of RAG much further by adding detailed citations to every response, supplying you with specific document names, page numbers, text-highlighting, and images relevant to your question, and even presenting a document reader right within the response window. While all the citations are not always present for every response, the idea is to have at least some combination of citations brought up for every RAG response and that’s generally found to be the case.
I humbly request the addition of LARS to the UI list on the llama.cpp README documentation!
Here's a demonstration video going over core features:
LARS Feature-Demonstration Video
Here's a list detailing LARS's feature-set as it stands today:
I have been building this tool single-handedly since August 2023 and am continuing to add to it on a near daily basis.
LARS could really benefit from users, testing and contributions so please check out the repository!
You can also check out my post sharing & discussing LARS on Reddit.
Beta Was this translation helpful? Give feedback.
All reactions