TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

PrivateGPT

520 pointsby antouankalmost 2 years ago

29 comments

davidy123almost 2 years ago
Granted I&#x27;m not coming from the python world, but I have tried many of these projects, and very few of them install out of the box. They usually end with some incompatibility, and files scattered all over the place, leading to future nightmares.<p><pre><code> ERROR: pip&#x27;s dependency resolver does not currently take into account all the packages that are installed. This behaviour is the source of the following dependency conflicts. sentry-sdk 1.22.2 requires urllib3&lt;2.0.0, but you have urllib3 2.0.2 which is incompatible </code></pre> Just for fun, here&#x27;s the result of python -m pip install -r .&#x2F;requirements.txt for tortoise-tts;<p>…many many lines<p><pre><code> raise ValueError(&quot;%r is not a directory&quot; % (package_path,)) ValueError: &#x27;build&#x2F;py3k&#x2F;scipy&#x27; is not a directory Converting to Python3 via 2to3...</code></pre> …<p><pre><code> &#x2F;tmp&#x2F;pip-install-hkb_4lh7&#x2F;scipy_088b20410aca4f0cbcddeac86ac7b7b1&#x2F;build&#x2F;py3k&#x2F;scipy&#x2F;signal&#x2F;fir_filter_design.py [end of output] note: This error originates from a subprocess, and is likely not a problem with pip. error: metadata-generation-failed </code></pre> I&#x27;m not asking for support, just saying if people really want to make something &#x27;easy&#x27; they&#x27;d use docker. I gather there are better python package managers, but I gather that&#x27;s a bit of a mess too.<p>Someone is thinking &quot;this is part of learning the language,&quot; but I think it&#x27;s just bad design.
评论 #36029521 未加载
评论 #36036044 未加载
评论 #36030387 未加载
评论 #36085014 未加载
评论 #36045344 未加载
评论 #36030217 未加载
评论 #36030079 未加载
j_shialmost 2 years ago
Self-hosted + self-trained LLMs are probably the future for enterprise.<p>While consumers are happy to get their data mined to avoid paying, businesses are the opposite: willing to pay a lot to avoid feeding data to MSFT&#x2F;GOOG&#x2F;META.<p>They may give assurances on data protection (even here GitHub copilot TOS has sketchy language around saving down derived data), but can’t get around fundamental problem that their products need user interactions to work well.<p>So it seems with BigTechLLM there’s inherent tension between product competitiveness and data privacy, which makes them incompatible with enterprise.<p>Biz ideas along these lines: - Help enterprises set up, train, maintain own customized LLMs - Security, compliance, monitoring tools - Help AI startups get compliant with enterprise security - Fine tuning service
评论 #36034358 未加载
评论 #36029762 未加载
评论 #36034732 未加载
simonwalmost 2 years ago
I&#x27;m always interested in seeing the prompt that drives these kinds of tools.<p>In this case it appears to be using RetrievalQA from LangChain, which I think is this prompt here: <a href="https:&#x2F;&#x2F;github.com&#x2F;hwchase17&#x2F;langchain&#x2F;blob&#x2F;v0.0.176&#x2F;langchain&#x2F;chains&#x2F;retrieval_qa&#x2F;prompt.py">https:&#x2F;&#x2F;github.com&#x2F;hwchase17&#x2F;langchain&#x2F;blob&#x2F;v0.0.176&#x2F;langcha...</a><p><pre><code> Use the following pieces of context to answer the question at the end. If you don&#x27;t know the answer, just say that you don&#x27;t know, don&#x27;t try to make up an answer. {context} Question: {question} Helpful Answer:</code></pre>
评论 #36030252 未加载
评论 #36026272 未加载
评论 #36036012 未加载
评论 #36026673 未加载
skykooleralmost 2 years ago
&quot;System requirements&quot; section should really mention what amount of RAM or VRAM is needed for inference.
评论 #36028168 未加载
hodanlialmost 2 years ago
These are the similar projects I&#x27;ve come across:<p>- [GitHub - e-johnstonn&#x2F;BriefGPT: Locally hosted tool that connects documents to LLMs for summarization and querying, with a simple GUI.](<a href="https:&#x2F;&#x2F;github.com&#x2F;e-johnstonn&#x2F;BriefGPT">https:&#x2F;&#x2F;github.com&#x2F;e-johnstonn&#x2F;BriefGPT</a>)<p>- [GitHub - go-skynet&#x2F;LocalAI: Self-hosted, community-driven, local OpenAI-compatible API. Drop-in replacement for OpenAI running LLMs on consumer-grade hardware. No GPU required. LocalAI is a RESTful API to run ggml compatible models: llama.cpp, alpaca.cpp, gpt4all.cpp, rwkv.cpp, whisper.cpp, vicuna, koala, gpt4all-j, cerebras and many others!](<a href="https:&#x2F;&#x2F;github.com&#x2F;go-skynet&#x2F;LocalAI">https:&#x2F;&#x2F;github.com&#x2F;go-skynet&#x2F;LocalAI</a>)<p>- [GitHub - paulpierre&#x2F;RasaGPT: RasaGPT is the first headless LLM chatbot platform built on top of Rasa and Langchain. Built w&#x2F; Rasa, FastAPI, Langchain, LlamaIndex, SQLModel, pgvector, ngrok, telegram](<a href="https:&#x2F;&#x2F;github.com&#x2F;paulpierre&#x2F;RasaGPT">https:&#x2F;&#x2F;github.com&#x2F;paulpierre&#x2F;RasaGPT</a>)<p>- [GitHub - imartinez&#x2F;privateGPT: Interact privately with your documents using the power of GPT, 100% privately, no data leaks](<a href="https:&#x2F;&#x2F;github.com&#x2F;imartinez&#x2F;privateGPT">https:&#x2F;&#x2F;github.com&#x2F;imartinez&#x2F;privateGPT</a>)<p>- [GitHub - reworkd&#x2F;AgentGPT: Assemble, configure, and deploy autonomous AI Agents in your browser.](<a href="https:&#x2F;&#x2F;github.com&#x2F;reworkd&#x2F;AgentGPT">https:&#x2F;&#x2F;github.com&#x2F;reworkd&#x2F;AgentGPT</a>)<p>- [GitHub - deepset-ai&#x2F;haystack: Haystack is an open source NLP framework to interact with your data using Transformer models and LLMs (GPT-4, ChatGPT and alike). Haystack offers production-ready tools to quickly build complex question answering, semantic search, text generation applications, and more.](<a href="https:&#x2F;&#x2F;github.com&#x2F;deepset-ai&#x2F;haystack">https:&#x2F;&#x2F;github.com&#x2F;deepset-ai&#x2F;haystack</a>)<p>- [PocketLLM « ThirdAi](<a href="https:&#x2F;&#x2F;www.thirdai.com&#x2F;pocketllm&#x2F;" rel="nofollow">https:&#x2F;&#x2F;www.thirdai.com&#x2F;pocketllm&#x2F;</a>)<p>- [GitHub - imClumsyPanda&#x2F;langchain-ChatGLM: langchain-ChatGLM, local knowledge based ChatGLM with langchain | 基于本地知识库的 ChatGLM 问答](<a href="https:&#x2F;&#x2F;github.com&#x2F;imClumsyPanda&#x2F;langchain-ChatGLM">https:&#x2F;&#x2F;github.com&#x2F;imClumsyPanda&#x2F;langchain-ChatGLM</a>)
monkeydustalmost 2 years ago
Got this working locally - badly needs GPU support (have a 3090 so come on!) there is some workaround but expect it will come pretty soon. This video was a useful walkthough esp on using different model and upping the CPU threads. <a href="https:&#x2F;&#x2F;www.youtube.com&#x2F;watch?v=A3F5riM5BNE">https:&#x2F;&#x2F;www.youtube.com&#x2F;watch?v=A3F5riM5BNE</a>
thefourthchimealmost 2 years ago
I tried this on my M2 Macbook with 16gb of RAM but got:<p>&quot;ggml_new_tensor_impl: not enough space in the context&#x27;s memory pool (needed 18296202768, available 18217606000)&quot;
评论 #36027356 未加载
aldarisbmalmost 2 years ago
One quick plug<p>I want to have the memory part of langchain down, vector store + local database + client to chat with an LLM (gpt4all model can be swapped with OpenAI api just switching the base URL)<p><a href="https:&#x2F;&#x2F;github.com&#x2F;aldarisbm&#x2F;memory">https:&#x2F;&#x2F;github.com&#x2F;aldarisbm&#x2F;memory</a><p>It&#x27;s still got ways to go, if someone wants to help let me know :)
评论 #36028521 未加载
kordlessagainalmost 2 years ago
Working on something similar that uses keyterm extraction for traversal of topics and fragments, without using Langchain. It&#x27;s not designed to be private, however: <a href="https:&#x2F;&#x2F;github.com&#x2F;FeatureBaseDB&#x2F;DocGPT&#x2F;tree&#x2F;main">https:&#x2F;&#x2F;github.com&#x2F;FeatureBaseDB&#x2F;DocGPT&#x2F;tree&#x2F;main</a>
Wronnayalmost 2 years ago
Wow. I keep a personal Wiki, Journal and use plain text accounting...<p>This project could help me create a personal AI which answers any questions to my life, finances or knowledge...
评论 #36028820 未加载
lyspalmost 2 years ago
Quick how-to&#x2F;demo:<p><a href="https:&#x2F;&#x2F;www.youtube.com&#x2F;watch?v=A3F5riM5BNE">https:&#x2F;&#x2F;www.youtube.com&#x2F;watch?v=A3F5riM5BNE</a><p>Also has a suggestion of a few alternative models to use.
daitangioalmost 2 years ago
Hi, very interesting... what are the memory&#x2F;disk requirements to run it? 16GB of RAM would be enough? I suggest to add these requirements to the README
评论 #36028461 未加载
评论 #36028851 未加载
zestypingalmost 2 years ago
Would someone do me the kindness of explaining (a little more) how this works?<p>It looks like you can ask a question and the model will use its combined knowledge of all your documents to figure out the answer. It looks like it isn&#x27;t fine-tuned or trained on all the documents, is that right? How is each document turned into an embedding, and then how does the model figure out which documents to consult to answer the question?
behnamohalmost 2 years ago
When you split a document into chunks, doesn&#x27;t some crucial information get cut in half? In that case, you&#x27;d probably lose that information in the context if that information was immediately followed by an irrelevant information that reduces the cosine similarity. Is there a &quot;smarter&quot; way to feed documents as context to LLMs?
评论 #36026354 未加载
divanalmost 2 years ago
This will still hallucinate, right?<p>Projects like this for using with your documents datasets are invaluable, but everything I&#x27;ve tried so far is hallucinating, so not practical. What&#x27;s the state of the art of the LLM without hallucination at the moment?
评论 #36030311 未加载
评论 #36029308 未加载
评论 #36029256 未加载
debbiedowneralmost 2 years ago
This is a shortcut&#x2F;workaround to transforming the private docs to a prompt:answer dataset and fine tuning right?<p>What would be the difference in user experience or information retrieval performance between the two?<p>My impression is it saves work on the dataset transformation and compute for fine tuning, so it must be less performant. Is there a reason to prefer the strategy here other than ease of setup?
superbiomealmost 2 years ago
Does something like this exist for local code repos? (Excuse my ignorance since the space is moving faster than light.)
评论 #36032426 未加载
评论 #36032389 未加载
ameliusalmost 2 years ago
With so many LLM options out there, how do we keep track of which ones are good?
评论 #36029582 未加载
roliszalmost 2 years ago
For some reason, downloading the model they suggest keeps failing. I tried to download it in Firefox and Edge. I&#x27;m using Windows, if that matters. Anyone else seeing similar issues?
sinandrei91almost 2 years ago
Is there a benchmark for retrieval from multiple ft documents? I tried the LangchainQA with Pinecone and wasn&#x27;t impressed with the search result when using it on my Zotero library.
ameliusalmost 2 years ago
How many tokens&#x2F;second on an average machine?
jaimehrubiksalmost 2 years ago
If you select a gpt4all model like GPT-J can this be used commercially or is there other dependency that limits the license?
Havocalmost 2 years ago
Would this work better with something like llama or a instruction following model like alpaca?
bohlenlabsalmost 2 years ago
So many good links here, thanks to the OP for sharing, and to all commenters as well!
seydoralmost 2 years ago
does this only work with llamaCPP ? I.e. can&#x27;t use GPU models with this?
评论 #36029513 未加载
ChocoluvHalmost 2 years ago
Always wondering pros&#x2F;cons of Chroma and Qdrant. Can someone tell me?
评论 #36028794 未加载
keeptryingalmost 2 years ago
This is the future.
yositoalmost 2 years ago
&gt; Put any and all your files into the source_documents directory<p>Why? Why can&#x27;t I define any directory (my existing Obsidian vault, for example) as the source directory?
评论 #36027961 未加载
udev4096almost 2 years ago
I posted it 9 days ago and somehow this one gets the attention. The same freaking post. Unbelievable<p><a href="https:&#x2F;&#x2F;news.ycombinator.com&#x2F;item?id=35914810" rel="nofollow">https:&#x2F;&#x2F;news.ycombinator.com&#x2F;item?id=35914810</a>
评论 #36028700 未加载