TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

Giving GPT “Infinite” Knowledge

121 pointsby sudoappsabout 2 years ago

15 comments

furyofantaresabout 2 years ago
Embeddings-based search is a nice improvement on search, but it&#x27;s still search. Relative to ChatGPT answering on its training data, I find embeddings-based search to be severely lacking. The right comparison is to traditional search, where it becomes favorable.<p>It has the same advantages search has over ChatGPT (being able to cite sources, being quite unlikely to hallucinate) and it has some of the advantages ChatGPT has over search (not needing exact query) - but in my experience it&#x27;s not really in the new category of information discovery that ChatGPT introduced us to.<p>Maybe with more context I&#x27;ll change my tune, but it&#x27;s very much at the whim of the context retrieval finding everything you need to answer the query. That&#x27;s easy for stuff that search is already good at, and so provides a better interface for search. But it&#x27;s hard for stuff that search isn&#x27;t good at, because, well: it&#x27;s search.
评论 #35866477 未加载
评论 #35866784 未加载
评论 #35867454 未加载
评论 #35867310 未加载
评论 #35867138 未加载
评论 #35867492 未加载
ftxbroabout 2 years ago
&gt; &quot;Once these models achieve a high level of comprehension, training larger models with more data may not offer significant improvements (not to be mistaken with reinforcement learning through human feedback). Instead, providing LLMs with real-time, relevant data for interpretation and understanding can make them more valuable.&quot;<p>To me this viewpoint looks totally alien. Imagine you have been training this model to predict the next token. At first it can barely interleave vowels and consonants. Then it can start making words, then whole sentences. Then it starts unlocking every cognitive ability one by one. It begins to pass nearly every human test and certification exam and psychological test of theory of mind.<p>Now imagine thinking at this point &quot;training larger models with more data may not offer significant improvements&quot; and deciding that&#x27;s why you stop scaling it. That makes absolutely no sense to me unless 1) you have no imagination or 2) you want to stop because you are scared to make superhuman intelligence or 3) you are lying to throw off competitors or regulators or other people.
评论 #35867406 未加载
评论 #35869134 未加载
评论 #35867826 未加载
评论 #35867577 未加载
评论 #35867351 未加载
评论 #35867975 未加载
评论 #35867390 未加载
评论 #35882251 未加载
评论 #35867754 未加载
Der_Einzigeabout 2 years ago
I get annoyed by articles like this. Yes, it&#x27;s cool to educate readers who aren&#x27;t aware of embeddings&#x2F;embeddings stores&#x2F;vectorDB technologies that this is possible.<p>What these articles don&#x27;t touch on is what to do once you&#x27;ve got the most relevant documents. Do you use the whole document as context directly? Do you summarize the documents first using the LLM (now the risk of hallucination in this step is added)? What about that trick where you shrink a whole document of context down to the embedding space of a single token (which is how ChatGPT is remembering the previous conversations). Doing that will be useful but still lossey<p>What about simply asking the LLM to craft its own search prompt to the DB given the user input, rather than returning articles that semantically match the query the closest? This would also make hybird search (keyword or bm25 + embeddings) more viable in the context of combining it with an LLM<p>Figuring out which of these choices to make, along with an awful lot more choices I&#x27;m likely not even thinking about right now, is what will seperate the useful from the useless LLM + Extractive knowledge systems
评论 #35866700 未加载
评论 #35866762 未加载
评论 #35866776 未加载
评论 #35866690 未加载
orasisabout 2 years ago
One caveat about about embedding based retrieval is that there is no guarantee that the embedded documents will look like the query.<p>One trick is to have a LLM hallucinate a document based on the query, and then embed that hallucinated document. Unfortunately this increases the latency since it incurs another round trip to the LLM.
评论 #35867183 未加载
评论 #35872477 未加载
评论 #35867589 未加载
评论 #35866991 未加载
评论 #35868228 未加载
评论 #35866969 未加载
Beltirasabout 2 years ago
I&#x27;m working on something where I need to basically add on the order of 150,000 tokens into the knowledge base of an LLM. Finding out slowly I need to delve into training a whole ass LLM to do it. Sigh.
评论 #35870216 未加载
评论 #35870823 未加载
评论 #35867254 未加载
chartpathabout 2 years ago
Search query expansion: <a href="https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;Query_expansion" rel="nofollow">https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;Query_expansion</a><p>We&#x27;ve done this in NLP and search forever. I guess even SQL query planners and other things that automatically rewrite queries might count.<p>It&#x27;s just that now the parameters seem squishier with a prompt interface. It&#x27;s almost like we need some kind of symbolic structure again.
sudoappsabout 2 years ago
If you are wondering what the latest is on giving LLM&#x27;s access to large amounts of data, I think this article is a good start. Seems like this is a space where there will be a ton of innovation so interested to learn what else is coming.
jeffchuberabout 2 years ago
hi everyone, this is jeff from Chroma (mentioned in the article) - happy to answer any questions.
评论 #35868589 未加载
iot_devsabout 2 years ago
A similar idea is been developed in: <a href="https:&#x2F;&#x2F;github.com&#x2F;pieroit&#x2F;cheshire-cat">https:&#x2F;&#x2F;github.com&#x2F;pieroit&#x2F;cheshire-cat</a>
pbhjpbhjabout 2 years ago
&gt;There is an important part of this prompt that is partially cut off from the image:<p>&gt;&gt; “If you don&#x27;t know the answer, just say that you don&#x27;t know, don&#x27;t try to make up an answer”<p>&#x2F;&#x2F;<p>It seems silly to make this part of the prompt rather than a separate parameter, surely we could design the response to be close to factual. Then run a checker to ascertain a score for the factuality of the output?
评论 #35866503 未加载
nicoabout 2 years ago
Can we build a model based purely on search?<p>The model searches until it finds an answer, including distance and resolution<p>Search is performed by a DB, the query then sub-queries LLMs on a tree of embeddings<p>Each coordinate of an embedding vector is a pair of coordinate and LLM<p>Like a dynamic dictionary, in which the definition for the word is an LLM trained on the word<p>Indexes become shortcuts to meanings that we can choose based on case and context<p>Does this exist already?
评论 #35867403 未加载
m3kw9about 2 years ago
This is like asking gpt to summarize what it found on Google, this is basically what bing does when you try to find stuff like hotels and other recent subjects. Not the revolution we are all expecting
A_D_E_P_Tabout 2 years ago
&quot;Infinite&quot; is a technical term with a highly specific meaning.<p>In this case, it can&#x27;t possibly be approached. It certainly can&#x27;t be attained.<p>Borges&#x27; Library of Babel, which represents all possible combinations of letters that can fit into a 400-page book, only contains some 25^1312000 books. And the overwhelming majority of its books are full of gibberish. The amount of &quot;knowledge&quot; that a LLM can learn or describe is VERY strictly bounded and strictly finite. (This is perhaps its defining characteristic.)<p>I know this is pedantic, but I am a philosopher of mathematics and this is a matter that&#x27;s rather important to me.
评论 #35868461 未加载
nadermxabout 2 years ago
I think someone did this <a href="https:&#x2F;&#x2F;github.com&#x2F;pashpashpash&#x2F;vault-ai">https:&#x2F;&#x2F;github.com&#x2F;pashpashpash&#x2F;vault-ai</a>
评论 #35867810 未加载
flukeshottabout 2 years ago
I wonder how effectively compressed LLMs are going to become...