TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

Ask HN: How come LLM/GPT have analysis skills, if they are mere text predictor?

2 pointsby jsdeveloperabout 2 years ago
The more I have been using LLM&#x2F;GPT, I am highly doubtfull that the secrete sauce is just train it on large corpuse of text and codes. Various LLM have various kind of analytical skills. This analytical skills can&#x27;t be explained with &quot;Oh but they are just text predictor, nothing more than that&quot;.<p>I wanted to ask HN here, how this analytical skills are possible, if its mere doing text prediction.<p>I asked a LLM&#x2F;GPT model: &gt;&quot;I want a word similar to accomplice , but which is positive meaning. It replied.<p>&gt;&quot;A word that is similar to accomplice but has a positive connotation is &quot;ally&quot;. An ally is someone who supports and helps another person or group in achieving a common goal. This word is often used in the context of social justice movements or political campaigns, where individuals or groups work together to bring about positive change. Other words that you might consider include &quot;partner&quot;, &quot;collaborator&quot;, or &quot;co-conspirator&quot; (although this last one may still carry some negative connotations depending on the context).&quot;<p>Now I don&#x27;t belive that the large language corpus its been trained on, had some where this definition, which it vomited out. It exactly understood and analyszed what I wanted and gave me.<p>Morever I have given it simple code C++ functions and asked it if I give following inputs , what would be the output. And it gave me correct answer.<p>This &quot;General Analytics&quot; for anythings seems to me dubious and based on my programming knowledge, I don&#x27;t think mere curve fitting models can create this anaylitics skills.<p>Again repeating what I am seeking: I wanted to ask HN here, how this analytical skills are possible, if its mere doing text prediction.

2 comments

aniijbodabout 2 years ago
Yeah, it turns out that it is totally misleading to dismiss something that is based upon text prediction as being only capable of things that require text prediction. In order to be able to predict the &#x27;correct text&#x27;, you, as a human, or as a &#x27;device&#x27; are not just &#x27;looking for instances where &#x27;b follows a&#x27; in your training data. This would just amount to nothing more than running a query on a database, finding the data in there and then outputting it.<p>Sure, there is a bit of this going on with an LLM system.<p>But it turns out that both the finding and the outputting in a neural net are so different to what is happening in a database-driven solution, that although the information contained within the answer might not be something that would be impossible to derive from a &#x27;conventional&#x27; database and therefore still make it unrealistic to claim that a conventional database has any &#x27;understanding&#x27; of the query that it is attempting to execute, this turns out not to be the case with a neural net-based solution like ChatGPT.<p>In order to &#x27;predict the next word&#x27;, the neural net can&#x27;t just look it up, because it is unlikely to be there. The neural net needs to have used its training process to have built up what we might call &#x27;intuitions&#x27;. It isn&#x27;t &#x27;just&#x27; consulting its training data. It is mostly consulting its &#x27;intuition-base&#x27;. To all intents and purposes, it is consulting its &#x27;understanding&#x27; of the training data, rather than consulting the data itself. This intuition-base is so much more optimised than any conventional data base that lookup is near-instantaneous. That&#x27;s why, when we &#x27;try to figure something out&#x27; we can need to take a little time to &#x27;think things through&#x27;, but when we react to something, it&#x27;s instant, nee jerk response. Using our own minds, when we are instantly reacting to something, are using the brain&#x27;s neural net in the same way as ChatGPT, but if we are are figuring things out on a &#x27;if this then do that&#x27; basis, our minds are working more like a computer running lines of code. Most of ChatGPT is based on NOT &#x27;just&#x27; running lines of code, but there is a tad of code in there as well as a neural net. But to cut a long story short, yes, the intuition-base side gives ChatGPT its ability to analyse and understand and the word prediction aspect is jus ta deceptively misleading characterisation of what&#x27;s really going on.
hackernotengabout 2 years ago
the base LLM is token predictor - given some stream of tokens, guess the next token, and the next, etc. But ChatGPT is trained on top of the LLM with a whole bunch of example questions+answers, and also most likely they are augmenting your prompt with additional prefix&#x2F;suffix to nudge the model to output a chat-like answer&#x2F;response to your input... that&#x27;s my guess and you can read more in the papers on how they trained it.