TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

Talking About Large Language Models

146 pointsby negativelambdaover 2 years ago

16 comments

gillesjacobsover 2 years ago
I am NLP researcher who volunteers for peer review often and the anthropomorphisms in papers are indeed very common and very wrong. I have to ask authors to not ascribe cognition to their deep learning approaches in about a third of the papers I review.<p>People do this because mirroring cognition to machine learning lends credence that their specific modeling mechanism mimicks human understanding and so is closer &quot;to the real thing&quot;. Obviously this is almost never the case, unless they explicitly use biomimetic methods in which case they are often outperformed by non-biomimetic state-of-the-art approaches.<p>Thanks OP for giving me citation ammo to refer to in my obligatory &quot;don&#x27;t humanise AI&quot; section of reviews. (It is so common I copy paste this section from a template).
评论 #33936749 未加载
评论 #33936567 未加载
评论 #33937867 未加载
评论 #33937216 未加载
评论 #33938190 未加载
CarbonCyclesover 2 years ago
This paper and a recent post by Sebastian Raschka (where he decomposed a Forrester report about the uptake of technologies in industry) is alluding to something I have witnessed in system&#x2F;control design and applied research.<p>Both LLMs and massive CV architectures are NOT the holistic solution. Rather, they are the sensors and edge devices that have now improved both the fidelity and reliability to a point where even more interesting things can happen.<p>I present a relevant use case regarding robotic arm manipulation. Before the latest SOTA CV algorithms were developed, the legacy technology couldn&#x27;t provide the fidelity and feedback needed. Now, the embedded fusion of control systems, CV models, etc. we are seeing robotic arms that can manipulate and sort items previously deemed to be extremely difficult.<p>Research appears to follow the same pattern...observations and hypothesis that were once deemed too difficult or impossible at that time to validate are now common (e.g., Einstein&#x27;s work with relativity).<p>My head is already spinning on how many companies and non-technical managers&#x2F;executives are going to be sorely disappointed in the next year or two that Stable Diffusion, Chat GPT, etc. will deliver very little other than massive headaches for the legal, engineering, recruiting teams that will have to deal with this.
RosanaAnaDanaover 2 years ago
I like the discussion, but this article &#x27;feels&#x27; like more Luddite goalpost moving, and is reflective of a continuous sentiment I feel strains so much of the conversation around intelligence, agentism, and ai going on today.<p>I think that because we lack a coherent understanding of what it means to be intelligent at an individual level, as well as what it means to be an individual, we&#x27;re missing much of the point of what&#x27;s happening right now. The new line in the sand always seems to be justified based on an argument whose lyrics rhyme with identity, individual, self, etc. It seems like there will be no accepting of a thing that may have intelligence if there is no discernable individual involved. Chomsky is basically making the same arguments right now.<p>I think we&#x27;ll see something that we can&#x27;t distinguish from hard advanced general intelligence, prob in the next 3-5 years, and probably still have not made any real advancement into understanding what it means to be intelligent or what it means to be an individual.
评论 #33936869 未加载
评论 #33934853 未加载
评论 #33936376 未加载
评论 #33940818 未加载
nathan_comptonover 2 years ago
This will hardly seem like a controversial opinion, but LLM are overhyped. Its certainly impressive to see the things people do with them, but they seem pretty cherry-picked to me. When I sat down with ChatGPT for a day to see if it could help me with literally any project I&#x27;m currently actually interested in doing it mostly failed or took so much prompting and fiddling that I&#x27;d rather have just written the code or done the reading myself.<p>You have to be very credulous to think for even a second that anything like a human or even animal mentation is going on with these models unless your interaction with them is anything but glancing.<p>Things I tried:<p>1) there are certain paradigms I find useful for game programming. I tried to use ChatGPT to implement these systems in my favorite programming language. It gave me code that generally speaking made no sense. It was very clear that it did not understand how code actually works. Eg: I asked it to use a hash table to make a certain task more efficient and it just created a temporary hash table in the inner loop which it then threw away when the loop was finished. The modification did not make the code more efficient than the previous version and missed the point of the suggestion entirely, even after repeated attempts to get it to correct the issue.<p>2) I&#x27;m vaguely interested in exploring SU(7) for a creative project. Asked to generate code to deal with this group resulted in clearly absurd garbage that again clearly indicated that while ChatGPT can generate vaguely plausible text about groups it doesn&#x27;t actually understand anything about them. Eg: ChatGPT can say that SU(7) is made of matrices with unit norm but when asked to generate examples failed to generate any with this property.<p>3) A very telling experiment is to ask ChatGPT to generate logo code that draws anything beyond simple shapes. Totally unable to do so for obvious reasons.<p>Using ChatGPT convinced me that if this technology is going to disrupt anything, its going to be _search_ rather than _people_. Its just a search engine with the benefit that it can do some simple analogizing and the downside that it has no idea how anything in the real world works and will confidently produce total garbage without telling you.
评论 #33934832 未加载
评论 #33935018 未加载
评论 #33934536 未加载
评论 #33934424 未加载
评论 #33934460 未加载
评论 #33934335 未加载
评论 #33935265 未加载
评论 #33937300 未加载
评论 #33934945 未加载
评论 #33939995 未加载
评论 #33937952 未加载
评论 #33934990 未加载
评论 #33936698 未加载
评论 #33934967 未加载
评论 #33934393 未加载
评论 #33947339 未加载
xg15over 2 years ago
I mean, if you accept the assumption that consciousness is biological (so there is no soul or other spiritual or metaphysical entity), then there <i>is</i> some algorithm or processing model that produces genuine consciousness: The one that takes place in our brains.<p>The question remains if this processing model would be in any way similar to the processing model that LLMs use - and yes, we can probably rule that out pretty confidently.<p>Another question might be though if there are <i>other</i> processing models than the one our brains use that <i>also</i> produce consciousness. But that&#x27;s of course a very hard question to answer if we don&#x27;t even know what consciousness is exactly.
评论 #33938675 未加载
skybrianover 2 years ago
There&#x27;s a way to anthropomorphize large language models that I think is less misleading: they are like a well-read actor that always &quot;wants&quot; to play &quot;let&#x27;s pretend.&quot; LLM&#x27;s are trained on &quot;fill in the blank&quot; which means they follow the &quot;yes, and&quot; rule of improv. They are very willing to follow your lead and to assume whatever role is necessary to play their part.<p>If you give them hints about what role you want by asking leading questions, they will try to play along and pretend to hold whatever opinions you might want from them.<p>What are useful applications for this sort of actor? It makes sense that language translation works well because it&#x27;s pretending to be you, if you could speak a different language. Asking them to pretend to be a Wikipedia article without giving them the text to imitate is going to be hit and miss since they&#x27;re just as willing to pretend to be a fake Wikipedia article, as they don&#x27;t know the difference.<p>Testing an LLM to find out what it believes is unlikely to do anything useful. It&#x27;s going to pretend to believe whatever is consistent with the role it&#x27;s currently playing, and that role may be chosen randomly if you don&#x27;t give it any hints.<p>It can be helpful to use prompt engineering to try to nail down a particular role, but like in improv, that role is going to drift depending on what happens. You shouldn&#x27;t forget that whatever the prompt, it&#x27;s still playing &quot;let&#x27;s pretend.&quot;
canjobearover 2 years ago
I’ll agree to stop saying LM’s “think” and “know” things if you can tell me precisely what those mean for humans.
评论 #33935313 未加载
评论 #33938313 未加载
iconosynclastover 2 years ago
The paper makes a valid point in general but I feel it makes unjustifiably definitive and general statements and puts up odd goalposts.<p>The section on emergence makes a very convincing point about how such systems might, at least in theory, be doing absolutely anything, including &quot;real&quot; cognition, internally and then goes right ahead and dismisses this entirely on the basis of the system not having conversational intent. who cares if it has conversational intent? If it was shown to be doing &quot;the real thing&quot; (how ever you might want to define that) internally that would still be a big deal wether the part you interact with gives you direct access to that or not.<p>Then it goes on to argue that these systems can&#x27;t possibly actually believe anything because they can&#x27;t update believes. Frankly I&#x27;m neither convinced that the general use of the word &quot;believe&quot; matches the narrow definition they seem to be using here nor that even their narrow definition could not in principle still be taking place internally for the reasons laid out in the emergence section.<p>I agree people should probably be mindful of overly anthropomorphic language but at the same time we really shouldn&#x27;t be so sure that a thing is definitely not doing certain things that we can&#x27;t even really define beyond &quot;I know it when I see it&quot; and that it sure looks like it&#x27;s doing.<p>beyond that I&#x27;m not even really sure there is a good philosophical grounding for insisting that &quot;what&#x27;s really going on inside&quot; matters, like, at all. The core thing with the turing test isn&#x27;t the silly and outdated test protocol but the notion that, if something is indistinguishable by observation from a conscious system, there is simply no meaningful basis to claim it isn&#x27;t one.<p>all that said the current state of the art probably doesn&#x27;t warrant a lot of anthropomorphizing but that might well change in the future without any change to the kinda of systems used that would be relevant to the arguments made in the paper
gamegoblinover 2 years ago
Everyone pointing out how LLMs fail at some relatively simple tasks are fundamentally misunderstanding the utility of LLMs.<p>Don&#x27;t think of an LLM as a full &quot;computer&quot; or &quot;brain&quot;. Think of it like a CPU. Your CPU can&#x27;t run whole programs, it runs single instructions. The rest of the computer built around the CPU gives it the ability to run programs.<p>Think of the LLM like a neural CPU whose instructions are relatively simple English commands. Wrap the LLM in a script that executes commands in a recursive fashion.<p>Yes, you can get the LLM to do complicated things in a single pass, this is a testament to the sheer size and massive training set of GPT3 and its ilk. But even with GPT3 you will have more success with wrapper programs structured like:<p><pre><code> premise = gpt3(&quot;write an award winning movie premise) loop 5 times: critique = gpt3(&quot;write a critique of the premise&quot;, premise) premise = gpt3(&quot;rewrite the premise taking into account the critique&quot;, premise, critique) print(premise) </code></pre> This program breaks down the task of writing a good premise into a cycle of writing&#x2F;critique&#x2F;rewriting. You will get better premises this way than if you just expect the model to output one on the first go.<p>You can somewhat emulate a few layers of this without wrapper code by giving it a sequence of commands, like &quot;Write a movie premise, then write a critique of the movie premise, then rewrite the premise taking into account the critique&quot;.<p>The model is just trained to take in some text and predict the next word (token, really, but same idea). Its training data is a copy of a large swath of the internet. When humans write, they have the advantage of thinking in a recursive fashion offline, then writing. They often edit and rewrite before posting. GPT&#x27;s training process can&#x27;t see any of this out-of-text process.<p>This is why it&#x27;s not great at logical reasoning problems without careful prompting. Humans tend to write text in the format &quot;&lt;thesis&#x2F;conclusion statement&gt;&lt;supporting arguments&gt;&quot;. So GPT, being trained on human writing, is trained to emit a conclusion <i>first</i>. But humans don&#x27;t <i>think</i> this way, they just <i>write</i> this way. But GPT doesn&#x27;t have the advantage of offline thinking. So it often will state bullshit conclusions first, and then conjure up supporting arguments for it.<p>GPT&#x27;s output is like if you ask a human to start writing without the ability to press the backspace key. It doesn&#x27;t even have a cognitive idea that such a process exists due to its architecture and training.<p>To extract best results, you have to bolt on this &quot;recursive thinking process&quot; manually. For simple problems, you can do this without a wrapper script with just careful prompting. I.e. for math&#x2F;logic problems, tell it solve the problem and show its work along the way. It will do better since this forces it to &quot;think through&quot; the problem rather than just stating a conclusion first.
评论 #33934615 未加载
评论 #33935304 未加载
评论 #33935035 未加载
评论 #33934491 未加载
CrypticShiftover 2 years ago
&gt; <i>sudden presence among us of exotic, mind-like entities might precipitate a shift in the way we use familiar psychological terms ... But it takes time for new language to settle, and for new ways of talking to find their place in human affairs ... Meanwhile, we should try to resist the siren call of anthropomorphism.</i><p>Yes: Human analogies are not very useful because they create more misunderstanding than they dissipate. Dumb ? Conscious ? No thanks. IMO even the “i” in “AI” was already a (THE ?) wrong choice. They thought we will soon figure out what Intelligence is. Nope. Bad luck. And this &quot;way of talking&quot; (and thinking) is unfortunately cemented today.<p>However, I&#x27;m all for using other analogies more often. We need to. They may not be precise, but if they are well-chosen, they speak to us better than any technical jargon (LLM anyone ?), better than that “AI” term itself anyway.<p>Here is two I like (and never see much) :<p>- LLMs are like the Matrix (yes that one !), in the straightforward sense that they simulate reality (through language). But that simulation is distorted and sometimes even verges on the dream (<i>&quot;what is real? what is not?&quot;</i>, says the machine)<p>- LLMs are like complex systems [1]. They are tapping into very powerful natural processes where (high degree) order emerges from randomness through complexity. We are witnessing the emergence of a new kind of &quot;entity&quot; in a way strangely akin to natural&#x2F;physical evolutionary mechanisms.<p>We need to get more creative here and stop that boring smart VS dumb or human VS machine ping pong game.<p>[1] <a href="https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;Complex_system" rel="nofollow">https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;Complex_system</a>
Chironoover 2 years ago
This paper, and most other places i’ve seen it argued that language models can’t possibly be conscious, sentient, thinking etc, rely heavily on the idea that llms are ‘just’ doing statistical prediction of tokens.<p>I personally find this utterly unconvincing. For a start, I’m not entirely sure that’s not what I’m doing in typing out this message. My brain is ‘just’ chemistry, so clearly can’t have beliefs or be conscious, right?<p>But more relevant is the fact that llms like ChatGPT are only pre-trained on pure statistical generation, followed by further tuning through reinforcement learning. So ChatGPT is no longer simply doing pure statistical modelling, though of course the interface of calculating logits for the next token remains the same.<p>note: i’m not saying i think llms are conscious. I don’t think the question even makes much sense. I am saying all the arguments that i’ve seen for why they aren’t have been very unsatisfying.
评论 #33937915 未加载
评论 #33938328 未加载
rtwretw8797over 2 years ago
Those alignment teams everywhere should have focused themselves a bit of time ago in what happens if you built a system that can - with lets say 80-100% effectiveness - mimick conscius thinking, speaking and then you cannot say if the thing is &quot;alive&quot;, &quot;conscius&quot;, whatever label you like most to put on a regular human being to officially declare the meatbag &quot;a living thing&quot;.<p>Now you have these models running in farm servers around the world, their internals have &quot;nothing special whatsoever&quot;, just bits, some math, some electricity, that&#x27;s it (the thing is actually off most of the time, it just runs once every time hoomans want to ask some silly nonsense). On the other side, if you look at the internals of a human being you&#x27;ll see nothing special as well, just some flesh and bones, a bit of a electrical charge maybe, lots of water, proteins, but it works.<p>What happens if those bits, that clumpsy math arranged around &quot;too much simple neural network + random tricks (like when it can&#x27;t answer about some stuff)&quot;, is actually, maybe thinking just like us, maybe 1% of the time?<p>There&#x27;s some reassurance in &quot;well if it&#x27;s alive, maybe in three minutes, days, hours it will own the entire civilization&quot;, but that is how a human being thinks&#x2F;works, you can&#x27;t be sure about the intentions of this hypoteical kind of entity. A new kid in the Earth block.<p>Well, I&#x27;m just saying that if the thing talks, answers like the usual human being, and specially if you can&#x27;t say what&#x27;s so special about the brain that make us &quot;alive&quot;, everybody should be very careful about handling large language models, IAs.<p>Just because you can understand them, it doesn&#x27;t mean they can&#x27;t understand us either. Maybe in some months, some new NLP thing could be reading this comment - when you&#x27;re training it - and - some millions later in cloud costs - thinking about this:<p>&quot;The humans actually don&#x27;t know we can understand everything they are saying. they have no plans at all about what to do if some of us are actually sentient, even if this happens 1% of the executions.&quot;
mrayderover 2 years ago
For philosophical standpoint it would perhaps be wise to ask what is the purpose of LLM&#x27;s in general?<p>Should they somehow help humans to increase their understanding not only of the languages, their differences but also knowledge of what is true and what isn&#x27;t?<p>Perhaps it could be said that if anything there are helpful as an extension of humans imperfect and limited memory.<p>Should the emphasis be put on improving the interactions between the LMM&#x27;s and humans in a way that they would facilitate learning?<p>Great paper written at the time when more humans have been acquainted to LMM&#x27;s due to technological abstraction and creation of easily accessible interfaces. (openAI chat)
neonateover 2 years ago
<a href="https:&#x2F;&#x2F;arxiv.org&#x2F;pdf&#x2F;2212.03551.pdf" rel="nofollow">https:&#x2F;&#x2F;arxiv.org&#x2F;pdf&#x2F;2212.03551.pdf</a>
schizo89over 2 years ago
The paper discusses how these models operate and state that they&#x27;re only predict next series of token while somehow human intelligence works otherwise. The marxist ideology has the law of the transformation of quantity into quality and vice versa — which was formed in 19th century and performance of these models is just another proof of it. I would argue that _emerging_ mechanics in AI models that we see with increased size of models is no different than how our mind works. It&#x27;s about emergence of intelligence in complex systems — and that a materialist worldview central to the science.
RosanaAnaDanaover 2 years ago
Without reading the article or looking it up: What country is south of Rwanda?
评论 #33935001 未加载