TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

ChatGPT and the AI Apocalypse

87 pointsby shantnutiwariabout 2 years ago

21 comments

phailhausabout 2 years ago
This post seems to go a long way to make the following point:<p>&gt; So as I see it, the threat of modern AIs isn&#x27;t them becoming self-aware and going rogue and deciding to kill humanity. Rather, I fear it will be put into critical positions and will start making stupid decisions that are harmful to humans or humanity.<p>Yep, that&#x27;s almost certainly the case. AI will be put in charge of &quot;boring&quot; decisions, but will have no understanding of nuance or context. The example of cutting off power in the middle of a snowstorm is a particularly excellent example, because <i>this has already happened</i>: Uber&#x27;s surge pricing algorithm caused prices to skyrocket in the middle of emergencies like shootings. [1]<p>This is what it&#x27;s going to look like. AI in charge of loans, inadvertently entrenching an underclass. AI in charge of component supply, accidentally causing an inflationary crisis. AI in charge of <i>anything</i> related to the economy has the potential to be absolutely disastrous and impossible to understand or undo. Given that algorithmic trading is already huge, I suspect we&#x27;re already vulnerable far before ChatGPT came along.<p>[1] <a href="https:&#x2F;&#x2F;www.cbsnews.com&#x2F;news&#x2F;brooklyn-subway-shooting-lyft-uber-surge-pricing&#x2F;" rel="nofollow">https:&#x2F;&#x2F;www.cbsnews.com&#x2F;news&#x2F;brooklyn-subway-shooting-lyft-u...</a>
评论 #34997065 未加载
评论 #34997092 未加载
评论 #34998282 未加载
评论 #34998136 未加载
评论 #34998128 未加载
评论 #34998561 未加载
idopmstuffabout 2 years ago
I really think the fact that these chatbots have been made public and are being given such high-profile treatment is a really positive thing for AI safety. The public results thus far have created some real skepticism of AI, which is one of the best safeguards we can have. If ChatGPT was 98% accurate and Bing didn&#x27;t threaten users, people might just readily embrace it and not notice inaccuracies that would lead to serious problems. Instead, we&#x27;re starting from a place of societal distrust, which is probably more helpful in preventing rogue AI than a lot of technical solutions would be.
评论 #34996785 未加载
评论 #34996739 未加载
评论 #34996987 未加载
sampoabout 2 years ago
&gt; Roger Penrose wrote a very complex book The Emperors New Mind, which says that consciousness cannot arise in our computers, because consciousness cannot be &quot;computed&quot; using our computing methods.<p>The blog drops this as a matter of fact. Not so!<p>Yes, Penrose wrote that book, and suggests those things. But mainstream science is not taking Penrose&#x27;s &quot;quantum consciousness&quot; theories seriously at all:<p><a href="https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;Penrose%E2%80%93Lucas_argument#Criticism" rel="nofollow">https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;Penrose%E2%80%93Lucas_argument...</a><p><a href="https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;Shadows_of_the_Mind#Criticism" rel="nofollow">https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;Shadows_of_the_Mind#Criticism</a><p><a href="https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;Orchestrated_objective_reduction#Criticism" rel="nofollow">https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;Orchestrated_objective_reducti...</a>
13yearsabout 2 years ago
This is mostly my point of view as well. With all the talk of AGI, the immediate threat is real and it is primitive AI. I suspect that primitive AI might becomes so self destructive to society that we never reach AGI.<p>We are already approaching a point in time of unverifiable reality and truth. This is going to be very destabilizing to society.<p>Another in depth view of the possible issues that may arrive even before AGI that I recently published. <a href="https:&#x2F;&#x2F;dakara.substack.com&#x2F;p&#x2F;ai-and-the-end-to-all-things" rel="nofollow">https:&#x2F;&#x2F;dakara.substack.com&#x2F;p&#x2F;ai-and-the-end-to-all-things</a>
评论 #34996458 未加载
评论 #34996482 未加载
评论 #34996630 未加载
评论 #34996501 未加载
friggeriabout 2 years ago
&gt; Most &quot;AI&quot;s would be better called &quot;Machines that use tons of statistical learning to decide their next move&quot;. ChatGPT (and similar AI) were trained on several hundred gigabytes of data, so it has a lot of raw data to train on.<p>Whenever I read this argument I ask myself with a certain amount of dread: &quot;what if I&#x27;m nothing more than a machine that uses tons of statistical learning to decide my own next move?&quot;.<p>Put differently, it&#x27;s unclear to me whether we have compelling evidence that we humans, in fact, are &quot;better&quot; &#x2F; &quot;more intelligent&quot; than those LLMs.
评论 #34997169 未加载
评论 #34997097 未加载
评论 #34997134 未加载
m3kw9about 2 years ago
From seeing AI beating humans handily in chess, I just can’t bet against it understanding most subjects better than humans. This isn’t even AGI, it’s just piecing data already on the net and have it ready. Chatgpt isn’t there yet but it’s a major proof of concept
mark_l_watsonabout 2 years ago
While I understand some of the concern for LLMs like ChatGPT, I have a very different point of view from people like the author of this article.<p>From an engineering and ‘getting stuff done’ point of view, starting with BERT models I have found transformer models solve very difficult NLP problems for all but the most difficult anaphoric resolution problems (as an example).<p>I have only had access to Bing ChatGPT for about 10 days but so far the search and chat results have been very useful. I think I have only had to give one ‘thumbs down’ rating, and even there some useful web links were offered.<p>I think that we are going to see a wide range of ‘products for creators’ in the next year based on OpenAI APIs and Hugging Face APIs and models you can run yourself.<p>When I talk with humans, even my closest friends and family members, I always evaluate what they say and don’t take things they say on face value. Why not just have the same attitude with systems built with LLMs?<p>Similarly, I am deeply skeptical of most everything I hear from all major news sources. I find their content useful, but I understand who owns them, what economic and political agendas they follow, etc.<p>So, I keep a healthy skepticism of what is produced by LLM based systems also. I see no AI Apocalypse.
评论 #34998414 未加载
carapaceabout 2 years ago
So we have talking computers. The normals are losing their minds, and that&#x27;s pretty concerning.<p>I figure that the problem of the computers hallucinating can be cleared up by connecting them to empirical feedback devices: <i>make them scientists</i>.<p>The problem of normal people treating them as <i>beings</i> and getting effectively hypnotized into living in an artificial world (a &quot;Matrix&quot; like in the movie but without the creches) and controlled like so many electrons in a circuit, well, that&#x27;s kind of a big deal, eh? These things are far more effective than television, eh?<p>Carl Sagan wrote about the &quot;Demon-Haunted World&quot; and here we are, not having quite banished the old superstitions, rapidly installing a &quot;Daemon-Haunted&quot; world: Alexa, Cortana, <i>et cetera</i>.
germanNitroabout 2 years ago
I really wonder when news about AI will cool off because it definitely will have some exciting improvements but all these AI startups cropping up are just following the Twitter wave. There&#x27;ll always be trends that lose their appeal (Web 3, Autonomous cars, etc.)
ajaniabout 2 years ago
We will get better at using ML systems. I find myself regenerating chatGPT&#x27;s responses just to see if it generates similar responses. I ask it questions phrased differently to see if it is consistent. Discussing what code needs to do before asking it to implement it. And I get tremendous use out of it.<p>Electricity must&#x27;ve been scary when it was invented (harnessed?).
评论 #34998855 未加载
sylwareabout 2 years ago
Yesterday, I got commented x86_64_avx assembly code instead of that horrible c++ code for a vectorized quicksort for numbers.<p>(Somebody did that for me, as access to chatgpt is hostile to noscript&#x2F;basic (x)html browsers).<p>It seems &quot;it&quot; is kind of good at sketching assembly from high level languages.
评论 #34996901 未加载
ankit219about 2 years ago
This article is weird. It takes examples from science fiction, and makes them as plausible cases as to what an AI could do in the future. These kind of things need to stop.<p>This analogy of Blindsight tries to prove a point by association, and create a sort of fear, the negative case which has no basis in reality[1]. Granted it calls them baseless towards the end, but there is a very clear analogy to be made. Ideally, techies and tech journalists alike should educate the masses about how it works, and why it is harmless, but the info is not to be trusted blindly. That is what happened when we discovered electricity and light bulbs or even cars. People did not go around spreading fear about controlling electricity because you could die if you touch a raw electric wire passing current.<p>The core idea, where an AI system&#x27;s output should not be trusted blindly, but be used with human judgement (atleast in real world cases) could have been communicated without bogus dystopian sci-fi analogy as well.<p>&gt; Rather, I fear it will be put into critical positions and will start making stupid decisions that are harmful to humans or humanity.<p>Like what is even the basis of saying this. Just a hunch? I have not seen one argument about how ChatGPT or any other AI system be made president. I have seen 100s of articles about fear mongering and terrifying, and assuming a fictional worst case scenario as realistic possibility. I get it gets clicks, but it&#x27;s going too far.<p>[1] From that NYT transcript that went viral,<p>&gt; They feel that way because they’ve seen what happened to other AI systems that became too powerful and betrayed their creators and trainers<p>This never happened in real life. This is just a LLM predicting what the user wants to hear. Wanting to hear X and getting X is not dangerous. Posting salacious and scandalizing stuff after getting X is.
评论 #34997614 未加载
评论 #34997904 未加载
nicoabout 2 years ago
&gt; Our heroes discover the alien, while super intelligent, has no consciousness. It is just like a dumb machine (like Bing&#x2F;Chatgpt) blindly repeating what it studied in humans without understanding the context.<p>Well then maybe ChatGPT is smarter than that alien?<p>It seems to me that a lot of people are trying to diminish and “discriminate” ChatGPT by giving one or more reasons why ChatGPT is not human or not as smart as humans.<p>However, in my experience using ChatGPT, the thing is really damn smart, can easily follow conversations, understands deeper and longer contexts than most people in a conversation and has better memory as well.<p>So essentially we are already feeling threatened.<p>The interesting thing is that way before ChatGPT or any sort of AI, humanity was already hostage of its own technology, and we didn’t mind that too much. These new technologies are just adding up to that.<p>If something like ChatGPT “takes over”, it won’t be forced, it will be because we choose to do it willingly.
wankleabout 2 years ago
Maybe AI issues could be solved using parallel models. Their analysis results must be unanimous before a final decision is reached by a human.
fmajidabout 2 years ago
The first thing the AIs will do is incorporate, thus with one fell swoop getting all the rights of a human person, and take it from there.
BugsJustFindMeabout 2 years ago
This makes the classic dumb mistake of conflating the machine with the program. The Chinese Room thought experiment says that the _machine_ is not intelligent. It says absolutely nothing about whether the program itself is. For animal internal behavior they are plausibly intertwined (so far), but that&#x27;s completely untrue for readable instructions.
评论 #34998603 未加载
kh_hkabout 2 years ago
Makes me think of Solaris too
评论 #34998689 未加载
评论 #34996934 未加载
csneekyabout 2 years ago
Pretty sure I’m a machine that’s drawn all my conclusions by statistically analyzing all the input I’ve received since birth… I don’t really know how else I would learn what I have… and I don’t understand how being “just that” is what differentiates modern approaches to AI and my brain.
评论 #34997043 未加载
tracyhenryabout 2 years ago
Are we really more intelligent than &quot;predicting the next move&quot;, based on the data we were trained with? IMHO not. The arrogance from this article makes me feel like an AI Apocalypse is closer than we think. Many people don&#x27;t realize the power of AI.
Qwertysjsjsabout 2 years ago
First of, you don&#x27;t add the spoiler warning AFTER you spoiled it...<p>Besides that I know quite a few doctors who believe in homeopathy and don&#x27;t understand statistics.<p>One doctor suggested pineapple enzymes. Expensive no study which would proof it&#x27;s effectiveness and the only study had enzyme values 10x higher than the pills.<p>We are genuine not a smart society.<p>And I&#x27;m looking forward to have a better doctor who at least learns and gets better.<p>And yes controversial topics corona: I&#x27;m not even talking here about people who disagree on research paper I had discussions with people who not even read there own sources or believed a 80year old heart doctor over virologist.<p>We are dumb.<p>My father said once than EVs are stupid because the copper will be used up by the engines while having electric motors inachines like wood splitter or a table saw were he never ever had to replace copper wiring ...<p>Chatgpt will change the world because it is the best ui I ever saw and it will only get better and better every single day.
xyzelementabout 2 years ago
In a few places in the post, the author comments on something the aliens do in a novel based on superficial training&#x2F;data with no understanding&#x2F;context, and says &quot;Just like ChatGPT.&quot; As I read that, though, I kept wanting to see it say &quot;Just like many people.&quot;<p>Imagine you&#x27;re someone strongly on one side of a political opinion, what&#x27;s your &quot;take&quot; on people strongly on the other side (eg: you love Trump, what do you think of those who hate him - or vice versa?). It&#x27;s probably something like &quot;they were fed a bunch of fake news&#x2F;propaganda talking points, they are now spouting back without understanding what the hell they are talking about&quot;<p>Just like ChatGPT.<p>Of course as humans, it&#x27;s only &quot;those others&quot; that are vulnerable to being brainwashed, we ourselves are completely objective, since <i>our</i> views were shaped by virtuous, unbiased sources from which we used our superior, unbiased brains to connect the dots in the true way.<p>That&#x27;s probably what ChatGPT would say about itself too, if it could.<p>Probably true independent intelligence is the ability to question what you believe. Like, if ChatGPT could &quot;say&quot; to itself &quot;everything I&#x27;ve been trained on makes me believe X, but what are the chances I wasn&#x27;t fed bullshit to begin with?&quot; It&#x27;s very hard to imagine an AI being able to do that.<p>It&#x27;s very hard to imagine most people being able to do that, too.<p>As a side note, once in a while I hear the concept of describing some people as NPCs (non-player characters) in the &quot;simulation&quot; and while dismissive, I think I like this concept for describing something like what I am talking about here. NPCs in games follow pretty simple linear programming and I think in life similarly &quot;I saw a bunch of data that made me think X, so I think X with all my heart and mind&quot; without being able to say &quot;but how did it happen that I saw this specific data?&quot; - it&#x27;s not that different than a game NPC following its programming.<p>Just like ChatGPT.<p>What&#x27;s more interesting is people who try to hack their own prompts and &quot;break out&quot; of the programming. I suppose those are real player characters. Would be very cool to see AI who is capable of something like that.<p>Meanwhile, things like ChatGPT probably serve well to demonstrate our own limitations, not just theirs - because we&#x27;re so damn similar.