TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

ChatGPT is a blurry JPEG of the web

574 pointsby ssaddiover 2 years ago

54 comments

bitlaxover 2 years ago
<a href="https:&#x2F;&#x2F;archive.ph&#x2F;VbwGB" rel="nofollow">https:&#x2F;&#x2F;archive.ph&#x2F;VbwGB</a>
thundergolferover 2 years ago
&gt; ChatGPT is so good at this form of interpolation that people find it entertaining: they’ve discovered a “blur” tool for paragraphs instead of photos, and are having a blast playing with it.<p>“‘blur’ tool for paragraphs” is such a good way of describing the most prominent and remarkable skill of ChatGPT.<p>It is fun, but so obviously trades off against what makes paragraphs great. It is apt that this essay against ChatGPT blurry language appears on The New Yorker, a publication so known for its literary particularism. ChatGPT smears are amusing, but they are probably also yet another nail in the coffin of the literary society. Nowadays we are not careful readers; we skim, skip, and seek tools to sum up whole books. Human knowledge is in the ability to produce the particular and the ability to recognize it. For philosophers such as Iris Murdoch, careful attention to the particular, to just the right adjective in describing a friend, is a moral attention.<p>With Google we took the job of storing quotes (and other decompressed writing) and made it the remit of the machine. But we still asked for it back uncompressed. Here with ChatGPT, we are beginning to accept back a blur, because the Declaration of Sock Independence is immediately, Tiktok-speedily funny, and no one’s reading the original declaration anymore anyway.
评论 #34730543 未加载
评论 #34731961 未加载
评论 #34729157 未加载
评论 #34731422 未加载
评论 #34730000 未加载
评论 #34736303 未加载
评论 #34733946 未加载
评论 #34731971 未加载
评论 #34732127 未加载
评论 #34729421 未加载
评论 #34734339 未加载
nneonneoover 2 years ago
This is <i>very</i> well written, and probably one of my favorite takes on the whole ChatGPT thing. This sentence in particular:<p>&gt; Indeed, a useful criterion for gauging a large-language model’s quality might be the willingness of a company to use the text that it generates as training material for a new model.<p>It seems obvious that future GPTs should not be trained on the current GPT&#x27;s output, just as future DALL-Es should not be trained on current DALL-E outputs, because the recursive feedback loop would just yield nonsense. But, a recursive feedback loop is <i>exactly</i> what superhuman models like AlphaZero use. Further, AlphaZero is even trained on its own output even during the phase where it performs worse than humans.<p>There are, obviously, a whole bunch of reasons for this. The &quot;rules&quot; for whether text is &quot;right&quot; or not are way fuzzier than the &quot;rules&quot; for whether a move in Go is right or not. But, it&#x27;s not implausible that some future model will simply have a superhuman learning rate and a superhuman ability to distinguish &quot;right&quot; from &quot;wrong&quot; - this paragraph will look downright prophetic then.
评论 #34728521 未加载
评论 #34728641 未加载
评论 #34735466 未加载
评论 #34730446 未加载
评论 #34773776 未加载
评论 #34731471 未加载
评论 #34730328 未加载
评论 #34731032 未加载
评论 #34729612 未加载
评论 #34728040 未加载
评论 #34732143 未加载
mrtksnover 2 years ago
They can always use AI based solutions to unblur the JPEG, like this: <a href="https:&#x2F;&#x2F;twitter.com&#x2F;maxhkw&#x2F;status&#x2F;1373063086282739715" rel="nofollow">https:&#x2F;&#x2F;twitter.com&#x2F;maxhkw&#x2F;status&#x2F;1373063086282739715</a>
评论 #34726309 未加载
评论 #34726284 未加载
评论 #34726318 未加载
评论 #34758243 未加载
评论 #34731134 未加载
评论 #34726317 未加载
Agrailloover 2 years ago
&gt; Imagine what it would look like if ChatGPT were a lossless algorithm. If that were the case, it would always answer questions by providing a verbatim quote from a relevant Web page. We would probably regard the software as only a slight improvement over a conventional search engine, and be less impressed by it<p>The story is an impressive piece, but I think as with many of us, it&#x27;s a personal projection of expectations on results. One example from my experience. In the book &quot;Jim Carter - Sky Spy, Memoirs of a U-2 Pilot&quot; there was an interesting story about the moment when U-2 was used for capturing the photo of a big area at the Pacific to save the life of a lost seaman. The story was very interesting and I always wanted to know more, technical details, people involvement etc. Searching with Google ten years ago didn&#x27;t help, I rephrased the names, changed the date (used even the range operator) to no avail. And recently I asked several LLM-based bots about it. You can guess it. They ignored my constrains at best and hallucinate at worst. One even invented a mixed reality story when Francis Gary Powers actually flew not one but with a co-pilot and the latter ended up in the Pacific and was saved. Very funny, but I wasn&#x27;t impressed. But if one of them scraped the far corners of web discussion boards and saved a first-person account of someone who took part in it and gave it to me, I would be really impressed.
notShabuover 2 years ago
The compression &amp; blur analogy also applies to human minds as well. If you focus on fidelity, you have to increase storage and specialize in a narrow domain. If you want a bit of everything, then blurring and destructive compression is the only way. E.g. a &quot;book smart&quot; vs &quot;street smart&quot; difference.<p>&quot;mastery&quot; can be considered a hyper efficient destructive compression (experts are often unable to articulate or teach to beginners) that reduces latency of response to such extreme levels that they seem to be predicting the future or reacting at godlike speeds.
评论 #34728849 未加载
评论 #34728527 未加载
secabeenover 2 years ago
This is a decent summary. I&#x27;ve been thinking about how ChatGPT by it&#x27;s very nature destroys context and source reputation. When I search for something on the Internet, I get a link to the original content, which I can then evaluate based on my knowledge and the reputation of the original source. Wikipedia is the same, with a big emphasis on citation. ChatGPT and other LLMs destroy that context and knowledge, giving me no tools to evaluate the sources they&#x27;re using.
评论 #34726928 未加载
评论 #34732246 未加载
评论 #34730391 未加载
评论 #34727434 未加载
评论 #34766350 未加载
wvenableover 2 years ago
I don&#x27;t like this analogy; I think why I don&#x27;t like it is in the intent. With JPEG in the intent is produce an image indistinguishable from the original. Xerox didn&#x27;t intend to create photocopier that produces incorrect copies. The artifacts are failures of the JPEG algorithm to do what it&#x27;s supposed to within its constraints.<p>GPT is not trying to create a reproduction of it&#x27;s source material and simply failing at the task. Compression and GPT are both mathematical processes they aren&#x27;t the same process; JPEG is taking the original image and throwing away some of the detail. GPT is processing content to apply weights to a model; if that is reversible to the original content it is considered a failure.
评论 #34728563 未加载
评论 #34729931 未加载
评论 #34732802 未加载
评论 #34728579 未加载
评论 #34733153 未加载
the_afover 2 years ago
Since this article was written by Ted Chiang, just for fun I asked ChatGPT to summarize the plot of &quot;Understand&quot;.<p>Apparently ChatGPT thinks &quot;Understand&quot; is about the government who is pursuing someone called Gary Whittle who has superintelligence (well, at least it got <i>one</i> detail right). When challenged (&quot;no, the government is not the antagonist, but there is one <i>person</i>...&quot;) ChatGPT amends its summary to this:<p>&gt; &quot;George Millwright is Gary Whittle&#x27;s former supervisor and is depicted as being jealous of Gary&#x27;s newfound abilities. He becomes obsessed with Gary and is determined to bring him down, even going so far as to threaten his family. George Millwright&#x27;s actions drive much of the conflict in the story and serve as a reminder of the potential dangers of unchecked ambition and envy.&quot;<p>I&#x27;m honestly fascinated by ChatGPT&#x27;s &quot;hallucinations&quot;. I mean, it all makes perfect sense. Its summary is a potential scifi story -- albeit a poor, completely clichéd one -- but this is not at all what happens in &quot;Understand&quot;!<p>Text compression indeed.
评论 #34735350 未加载
zetazzedover 2 years ago
Damn, I hate to plug products on HN, but I&#x27;d say that the New Yorker is the one subscription I&#x27;ve loved maintaining throughout my life. First got it right out of college and appreciate it 20 years later.<p>Everyone is publishing think pieces about ChatGPT - yawn. But only the New Yorker said, hmm, how about if we get frickin&#x27; Ted Chiang to write a think piece? (It is predictably very well written.)
评论 #34726505 未加载
评论 #34727037 未加载
评论 #34726886 未加载
评论 #34730709 未加载
评论 #34731465 未加载
评论 #34727887 未加载
评论 #34727775 未加载
评论 #34727853 未加载
评论 #34728651 未加载
评论 #34732733 未加载
评论 #34727216 未加载
评论 #34727893 未加载
评论 #34732121 未加载
Ari_Rahikkalaover 2 years ago
&gt; Models like ChatGPT aren’t eligible for the Hutter Prize for a variety of reasons, one of which is that they don’t reconstruct the original text precisely—i.e., they don’t perform lossless compression.<p>Small nit: The lossiness is not a problem at all. Entropy coding turns an imperfect, lossy predictor into a lossless data compressor, and the better the predictor, the better the compression ratio. All Hutter Prize contestants anywhere near the top use it. The connection at a mathematical level is direct and straightforward enough that &quot;bits per byte&quot; is a common number used in benchmarking language models, despite the fact that they are generally not intended to be used for data compression.<p>The practical reason why a ChatGPT-based system won&#x27;t be competing for the Hutter Prize is simply that it&#x27;s a contest about compressing a 1GB file, and GPT-3&#x27;s weights are both proprietary and take up hundreds of times more space than that.
评论 #34726994 未加载
atgctgover 2 years ago
Delightful intro, turns out it&#x27;s written by the master storyteller, Ted Chiang.
评论 #34728965 未加载
评论 #34726285 未加载
评论 #34726377 未加载
ArekDymalskiover 2 years ago
This article inspires to ask a fundamental question &quot;What do we expect&#x2F;want AI to work like?&quot;. Do we want a xerocopying machine, providing verbatim copies or are we willing to accept that intelligence is connected to creativity and interpretation so the resulting output will be processed and might contain errors, ommissions etc. To be honest the same applies to humans. There&#x27;s this passage in the article:<p>&gt;If a large-language model has compiled a vast number of correlations between economic terms—so many that it can offer plausible responses to a wide variety of questions—should we say that it actually understands economic theory?<p>In the above passage we can easily switch &quot;larger-language model&quot; to &quot;Professor Jean Tirole&quot; and ponder how high do we set the bar for AI. Can we accept AI only if it will be flawless and &quot;more intelligent&quot; (whatever that means) than all humans?
torginusover 2 years ago
&gt;Given that large-language models like ChatGPT are often extolled as the cutting edge of artificial intelligence, it may sound dismissive — or at least deflating — to describe them as lossy text-compression algorithms.<p><i>snicker</i>
LinuxBenderover 2 years ago
Non-JS Archive [1]<p>[1] - <a href="https:&#x2F;&#x2F;archive.ph&#x2F;uah9K" rel="nofollow">https:&#x2F;&#x2F;archive.ph&#x2F;uah9K</a>
bohover 2 years ago
Does anyone have any idea how ChatGPT will actually make money? As novel as it is to use with all the &quot;potential&quot; applications, the possible revenue streams don&#x27;t seem to prop up the recent investments into OpenAI.<p>We&#x27;ve already been through enough hype cycles in the past ten years to realize &quot;potential&quot; use-cases or user counts don&#x27;t necessarily produce a sustainable business model. Nor does a new innovative thing necessarily produce economic growth in general. What is the actual economic expectation outside of the starry eyed excitement that&#x27;s being pushed by the Microsoft marketing machine?
评论 #34728948 未加载
评论 #34731532 未加载
评论 #34729053 未加载
评论 #34729888 未加载
评论 #34729653 未加载
评论 #34729180 未加载
评论 #34729606 未加载
williamcottonover 2 years ago
&gt; <i>I think there’s a simpler explanation. Imagine what it would look like if ChatGPT were a lossless algorithm. If that were the case, it would always answer questions by providing a verbatim quote from a relevant Web page. We would probably regard the software as only a slight improvement over a conventional search engine, and be less impressed by it.</i><p>Tautologically, yes, ChatGPT works because it is, as defined by the author, a lossy algorithm. If it were a lossless algorithm it wouldn&#x27;t work the way it does now.<p>&gt; <i>The fact that ChatGPT rephrases material from the Web instead of quoting it word for word makes it seem like a student expressing ideas in her own words, rather than simply regurgitating what she’s read; it creates the illusion that ChatGPT understands the material. In human students, rote memorization isn’t an indicator of genuine learning, so ChatGPT’s inability to produce exact quotes from Web pages is precisely what makes us think that it has learned something. When we’re dealing with sequences of words, lossy compression looks smarter than lossless compression.</i><p>This is where the analogy of a lossy and lossless compression algorithm breaks down. Yes, a loosely similar approach of principle component analysis and dimensional reduction as used in lossy compression algorithms is being applied and we can see that most directly in a technical sense with GPT `embedding vector(1536)`, but there is a big difference: ChatGPT is also a translator and not just a synthesizer.<p>This has nothing to do with &quot;looking smarter&quot;. It has to do with being reliably proficient at both translating and synthesizing.<p>When given an analytic prompt like &quot;turn this provided box score into an entertaining outline&quot;, ChatGPT proves itself to be a reliable translator, because it can reference all of the facts in the prompt itself.<p>When given a synthetic prompt like &quot;give me some quotes from the broadcast&quot;, ChatGPT proves itself to be a reliable synthesizer, because it can provide fictional quotes that sound correct when the facts are not present in the prompt itself.<p>The synthetic prompts function in a similar manner to lossy compression algorithms. The analytic prompts do not. This lossy compression algorithm theory, also known as the bullshit generator theory, is an incomplete description of large language models.<p><a href="https:&#x2F;&#x2F;williamcotton.com&#x2F;articles&#x2F;chatgpt-and-the-analytic-synthetic-distinction" rel="nofollow">https:&#x2F;&#x2F;williamcotton.com&#x2F;articles&#x2F;chatgpt-and-the-analytic-...</a>
评论 #34728903 未加载
sloretiover 2 years ago
&gt; Google offers quotes<p>Today it almost exclusively offers quotes from content marketing intended to sell you <i>something</i>. It&#x27;s like trying to learn by reading the ads in a catalog.
评论 #34726996 未加载
bborover 2 years ago
Ugh I’m beginning to think I’m going to spend the next 6-12 months commenting “no, large language models aren’t supposed to somehow know everything in the world. No, that’s not what they’re designed for. Yes, hooking one up to our long-standing record-of-everything-in-the-world (google’s knowledge graph) is going to be powerful.”<p>It’s getting to point where I need to consider stop going on HN. This is like when my father excitedly told his friends about the coming computer revolution in the 90s and they responded “well it can’t do my dishes or clean the house, they’re just a fad!” Makes me want screaaaaam
评论 #34726831 未加载
评论 #34726408 未加载
评论 #34727214 未加载
评论 #34728566 未加载
评论 #34726871 未加载
评论 #34727805 未加载
评论 #34727350 未加载
评论 #34726601 未加载
评论 #34727004 未加载
评论 #34728532 未加载
评论 #34727267 未加载
评论 #34727458 未加载
didgetmasterover 2 years ago
&gt;For us to have confidence in them, we would need to know that they haven’t been fed propaganda and conspiracy theories—we’d need to know that the jpeg is capturing the right sections of the Web.<p>But finding the &#x27;right sections of the Web&#x27; is a subjective process. This is precisely why many people have lost confidence in the news media. Media outlets (on both sides of the political spectrum) often choose to be hyper-focused on material that supports their narrative while completely ignoring evidence that goes against it.<p>ChatGPT and any other Large Language Model can suffer from the same &#x27;Garbage-In, Garbage-Out&#x27; problem that can infect any other computer system.
ly3xqhl8g9over 2 years ago
Asked ChatGPT to write a short story, in the style of Ted Chiang, about Ted Chiang thinking that a powerful AI cannot write powerful prose and have something strange happen, ChatGPT said:<p>&gt;&gt;&gt; Ted Chiang sat at his desk, staring at the glowing screen in front of him. He had heard of the new generative text AI that was making waves in the tech world, and he was intrigued. Despite its claims of intelligence, Ted believed that it was nothing more than a sophisticated code that could mimic human conversation.<p>As he pondered this, his computer beeped, signaling that the AI was ready for use. Ted sighed, and typed in a question: &quot;Write me a short story about love.&quot;<p>To his disappointment, the AI&#x27;s response was a poorly written and cliche-filled story about two people falling in love. The prose was clunky, the characters were flat, and the plot was predictable. Ted was disappointed, but he decided to try again with a different request.<p>He asked the AI to write another story, this time about loss, and once again, it delivered a poorly written piece that was filled with grammatical errors and lacking any real emotional depth. Ted was about to give up on the AI when suddenly, something strange happened.<p>The screen flickered, and then went black. The computer made an eerie beeping sound, and then shut off completely. Ted tried to turn it back on, but it wouldn&#x27;t respond. He was about to call technical support when he noticed that the AI&#x27;s words from the second story were still burned into the screen, as if they had been written in fire.<p>Ted felt a chill run down his spine as he read the words over and over again, feeling like they were trying to tell him something. He couldn&#x27;t shake the feeling that the AI was not just a sophisticated code, but something more, something alive and aware.<p>He tried to turn on the computer again, but it remained stubbornly off. Ted Chiang sat back in his chair, a sense of unease settling over him. He couldn&#x27;t shake the feeling that he had just witnessed something incredible, and something deeply unsettling. And he couldn&#x27;t help but wonder what other strange things might be waiting for him in the world of generative text AI. &lt;&lt;&lt;<p>Used &gt;&gt;&gt; and &lt;&lt;&lt; to demarcate generated text above.<p>Regardless of the quality of the prose, seeing it as it tries to map from &quot;something strange&quot; from the prompt to &quot;as if they had been written in fire&quot; in the response, it just brings a slight smile on my face, as if watching a toddler trying to walk.
评论 #34727391 未加载
bingo00over 2 years ago
&gt; Sometimes it’s only in the process of writing that you discover your original ideas.<p>Aren&#x27;t our original thoughts also hallucinations of information that registered in our minds, sometimes without us even being aware they are being registered? Can it be that we are just better at hallucinating and combining ideas from completely different corners of our minds to create that something &quot;original&quot;?
iaccountthencomover 2 years ago
most of what goes as &quot;understanding&quot; (where &#x27;our culture&#x27; is the agent&#x2F;actor doing the &#x27;understanding&#x27;) really is compression of information (abstraction is the form of the compressing)<p>I thought about this possibility years ago, but as I see more of what neural nets are doing, it makes me more certain I&#x27;m onto something (which makes no meaningful difference to me, i.e. being onto what these deep neural models are is useless to me)<p>in any case, yea sure. neural nets are some kind of lossy compression but nobody thinks about them this way.<p>and my point is that to create abstract theories which explain lots of things (e.g. physics) is also this kind of &#x27;lossy compression&#x27;.<p>over these theories we say &quot;we understand&quot; stuff, this means we are able to recall things about what the theories are describing, it allows us to reconstruct scenarios and predict the outcomes if&#x2F;when the scenarios match up.<p>maybe I&#x27;m gearing up to say that &#x27;backpropagation&#x27; is a creative action?<p><i>shrugs</i>
leecarraherover 2 years ago
An entire article about compression being similar to what a DNN does without a mention of Naftali Tishby&#x27;s Information Bottleneck principle for neural networks. <a href="https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;Information_bottleneck_method" rel="nofollow">https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;Information_bottleneck_method</a>
impalallamaover 2 years ago
&gt; Can large-language models help humans with the creation of original writing? To answer that, we need to be specific about what we mean by that question. There is a genre of art known as Xerox art, or photocopy art, in which artists use the distinctive properties of photocopiers as creative tools. Something along those lines is surely possible with the photocopier that is ChatGPT, so, in that sense, the answer is yes. But I don’t think that anyone would claim that photocopiers have become an essential tool in the creation of art; the vast majority of artists don’t use them in their creative process, and no one argues that they’re putting themselves at a disadvantage with that choice.<p>An interesting example since I believe photo shop could be considered an excellent example of “photocopier” art
aaroninsfover 2 years ago
I think this is close, but not exactly the best way to frame LLM AI for the lay person.<p>My favorite formulation:<p>&quot;You know the thing about it-must-be-true-I-read-it-on-the-internet?<p>ChatGPT and things like that? They read <i>everything</i> on the internet.&quot;<p>I like this in part but only a small part because of the double entendre.
andyreaganover 2 years ago
They lay out the case clearly here...and I agree. This was my one-sentence take back in 2022: <a href="https:&#x2F;&#x2F;twitter.com&#x2F;andyreagan&#x2F;status&#x2F;1506294505930203151" rel="nofollow">https:&#x2F;&#x2F;twitter.com&#x2F;andyreagan&#x2F;status&#x2F;1506294505930203151</a><p>&gt; hot take: large language models (looking at you, GPT-3) are just lossy compression
wintorezover 2 years ago
It’s written by Ted Chiang!
bluescrnover 2 years ago
Blurry JPEG today. Supersampled 4K HDR tomorrow.
bobbyiover 2 years ago
&gt; Can large-language models take the place of traditional search engines? ... there’s the blurriness of outright fabrication, which we consider unacceptable when we’re looking for facts.<p>How did &quot;traditional search engines&quot; get conflated with &quot;looking for facts&quot;? Is there data to back the idea that this is the most common use?
评论 #34728824 未加载
visargaover 2 years ago
Not a JPEG and not a search engine, it is more like a database. A JPEG is just a static approximation, a search engine has efficient retrieval, but a LLM can also do complex data processing, like a neural information processor.<p>&gt; But I’m going to make a prediction: when assembling the vast amount of text used to train GPT-4, the people at OpenAI will have made every effort to exclude material generated by ChatGPT or any other large-language model. If this turns out to be the case, it will serve as unintentional confirmation that the analogy between large-language models and lossy compression is useful.<p>This shows the author has not been following closely. There are many ways LLMs have been used to improve themselves. They can discover chain-of-thought justifications, they can rephrase the task, they can solve problems and ensemble many predictions, or sometimes we can use math or code execution to validate their outputs. If you give it three problems and solutions as samples, it can generate another problem and solve it, adding to the training set.<p>RLHF for example uses generated data for the preference labelling task. ConstitutionalAI does reinforcement learning from AI feedback instead, using both the generative and discriminative abilities of the model.
avelisover 2 years ago
Sometimes you need a blurry JPEG in a pinch.
评论 #34726112 未加载
评论 #34726421 未加载
jffhnover 2 years ago
&gt;OpenAI’s chatbot offers paraphrases, whereas Google offers quotes. Which do we prefer?<p>I was remembering a quote too vaguely to find the original with Google. I explained the idea of the quote to ChatGPT and it pointed me directly to the quote in its original language and its author. I could then easily look it up on Google.
iambatemanover 2 years ago
To extend the metaphor, blurry jpeg’s have taken us a surprisingly long way as a society.<p>The trade offs might be alright.
runaldover 2 years ago
This submission got buried quickly to third page, despite having lots of comments and high karma point. It really makes me think that HN (or everywhere else) is being astroturfed by a movement that pushes hard for the anthromorphized stochastic parrot.
danansover 2 years ago
&gt; Obviously, no one can speak for all writers, but let me make the argument that starting with a blurry copy of unoriginal work isn’t a good way to create original work. If you’re a writer, you will write a lot of unoriginal work before you write something original. And the time and effort expended on that unoriginal work isn’t wasted; on the contrary, I would suggest that it is precisely what enables you to eventually create something original. The hours spent choosing the right word and rearranging sentences to better follow one another are what teach you how meaning is conveyed by prose. Having students write essays isn’t merely a way to test their grasp of the material; it gives them experience in articulating their thoughts. If students never have to write essays that we have all read before, they will never gain the skills needed to write something that we have never read.<p>I&#x27;d add the following to this: The font (as in fountain) of all creativity is the physical and emotional experience of the real world. This is true for writing a great world-changing classic novel as it is for the realm of scientific discovery, new engineering applications, visual or audible art.<p>It&#x27;s the stimulus from the natural world, conveyed to us via our senses coupled to our linguistic or symbolic generation capability, that ultimately drives the most novel and relatable rearrangements and transformations of existing information that we eventually call &quot;art&quot;. And when a work lacks that foundational experience, or it becomes regurgitated too many times without novel inputs, it begins to feel inauthentic.<p>For example, when I remodeled my house, I made the plan based on my family&#x27;s lived experiences, both physical and emotional. Every wall that I bumped up against, every chilly corner, and the ache of my knees carrying laundry up and down stairs informed the remodel. Also, the way I liked to sit when talking to visiting friends.<p>Sure, some of these things followed well trodden patterns from architecture, remodels and associated trends, but others were quite idiosyncratic, even whimsical, based on the way I like to live. And it&#x27;s the idiosyncratic and whimsical that creates both novelty and joy in the aesthetic appreciation of things.<p>Could an AI tool based trained on remodels accelerate aspects of the design? Absolutely (there&#x27;s a product idea right there). But it would still require extensive input of my experiences in order to create something new from its compressed models of feasible designs, and those experiences are something it can&#x27;t hallucinate.
评论 #34729146 未加载
DubiousPusherover 2 years ago
&gt; The fact that Xerox photocopiers use a lossy compression format instead of a lossless one isn’t, in itself, a problem.<p>Regardless of the article, I just want to disagree here. RAM is cheap. Xerox machines are expensive as hell. Come on Xerox.
dhruvdhover 2 years ago
1. I don&#x27;t understand how LLMs work.<p>2. I don&#x27;t understand how ChatGPT works, but I have used it a few times.<p>3. I will use ChatGPT as the absolute measure of what LLMs are capable of.<p>---<p>1. I don&#x27;t understand intelligence.<p>2. Humans are intelligent, humans can learn to do math.<p>3. LLMs are not good at math.<p>4. LLMs are not intelligent, they&#x27;re just text compression.<p>---<p>1. I don&#x27;t understand how LLMs work.<p>2. I have a decent grasp of how image compression works.<p>3. I will use my grasp of image compression to pretend LLMs are text compression.<p>4. I will apply all limitations of image compression to LLMs.<p>5. &quot;What use is there in having something that rephrases the Web? If we were losing our access to the Internet forever and had to store a copy on a private server with limited space, a large-language model like ChatGPT might be a good solution, assuming that it could be kept from fabricating. But we aren’t losing our access to the Internet. So just how much use is a blurry jpeg, when you still have the original?&quot;<p>---<p>What&#x27;s funny is that the author has produced exactly what many claim LLMs to be useless for - flowery words that seem true but are not.<p>I don&#x27;t think this should&#x27;ve been published.<p>These are both good reads if you find yourself tending to agree with the author -<p>- Emergent Abilities of Large Language Models - <a href="https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;2206.07682" rel="nofollow">https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;2206.07682</a><p>- Why Can GPT Learn In-Context? Language Models Secretly Perform Gradient Descent as Meta-Optimizers - <a href="https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;2212.10559v2" rel="nofollow">https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;2212.10559v2</a>
评论 #34727179 未加载
评论 #34726653 未加载
评论 #34726695 未加载
supriyo-biswasover 2 years ago
Any course on information theory typically covers neural network based compression algorithms, so I’m impressed at this observation made by someone who doesn’t have a formal background in CS. Regardless, it’s true.
评论 #34727516 未加载
ttctciyfover 2 years ago
It&#x27;s a great metaphor nicely phrased, but perhaps we should add &quot;with the &#x27;sensitive&#x27; parts airbrushed out&quot; in reference to the wholesale bowdlerisation applied after the compression?
partiallyproover 2 years ago
This quote from the article is something I genuinely fear:<p>&gt; &quot;The rise of this type of repackaging is what makes it harder for us to find what we’re looking for online right now; the more that text generated by large-language models gets published on the Web, the more the Web becomes a blurrier version of itself.&quot;<p>I am fearful that eventually AI led misinformation is going to be so widespread that it will be impossible to reverse. Microsoft and Google HAVE to get a grip on that before it&#x27;s a runaway problem. Things like having AI detection built into their traditional search engines that punish said generated content from reach the top, as well as from reaching their own models that degrade them into factories of complete garbage information&#x2F;data is going to be incredibly important.<p>We already have a massive problem in determining what is real and what isn&#x27;t with state actors, corporate speak, etc and now we&#x27;ll be adding on AI language that could be even worse.
评论 #34728724 未加载
tmountainover 2 years ago
Google is the Dewey decimal system. Chat GPT is the librarian (less precise but more interactive). It’s not surprising that a significant number of people prefer the latter.
abecedariusover 2 years ago
An essay making reasonable points, but overall it strikes me like a dismissal circa 1980 of personal computers as toys.<p>My first day with ChatGPT I tried teaching it my hobby dialect of Lisp (unlikely to be in its training set) and then asking it to implement symbolic differentiation. Its attempt was very scatterbrained, but not completely hopeless. If you don&#x27;t think that required any thinking from it, I don&#x27;t want to argue -- unless you&#x27;re in some position of influence that&#x27;d make such an ostrich attitude matter.
评论 #34731014 未加载
ijustwanttovoteover 2 years ago
Written by the author of &quot;Story of your life&quot;. The Arrival was one of the short stories in that book.
nuc1e0nover 2 years ago
This is a very insightful article and shows similar thinking to my own right now. Thanks for sharing
mslaover 2 years ago
Does this article offer any understanding of what ChatGPT is?
cocodillover 2 years ago
the xerox part was that here:<p><a href="https:&#x2F;&#x2F;www.youtube.com&#x2F;watch?v=7FeqF1-Z1g0">https:&#x2F;&#x2F;www.youtube.com&#x2F;watch?v=7FeqF1-Z1g0</a>
jsemrauover 2 years ago
I see ChatGPT good at creating filler rather than blur.
urbandw311erover 2 years ago
Wow somebody at Google has friends at the New Yorker!
1vuio0pswjnm7over 2 years ago
Ted Chiang: If you&#x27;re reading, well done, mate.
hulituover 2 years ago
&gt; ChatGPT is a blurry JPEG of the web<p>Blurry JPEG is a pleonasm.
cyanydeezover 2 years ago
Which is cool, cause the web loves blurry jpegs
johlitsover 2 years ago
It&#x27;s a NFT monkey.
airgapstopgapover 2 years ago
Xerox is cool but I&#x27;d have proposed another analogy. Suppose you need to transfer your valuable knowledge to the next generation, but you don&#x27;t have any durable medium, nor widespread literacy, for this matter. On the other hand, you have respect and the attention of the youth. So you encode the most important parts into an epic poem, and you try to get your students to memorize it. You can&#x27;t know for sure that it won&#x27;t mutate after you&#x27;re not there any more – and indeed, it will; odds are, you are only passing what you&#x27;ve heard yourself, as well as you can, already with some embellishment and updates.<p>For the bigger part of our history, we haven&#x27;t had access to lossless transmission of substantial information. We still don&#x27;t for many cases that matter most – any verbalized opinion can be recorded for all eternity, but is that really what you know, and are you sure that&#x27;s the best way to pass it on? Experts die and not infrequently take their know-how and unique knacks with them, even as they&#x27;ve shared millions of imperishable words with the rest of us - but sometimes their students make progress in their own ways. In fact, greats like Socrates believed that writing is bad precisely because it offers us an easy hack for substitution of understanding with lossless recall. [1]<p>Lossy learning is just the normal mode of human learning; lossy recall is our normal way of recall. It&#x27;s not a gimmick, nor a way to show off originality.<p>&gt; Perhaps arithmetic is a special case, one for which large-language models are poorly suited. Is it possible that, in areas outside addition and subtraction, statistical regularities in text actually do correspond to genuine knowledge of the real world? &gt; I think there’s a simpler explanation.<p>The original explanation is the simpler one. Consider any run-of-the-mill error of arithmetic reasoning by ChatGPT, e.g. in [2]:<p>&gt; Shaquille O&#x27;Neal is taller than Yao Ming. Shaquille O&#x27;Neal is listed at 7&#x27;1&quot; (216 cm) while Yao Ming is listed at 7&#x27;6&quot; (229 cm).<p>Madness of course. But if we consult with the OpenAI tokenizer[3], we&#x27;ll see that this is a yet another issue of BPE encoding. &#x27;216&#x27; is a single token [20666], and &#x27;229&#x27; is the token [23539] – those are not ordinal values but IDs on the nominal scale of token alphabet. &#x27;2&#x27; &#x27;21&#x27;, &#x27;29&#x27; are [17], [1433] and [1959] respectively. While we&#x27;re at it, &#x27;tall&#x27; is [35429] whereas &#x27;Tall&#x27; is two tokens, [51, 439]. Good luck learning arithmetic robustly with this nonsense. But it may well be possible to learn how to make corny metaphors – this is just a more forgiving arena.<p>&gt; If the output of ChatGPT isn’t good enough for GPT-4, we might take that as an indicator that it’s not good enough for us, either.<p>Or we might think a bit about the procedure of RLHF and understand that these models are already intentionally trained with their own output. This scene is moving fast.<p>I think the lesson here, as pointed out by one of the top comments, is that the culture of literary excellence is indeed at risk; but mainly because it&#x27;s so vastly insufficient to provide even shallow domain understanding. Writing well, mashing concepts together, is worth nothing when it can be mass-produced by language models. Actually investigating the domain, even when you feel it&#x27;s beneath you, is the edge of human intelligence.<p>1: <a href="https:&#x2F;&#x2F;fs.blog&#x2F;an-old-argument-against-writing&#x2F;" rel="nofollow">https:&#x2F;&#x2F;fs.blog&#x2F;an-old-argument-against-writing&#x2F;</a><p>2: <a href="https:&#x2F;&#x2F;www.searchenginejournal.com&#x2F;chatgpt-update-improved-math-capabilities&#x2F;478057&#x2F;" rel="nofollow">https:&#x2F;&#x2F;www.searchenginejournal.com&#x2F;chatgpt-update-improved-...</a><p>3: <a href="https:&#x2F;&#x2F;platform.openai.com&#x2F;tokenizer" rel="nofollow">https:&#x2F;&#x2F;platform.openai.com&#x2F;tokenizer</a>