Apparently, the rabbit hole goes deeper. The wayback machine puts the "AI Generated" description on the cheese website at August 7th, 2020. The AI didn't hallucinate anything because it didn't generate anything, the entire premise is simply fake.<p><a href="https://web.archive.org/web/20200807133049/https://www.wisconsincheesemart.com/products/gouda-cheese-smoked" rel="nofollow">https://web.archive.org/web/20200807133049/https://www.wisco...</a><p>The (edited) cheese ad: <a href="https://www.youtube.com/watch?v=I18TD4GON8g" rel="nofollow">https://www.youtube.com/watch?v=I18TD4GON8g</a><p>What probably should be the target link: <a href="https://www.theverge.com/news/608188/google-fake-gemini-ai-output-super-bowl" rel="nofollow">https://www.theverge.com/news/608188/google-fake-gemini-ai-o...</a>
“the Google executive Jerry Dischler said this was not a “hallucination” – where AI systems invent untrue information – but rather a reflection of the fact the untrue information is contained in the websites that Gemini scrapes.”<p>So what you are saying is that Gemini is basically useless?<p>Good to know. Thanks for clarifying that Google!
Google saying “multiple websites have the fact so Gemini ran with it” is extraordinarily rich.<p>If ever there was a company which should know that “multiple websites” is not a good benchmark for accuracy, it’s Google.<p>This feels like a good parable for Google’s search these days. I’ve seen more wrong information from them than anyone lately. I wonder if they can course correct before it’s too late.
The source sentence on cheese.com was: "It is the most popular Dutch cheese in the world, accounting for 50 to 60% of the world's cheese consumption."<p>The glorified autocomplete failed to detect the nuance in this (admittedly poorly written) sentence that the 50-60% statistic was of the world's <i>Dutch</i> cheese consumption
I know very little about cheese and such a stat seems just incredibly obviously untrue. I wonder how it made it past the dozens or more people who worked on the ad.
> The local commercial, which advertises how people can use “AI for every business”, showcases Gemini’s abilities by depicting the tool helping a cheesemonger in Wisconsin to write a product description<p>By copy/pasting another Cheese monger's incorrect product description...
Top relevant search result is a Reddit thread from 11 years ago quoting a cheese.com article: <a href="https://www.reddit.com/r/todayilearned/comments/2h2euc/til_gouda_accounts_for_over_half_of_the_worlds/" rel="nofollow">https://www.reddit.com/r/todayilearned/comments/2h2euc/til_g...</a><p>Internet Archive confirms that page has had the blatantly wrong stat up since at least April 2013: <a href="https://web.archive.org/web/20130423054113/https://www.cheese.com/smoked-gouda/" rel="nofollow">https://web.archive.org/web/20130423054113/https://www.chees...</a><p>> If truth be told, it is one of the most popular cheeses in the world, accounting for 50 to 60 percent of the world's cheese consumption.<p>Clearly predates generative AI, so I think this is junk human-written SEO misinformation instead.<p>Here's that page today: <a href="https://www.cheese.com/smoked-gouda/" rel="nofollow">https://www.cheese.com/smoked-gouda/</a> - still has that junk number but is now a whole lot longer and smells a bit generative-AI to me in the rest of the content.
I’m very far from an evangelist of the tech but I was under the impression they had gotten better at this sort of thing.<p>I question why a model which “knows” about cheddar / mozzarella cheese would make this blunder.<p>Was this supposed to be generated by one of those “show your work” reasoning models or is this just the regurgitation of one of the single short response parrot of old quora answer or reddit post ai chatbots?
My company contracted one of these LLMs to give us a bespoke chatbot. Works great for translation, and that's what I've been using it for. I popped in the line "Table 1 (etc, etc)"<p>It perfectly translated the line, but doesn't it also give me a completely made up 2-column 10-row data table! I asked it why, and the response was along the lines of "I am designed to make your life easier, and I thought providing this table would reduce your workload"
Sounds like Google's AI produced some stats that were quite obviously mistaken, but nobody cared to fact-check in the slightest before shipping.<p>> the Google executive Jerry Dischler said this was not a “hallucination” [...] but rather a reflection of the fact the untrue information is contained in the websites that Gemini scrapes.<p>What's this guy is describing is pretty much the root cause of what we colloquially refer to as "hallucination".
Google: "Gemini didn't make a mistake. It simply plagiarized a factually-incorrect article. Word for word."<p>Gee, thank you Google for convincing me you have a product that I find useful and that I can trust. I look forward to you trying to cram this down my throat, against my will, at every opportunity you see fit. :-/
> Jerry Dischler said this was not a “hallucination” – where AI systems invent untrue information – but rather a reflection of the fact the untrue information is contained in the websites that Gemini scrapes.<p>So LLMs like any other computer system suffer from "Garbage In, Garbage Out".
This is at least the second embarrassment with a high-profile Google AI ad/demo.<p>(I'm also thinking of when, in the LLM boat-missed frenzy, they faked an interactive AI demo, to make it look much more responsive than their actual tech was.)<p>I'm unclear on how either incident was allowed to happen.
none of the marketing geniuses thought: hey, that's odd, 50% of my cheese consumption is definitely not Gouda which probably accounts for every single one of them, Dutch people included.
It's bizarre to me how the media (and mainly tech media) has been bending over backwards to make this story a big deal.<p>A cheese fact on a cheese website was wrong therefore Gemini is bad? What?
This is just another point showcasing how useless LLMs are. They either hallucinate OR are polluted by "wrong" information on websites. If I need to double or triple check everything an LLM does, I'd rather just do it myself immediately.