I actually experienced this the other day. Bought the new Baldur's Gate and was wondering what items to keep or sell (don't judge me, I'm a pack rat in games!)<p>I had found some silver ingots. The top search result for "bg3 silver ingot" is a content farm article that very confidently claims you can use them at a workbench in Act 3 to upgrade your weapons.<p>Except this is a complete fabrication: silver ingots exist only to sell, and there is no workbench. There is no mechanic (short of mods) that allows you to change a weapon's stats.<p>I'm pretty sure an LLM "helped" write the article because it's a lot of trouble to go through just to be straight up wrong - if you're a low effort content farm, why in the world would you go through the trouble if fabricating an entire game mechanic instead of taking the low effort "They exist only to be sold" road?<p>This experience has caused me to start checking the date of search results: if it's 2022 and before, at least it's written by a human. If it's 2023 and on, I dust off my 90's "everything on the World Wide Web is wrong" glasses.
I wonder if there will be a human information/knowledge equivalent of low-background steel (pre-WWII/nukes). Data from before a certain point won't be 'contaminated' with LLM stuff, but it'll be everywhere after that.<p><a href="https://en.wikipedia.org/wiki/Low-background_steel" rel="nofollow noreferrer">https://en.wikipedia.org/wiki/Low-background_steel</a>
I think it'd be kind of neat in a backward way if we went back to the 'specialized encyclopedia' days of the 90s.<p>Web directories , 'Who's Who in Engineering' type lists, etc.<p>It's a step back from universal search engines being able to find stuff, but it's a step forward with regards to curation and quality of results; so i'm not sure if it's entirely a downgrade.<p>The early 90s 'website phonebook' type encyclopedias were interesting[0], but I always had to remind my mom "No, this isn't the entire internet, it's just a bunch of places that people like; the secret ones are 'unlisted'."<p>Note: I never say this is better than a search engine, it's just an interesting end-result after search engines got polluted and modified til the point of uselessness that we're at now with Google.<p>[0]: <a href="https://www.amazon.com/Internet-Directory-Guide-Usenet-Bitnet/dp/0449908984" rel="nofollow noreferrer">https://www.amazon.com/Internet-Directory-Guide-Usenet-Bitne...</a>
Recently an article came out where someone said that the company I work for is a big user of WebAssembly, but the reality is that we don't use it.<p>After finding the contributed article (on a well-known news site, not Wired though), it looks like a tech founder might've been using ChatGPT to write an article about the uses for WASM. The arguments were generally sound, but I don't think that anyone did the work to manually check any of the facts they presented in it.
More amusing and frightening is when people search about themselves
and turn up AI generated crap. Googling yourself was always a lucky
grab bag, with the possibility of long-forgotten embarrassments being
dragged up. But at least you'd have to face facts.<p>Now I hear of people discovering they're in prison, married to random
people they've never met, or are actually already dead.<p>What is this going to do to recon on individuals (for example by
employers, border agents or potential romantic partners) when there's
a good chance the reputation raffle will report you as a serial
rapist, kiddy-fiddler or Tory politician?
The SEO garbage has been poisoning the search for years. Even before the chatbots it got to the point when most top results are crap. The LLM's can surely make it much worse, though.
I started to go down a line of thinking where I think we might see a return to books in the next 3-5 years. The reason is that with a book it’s a big collection of knowledge and people can post reviews about the quality of the book whereas on the web you have no way of knowing what quality of an article will be anymore.
Here's what people don't understand: this is mostly good for google.<p>The worse organic results are, the more people will click on paid links. This is WHY everyone on HN is complaining about search results, because google doesn't really have an incentive to give you really good results. They only need to be good enough to keep 95% of the population still using google, but mostly expecting the good results to be ads.<p>Google ads are the equivalent of verification on FB and X. They just call it something different. The verified, high quality results will be paid.
it's almost like AI just repeats data its fed on, even incorrect data, without any real intelligence to determine if the data is correct.... /s<p>Its not simply garbage in garbage out. There is no logic to verify and analyze the data. You are simply told what is popular in the data.
The correct term is spamming. People are using these text generators to spam everyone and everything under the sun. It will be detrimental to the internet as many people will just give on this huge pile of ... spam.
Without naming the company, I have seen specific examples of blog posts being written by AI, hallucinating a "fact", and then that "fact" re-surfacing inside of Bard.<p>Its xkcd's Citogenesis automated and at internet scale <a href="https://xkcd.com/978/" rel="nofollow noreferrer">https://xkcd.com/978/</a>
I have to say--the opening paragraph doesn't describe a reality I'm familiar with:<p>>Web search is such a routine part of daily life that it’s easy to forget how marvelous it is. Type into a little text box and a complex array of technologies—vast data centers, ravenous web crawlers, and stacks of algorithms that poke and parse a query—spring into action to serve you a simple set of relevant results.<p>Web search has, for me, become a nasty twisted hall of mirrors well before generative AI. I almose never get fed relevant results, I alsmost always have to go back and quote all my search terms because the search engine decided it didn't really need to use all of them (usually just one.) The only difference is the poison was human generated. generative AI will simply erase the 5% of results that might give me an answer quickly.
The signal to noise ratio of web search results has been trending toward utter uselessness for years; so while AI content will make it worse, it won't make it dramatically worse. We'll just advance toward useless at a higher rate.
You should never have had been trusting what you find on the web, let alone other media. I hope widespread usage of generative AIs including deepfakes will finally force the masses to start thinking more critically.
«Could»? Google has already been doing this for quite some time, at least in my region (Norway), and I’d say more than half of the suggestions Google provides as top results are false.
I think the insufficient accuracy in the output of LLMs is going to lead them to be a lot more niche then the current hype is hoping for. I think most people care that someone is taking accountability for what they are reading - not that it is necessarily correct but at least that someone thinks it is correct (and that someone can be taken to task if it is inaccurate).<p>If LLM usage in media becomes widespread, I'd pay for a service that identifies and hides the LLM shit for me.
It's great to hear that someone finds Replit's AI capabilities useful for accelerating their learning... However, I must point out that the post sounds like an advertisement for an overpriced product. Not poking the AI assistance in coding, it definitely can be valuable, but the effectiveness vs. cost-efficiency varies. 20 dollars per month for coding a chatbot, not worth it for me...
Surely, just as content farms have gradually trashed the quality of search results on major platforms. There's also an over-reliance on raw quantities; I often get irrelevant and unwanted news articles from India simply because the huge population of that country coupled with widespread use of English outweighs US content on the social graph.
This is a pretty intractable problem and my app is in the alpha-est of stages, but I built something for this purpose. It maps creators on a 2D grid (using React-flow) based on subject and lets users vote on their trustworthiness. <a href="https://www.graphting.org" rel="nofollow noreferrer">https://www.graphting.org</a>
It'll make it easier to trust if you have your own index of documents. That's why I built this: <a href="https://mitta.us/" rel="nofollow noreferrer">https://mitta.us/</a><p>There still exists a problem that users need to run and manage their own indexes, at times.
Wasnt there a paper a few months back, Textbooks are all you need. yes found it <a href="https://arxiv.org/abs/2306.11644" rel="nofollow noreferrer">https://arxiv.org/abs/2306.11644</a><p>So search engines in their traditional sense will be obsolete anyway.<p>1) GPT-4 and other such LLMs will generate textbooks and manuals for every conceivable topic.<p>2) These textbooks will be 'dehallucinated' and curated by known experts on particular topics, who have reputations to maintain. The experts' names will be advertised by the LLM provider.<p>3) People will search for stuff by chatting with the LLMs, which will in turn provide citations for the chat output from the curated textbooks.
Well, did you have to use Microsoft documentation ? Imagine whole Internet presenting that content, with random flashes of animations and html5 whistles
Just another reason that I consider generative AI to be a lot like crypto. A lot of talk about it being the future but really only turns out to be dangerous or useless. I find it incredibly irresponsible that companies are shoving their latest AI tech into all their products when it's still unproven.
It’s especially terrifying that misinformation compounds multiplicatively with AI because it happens in 2 layers - once at the retrieval layer (where AI-generated content is worsening the problem of bad SEO content) and again at the retrieval augmented generation (RAG) LLM layer.<p>(shameless plug) At Metaphor (<a href="https://platform.metaphor.systems/">https://platform.metaphor.systems/</a>), we’re building a search engine that avoids SEO content by relying on human curation + neural embeddings for our index + retrieval algorithm. Our mission is to ensure that the information we receive is as high quality and truthful as possible as AI adoption marches onwards. You (or your LLM) can feel free to give it a try :)
At first I thought the article was going to be about human-led misinformation but I wonder whether with both hallucinations and human-fed misinformation (AI-helped or not!) whether we can use AI to fact/self check results (both AI generated and human ones) and prompt us about potential misinformation and link to relevant sources? That way AI could actually help solve the trust issue.
Why are people calling them hallucinations and not just errors, flaws or bugs? You can't hallucinate if all of your perception is one internal state. Chatbots don't dream of electric sheep.
Leaving aside the article to discuss the source for a moment. When did Wired become so antitech?<p>There are good critical viewpoints but most of the articles they are putting out at this point read like bitter diatribes. Which is a shame because they used to be an excellent publication.