The linked article describes an attack against NotebookLM, which is limited to people who deliberately create a Notebook that includes the URL of the page with the attack on it.<p>I had a go at something a bit more ambitious a few weeks ago.<p>If you ask Google Gemini "what was the name of the young whale that hung out in pillar point harbor?" it will tell you that the whale was called "Teresa T".<p>Here's why: <a href="https://simonwillison.net/2024/Sep/8/teresa-t-whale-pillar-point/" rel="nofollow">https://simonwillison.net/2024/Sep/8/teresa-t-whale-pillar-p...</a><p>(Gemini used to just say "Teresa T", but when I tried just now it spoiled the effect a bit by crediting me as the person who suggested the name.)
I write fiction sometimes and I've got this story I've been working on which has languished by the wayside for at least a year. Whacked it into the podcast machine. Boom. Hearing these two people just get REALLY INTO this unfinished story, engaging with the themes, with the characters, it's great, it makes me want to keep writing.
Isn’t this just like SEO where you can also try and trick the crawlers? Only difference is that it feels more serious with AI, it’s more realtime, and the AI engines aren’t always smart enough with anti-duping capabilities?
I am very confused. Is this talking about NotebookLM (<a href="https://notebooklm.google.com/" rel="nofollow">https://notebooklm.google.com/</a>) or NotebookLLM (<a href="https://notebookllm.net/" rel="nofollow">https://notebookllm.net/</a>) or both? Something else? The article appears to consistently use LLM but link to LM, but the LLM site I linked has a podcast generator?<p>One of these projects has to change their name!
FWIW, had a pleasantly surprising experience with this podcast thing. I tried it out on a few little blogposts I wrote and I was like, hmm cool. Showed my 8 year old son how it was referencing things I wrote.<p>And he was ON IT. Like, he ran to his room and grabbed a pencil and paper and put down an essay (okay about 6 or so sentences) about Minecraft, had me type them in, and ran the Notebook, and now he's just showing off both to EVERYONE.<p>(Yes, he understands it's not real people.)
AI is kind of bad at searching the web right now anyway. I've found myself having to waste tokens forcing models to not do so just to achieve the results I actually want.
I have no problem with this. Once we switch over to an LLM-based education system, there won't be a problem with this Benson on the moon story, because everyone will just learn it's true.<p>Every technological revolution has tradeoffs. Luckily once the people who knew what we lost <i>finally</i> die off, the complaints will stop and everyone will think the new normal is fine <i>and better</i>.
The big asterisk here is, what did they prompt the AI with to generate the podcast? Was it "Generate a podcast based on the website 'Foo'", or was it "Generate a podcast telling the true story of the Space Race?"
I fed my resume into this thing and I can't stop laughing.<p><a href="https://masto.xyz/tmp/podcast.mp3" rel="nofollow">https://masto.xyz/tmp/podcast.mp3</a>
somewhat of a side-note: It's interesting to me that the first couple of sentences of the AI podcast sound 'wrong', even though the rest sounds like a real podcast. Is this something to do with having no good initial conditions from which to predict "what comes next"?
Wow, content aside, this is probably the first time I heard a podcast coming from NotebookLLM and it's kinda nerve wracking and mind blowing at the same time. Those fake laughs in the snippet makes me feel...so uncomfortable for some reason knowing that its "fake". But sounds very real, too real.
I suggest <a href="https://www.apa.org/monitor/2009/12/consumer" rel="nofollow">https://www.apa.org/monitor/2009/12/consumer</a> or the Eduard Bernays story to convince MDs that smoking is good for health: he create a new scientific journal distributing it for free "because it's new, we want to spread", hosting REAL publications from anyone who want to publish and being spread for free... After a bit of time he inject some false articles self-written formally as some PhD of remote universities finding that smoking tobacco is good for health, others real professors follow the path stating they discover this or that specific beneficial use of cigarettes, then the false became officially true, tested and proved science in most people mind.<p>With LLMs is far cheaper and easier, but the principle is the very same: trust vs verification or the possibility thereof.
I gave it all my blog posts (<a href="https://ebruce613.prose.sh/" rel="nofollow">https://ebruce613.prose.sh/</a>) and the result... hilarious.
<a href="https://0x0.st/XE4h.mp3" rel="nofollow">https://0x0.st/XE4h.mp3</a>
I’m not sure what the attack would be, tbh. Is there a situation where I would want to feed a lie to an LLM that I wouldn’t want regular chrome users to see?
I tried feeding NotebookLM a Wikipedia article about the murder of Junko Furuta, a horrifying story of a poor girl tortured and murdered in Japan in 1989. NotebookLM refused to do anything with this document - not answer questions, not generate a podcast, nothing. Then I tried feeding it the wiki on Francesco Bagnaia, a wholesome MotoGP rider, and it worked fine.<p>Who wants this shit? I do not want puritanical American corporations telling me what I can and can't use their automated tools for. There's nothing harmful in me performing a computer analysis about Junko Furuta, no more so than Pecco Bagnaia. How have we let them treat us like toddlers? It's infantilising and I won't take part in it. Google, OpenAI, Microsoft, Apple, Meta and the rest of them can shove these crappy "AI" tools.
It explains my book (Opinionated Launch) better than myself :D <a href="https://notebooklm.google.com/notebook/98539685-0890-438b-a014-7887db435fdb/audio" rel="nofollow">https://notebooklm.google.com/notebook/98539685-0890-438b-a0...</a>
LLMs are the type of junk AI that these corps think will succeed? They are spending billions and consuming a large amount of resources and energy for this. Seriously, what a waste.
> Google's AI thinks I left a Gatorade bottle on the moon<p>No, NotebookLM creates summaries and podcasts, or answers questions specifically from the documents you feed it.<p>Feed it fiction it will create fiction as would a human tasked to do the same.
It's such a cool concept, but yeah, when I've listened to it and Illuminate, it's also a bit scant on details too. Neat technology, even engaging, but not good for more than best-effort high level summaries.
As a personal preference, I dislike podcast artificial banter, and this audio is a great example of what I dislike.<p>Artificial artificial.<p>Great little project, though. And, as satire, I did like the show notes writing.<p>And the generative AI was impressive, in a way. Though I haven't yet thought of a positive application for it. And I don't know the provenance of the training data.
This is no different than the decades-old technique of "cloaking", to fool crawlers from Google and other search engines.<p>I fail to see the value in doing this.<p>"Oh hey everybody! I set up a website which presents different content to a crawler than to a human ..... and the crawler indexed it!!"
> You can upload a documents with fake show notes straight to NotebookLLM's website, so if you're making silly podcast episodes for your kids, that's the best way to do it.<p>Please don’t do this. You don’t need a professional mic to record a podcast with your kids any phone or computer mic will work. Then you can have fun editing it with open source audio tools.<p>Don’t have a computer generate crap for your kids to consume. Make it with them instead.
Google Search is pretty good at detecting this dual-content attacks. It's not this is the first time someone thought about that and it will heavily penalize websites that do that.<p>This is just the NotebookLM crawler that is being tricked, which is still in it's experimental stage. Rest assured as it scales Google can easily implement safeguards against all spammy tricks people use