Does copyright become obsolete now that large language models (LLMs) are scraping and ingesting the content we write, and then use it to train themselves? What's the point of being a volunteer content creator for OpenAI or Google if LLMs scrape everything you put online? Why even try?
This was the conundrum I faced when I learned that everything was being scraped to train LLMs. I haven't figured out a reasonable solution to this at all and was forced to remove my websites from being accessible to random web visitors.<p>I'm very eager to find a way that I can make my sites accessible again.
Well there is some flag that you can set in robots.txt of your website. But however I am sure, not many AI vendors respect this flag.<p>Which is why Google Gemini doesn't works to that level of magical greatness, I guess. They go with the ethical route of data collection.