For many folks writing or sharing art on the open web in 2024, the rise of corporate theft under the guise of "AI" has become a real sticking point. I share these sentiments, though have yet to start taking any real action against them. Still, when I do consider the actions I could take, I find my thoughts echoing those that Jeremy has shared. Beyond merely requesting to be kept out of "training datasets", can we somehow punish or poison them for daring to steal our words, images, and videos without our consent?
I certainly like the idea π
On the rational response to LLMs β fight back:
If my words are going to be snatched away, I want them to be poison pills.
On techniques people are using to make their content dangerous for LLMs to "consume":
Smarter people than me are coming up with ways to protect content through sabotage: hidden pixels in images; hidden words on web pages.
If enough people do this weβll probably end up in an arms race with the bots. Itβll be like reverse SEO. Instead of trying to trick crawlers into liking us, letβs collectively kill βem.