In particular, many have griped over their original work being used to train these AI models — a use they never opted into, and for which they’re not compensated.
But what if artists could “poison” their work with a tool that alters it so subtly that the human eye can’t tell, while wreaking havoc on AI systems that try to digest it?
That’s the idea behind a new tool called “Nightshade,” which its creators say does exactly that. As laid out in a yet-to-be-peer-reviewed paper spotted by MIT Technology Review, a team of researchers led by University of Chicago professor Ben Zhao built the system to generate prompt-specific “poison samples” that scramble the digital brains of image generators like Stable Diffusion, screwing up their outputs.
Comments are closed.