This new tool lets artists ‘poison’ their artwork to deter AI companies from using it to train their models—here’s how it works
Artists who want to share their artwork often face a tough choice: keep it offline or post it on social media and risk having it used to train data-hungry AI image generators.
But a new tool may soon be able to help artists deter AI companies from using their artwork without permission.
It’s called “Nightshade” and was developed by a team of researchers at the University of Chicago. It works by “poisoning” an artist’s creation by subtly changing the pixels of the image so that AI models aren’t able to accurately determine what the image is depicting, according to MIT Technology Review.
While the human eye isn’t able to detect these small changes, they aim to cause a machine-learning model to mislabel the picture as something other than what it is. Since these AI models rely on accurate data, this “poisoning” process would essentially render the image useless for the purposes of training.
If enough of these “poisoned” images are scraped from the web and used to train an AI image generator, the AI model itself may no longer be able to produce accurate images.
For example, researchers fed Stable Diffusion, an AI image generator, and an AI model they created themselves 50 “poisoned” images of dogs, then asked it to generate new pictures of dogs. The generated images featured animals with too many limbs or cartoonish faces that only somewhat resembled a dog, per MIT Technology Review.
After researchers fed Stable Diffusion 300 “poisoned” images of dogs, it eventually began producing images of cats. Stable Diffusion did not respond to CNBC Make It’s request for comment.