MIT Technology Review Subscribe

Artists can now opt out of the next version of Stable Diffusion

The move follows a heated public debate between artists and tech companies over how text-to-image AI models are trained.

Artists will have the chance to opt out of the next version of one of the world’s most popular text-to-image AI generators, Stable Diffusion, the company behind it has announced

Stability.AI will work with Spawning, an organization founded by artist couple Mat Dryhurst and Holly Herndon, who have built a website called HaveIBeenTrained that allows artists to search for their works in the data set that was used to train Stable Diffusion. Artists will be able to select which works they want to exclude from the training data.

Advertisement

The decision follows a heated public debate between artists and tech companies over how text-to-image AI models should be trained. Stable Diffusion is based on the open-source LAION-5B data set, which is built by scraping images from the internet, including copyrighted works of artists. Some artists’ names and styles have become popular prompts for wannabe AI artists

This story is only available to subscribers.

Don’t settle for half the story.
Get paywall-free access to technology news for the here and now.

Subscribe now Already a subscriber? Sign in
You’ve read all your free stories.

MIT Technology Review provides an intelligent and independent filter for the flood of information about technology.

Subscribe now Already a subscriber? Sign in

Dryhurst told MIT Technology Review that artists have “around a couple of weeks” to opt out before Stability.AI starts training its next model, Stable Diffusion 3. 

The hope, Dryhurst says, is that until there are clear industry standards or regulation around AI art and intellectual property, Spawning’s opt-out service will augment legislation or compensate for its absence. In the future, Dryhurst says, artists will also be able to opt in to having their works included in data sets.

A spokesperson for Stability.AI told MIT Technology Review: ”We are listening to artists and the community and working with collaborators to improve the dataset. This involves allowing people to opt out of the model and also to opt in when they are not already included.”

But Karla Ortiz, an artist and a board member of the Concept Art Association, an advocacy organization for artists working in entertainment, says she doesn’t think Stability.AI is going far enough.

The fact that artists have to opt out means “that every single artist in the world is automatically opted in and our choice is taken away,” she says.

“The only thing that Stability.AI can do is algorithmic disgorgement, where they completely destroy their database and they completely destroy all models that have all of our data in it,” she says. 

The Concept Art Association is raising $270,000 to hire a full-time lobbyist in Washington, DC, in hopes of bringing about changes to US copyright, data privacy, and labor laws to ensure that artists’ intellectual property and jobs are protected. The group wants to update laws on intellectual property and data privacy to address new AI technologies, require AI companies to adhere to a strict code of ethics, and work with labor unions and industry groups that deal with creative work. 

Advertisement

“It just truly does feel like we artists are the canary in the coal mine right now,” says Ortiz. 

Ortiz says the group is sounding the alarm to all creative industries that AI tools are coming for creative professions “really fast,” and “the way that it’s being done is extremely exploitative.” 

This is your last free story.
Sign in Subscribe now

Your daily newsletter about what’s up in emerging technology from MIT Technology Review.

Please, enter a valid email.
Privacy Policy
Submitting...
There was an error submitting the request.
Thanks for signing up!

Our most popular stories

Advertisement