Nightshade: A New Tool to Protect Artists' Work from AI Copyright Infringement
This new data poisoning tool lets artists fight back against generative AI 🔗
A new data poisoning tool called Nightshade allows artists to add invisible changes to their art before uploading it online, aiming to disrupt the training data used by image-generating AI models. This tool is intended to combat AI companies that use artists' work without permission, potentially rendering some AI outputs useless. The tool's creators also developed Glaze, which allows artists to mask their personal style to prevent it from being scraped by AI companies. While Nightshade could be used for malicious purposes, it could also serve as a powerful deterrent against disrespecting artists' copyright and intellectual property. The tool exploits a security vulnerability in generative AI models, causing them to malfunction when fed with poisoned training data. The research has received positive feedback from experts, suggesting that it could have a significant impact on the relationship between AI companies and artists.