Protective Poison: An Alternative Method to Copyright Protection

The MIT Review released an article discussing Nightshade, a tool for artists to protect their images from web scraping. The research was conducted by Ben Zhao’s team at the University of Chicago. Nightshade was created as a tool that artists can use to stand up against AI companies that use images to train models without the artist’s permission.

According to the creators of Nightshade, the goal of the tool is to prevent AI companies from using images scraped from the internet without proper consent. Currently, several large AI companies, such as OpenAI and Stability AI, have copyright infringement cases against them, alleging the unauthorized use of images and works. It aims to shift the power back to artists and creators by deterring AI companies from including the images in their training models because the images are “poisoned”.

Nightshade poisons datasets when images that were processed by the tool are included in the dataset. When “poisoned” images are included in a dataset, the AI model will learn inaccurate labels, resulting in processed images that do not match the intended prompt. Nightshade introduces changes to the pixels of digital art, which do not impact what a human sees, but will cause an AI model to label the image incorrectly. The article states that images of plants may be labeled as animals, dogs as cats, cars as cows, hats as cakes, etc. As more and more poisoned images are included in a dataset, the farther the resulting images will be from its intended output.

Nightshade will likely impact the way AI companies collect data and train their models. It is likely that AI companies will shift towards methods of decreased indiscriminate web scraping. If there is a threat that datasets may be poisoned by certain images, it would be more favorable for AI companies to reach an agreement with creators to avoid poisoned datasets by meeting creators’ demands for consent and adequate compensation. Without further evaluation and discussion with creators, AI companies run the risk of using images that may harm their models.

Impact and Risks

This tool appears to be an alternative method for creators to protect their intellectual property. Instead of waiting for more stringent regulations or litigating against AI companies for copyright infringement, creators can use Nightshade as a preventative method as AI companies may exercise caution from conducting indiscriminate web scraping out of fear that their products will be harmed. Although it does not stop an image from being included in a dataset, it will instead hide the concept and style of an image.

Nightshade appears to be quite beneficial for creators; however, it’s important to note that there are several risks in using this novel tool. There is a high chance that poisoned images are unintentionally introduced in AI models that have consent from creators. Since the changes in the poisoned images are not noticeable by the human eye, it may be difficult to differentiate between images. Once poisoned images are included in a dataset, it can jeopardize AI models that require a significant amount of time, money, and resources to create. Additional resources may be needed to remove the poisoned images. Not only does this impact the poisoned data set, but it may also stifle further developments and projects when resources are diverted to resolving issues.

Additionally, there’s a chance that the creation of tools like Nightshade will only incentivize AI companies to create smarter internet scrapers or AI models. Because the Nightshade developers made the tool open-sourced, AI companies will have access to this tool and may find ways to reverse engineer the process that was used. From there, it will be a race between AI companies and developers of tools like Nightshade.

Innovation and Moral Rights

In considering the purpose of copyright protection, the Court in Cinar Corporation v. Robinson, 2013 SCC 73, [2013] 3 S.C.R. 1168 (“Cinar”) emphasizes the importance of striking a balance between protecting the skill and judgment of creators and leaving ideas free for the public to draw upon. In the context of Nightshade, one can argue that this tool provides an appropriate balance of these elements. The tool protects creators’ works by incentivizing consent while still allowing for the use of AI models. The general public can still use AI models for artwork. However, it can also be argued that the burden to ensure the accuracy of AI models is on the AI companies. This burden may potentially limit innovation and development. Nightshade will inevitably lead to the need to use additional resources, either to prevent poisoning or resolve issues if datasets are affected. Additionally, if datasets are successfully poisoned and AI companies fail to resolve the issues, the public will need to contend with navigating AI models that may be inaccurate and unreliable.

On its face, the tool appears to meet the balance discussed in Cinar. However, this tool is still at its infancy. It is difficult to predict how AI companies will respond to this tool. As there are a multitude of potential scenarios, an analysis balancing the elements may not be possible until the tool is out for creators to use.

Regarding moral rights, the Copyright Act states that the author of a work has the right to integrity of the work. In accordance with section 28.2, the right to integrity is only infringed when it prejudices the author’s honour or reputation, including distortion, mutilation, or modification. Moral rights in this context are quite interesting because creators are intentionally modifying and distorting their work as a form of protection. This distortion, however, does not impact the creator’s honour or reputation. In this context, the definition of moral rights can not be easily applied, particularly from the perspective of the creator.

Examining moral rights from a corporate standpoint adds an interesting dimension to the discussion, specifically exploring whether companies can leverage moral rights. Can companies assert that their AI models’ integrity may be compromised by poisoned images? Although making such an argument may face challenges, one could argue that the poisoning of AI models might result in distorted outcomes, leading to unreliable results. This could impact the company’s honor and reputation. However, a counterargument is that AI models do not possess the requisite skill and judgment to warrant copyright protection. Despite the potential for distortion and subsequent reputational consequences, AI models currently lack rights afforded to humans in terms of copyright protection.

Overall, the development and launch of this tool will provide further complexity in the evolving world of artificial intelligence and copyright image protection. At this time, it is difficult to predict how large AI companies will respond to the image poisoning tools.

Source
Heikkila, Melissa. “This new data poisoning tool lets artists fight back against Generative AI”, (24 October 2023), online: MIT Technology Review <https://www.technologyreview.com/2023/10/23/1082189/data-poisoning-artists-fight-generative-ai/>.