High quality media assets are an extremely valuable resource for AI training (as compared to the content of the internet at large, for example). The past year has seen an increase in licensing deals in which media rightsholders permit AI developers to use their IP-protected assets to train AI tools. During this time, Axel Springer, News Corp and the Financial Times have all entered into high-profile deals with OpenAI, reportedly for large sums of money. Image database Shutterstock has entered into a number of licensing deals with OpenAI, Meta, Apple and others and is reported to have generated over $100 million in revenue from AI licensing in the year to June 2024.
Media businesses are increasingly considering partnerships with AI developers for a number of reasons other than simply earning significant licensing revenue. Importantly, licensing allows media businesses a measure of control in the way that an AI tool may be trained and may react to certain user queries. For example, a licence agreement could have built into it provisions that the rightsholder's content will only form a certain percentage of the material the model is trained on. It may allow the rightsholder to withdraw content and oblige the AI developer to stop using that content if certain circumstances occur, eg litigation against the AI developer. It could also provide obligations for the AI developer to hard code the AI model to return specific responses to user attempts to prompt engineer in order to access the rightsholder's content.
Media businesses are also increasingly deploying AI in their own businesses, for example, to improve user experience or content discovery, to offer tools for user creativity, or to generate creative assets themselves. Partnering with an AI developer can provide favourable terms on which to access the developer's services. Before entering a licensing deal of this nature, the rightsholder will need first to be sure of its own rights position from creators, which can be a complicated assessment where rights are held multi-nationally.
Despite an uptick in licensing activity, disagreements and disputes regarding the use of copyright-protected works for AI training remain rife. Where no specific licensing deal has been agreed in relation to AI training, rightsholders are now increasingly amending their existing and template distribution agreements to expressly exclude certain AI use cases. Rightsholders are also taking steps to avail themselves of the ability to "opt-out" of the EU's text and data mining exception to copyright and database rights infringement. For example, Sony Music and Warner Music Group have issued letters to the world at large stating their position, and other rightsholders are taking steps to amend website terms and conditions and copyright notices.
Questions remain around the methods rightsholders can use to effectively opt-out, due to the requirement in the EU Digital Single Market Directive for the opt-out to be "machine-readable". It is currently untested whether methods such as sending letters can be effective. This question around the form of the opt-out, and specifically whether an opt-out in website terms and conditions can be sufficient, is the subject of litigation in Germany in the LAION case, which came to an oral hearing on 11 July 2024.
In the UK, there is no general text and data mining exception to copyright infringement, but an exception does apply to text and data mining for non-commercial research and there is also a temporary copying exception (that also exists in the EU). Both are untested in this context. The Getty v Stability AI case may not consider the application of these exceptions. Stability AI - in its defence to Getty's claims regarding the training and development of the Stable Diffusion model - is largely relying on the fact that training took place outside of the UK and therefore no copyright-relevant acts took place in the UK. This will need to be argued and decided as a matter of fact rather than law. However, Stability AI does argue, in the context of Getty's secondary infringement claim, that, if the Stable Diffusion model as a whole is an infringing copy of Getty's works used for training, then the making of the model would not have infringed copyright in the UK. If these arguments are expanded then the case may yet address questions of whether AI training infringes copyright.