Why Investors Are Betting Big on Cognition’s AI Coding Tech
Getting Data
Loading...

Apple faces lawsuit for unauthorized book use in AI models

Apple is sued for using pirated books to train its AI models without permission or compensation.

AvatarJR

By Jace Reed

3 min read

Image for illustrative purpose.
Image for illustrative purpose.

Apple has been sued by authors Grady Hendrix and Jennifer Roberson for allegedly using pirated books to train its OpenELM AI models without permission or compensation.

The lawsuit was filed in federal court in Northern California and aims to represent other authors affected by similar unauthorized use.

This lawsuit coincides with a recent $1.5 billion settlement from AI company Anthropic over similar claims, highlighting a growing legal focus on the use of copyrighted materials in AI training datasets.

What is the basis of the Apple lawsuit?

The lawsuit claims Apple used the Books3 dataset, containing over 196,000 pirated books from shadow libraries, to develop its AI models.

The authors argue Apple did not seek permission, provide credit, or offer compensation for their works incorporated into the dataset.

This unauthorized use allegedly violates copyright laws and harms authors who rely on royalties from their works.

The complaint emphasizes Apple's failure to recognize authors' rights in this lucrative AI venture.

Did you know?
Apple was founded on April 1, 1976, by Steve Jobs, Steve Wozniak, and Ronald Wayne.

How does Apple’s AI model use copyrighted books?

Apple’s OpenELM AI models power some features within the Apple Intelligence suite. These models were trained on large datasets, including pirated books, to improve language understanding and generation.

The authors assert their copyrighted books were part of the pirated dataset, enabling the AI to learn from protected content without authorization. This raises questions about ethical AI training and data sourcing.

AI companies face increasing litigation over training data sourced from copyrighted works. Anthropic’s recent settlement is the largest copyright recovery case in U.S. history, signaling heightened scrutiny.

However, Meta recently secured a legal victory affirming the fair use of copyrighted books for AI training when legally obtained. The distinction often lies in whether the data was pirated or legally purchased.

ALSO READ | Google Photos Upgrades Photo-to-Video with Veo 3 AI

What risks does the lawsuit pose to Apple?

Apple faces significant financial liabilities and reputational damage if found guilty of using pirated materials. The lawsuit could trigger costly settlements or court rulings against Apple’s AI practices.

The case challenges Apple’s image as a privacy-conscious company and adds to pressures on tech giants to adopt ethical data sourcing for AI development and avoid copyright infringement.

How could this case influence future AI development?

This lawsuit may shape legal standards requiring AI developers to secure proper licenses for copyrighted content before training models. It highlights growing concerns about respecting intellectual property in AI progression.

Global legal disputes could result in increased transparency and licensing frameworks, which would ensure creators receive compensation and promote responsible AI innovation.

Forward-looking, the resolution of this case could set crucial precedents for the AI industry's development, balancing technological advancement with creators’ rights.

Should tech companies pay authors for AI training data?

(0)

Please sign in to leave a comment

Related Articles

MoneyOval

MoneyOval is a global media company delivering insights at the intersection of finance, business, technology, and innovation. From boardroom decisions to blockchain trends, MoneyOval provides clarity and context to the forces driving today’s economic landscape.

© 2025 MoneyOval.
All rights reserved.