Apple in trouble with a lawsuit as authors file for pirated books for Apple Intelligence

Apple Hit With Authors’ Lawsuit Over Alleged Pirated-Book Training, Challenging Its Privacy-First Image

Apple’s privacy-first image is facing a high-stakes test. Two neuroscience authors have filed a class-action lawsuit alleging the company used pirated copies of their books to train Apple Intelligence, its new AI initiative. For a brand that has long marketed itself as the guardian of user privacy and ethics, the claim strikes at its core message just as it moves deeper into artificial intelligence.

The complaint centers on Books3, a controversial dataset contained within The Pile, which aggregates large volumes of text sourced from shadow libraries. According to the lawsuit, Apple relied on Books3 during training, capturing works by Susana Martinez-Conde and Stephen Macknik, including Sleights of Mind: What the Neuroscience of Magic Reveals About Our Everyday Deceptions and Champions of Illusion: The Science Behind Mind-Bending Magic Tricks. The filing further claims Apple had ties to Books3-linked data and later pivoted away from it in 2023 amid rising copyright concerns.

What makes the accusation particularly damaging is the perceived gap between Apple’s brand promises and the practices alleged in the suit. Apple has consistently set itself apart as a company that protects data and acts with higher ethical standards. If the claims hold up, critics argue it could undercut the company’s credibility at a pivotal moment for its AI ambitions.

This legal battle could have industry-wide consequences. Tech leaders like OpenAI, Google, and Meta have all faced scrutiny over how copyrighted material is used in training large language models. A ruling against Apple could accelerate a shift toward stricter compliance: licensed datasets, paid access to copyrighted works, improved provenance tracking, and more transparent model documentation. It could also nudge AI developers to adopt clearer opt-out mechanisms for creators and to conduct third-party audits of training data sources.

For now, these are allegations, not established facts. The case is in its early stages, and there has been no finding of liability. Whether Apple trained Apple Intelligence on pirated books remains unproven—and the outcome could define not only Apple’s AI roadmap but also the standards for responsible AI training across the industry.

Can Apple preserve its privacy-first reputation while competing in the AI race, or will this lawsuit mark a turning point for how tech giants source the data that powers their models?