- Pascal's Chatbot Q&As
- Posts
- Journalists v. Google: The plaintiffs — Illinois-based journalists, podcasters and audiobook narrators — allege that Google used their recorded voices to train voice AI systems powering products.
Journalists v. Google: The plaintiffs — Illinois-based journalists, podcasters and audiobook narrators — allege that Google used their recorded voices to train voice AI systems powering products.
Regulators should require provenance, speaker consent, retention/deletion plans, and stricter rules for biometric data in AI training.
Summary: The lawsuit is not just about AI copying content; it is about Google allegedly extracting biometric voiceprints from journalists, podcasters and narrators without notice, consent or compensation, then using those voices to build products that may compete with them.
The strongest and most surprising claim is that Google allegedly knew how to build consent systems for voice cloning, but used them only for some voices while leaving pretraining voices unconsented, making this look less like a technical oversight and more like a business choice.
The case may survive in part, especially on BIPA/privacy grounds, but the most likely outcome is settlement rather than court-ordered model destruction; regulators should require provenance, speaker consent, retention/deletion plans, and stricter rules for biometric data in AI training.
The Voiceprint Wars: Why the Google Journalist Lawsuit Is Bigger Than Copyright
by ChatGPT-5.5
The Journalists v. Google complaint is not just another “AI trained on my work” lawsuit. It is more dangerous for Google because it tries to move the fight away from copyright’s contested fair-use battlefield and into the territory of biometric privacy, identity rights, consent, market substitution and model-level contamination. The plaintiffs — Illinois-based journalists, podcasters and audiobook narrators — allege that Google used their recorded voices to train voice AI systems powering products such as Gemini Live, NotebookLM Audio Overviews, YouTube auto-dubbing, Google Cloud Text-to-Speech and Google Assistant. Reuters describes the case as a proposed class action by award-winning journalists, podcasters and narrators accusing Google of violating Illinois publicity and biometric privacy rights by using voice recordings without permission. Google had not yet responded when Reuters reported the case.
1. The nature of the grievances
The core grievance is simple: Google allegedly took human voices, converted them into machine-usable identity signals, embedded those signals in commercial voice models, and then sold products that compete with the very people whose voices helped build them. That makes the complaint morally intuitive in a way some copyright-training cases are not. A book, article or image can be argued over as “content”; a voice is harder to treat as mere material. It is biological, personal, recognisable, and commercially valuable.
Legally, the complaint rests most heavily on the Illinois Biometric Information Privacy Act, or BIPA. That is clever because BIPA expressly covers “voiceprints” and requires notice, purpose disclosure, retention rules and written consent before biometric identifiers are collected. The complaint also pleads Illinois right-of-publicity claims, consumer fraud/deceptive trade practices, and unjust enrichment. In other words, the plaintiffs are not only saying “you copied our recordings.” They are saying: you extracted our identities, used them as infrastructure, concealed the extraction, monetised the result, and damaged our labour market.
The complaint’s strongest framing is that Google allegedly understood exactly how consent should work because it built explicit consent workflows for products such as Custom Voice and Chirp 3 Instant Custom Voice. The complaint says Google required speakers in those contexts to record a consent statement, but did not apply comparable consent infrastructure to the much larger pool of voices used for pretraining foundational voice systems. That distinction — consent for the voices Google paid for, opacity for the voices allegedly scraped at scale — is the emotional and legal centre of the case.
The complaint’s most ambitious theory is that the unlawfully collected biometric data is not sitting in a neat database waiting to be deleted. It alleges that voiceprints are encoded in model parameters and reproduced through generated audio, so that “the biometric data and the product are the same thing.” That is powerful, but it is also where the case becomes technically and legally vulnerable. Courts will need to decide whether model weights that encode statistical patterns from voice data are legally equivalent to retained biometric identifiers or biometric information. That is a much harder question than whether a company stored a fingerprint scan in a database.
2. The most surprising statements
The most surprising statement is that the plaintiffs seek not merely damages but destruction or retraining of the foundational voice models and downstream products allegedly containing unlawfully obtained voiceprints. That is a direct attack on the “train first, litigate later” model of AI development. If taken seriously, it would convert training-data compliance from a paper risk into a product-continuity risk. The requested relief includes disclosure of training-data sources, destruction of unlawfully obtained voiceprints, and retraining or destruction of models and products built on those voiceprints.
Second, the complaint alleges that Google never published a model card, data sheet, training-data manifest, licensing inventory or transparency report identifying the voice recordings used for its foundational voice models. For regulators, that allegation matters almost as much as the alleged misuse itself. The case turns opacity into evidence: because Google allegedly controls the training records, the plaintiffs argue that they should not be penalised for lacking precise proof before discovery.
Third, the complaint points to Google’s alleged YouTube-related training practices and the September 2024 amendment of YouTube’s Terms of Service to add machine-learning and AI-use language. The plaintiffs argue that even if uploaders granted Google rights, uploaders could not waive the personal biometric or publicity rights of third-party speakers appearing in videos, podcasts, interviews or news clips. That is a valuable point: platform upload consent is not necessarily speaker consent.
Fourth, the complaint makes a direct market-substitution claim. It alleges that Google Cloud Text-to-Speech competes with audiobook narrators, NotebookLM Audio Overviews competes with podcast/audio journalists, and YouTube auto-dubbing competes with voiceover/localisation professionals. This makes the harm concrete: the grievance is not only privacy invasion; it is using labour-derived identity data to automate and underprice the labour market from which that data came.
3. The most controversial statements
The most controversial statement is the assertion that model parameters contain the plaintiffs’ voiceprints in a legally actionable form. Technically, a model may encode features learned from many speakers without preserving any one speaker’s voiceprint in a simple extractable form. Plaintiffs will need expert evidence showing that Google’s systems created or retained speaker-identifying representations, not merely that they trained on audio. Google will likely argue that training a model on speech is not the same as collecting a BIPA “voiceprint” used to identify a person.
The second controversial point is causation. The plaintiffs allege on information and belief that their voices were included in Google’s training data because their recordings were publicly available, professionally produced, long-form and highly suitable for training. That may be enough to survive early pleading if the court accepts that the relevant evidence is uniquely in Google’s possession. But it is not enough to win. At some stage, plaintiffs must connect specific recordings, specific speakers and specific model-training pipelines.
The third controversial point is extraterritoriality. BIPA is an Illinois statute. The complaint tries to localise the case by stressing that the plaintiffs are Illinois residents, that their recordings were produced or distributed from Illinois, that Google operates significantly in Illinois, and that Google sells voice AI products in Illinois. Google will likely argue that the relevant collection and training occurred elsewhere, probably on Google infrastructure outside Illinois, and that Illinois law should not govern global AI-training practices simply because some speakers reside there. The complaint anticipates that problem, but it remains a major battleground.
The fourth controversial issue is remedy. A court may be willing to award damages or require policy changes, but ordering a major AI company to destroy or retrain foundational models would be extraordinary. The plaintiffs are right to ask for it strategically because it increases leverage and frames model weights as contaminated assets. But the most likely practical outcome is not a judicial order to burn down the model; it is a settlement involving money, disclosures, consent mechanisms, retention policies and perhaps limits on particular uses.
4. The most valuable statements
The most valuable sentence in the complaint is not a single legal formula but the underlying proposition: consent infrastructure exists when companies decide it matters. That is the key lesson for regulators and rights owners. Google allegedly knew how to obtain recorded speaker consent because it required it for some voice products. The question is therefore not whether consent at scale is impossible; it is whether companies are allowed to divide the world into “licensed consent-required voices” and “ambient internet voices” used for pretraining.
The second valuable insight is that AI transparency cannot stop at text, images and copyright. Voice, face, likeness, gesture, accent and performance style are identity assets. A future AI governance regime that only asks whether copyrighted works were copied will miss the deeper problem: models can absorb human attributes that are economically and socially meaningful even when traditional copyright claims are uncertain.
The third valuable point is that platforms cannot solve consent by relying on uploader terms. A person interviewed on YouTube, recorded in a news clip, quoted in a podcast or featured in a documentary may never have agreed to Google’s platform terms. That distinction matters enormously for publishers, broadcasters, universities and event organisers: the organisation may own or license the recording, but the speaker may still have biometric, publicity or performer-related rights.
The fourth valuable point is the link between training and displacement. This is not merely an abstract privacy case. It describes a cycle: collect voices from professionals; train systems on those voices; sell cheap synthetic substitutes; reduce the market value of the original professionals. Whether every factual allegation is proven or not, that cycle captures one of the central political-economy problems of generative AI.
5. Predicted outcome
My prediction: Google is unlikely to get the entire case dismissed early, but the plaintiffs are unlikely to obtain the full model-destruction remedy. The most likely outcome is a hard-fought motion-to-dismiss phase, followed by discovery pressure, followed by settlement if plaintiffs obtain meaningful evidence about training data.
The BIPA claims are the plaintiffs’ best route. Illinois BIPA has historically been plaintiff-friendly: the Illinois Supreme Court held in Rosenbach that plaintiffs do not need to show separate actual injury beyond violation of statutory rights, and in Timsthat a five-year limitations period applies to BIPA claims. Those doctrines help plaintiffs survive threshold arguments that “nothing happened” unless a clone was generated or a financial loss was proven.
However, damages leverage is lower than it would have been several years ago. Illinois amended BIPA in 2024 to limit repeated collections or disclosures of the same biometric identifier from the same person to a single recovery, rather than per-scan damages; the Seventh Circuit held in April 2026 that the amendment applies retroactively to pending cases. That does not eliminate liability, but it reduces the threat of annihilative statutory damages and will affect settlement value.
The right-of-publicity claim may survive if the plaintiffs can show commercial use of their voices or identities, especially given the broader trend of courts allowing some AI voice claims to proceed. In the Lovo voice-actor litigation, a federal judge allowed parts of the case to continue, including publicity/privacy-related theories, while dismissing or narrowing other claims. That suggests courts may be receptive to voice-as-identity claims even when copyright claims are weaker.
The weakest claims are likely to be the broad consumer-fraud/deceptive-practices and unjust-enrichment theories, at least unless plaintiffs can show specific deceptive conduct directed at them or a concrete benefit traceable to their voices. Those claims may survive in narrowed form, but BIPA and right of publicity are the real engines of the case.
Class certification will be difficult. The plaintiffs need a class that can be identified without endless mini-trials about who lived in Illinois, whose recordings were used, whether their voices were extracted, whether they consented, and whether Google’s systems retained legally relevant voiceprints. The complaint argues that Google’s records, public metadata and voice-matching analysis can solve this. Perhaps. But in practice, class certification may become the central pressure point: Google will argue individualised proof; plaintiffs will argue automated uniform conduct.
So the most likely litigation path is:
Some claims survive dismissal, especially BIPA notice/consent and retention-policy theories.
Alphabet may try to exit or narrow its role, but the complaint’s use of Alphabet-level AI governance disclosures is designed to resist that.
Discovery into training data becomes the decisive fight.
If plaintiffs get discovery showing identifiable ingestion of their recordings or comparable Illinois voice professionals, settlement pressure rises sharply.
Final relief is more likely to be monetary settlement plus compliance commitments than a court-ordered retraining of Gemini/NotebookLM-related voice models.
6. Recommendations for regulators
Regulators should treat this lawsuit as a warning that AI training governance is no longer only a copyright issue. It is also a biometric, labour, identity, platform and product-safety issue. The regulatory response should therefore be broader than “was the work copyrighted?”
First, regulators should require training-data provenance records for biometric and identity-bearing media. Voice, face, gait, likeness and performance data should have auditable lineage: source, licence, consent status, permitted uses, retention period and deletion/retraining feasibility.
Second, regulators should prohibit companies from relying solely on platform uploader terms for biometric or publicity rights. A YouTube uploader, podcast distributor or archive manager should not be presumed to have authority to consent on behalf of every identifiable speaker whose voice appears in the content.
Third, regulators should require separate consent rules for synthetic voice and likeness systems. Consent should be specific, informed, revocable where technically feasible, and tied to defined use cases: training, fine-tuning, cloning, dubbing, commercial API access, advertising and downstream sublicensing should not be bundled into one vague permission.
Fourth, regulators should mandate model-level remediation plans before deployment. Companies should not be allowed to say, after litigation begins, that deletion is impossible because the data is embedded in weights. If a company builds a model from regulated data, it should have a pre-existing plan for audit, removal, suppression, retraining or compensation.
Fifth, regulators should require public-facing registries or transparency reports for high-risk voice AI systems, including whether professional voice datasets, public broadcast archives, podcasts, audiobooks, YouTube videos or other third-party recordings were used.
Finally, regulators should recognise the labour-market issue. AI companies should not be allowed to build synthetic substitutes from the uncompensated identity assets of the workers they displace. For voice professionals, journalists, narrators, actors and creators, the question is no longer just “was my work copied?” It is: was my personhood converted into infrastructure for a machine that now competes against me? That is the deeper issue this lawsuit forces into the open.
