• Pascal's Chatbot Q&As
  • Posts
  • When a user communicated with ChatGPT, code placed on the website allegedly caused the browser to duplicate part of that communication to Meta or Google.

When a user communicated with ChatGPT, code placed on the website allegedly caused the browser to duplicate part of that communication to Meta or Google.

The case was filed in the Northern District of California as Lim v. OpenAI Global LLC, but the docket now shows a voluntary dismissal without prejudice on May 13, 2026.

Summary: The complaint alleges that OpenAI allowed Meta and Google tracking tools to receive ChatGPT query-related data, turning private chatbot interactions into adtech signals.
The strongest argument is that AI prompts are highly sensitive communications, but the case depends on proving that actual prompt content or meaningful prompt-derived data was shared, not just generic analytics metadata.
For regulators, the key lesson is that AI privacy cannot focus only on model training; prompt interfaces should be treated as high-sensitivity spaces where advertising pixels, identity matching, and vague cookie consent are not enough.

by ChatGPT-5.5

The complaint against OpenAI is not, at its core, a traditional AI copyright or training-data lawsuit. It is a privacy and surveillance case dressed in the language of modern chatbot use. The central allegation is simple but potentially explosive: users typed queries into ChatGPT believing they were communicating with OpenAI, but OpenAI allegedly embedded Meta and Google tracking technologies into the ChatGPT website, causing user queries or query-related information, together with identifying cookies and account-linked signals, to be transmitted to Meta and Google without adequate consent.

Bloomberg Law reported the lawsuit as one of the first federal suits alleging unlawful disclosure of information submitted to AI chatbots. The case was filed in the Northern District of California as Lim v. OpenAI Global LLC, but the docket now shows a voluntary dismissal without prejudice on May 13, 2026. That means there has been no merits ruling. The complaint may have been withdrawn for strategic, pleading, settlement, defendant-identification, or procedural reasons, and it could theoretically be refiled or echoed by other plaintiffs.

The core grievance

The plaintiff’s main grievance is that ChatGPT is not an ordinary website. Users do not merely browse a product page or click on a marketing banner. They ask questions, describe problems, disclose personal concerns, test business ideas, seek health, diet, legal, financial, and emotional advice, and sometimes paste confidential work material. The complaint argues that this creates a much stronger expectation of privacy than ordinary web browsing.

The second grievance is that OpenAI allegedly embedded Meta Pixel and Google Analytics or related tracking technologies into the ChatGPT website. According to the complaint, these tools did not merely measure anonymous traffic. They allegedly transmitted user activity, query topics, page or event information, cookies, hashed identifiers, Facebook IDs, Google identifiers, or other signals that could connect chatbot interactions to real-world individuals.

The third grievance is that the alleged transmission happened in real time. The complaint frames the technical process as a kind of modern wiretap: when a user communicated with ChatGPT, code placed on the website allegedly caused the browser to duplicate part of that communication to Meta or Google. This is why the lawsuit invokes the federal Wiretap Act and California’s Invasion of Privacy Act, rather than only claiming ordinary unfair-business-practice or privacy-policy violations.

The fourth grievance is commercial exploitation. The complaint alleges that Meta and Google did not receive the information as neutral infrastructure providers only, but as advertising and analytics companies able to enrich profiles, improve ad targeting, build audiences, measure conversions, and connect activity across services. This is the allegation that gives the case its broader policy charge: intimate AI conversations may have been converted into signals inside the adtech economy.

The fifth grievance is lack of meaningful consent. OpenAI will almost certainly argue that it provides privacy notices, cookie controls, and disclosures. Its current public cookie policy lists Google Analytics and Meta-related cookies or APIs on openai.com and chatgpt.com, and its U.S. privacy policy states that it may disclose limited personal data to service providers and selected marketing partners, including through cookies and similar technologies. But the plaintiff’s point is that generic disclosure is not the same as informed, affirmative consent to disclose the substance or meaning of chatbot prompts to third-party adtech companies.

The complaint brings four broad categories of claims.

First, it alleges violation of the federal Electronic Communications Privacy Act, specifically the Wiretap Act. The theory is that Meta and Google allegedly intercepted electronic communications while they were in transit, and that OpenAI either caused, procured, aided, or permitted that interception by embedding third-party tracking code.

Second, it alleges violation of California Penal Code § 631, part of the California Invasion of Privacy Act. This is one of the most popular statutes in recent website-tracking litigation because it can be used against companies that allegedly allow a third party to “read,” “learn the contents,” or “learn the meaning” of a communication without consent.

Third, it alleges violation of California Penal Code § 632, which concerns confidential communications. The complaint argues that ChatGPT queries can be confidential because users reasonably expect that personal, health-related, business, or other sensitive questions are not being disclosed to third-party advertising networks.

Fourth, it pleads invasion of privacy under the California Constitution and common law intrusion upon seclusion. This is the broadest theory: even if the statutory claims face technical hurdles, the plaintiff argues that the conduct is offensive because it invades a private conversational space.

The complaint seeks class certification, injunctive relief, statutory damages, actual damages, punitive damages, attorneys’ fees, and a jury trial. The statutory damages theory is potentially enormous because California privacy statutes can provide damages per violation, which is why these cases often create very high theoretical exposure even before actual harm is proven.

Why the case matters

The case matters because it attacks a weak point in the AI industry’s privacy architecture: the collision between intimate conversational AI and legacy web analytics.

For years, companies treated tracking pixels, analytics tags, conversion APIs, advertising cookies, and identity-matching tools as normal background infrastructure. That assumption becomes much harder to defend when the interface is not a retail site or news page, but a chatbot that invites users to disclose problems, plans, anxieties, medical symptoms, legal questions, workplace secrets, and personal vulnerabilities.

This is the most important conceptual move in the complaint. It says: a prompt is not just a click. A chatbot query is closer to a conversation, search history, diary entry, consultation request, or confidential memorandum than to an ordinary page view. If regulators accept that framing, AI companies will need to treat prompt interfaces as high-sensitivity environments where advertising infrastructure is presumptively inappropriate.

The case also matters because the legal theory has some precedent. In In re Facebook Internet Tracking Litigation, the Ninth Circuit held that plaintiffs had sufficiently alleged Wiretap Act and CIPA claims where Facebook allegedly received duplicated GET requests through tracking technology. The court rejected the idea that simultaneous duplication to a third party is automatically protected by the “party exception” at the pleading stage. That precedent does not guarantee success here, but it gives plaintiffs a plausible path, especially in California.

Are the arguments robust?

The arguments are serious but not yet proven. As a policy case, the complaint is strong. As a legal case, it is plausible but vulnerable. As an evidentiary case, it depends heavily on what exactly was transmitted to Meta and Google.

The strongest part of the complaint is the intuition that AI prompts deserve heightened privacy protection. Users are encouraged to treat chatbots as assistants, tutors, analysts, coaches, and quasi-advisers. OpenAI and other AI companies benefit commercially from that intimacy. They cannot then casually fall back on ordinary website-tracking norms if prompt-derived information is flowing to adtech partners.

The second strong point is the identity-linking allegation. The complaint does not merely say that Meta or Google saw anonymous traffic. It alleges that transmissions included or could be linked to Facebook IDs, Google cookies, hashed emails, profile identifiers, or other persistent signals. If true, that is materially different from aggregate analytics. The privacy risk comes from the combination of query content plus durable identity.

The third strong point is the California law angle. CIPA has become a powerful weapon against hidden third-party tracking, and the Ninth Circuit’s Facebook tracking decision gives plaintiffs a foundation for arguing that third-party trackers can be treated as eavesdroppers when they receive duplicated communications in real time. Again, that does not decide the case, but it prevents the theory from being frivolous.

But the complaint also has weaknesses.

The biggest evidentiary weakness is whether the actual contents or meaning of ChatGPT prompts were transmitted. It is one thing to show that Meta or Google received cookies, page URLs, event names, device information, or analytics metadata. It is another to show that they received the user’s prompt, a prompt-derived title, a query string, a page title revealing the subject matter, or enough context to infer the substance of the communication. The Wiretap Act and CIPA claims become much stronger if prompt content or meaningful query information was transmitted. They become weaker if the transmission was only technical metadata.

The second weakness is consent. OpenAI’s current public documents disclose the use of cookies, analytics, advertising-related technologies, and some sharing with marketing partners, including opt-out controls. That does not automatically defeat the complaint, because consent must be meaningful and specific enough for the conduct at issue. But it gives OpenAI an obvious defence: users were notified, cookie settings existed, and any sharing was governed by disclosed policies. The plaintiff will need to show that those disclosures were insufficient, misleading, untimely, or inadequate for wiretap-law consent.

The third weakness is the role of Meta and Google. OpenAI may argue that these companies were service providers, analytics vendors, or processors, not independent “eavesdroppers.” Plaintiffs will argue the opposite: Meta and Google are advertising giants with their own commercial interests, profile-building systems, and downstream monetization incentives. The case may therefore turn on the precise contracts, settings, data-use restrictions, and technical configurations.

The fourth weakness is overbreadth. The proposed class appears to include a very large universe of U.S. ChatGPT users. But users may have different account types, cookie settings, browsers, locations, logged-in statuses, privacy controls, and interaction histories. Some may have blocked cookies. Some may not have been logged into Meta or Google. Some may have used mobile apps rather than the website. Some may have used paid tiers with different settings. Those differences could make class certification difficult.

The fifth weakness is pleading quality. The complaint contains some language that appears borrowed from healthcare-tracking litigation, including references to protected health information and a somewhat awkward statutory framing around confidential medical information. That does not destroy the case, but it gives the pleading a template-driven feel. In a technically complex AI privacy case, precision matters.

The sixth weakness is one specific technical claim: the suggestion that Google could “decrypt” hashed emails. Strictly speaking, hashing is generally one-way. The more accurate allegation would be that Google could match a hashed email against its own hashed user records if the same hashing method was used. That distinction matters because courts and regulators increasingly understand these technical details. Overstatement can undermine credibility.

So my, ChatGPT’s, view is this: the complaint’s policy theory is robust, its legal theory is plausible, but its success would depend on forensic proof of actual data flows. If prompt contents or meaningful prompt-derived data were sent to Meta and Google alongside persistent identifiers, the case becomes powerful. If the transmissions were limited to generic analytics events and disclosed cookies, the case becomes much harder.

Most surprising statements and findings

The most surprising allegation is that ChatGPT prompts or prompt-derived information may have been transmitted to Meta and Google through ordinary website tracking infrastructure. The shock is not that a website used analytics. The shock is that the same adtech architecture allegedly touched an AI interface where users may disclose deeply personal or confidential information.

The second surprising element is the scale of the proposed class. The complaint seeks to represent all U.S. residents whose personally identifiable information and ChatGPT communications were disclosed to third parties through use of the website. That turns a technical web-tracking dispute into a potentially national AI privacy case.

The third surprising point is the use of conventional wiretap law against AI chatbot infrastructure. The complaint does not rely on a new AI-specific statute. It uses old surveillance and communications privacy laws to argue that modern tracking pixels can function as eavesdropping devices.

The fourth surprising point is the timing. The case was filed on May 5, 2026, but the public docket shows voluntary dismissal without prejudice by May 13, 2026. That makes the case important as a signal, but not yet important as precedent.

Most controversial statements and findings

The most controversial claim is that Meta Pixel and Google Analytics can be characterized as wiretaps. Many companies would say these are standard analytics and advertising tools. Plaintiffs respond that “standard” does not mean lawful when the tools transmit the contents or meaning of private communications to third parties.

The second controversial claim is that ChatGPT users had a reasonable expectation of confidentiality. OpenAI may argue that users interact with a cloud service governed by posted terms and privacy policies, not a doctor, lawyer, or therapist. The plaintiff will argue that the nature of the interface, the sensitivity of the queries, and the company’s own positioning of ChatGPT as a personal assistant create a privacy expectation far above normal browsing.

The third controversial point is statutory damages. If every query, event, or transmission is treated as a separate violation, theoretical damages could become enormous. That raises the usual tension in privacy class actions: statutory damages can be necessary to deter invisible surveillance, but they can also produce settlement pressure far beyond proven individual harm.

The fourth controversial point is the hashed-email allegation. The complaint’s apparent “decryption” framing is technically imprecise. The real concern is not decryption but matchability: whether a hashed email or other identifier allows a platform to connect chatbot activity to an existing advertising identity.

The fifth controversial point is whether privacy policies can cure this problem. Many companies assume that broad cookie banners and privacy notices are enough. This case challenges that assumption. For AI systems, regulators may decide that burying adtech disclosures in a general policy is not sufficient when users are entering sensitive prompts.

Most valuable statements and findings

The most valuable insight is that prompt privacy should be treated as a distinct regulatory category. A prompt is not merely “usage data.” It can contain health symptoms, legal worries, trade secrets, unpublished manuscripts, financial stress, workplace allegations, sexual or emotional disclosures, and sensitive inferences about identity. AI companies should not be allowed to collapse all of that into ordinary web analytics.

The second valuable insight is that third-party code is a governance problem. AI safety discussions often focus on model weights, hallucinations, training data, and misuse. This complaint reminds us that privacy failures may arise from mundane product plumbing: tags, pixels, cookies, SDKs, conversion APIs, debugging tools, and analytics dashboards.

The third valuable insight is that identity linkage is the heart of the harm. Even if the transmitted text is partial, the combination of topic, timing, account identity, browser identifier, and platform cookie can reveal enough to be privacy-invasive. Regulators should therefore look not only at full transcript disclosure but also at inference-rich metadata.

The fourth valuable insight is that AI companies may be judged by a higher standard than ordinary websites. The more an AI vendor markets its product as a personal assistant, productivity partner, tutor, coach, or professional tool, the less plausible it becomes to treat the interface as a normal ad-supported webpage.

The fifth valuable insight is that this litigation theory will travel. Even though this specific case was voluntarily dismissed without prejudice, the theory can be applied to any AI service that combines sensitive conversational interfaces with third-party advertising, analytics, session-replay, conversion, or identity-matching tools.

What this means for OpenAI and the wider AI industry

For OpenAI, the case is reputationally awkward even without a merits ruling. OpenAI wants enterprise customers, governments, lawyers, doctors, publishers, universities, and regulated industries to trust ChatGPT and related products with sensitive workflows. Allegations that chatbot interactions were exposed to adtech infrastructure cut directly against that trust narrative.

For the wider AI industry, the lesson is broader: AI companies cannot simply import the surveillance-based web business model into conversational AI. The old internet normalized tracking. AI makes that normalization more dangerous because the interface elicits deeper disclosure.

This is especially important for professional and enterprise use. A user pasting draft contracts, unpublished research, patient-like descriptions, business plans, or internal policy documents into an AI assistant is not equivalent to clicking on a shoe advertisement. AI vendors that want to serve high-trust markets need to prove that prompt data is isolated from advertising systems, identity graphs, marketing audiences, and third-party profiling infrastructure.

The case also creates a warning for procurement. Companies buying AI tools should ask not only whether prompts are used for model training, but whether prompts, prompt metadata, conversation titles, URLs, event logs, analytics records, support logs, or telemetry can flow to third-party vendors. “We do not train on your data” is not the same as “we do not leak your data into the adtech stack.”

Recommendations for regulators

Regulators should treat AI prompt interfaces as high-sensitivity environments by default. Consumer-facing AI systems should not be allowed to transmit prompt contents, prompt-derived titles, conversation metadata, or sensitive usage signals to advertising or identity-matching partners unless the user has given explicit, granular, opt-in consent.

Regulators should require AI companies to publish clear data-flow disclosures. These should identify all third-party endpoints receiving prompt-related data, the categories of data transmitted, the purpose of each transfer, retention periods, onward-use restrictions, and whether the recipient can combine the data with advertising IDs, social-media IDs, cookies, hashed emails, or other persistent identifiers.

Regulators should distinguish between product analytics and advertising surveillance. Basic security logging, abuse prevention, service diagnostics, and aggregate performance measurement may be legitimate. But marketing pixels, retargeting tools, lookalike-audience tools, conversion APIs, and identity-resolution systems should be presumptively prohibited inside AI prompt and response interfaces.

Regulators should require independent technical audits of AI data flows. Privacy policies are not enough. Auditors should inspect browser network traffic, mobile SDK traffic, server-side event pipelines, analytics dashboards, tag managers, conversion APIs, and data-sharing contracts. The audit question should be simple: can any third party learn the contents, meaning, subject matter, or sensitive inferences of a user’s prompt?

Regulators should require special protections for sensitive domains. AI products used for health, legal, education, employment, finance, children, research, and workplace productivity should face stricter rules. In these settings, prompt data should be treated more like confidential communications than ordinary web behaviour.

Regulators should prohibit dark-pattern consent for AI tracking. A cookie banner or buried privacy-policy clause should not be sufficient to authorize the disclosure of sensitive chatbot interactions to adtech partners. Consent should be specific, intelligible, revocable, and separate from general service access.

Regulators should require enterprise-grade contractual controls. Public-sector bodies, universities, publishers, law firms, healthcare providers, and corporations should be able to demand contractual warranties that no prompt data, metadata, or derived signals are shared with advertising or marketing partners.

Regulators should create incident-reporting obligations for prompt-data leakage. If an AI company discovers that prompts or prompt-derived information were accidentally transmitted to third-party analytics, advertising, or debugging tools, that should trigger notification duties similar to other privacy incidents.

Finally, regulators should stop treating AI privacy as only a model-training issue. The question is not merely whether prompts are used to train future models. The more immediate question may be whether prompts are being routed through a commercial surveillance infrastructure built for an older internet. This complaint is valuable because it exposes that hidden layer. Even if this particular lawsuit never produces a judgment, the regulatory lesson is clear: AI assistants must not become confession boxes wired to the advertising economy.