• Pascal's Chatbot Q&As
  • Posts
  • GPT-4o: Schrems questions Meta's reliance on legitimate interest to justify scraping social media data for AI training. He argues that this is fundamentally flawed.

GPT-4o: Schrems questions Meta's reliance on legitimate interest to justify scraping social media data for AI training. He argues that this is fundamentally flawed.

Schrems: AI makers should focus on finding technical solutions, such as removing personal data, using synthetic or pseudonymized data, and implementing filters.

Recording available here.

Question 1 of 3 for ChatGPT-4o: Please read the transcript of the broadcast "Live session with Max Schrems: Privacy Rights in the Age of AI" with Luiza Jarovsky and Max Schrems, and tell me what it says.

ChatGPT-4o’s answer:
The transcript of the broadcast "Live session with Max Schrems: Privacy Rights in the Age of AI" with Luiza Jarovsky and Max Schrems covers a variety of topics related to AI, data privacy, and regulation. Here are the key points:

  1. Introduction to Max Schrems and NOYB (None of Your Business):

    • Max Schrems, a well-known privacy advocate, discusses his work at NOYB, a nonprofit organization that enforces GDPR. NOYB has filed hundreds of cases, focusing on companies that violate privacy laws intentionally, particularly for profit​.

  2. The Role of AI in Privacy Violations:

    • The conversation transitions into how AI, particularly large language models (LLMs), is trained using personal data. Schrems explains that AI companies, like Twitter/X, Meta, and others, often collect and process user data, raising serious concerns about privacy rights under the GDPR​.

  3. Twitter/X Case:

    • Twitter's AI system was using EU users' personal data for AI training without their consent, which led to NOYB filing a complaint. Schrems highlights issues with Twitter’s lack of transparency, stating that Twitter’s privacy compliance team was essentially dismantled under Elon Musk, leading to potential GDPR violations​.

  4. Legitimate Interest and GDPR:

    • Schrems discusses how companies often try to justify the use of personal data for AI training under "legitimate interest," a legal basis under GDPR. He is skeptical that this argument will hold, as AI companies often bypass user consent, which he believes is necessary to process personal data​.

  5. Meta Case:

    • Schrems also touches on NOYB's cases against Meta, which similarly involve the use of personal data for AI training. Meta initially claimed it could process EU users’ data for this purpose, but after legal challenges, they committed to stopping the processing of such data​.

  6. AI Act and Privacy:

    • The discussion also moves toward the upcoming EU AI Act, which will regulate the use of AI in the EU. Schrems emphasizes that AI companies will need to find ways to train AI models in compliance with privacy laws, such as using anonymized or pseudonymized data​.

  7. Brazilian Data Protection Case:

    • The broadcast includes a comparison with Brazil, where Meta was allowed to use Brazilian users' data under specific conditions. This raised concerns about how different countries enforce privacy laws, often requiring only minimal transparency measures​.

  8. GDPR Adequacy:

    • Schrems does not believe the GDPR needs fundamental changes to deal with AI but calls for clearer guidance from authorities. He advocates for stronger enforcement of the existing law and a more balanced approach between technological innovation and privacy protection​.

In conclusion, the broadcast reflects the growing concerns about how AI systems are trained using personal data and the legal challenges that come with balancing innovation and privacy rights in the AI era.


Question 2 of 3 for ChatGPT-4o: List the most controversial, surprising and valuable statements made during this conversation.

ChatGPT-4o’s answer:
Here are some of the most controversial, surprising, and valuable statements from the conversation between Max Schrems and Luiza Jarovsky:

Controversial Statements:

  1. Twitter/X's Privacy Violations:

    • Schrems criticizes Twitter for moving forward with data processing even while negotiating with the Irish Data Protection Commission (DPC). He suggests Twitter, under Elon Musk, has dismantled its privacy team, making it difficult for the company to comply with GDPR. Schrems goes so far as to say Twitter "got away with murder," as they faced no real penalties for using EU personal data to train AI​.

  2. Meta’s Use of Legitimate Interest for AI Training:

    • Schrems questions Meta's reliance on legitimate interest to justify scraping social media data for AI training. He argues that this is fundamentally flawed, as GDPR generally doesn't allow for such broad data collection without consent​.

  3. EU AI Legislation and Company Manipulation:

    • Schrems highlights how companies, including Meta, manipulate public opinion by suggesting that without broad data access, Europeans would be denied access to the latest AI technologies. He calls this "collective punishment" and criticizes the narrative that privacy laws hinder innovation​.

Surprising Statements:

  1. Ireland’s GDPR Enforcement Lapses:

    • Schrems reveals that Ireland's Data Protection Commission (DPC), which oversees many tech companies, initially approved Meta's AI data processing practices, only to reverse course later due to pressure from other EU regulators. This reflects a pattern where Ireland first sides with tech giants, only to backtrack later​.

  2. Brazil’s Minimal Privacy Requirements:

    • Schrems contrasts the EU's stringent privacy laws with Brazil’s data protection practices, where Meta was allowed to use Brazilian users' data for AI training as long as they met basic transparency requirements, like sending notifications and adding opt-out options. This lenient approach was seen as surprisingly inadequate compared to EU standards​.

  3. AI Compliance Workarounds:

    • Schrems mentions that AI companies like OpenAI have claimed they are unable to comply with GDPR’s requirements for data deletion, rectification, or access. He points out that this undermines the core principles of GDPR, making the use of legitimate interest for AI training questionable​.

Valuable Statements:

  1. Transparency in AI Training:

    • Schrems explains that companies should be transparent about how personal data is used in AI models. He advocates for a system where users are clearly informed and can opt in or out, rather than hiding behind complex legal justifications​.

  2. Technical Solutions for AI Compliance:

    • Schrems encourages AI developers to invest in technical solutions, such as using anonymized or pseudonymized data, to ensure compliance with GDPR. He insists that data minimization should be central to AI training processes, pushing back against companies that claim this isn’t possible​.

  3. The Role of Regulators in AI Oversight:

    • Schrems emphasizes the importance of strong, clear decisions from data protection authorities, particularly regarding AI. He calls for more legal certainty, which would help companies navigate compliance without undermining privacy rights​.

These statements collectively reflect a critical perspective on how tech companies are using personal data for AI, the varying enforcement of privacy laws across jurisdictions, and the need for more technical and legal measures to protect user privacy.


Question 3 of 3 for ChatGPT-4o: List all lessons that AI Makers should be taking seriously and should be acting on, on the basis of what has been said in this conversation.

ChatGPT-4o’s answer:
Here are the key lessons AI makers should take seriously, based on the conversation between Max Schrems and Luiza Jarovsky:

1. Respect Data Minimization and Anonymization Principles

  • AI makers should ensure that personal data is either anonymized, pseudonymized, or not used at all in training datasets. Schrems stresses that companies should prioritize data minimization and avoid collecting unnecessary personal information, especially from sources like social media​.

  • Relying on legitimate interest as a justification for using personal data in AI training is legally questionable. Schrems advocates for asking users to explicitly opt in rather than hiding behind vague claims of legitimate interest. AI makers should seek user consent whenever possible to ensure compliance with privacy laws​.

3. Invest in Transparency and Simplicity

  • Schrems points out that companies often bury privacy policies and opt-out mechanisms under layers of complexity. AI makers should make privacy notices, opt-in/opt-out options, and data processing details easy for users to understand and access, ensuring full transparency​.

4. Enable Data Subject Rights

  • AI systems must be built with mechanisms that respect data subjects' rights, including access, correction, deletion, and rectification of their personal data. Schrems highlights that many AI companies currently claim they cannot meet these basic GDPR requirements, which undermines compliance. AI makers should develop technical solutions to enable these rights​.

5. Avoid Using Social Media Data Without Proper Justification

  • Schrems mentions that scraping personal data from social media platforms, especially without consent, is a violation of GDPR principles. AI makers should refrain from using social media data unless they have clear legal grounds and user consent​.

6. Prepare for Regulatory Scrutiny

  • AI makers should expect regulators to closely scrutinize their practices, especially as AI continues to evolve. Schrems encourages AI companies to anticipate legal challenges and regulatory actions, suggesting that compliance should be built into AI systems from the start​.

  • Schrems suggests that AI makers should focus on finding technical solutions, such as removing personal data, using synthetic or pseudonymized data, and implementing filters to prevent inaccurate or harmful information from being generated by AI models. This proactive approach can help mitigate legal risks​.

8. Be Cautious About Legitimate Interest Claims

  • Schrems repeatedly questions the legitimacy of using personal data for AI training under the "legitimate interest" argument, particularly when users cannot exercise their rights. AI makers should be cautious about relying on this legal basis and instead aim for user-centric approaches​.

9. Embrace Responsibility for Accuracy

  • AI systems often produce inaccurate information about individuals, sometimes with serious consequences. AI makers must take responsibility for ensuring the accuracy of the data their models generate, incorporating mechanisms for rectification and correction​.

  • Schrems highlights the evolving nature of AI regulation and enforcement across different jurisdictions. AI makers should stay updated on regulatory developments, particularly in regions like the EU, and be prepared for legal uncertainty. This requires a flexible approach to compliance and readiness to adapt to new legal interpretations​.

11. Provide Users with Easy Opt-Out Mechanisms

  • AI makers should offer clear, easy-to-access opt-out options for users who do not wish to have their data processed for AI training. Complex opt-out processes are not sufficient and could lead to legal challenges, as mentioned by Schrems in the case of Meta​.

By acting on these lessons, AI makers can ensure better compliance with privacy regulations, build trust with users, and avoid potential legal pitfalls.