CIPL Publishes Discussion Paper on Applying Data Protection Principles to Generative AI
5 Minute Read
December 16, 2024
In December 2024, the Centre for Information Policy Leadership (“CIPL”) at Hunton Andrews Kurth published a discussion paper titled, “Applying Data Protection Principles to Generative AI: Practical Approaches for Organizations and Regulators” (“Discussion Paper”).
The Discussion Paper considers the following privacy and data protection concepts and explores how they can be effectively applied to the development and deployment of generative AI (“genAI”) models and systems:
- fairness;
- collection limitation;
- purpose specification;
- use limitation;
- individual rights;
- transparency;
- organizational accountability; and
- cross-border data transfers.
CIPL presents the following recommendations, discussed in greater detail in the Discussion Paper, to organizations and regulators:
- To enable beneficial development and use of AI technologies in the modern information age, laws and regulatory guidance should facilitate lawful mechanisms for the use of personal data in model training. Lawmakers and regulators should avoid legal interpretations that are unduly restrictive regarding the use of personal data in AI model training, development and deployment.
- Different data privacy rules, considerations and mitigations apply in different phases of the AI lifecycle—data collection, model training, fine tuning and deployment. Regulators and organizations should interpret data protection principles separately in the context of each relevant phase of the AI technologies.
- Organizations should be able to rely on the “legitimate interests” legal basis for processing publicly available personal data collected through web scraping and personal data that they already have in their possession and control (first-party data) for genAI model training, as long as the interest concerned (which could be the controller’s, users’ or society’s at large) is not outweighed by the fundamental rights of individuals and appropriate, risk-based mitigation measures are put in place.
- Laws and regulatory guidance should be drafted or interpreted to recognize and enable the processing and retention of sensitive personal data for AI model training, as this is necessary to avoid algorithmic bias or discrimination and ensure content safety. In addition, sensitive personal data may be necessary for the training and development of certain AI systems whose sole purpose is based on the processing of sensitive personal data or to deliver benefits to protected categories of individuals (such as accessibility tools, or health systems).
- Developers should explore opportunities to employ privacy-enhancing and privacy-preserving technologies (“PETs/PPTs”), such as synthetic data and differential privacy. This would enable genAI models to have the rich datasets they need during training while reducing the risks associated with the use of personal data. Laws and regulatory guidance should encourage the use of and acknowledge the need for continued research and investment in PETs/PPTs.
- The fairness principle is useful in the genAI context and should be interpreted to facilitate personal data processing in genAI model development to train accurate and accessible models that do not unjustly discriminate. Considerations of fairness also need to take into account the impact on the individual or society of not developing a particular AI application.
- Data minimization should be understood contextually as limiting the collection and use of data that is necessary for the intended purpose (g., model training, model fine-tuning, or model deployment for a particular purpose). Data minimization should not stand in the way of enabling the collection and use of data that is necessary and appropriate for achieving a robust and high-quality genAI model. As such, this principle does not prohibit or conflict with the collection and use of large volumes of data.
- Training general-purpose AI models should be recognized as a legitimate and permissible purpose in itself, so long as appropriate accountability measures and safeguards are reasonably and sufficiently implemented.
- Purpose or use limitation principles should be sufficiently flexible. In the context of genAI, purpose limitation principles in laws and regulations should allow organizations to articulate data processing purposes that are sufficiently flexible for the range of potentially useful applications for which genAI models may be used. Furthermore, processing personal data for the development of a genAI model should be treated as a separate purpose from processing personal data for the development, deployment or improvement of a specific application that uses a genAI model.
- The responsibility to inform individuals about the use of their data should fall to the entity closest to the individual from whom the data is collected. Where data is not collected directly from individuals, organizations should be able to fulfil transparency requirements through public disclosures or other informational resources.
- Where appropriate and practicable, individuals should be able to request that their input prompts and model output responses not be included in genAI model fine-tuning, especially if such prompts include personal or sensitive data.
- Transparency in the context of genAI models should be contextually appropriate and meaningful, while also fulfilling transparency requirements under applicable laws and regulations. Transparency should not come at the expense of other important factors, such as usability, functionality, and security, or create additional burdens for users. Organizations should also consider transparency in the wider sense, beyond individuals and users—to regulators, auditors and red-team experts.
- Lawmakers and regulators should consult with developers and deployers of genAI systems to clarify the distinctions in duties and responsibilities across the phases of genAI development.
- Organizations developing and deploying genAI models and systems must invest in comprehensive and risk-based AI and data privacy programs, continually improving and evolving their controls and best practices. Lawmakers and regulators should encourage and reward organizational accountability in development and deployment of AI, including the existence and demonstration of AI and data privacy management programs.
Tags: Accountability, Artificial Intelligence, Automated Decisionmaking, Big Data, Fundamental Rights, Personal Information, Privacy By Design, Privacy Rights, Personal Data
Search
Recent Posts
- CIPL Publishes Discussion Paper on Applying Data Protection Principles to Generative AI
- GoodRx Agrees to Pay $25 Million Settlement for Privacy Violations
- D.C. Circuit Upholds Protecting Americans’ Data from Foreign Adversary Controlled Applications Act as TikTok Ban Dispute Edges Closer to Supreme Court
Categories
- Behavioral Advertising
- Centre for Information Policy Leadership
- Children’s Privacy
- Cyber Insurance
- Cybersecurity
- Enforcement
- European Union
- Events
- FCRA
- Financial Privacy
- General
- Health Privacy
- Identity Theft
- Information Security
- International
- Marketing
- Multimedia Resources
- Online Privacy
- Security Breach
- U.S. Federal Law
- U.S. State Law
- Workplace Privacy
Tags
- Aaron Simpson
- Accountability
- Adequacy
- Advertisement
- Advertising
- American Privacy Rights Act
- Anna Pateraki
- Anonymization
- Anti-terrorism
- APEC
- Apple Inc.
- Argentina
- Arkansas
- Article 29 Working Party
- Artificial Intelligence
- Australia
- Austria
- Automated Decisionmaking
- Baltimore
- Bankruptcy
- Belgium
- Biden Administration
- Big Data
- Binding Corporate Rules
- Biometric Data
- Blockchain
- Bojana Bellamy
- Brazil
- Brexit
- British Columbia
- Brittany Bacon
- Brussels
- Business Associate Agreement
- BYOD
- California
- CAN-SPAM
- Canada
- Cayman Islands
- CCPA
- CCTV
- Chile
- China
- Chinese Taipei
- Christopher Graham
- CIPA
- Class Action
- Clinical Trial
- Cloud
- Cloud Computing
- CNIL
- Colombia
- Colorado
- Committee on Foreign Investment in the United States
- Commodity Futures Trading Commission
- Compliance
- Computer Fraud and Abuse Act
- Congress
- Connecticut
- Consent
- Consent Order
- Consumer Protection
- Cookies
- COPPA
- Coronavirus/COVID-19
- Council of Europe
- Council of the European Union
- Court of Justice of the European Union
- CPPA
- CPRA
- Credit Monitoring
- Credit Report
- Criminal Law
- Critical Infrastructure
- Croatia
- Cross-Border Data Flow
- Cyber Attack
- Cybersecurity
- Cybersecurity and Infrastructure Security Agency
- Data Brokers
- Data Controller
- Data Localization
- Data Privacy Framework
- Data Processor
- Data Protection Act
- Data Protection Authority
- Data Protection Impact Assessment
- Data Transfer
- David Dumont
- David Vladeck
- Delaware
- Denmark
- Department of Commerce
- Department of Health and Human Services
- Department of Homeland Security
- Department of Justice
- Department of the Treasury
- District of Columbia
- Do Not Call
- Do Not Track
- Dobbs
- Dodd-Frank Act
- DPIA
- E-Privacy
- E-Privacy Directive
- Ecuador
- Ed Tech
- Edith Ramirez
- Electronic Communications Privacy Act
- Electronic Privacy Information Center
- Elizabeth Denham
- Employee Monitoring
- Encryption
- ENISA
- EU Data Protection Directive
- EU Member States
- European Commission
- European Data Protection Board
- European Data Protection Supervisor
- European Parliament
- Facial Recognition Technology
- FACTA
- Fair Credit Reporting Act
- Fair Information Practice Principles
- Federal Aviation Administration
- Federal Bureau of Investigation
- Federal Communications Commission
- Federal Data Protection Act
- Federal Trade Commission
- FERC
- FinTech
- Florida
- Food and Drug Administration
- Foreign Intelligence Surveillance Act
- France
- Franchise
- Fred Cate
- Freedom of Information Act
- Freedom of Speech
- Fundamental Rights
- GDPR
- Geofencing
- Geolocation
- Georgia
- Germany
- Global Privacy Assembly
- Global Privacy Enforcement Network
- Gramm Leach Bliley Act
- Hacker
- Hawaii
- Health Data
- Health Information
- HIPAA
- HIPPA
- HITECH Act
- Hong Kong
- House of Representatives
- Hungary
- Illinois
- India
- Indiana
- Indonesia
- Information Commissioners Office
- Information Sharing
- Insurance Provider
- Internal Revenue Service
- International Association of Privacy Professionals
- International Commissioners Office
- Internet
- Internet of Things
- IP Address
- Ireland
- Israel
- Italy
- Jacob Kohnstamm
- Japan
- Jason Beach
- Jay Rockefeller
- Jenna Rode
- Jennifer Stoddart
- Jersey
- Jessica Rich
- John Delionado
- John Edwards
- Kentucky
- Korea
- Latin America
- Laura Leonard
- Law Enforcement
- Lawrence Strickling
- Legislation
- Liability
- Lisa Sotto
- Litigation
- Location-Based Services
- London
- Madrid Resolution
- Maine
- Malaysia
- Markus Heyder
- Maryland
- Massachusetts
- Meta
- Mexico
- Microsoft
- Minnesota
- Mobile App
- Mobile Device
- Montana
- Morocco
- MySpace
- Natascha Gerlach
- National Institute of Standards and Technology
- National Labor Relations Board
- National Science and Technology Council
- National Security
- National Security Agency
- National Telecommunications and Information Administration
- Nebraska
- NEDPA
- Netherlands
- Nevada
- New Hampshire
- New Jersey
- New Mexico
- New York
- New Zealand
- Nigeria
- Ninth Circuit
- North Carolina
- Norway
- Obama Administration
- OECD
- Office for Civil Rights
- Office of Foreign Assets Control
- Ohio
- Oklahoma
- Opt-In Consent
- Oregon
- Outsourcing
- Pakistan
- Parental Consent
- Payment Card
- PCI DSS
- Penalty
- Pennsylvania
- Personal Data
- Personal Health Information
- Personal Information
- Personally Identifiable Information
- Peru
- Philippines
- Phyllis Marcus
- Poland
- PRISM
- Privacy By Design
- Privacy Policy
- Privacy Rights
- Privacy Rule
- Privacy Shield
- Protected Health Information
- Ransomware
- Record Retention
- Red Flags Rule
- Regulation
- Rhode Island
- Richard Thomas
- Right to Be Forgotten
- Right to Privacy
- Risk-Based Approach
- Rosemary Jay
- Russia
- Safe Harbor
- Sanctions
- Schrems
- Scott H. Kimpel
- Scott Kimpel
- Securities and Exchange Commission
- Security Rule
- Senate
- Serbia
- Service Provider
- Singapore
- Smart Grid
- Smart Metering
- Social Media
- Social Security Number
- South Africa
- South Carolina
- South Dakota
- South Korea
- Spain
- Spyware
- Standard Contractual Clauses
- State Attorneys General
- Steven Haas
- Stick With Security Series
- Stored Communications Act
- Student Data
- Supreme Court
- Surveillance
- Sweden
- Switzerland
- Taiwan
- Targeted Advertising
- Telecommunications
- Telemarketing
- Telephone Consumer Protection Act
- Tennessee
- Terry McAuliffe
- Texas
- Text Message
- Thailand
- Transparency
- Transportation Security Administration
- Trump Administration
- United Arab Emirates
- United Kingdom
- United States
- Unmanned Aircraft Systems
- Uruguay
- Utah
- Vermont
- Video Privacy Protection Act
- Video Surveillance
- Virginia
- Viviane Reding
- Washington
- Whistleblowing
- Wireless Network
- Wiretap
- ZIP Code