On August 6, 2025, the Personal Information Protection Commission (the “PIPC”) released the Guideline on Processing of Personal Information in Developing and Using Generative AI (the “Guideline”) (available in Korean, Link).
The Guideline divides the lifecycle of developing and using generative AI into four stages: (i) purpose setting, (ii) strategy setting, (iii) AI training and development, and (iv) system application and management. It provides considerations for service providers developing and using generative AI to protect personal information at each stage. The Guideline also presents examples of investigations and administrative actions related to AI-based personal information processing, as well as examples of innovation support systems, to assist service providers in better understanding these issues.
The PIPC describes the Guideline as a compilation of experiences accumulated through preliminary inspections, regulatory sandboxes, and preliminary adequacy reviews, as well as insights from existing publications, such as the Guideline on Processing Publicly Available Data for AI Development and Services and the AI Privacy Risk Management Model.
Below is a summary of the four stages of developing and using generative AI, including the tasks and considerations for each stage as presented in the Guideline:
Stage |
Tasks and Considerations |
||||||
Purpose Setting |
|
||||||
Strategy Setting |
|
||||||
AI Training and Development |
|
||||||
System Application and Management |
|
Key Considerations in Each Stage of Developing and Using Generative AI
1. |
Stage 1: Setting the Purpose of Using Generative AI and Identifying a Lawful Basis for Processing Personal Information |
(1) |
Collecting and using publicly available data: The legitimate interest clause under Article 15, Paragraph (1), Item 6 of the Personal Information Protection Act (the “PIPA”) may serve as a lawful basis for the collection and use of personal information. To this end, it is essential to minimize the risk of infringing data subjects’ rights by establishing the legitimacy of the purpose, demonstrating the necessity of processing publicly available data, and implementing the technical, managerial safeguards, as well as measures to ensure data subjects’ rights.[1] |
(2) |
Training or developing AI by reusing data subjects’ personal data: As detailed in the table below, the lawful basis for processing such data varies depending on the relevance of the AI training and development to the original purpose of personal data collection, as well as the nature of the personal data. |
Case |
Details |
Where data processing is aimed at improving or enhancing services within the scope of the original purpose of collection |
|
Where data processing is reasonably relevant to the original purpose of collection |
|
Where data is used to develop new services apart from the original purpose of collection |
|
Where processing involves processing of sensitive information or unique identification information |
|
2. |
Stage 2: Establishing Strategies for Generative AI Development, Use and Risk Management |
Classification |
Issues and Considerations |
LLM-as-a-Service |
|
Ready-Made LLM |
|
Self-Development |
|
3. |
Stage 3: Safeguards at the Stage of Training and Development of Generative AI |
4. |
Stage 4: System Application and Management, Such As Monitoring Infringement of Data Subjects’ Rights |
Establishing AI Privacy Governance
The Guideline emphasizes the need for companies and organizations to establish and operate an internal management system led by the CPO, who oversees compliance with privacy laws and risk management, as management of risks associated with data processing by generative AI becomes increasingly important.
By establishing such governance, the CPO will be able to oversee and manage the entire process—from defining the objectives of generative AI to its implementation and management—ensuring the legality and security of personal information processing. The key processes presented in the Guideline, grounded in AI privacy governance, are as follows.
-
Continuous privacy risk assessment using assessment tools, such as privacy impact assessment
-
Multi-layered safeguards to mitigate privacy risks
-
Systematic documentation of privacy risk management policies
-
Monitoring, assessment, and reporting of vulnerabilities in personal information
-
Support for data subjects’ exercise of their rights
Implications of the Guideline and Prospects
The Guideline is significant in that the PIPC systematically compiled and presented key safeguards for generative AI service providers at each stage of generative AI development and use, based on the PIPC’s other guidelines and accumulated enforcement cases and policies related to AI services.
While primarily focused on language model-based generative AI, the Guideline is expected to gradually expand to multimodal and agentic AI, processing various types of information, such as voice, images, and video, making close monitoring of upcoming guidelines important.
The Guideline places special emphasis on establishing AI privacy governance to address legal risks associated with generative AI. Accordingly, companies using publicly available personal information or previously collected personal information for service development are advised to map and assess risks that may arise at each stage of development and use and to establish and implement risk management policies and measures under the leadership of the CPO.
[1] For more information regarding the processing of publicly available data, please refer to the Guideline on Processing Publicly Available Data for AI Development and Services, released by the PIPC on July 17, 2024, and our newsletter (Link).