Privacy-Enhancing Technologies
The Open Loop Brazil program was launched in tandem with a twin policy prototyping program in Uruguay, with the aim of guiding and enabling companies in Brazil to leverage and apply privacy-enhancing technologies (PETs) to help deidentify data and mitigate privacy-related risks. In this initiative, nine organizations in Brazil engaged in testing a prototype PETs Playbook devised to help organizations connect data protection expectations with the selection of suitable PETs. The program was a collaborative effort between Meta and the Instituto Liberdade Digital of Brazil, in collaboration with the Brazilian Data Protection Authority – ANPD and the Executive Secretariat of the National AI Strategy – EBIA (at the Ministry of ICTs) as observers.
Deployment Period | September 2022 - April 2023
Read the report now!
This report presents the findings and recommendations of the Open Loop Brazil program. Through desk research, interviews, surveys and workshops, the policy prototyping program investigated:
Download the report:
PROGRAM DETAILS
Main Findings & Recommendations
The program’s outcomes resulted in several notable recommendations to guide and enable companies in Brazil to leverage and select privacy-enhancing technologies, including:
A flexible, risk-based approach to anonymization
Measuring the level of risk should be a fact-specific assessment and should focus on whether parties who might realistically get access to the data could re-identify the data.
Processing data
Policymakers should clarify that entities can process data for the purpose of reducing the risk of identifiability.
Advancing multi-stakeholder dialogues
Not only could these conversations help to build entities’ capacities to deploy PETs, but they could also make progress on developing a shared understanding of PETs.
Direct investment in PETs research and development
Policymakers could also fund R&D into open-source PETs implementations, which could be more readily used off-the-shelf by small and medium entities.
Regulatory sandboxes
Policymakers are encouraged to explore the above topics more thoroughly through regulatory sandboxes.
Partners & Observers
The program was a collaborative effort between Meta and the Instituto Liberdade Digital of Brazil, in collaboration with the Brazilian Data Protection Authority – ANPD and the Executive Secretariat of the National AI Strategy – EBIA (at the Ministry of ICTs) as observers.
Explore other programs
Competition in AI Foundation Models
Meta’s Open Loop program is excited to have launched its first policy prototyping program in the United Kingdom, which is focused on testing the Competition and Markets Authority (CMA) AI Principles to ensure that they are clear, implementable and effective at guiding the ongoing development and use of AI Foundation Models, while protecting competition and consumers.
Generative AI Risk Management
Meta’s Open Loop launched its first policy prototyping research program in the United States in late 2023, focused on testing the National Institute of Standards and Technology (NIST) AI Risk Management Framework (RMF) 1.0. This program gave participating companies the opportunity to learn about NIST's AI RMF and subsequent “Generative AI Profile” (NIST AI 600-1), and to understand how this guidance can be applied to developing and deploying generative AI systems. At the same time, the program gathered evidence on current practices and provided valuable insights and feedback to NIST, which can inform future iterations of the RMF and Gen AI profile.
Artificial Intelligence Act
The EU AI Act program is the largest policy prototyping initiative to date, engaging over 60 participants from more than 50 companies developing AI and ML products. The program was structured into three pillars, each focusing on key articles of the EU proposal and assessing and scrutinizing them.
Human-centric AI
The Open Loop India program was a collaborative effort between Meta, ArtEZ University of the Arts and The Dialogue, to develop a stakeholder engagement framework that operationalizes the principle of human-centered AI.
AI Impact Assessment
This program aimed to develop and test a risk assessment framework, called ADIA (Automated Decision Impact Assessment), for AI applications deployed in Europe.
GET INVOLVED
Do you have innovative ideas on how to govern emerging technologies?
Do you want to co-develop and test new policy ideas?
We want to hear from you!