EU AI Act Compliance Checklist for Developers
A practical checklist for developers building AI applications under the EU AI Act. Covers risk classification, transparency requirements, and implementation steps.
The EU AI Act is the world's first comprehensive legal framework for artificial intelligence. Signed into law and entered into force in August 2024, it establishes binding rules for anyone who develops, deploys, or distributes AI systems within the European Union — regardless of where the company is headquartered. If your AI application is accessible to EU users, the Act applies to you.
Enforcement follows a phased timeline through 2027, giving developers time to prepare. But the penalties for non-compliance are among the steepest in regulatory history: up to €35 million or 7% of global annual turnover, whichever is higher. For prohibited AI practices, fines scale even further. The message is clear: the EU expects AI developers to take compliance seriously from the start.
This article provides a practical, developer-focused checklist for EU AI Act compliance. It covers risk classification, transparency obligations, data governance, and concrete implementation steps — including how PII protection fits into the broader compliance picture.
Understanding Risk Classification
The EU AI Act organizes AI systems into four risk tiers. Your compliance obligations depend entirely on which tier your system falls into. Getting the classification right is the first and most important step.
Unacceptable Risk (Banned)
These AI practices are prohibited outright. They include social scoring systems used by governments, real-time biometric identification in public spaces (with narrow law enforcement exceptions), manipulation techniques that exploit vulnerabilities, and systems that infer emotions in workplaces or educational institutions. If your application falls here, it cannot be offered in the EU at all.
High-Risk
AI systems used in critical sectors carry the heaviest compliance burden. This includes AI for medical device diagnostics, credit scoring, recruitment and HR decisions, education admissions, law enforcement, and critical infrastructure management. High-risk systems must undergo conformity assessments, maintain detailed technical documentation, implement human oversight, and register in the EU database.
Limited Risk
Systems that interact directly with people or generate content fall here. Chatbots, AI-powered customer support, content generation tools, and emotion recognition systems (where permitted) all carry transparency obligations. Users must be informed they are interacting with AI, and AI-generated content must be labeled as such.
Minimal Risk
AI systems that pose little or no risk — spam filters, AI-powered search, recommendation engines, inventory optimization — face no specific obligations under the Act beyond voluntary codes of conduct. Most backend AI tooling falls here.
For most developers: If you are building chatbots, content generation tools, summarization pipelines, or analysis features powered by LLMs, your system likely falls into the Limited Risk category. This means transparency requirements apply, and data governance best practices are strongly recommended, but you do not need to undergo conformity assessments.
Transparency Requirements
Transparency is a cross-cutting obligation in the EU AI Act. The specific requirements vary by risk tier, but certain rules apply broadly to any AI system offered in the EU.
- AI interaction disclosure: Users must be clearly informed when they are interacting with an AI system. This applies to chatbots, virtual assistants, and any interface where a user might otherwise believe they are communicating with a human.
- AI-generated content labeling: Text, images, audio, and video produced by AI must be marked as machine-generated. This includes deepfakes and synthetic media, which must be labeled in a machine-readable format.
- Technical documentation: High-risk systems require detailed documentation covering the system's purpose, capabilities, limitations, training data characteristics, performance metrics, and known risks. Limited-risk systems benefit from maintaining similar documentation even where it is not strictly required.
- General-purpose AI model providers (such as companies offering foundation models) must publish model cards, document training data, and comply with EU copyright law.
Data Governance for AI Systems
Article 10 of the EU AI Act requires “appropriate data governance and management practices” for high-risk AI systems. While this obligation formally applies to the high-risk tier, the principles represent best practices that regulators expect across all risk levels — particularly when personal data is involved.
The Act's data governance requirements cover several areas that directly affect how developers handle user data in AI pipelines:
- Training data quality: Data used to train or fine-tune models must be relevant, representative, and free from errors to the extent possible. Datasets must be examined for biases that could lead to discriminatory outcomes.
- Bias detection and mitigation: Developers must actively test for and address biases in both training data and model outputs, especially for high-risk applications affecting people's rights or access to services.
- Privacy-by-design: The Act reinforces GDPR's data protection by design principle. AI systems must incorporate data protection measures from the architecture level, not as an afterthought. This includes minimizing the personal data processed, implementing technical safeguards, and ensuring data is handled in accordance with its intended purpose.
- Data flow documentation: Developers should maintain clear records of how data moves through their AI pipeline — from user input to model processing to output — including which third-party services receive data and what protections are in place.
PII handling sits at the intersection of all four requirements. When personal data flows into an AI model without protection, it creates bias risks (models may memorize personal data), quality issues (PII in training data degrades generalization), and privacy violations. A systematic approach to PII protection addresses multiple data governance requirements simultaneously.
PII Protection as a Compliance Building Block
The EU AI Act does not prescribe specific technical solutions for data governance, but it does require that developers implement “appropriate measures” to protect personal data in AI systems. Tokenization — detecting PII in text and replacing it with structured placeholder tokens before the text reaches any AI model — is one of the most effective ways to satisfy this requirement.
When you tokenize PII before AI processing, you achieve several compliance objectives at once:
- Privacy-by-design: PII is stripped from text at the architecture level, before any external service processes it. The AI model never sees real names, email addresses, or other identifiers.
- Data minimization: The model receives only the information it needs to generate a useful response — the structure and context of the text — without the personal identifiers.
- Data governance documentation: Tokenization operations generate audit logs that demonstrate systematic data handling, supporting your compliance records.
- GDPR alignment: Since the EU AI Act explicitly builds on GDPR obligations, tokenization satisfies requirements under both regulations simultaneously.
Here is how this looks in practice using Blindfold with the gdpr_eu policy and EU region processing:
from blindfold import Blindfold from openai import OpenAI # Initialize with EU region — PII processing stays in Europe bf = Blindfold(api_key="your-api-key", region="eu") openai_client = OpenAI() # Tokenize PII before sending to the AI model user_input = "My name is Sophie Laurent and my email is sophie.laurent@example.fr" protected = bf.tokenize(user_input, policy="gdpr_eu") # protected.text → "My name is <Person_1> and my email is <Email Address_1>" # The AI model never sees the real name or email response = openai_client.chat.completions.create( model="gpt-4", messages=[{"role": "user", "content": protected.text}] ) # Restore original data in the response for the end user final = bf.detokenize(response.choices[0].message.content, protected.mapping)
With three lines of integration — initialize, tokenize, detokenize — your AI pipeline gains a systematic data governance layer. The AI model works with tokens like <Person_1> and <Email Address_1> instead of real personal data, and every operation is logged for audit purposes.
The EU AI Act Compliance Checklist
Use this checklist as a practical roadmap for bringing your AI application into compliance. Items are ordered by priority — start with classification and work your way down.
- Classify your AI system's risk level. Determine whether your application is Minimal, Limited, High-Risk, or Unacceptable. This determines every other obligation. Review Annex III of the Act for the full list of high-risk categories.
- Document all data flows and processing activities. Map how user data enters your system, which services process it (including AI providers), what transformations occur, and where data is stored. This documentation is required for high-risk systems and strongly recommended for all tiers.
- Implement PII protection before AI processing. Tokenize personal data before it reaches any AI model. Use a policy-based approach (such as
policy="gdpr_eu") to ensure comprehensive detection of all relevant entity types. This satisfies privacy-by-design and data minimization requirements. - Add transparency labels for AI-generated content. Ensure users know when they are interacting with an AI system. Label AI-generated text, images, and media clearly. For chatbots, include a visible indicator that the conversation is AI-powered.
- Log all AI decisions and inputs. Maintain audit trails that record what data was processed, when, and what the system produced. Audit logs should capture entity types detected, policies applied, and processing regions — without storing the actual personal data.
- Conduct bias and fairness assessments. Test your AI system for discriminatory outcomes across demographic groups. Document the methodology, results, and any mitigation steps. This is mandatory for high-risk systems and best practice for all AI applications.
- Prepare technical documentation. Document your system's intended purpose, capabilities, known limitations, training data characteristics (if applicable), performance benchmarks, and risk mitigation measures. The Act specifies detailed documentation requirements for high-risk systems in Annex IV.
- Implement human oversight mechanisms. Design your system so that a human can intervene, override, or shut it down when needed. For high-risk systems, this must be built into the system design. For limited-risk systems, ensure operators can review and correct AI outputs.
- Conduct a conformity assessment (high-risk only). High-risk AI systems must undergo a conformity assessment before deployment. Depending on the category, this may be a self-assessment or require involvement of a notified body.
- Register high-risk systems in the EU database. High-risk AI systems must be registered in the EU-wide database before being placed on the market. The registration includes information about the system, its provider, and its intended use.
- Appoint an AI compliance officer or function. Designate a person or team responsible for overseeing AI Act compliance. This role coordinates documentation, risk assessments, incident response, and communication with regulators. For smaller teams, this can be combined with an existing data protection role.
- Establish monitoring and incident response procedures. Implement post-deployment monitoring to track your AI system's performance and detect issues. Define a clear process for reporting serious incidents to the relevant authority — mandatory for high-risk systems within defined timelines.
How Blindfold Helps
Several checklist items map directly to capabilities that Blindfold provides out of the box. Here is how the tool fits into your compliance architecture:
PII Protection Before AI Processing (Item 3)
Blindfold's tokenization API detects and replaces PII in text before it reaches any AI model. The gdpr_eu policy covers 15+ entity types relevant under GDPR and the AI Act — names, emails, phone numbers, IBANs, national IDs, health data, and more. This implements privacy-by-design at the architecture level.
Data Flow Documentation and Audit Logs (Items 2 and 5)
Every tokenization and detokenization operation generates an audit log entry. These logs record timestamps, entity types detected, the policy used, and the processing region — providing the documentation trail you need for compliance records and regulatory inquiries, without storing any actual personal data.
Systematic PII Handling via Policies (Item 3)
Instead of manually configuring which entity types to detect, Blindfold's pre-built policies ensure consistent, comprehensive PII handling across your entire application. The gdpr_eu policy is maintained and updated to reflect current regulatory requirements, reducing the risk of missed entity types.
Regional Processing for Data Residency
Setting region="eu" routes all PII processing to EU-based infrastructure. Personal data never leaves European borders during the tokenization step, satisfying data residency requirements under both the AI Act and GDPR. Combined with tokenization, this means no personal data reaches any non-EU service at any point in the pipeline.
from blindfold import Blindfold # EU region + GDPR policy = full data governance coverage bf = Blindfold( api_key="your-api-key", region="eu" # Routes to eu-api.blindfold.dev ) # Every call generates an audit log entry protected = bf.tokenize(text, policy="gdpr_eu") # Inspect detected entities for entity in protected.entities: print(f"{entity.type}: {entity.token}") # Person: <Person_1> # Email Address: <Email Address_1>
Timeline and Next Steps
The EU AI Act uses a phased enforcement schedule. Here are the key dates developers need to know:
- February 2025: Prohibitions on unacceptable-risk AI practices take effect. Systems involving social scoring, manipulative techniques, and most real-time biometric identification must be discontinued.
- August 2025: Obligations for general-purpose AI models apply. Providers of foundation models must publish model cards and comply with transparency requirements.
- August 2026: Full enforcement for high-risk AI systems. Conformity assessments, technical documentation, human oversight, and database registration all become mandatory.
- August 2027: Extended deadline for high-risk AI systems that are components of products already subject to EU product safety legislation.
What to Do Now
Even if full enforcement for your risk category is still months away, the preparation work takes time. Here is what to prioritize today:
- Classify your system. If you have not yet determined your risk tier, do it now. This unblocks every subsequent compliance decision.
- Implement PII tokenization. This is a quick, high-impact step that addresses data governance, privacy-by-design, and GDPR requirements simultaneously. It can be integrated in a single afternoon.
- Start documenting. Begin building your technical documentation and data flow maps. These take time to assemble and are much easier to create while the architecture decisions are fresh.
- Add transparency labels. If your application uses AI-generated content, add labels now. This is a straightforward UI change with no downside.
Bottom line: The EU AI Act adds a new layer of obligations on top of existing GDPR requirements. The good news for developers who already handle PII responsibly is that much of the groundwork carries over. By implementing tokenization, maintaining audit logs, and using EU region processing, you address the data governance requirements of both regulations with a single, consistent approach.
Try It Yourself
Clone a complete working example from our cookbook and run it in minutes:
- GDPR + OpenAI Python — EU region,
gdpr_eupolicy, batch processing - E2B Data Analyst — AI writes analysis code from tokenized data, sandbox executes on real data
- All cookbook examples — OpenAI, LangChain, FastAPI, Express, HIPAA, and more
Start protecting sensitive data
Free plan includes 500K characters/month. No credit card required.