EU AI Act: a practical field guide for 2025
7/13/2025
Leaders now need a practical way to map their AI systems to the EU AI Act and make steady progress without slowing core work. This guide focuses on classification, prioritization, and execution for 2025, reflecting the EU’s goal of setting a global standard for trustworthy AI that protects fundamental rights and safety.
What the law does in brief
The EU AI Act regulates AI based on risk level. It creates broad categories:
- Prohibited practices: A small set of AI uses that are outright banned (e.g. social scoring, real-time biometric identification in public spaces, certain manipulative techniques). These are “no-go” zones - if your AI project falls here, you must stop or radically change it.
- High-risk systems: AI systems in sensitive areas (like those affecting safety, livelihoods, or rights - for example, CV-scanning tools for hiring, credit scoring, medical devices with AI) are allowed but face strict obligations (on data governance, documentation, human oversight, etc.). These will require the most work to comply.
- Limited-risk (transparency-required) systems: This includes things like chatbots or deepfake generators where the Act might require you to inform users that they’re interacting with AI or label AI-generated content. These obligations are lighter (mostly notification/transparency).
- Minimal or low-risk systems: For most AI (say an AI-enabled spreadsheet autofill or a shopping website recommendation engine), the Act imposes very little - basically encouraging voluntary codes but no heavy requirements.
In sum, the law draws lines based on risk: unacceptable = banned, high-risk = heavily controlled, general-purpose or low-risk = monitoring and best practices, but not much red tape.
Start with an inventory
You can’t comply if you don’t know what you have. Create a living inventory of all AI systems or AI components in your organization, both customer-facing products and internal tools: For each, note the purpose (what is it used for and criticality), inputs (data it takes), outputs (decisions or predictions it makes), human involvement (is there a human reviewing outputs or is it fully automated?), origin of model (did you build it? fine-tune an existing model? using a vendor API?), and user base (who is affected – employees, customers, the public?). Assign an owner for each system – someone responsible for its compliance and performance. Even if it’s a small feature, give it a name on your list. Make this inventory cross-functional: involve IT, the data science/AI team, and business unit heads. Often they’ll surface “oh, I didn’t think that little script was AI, but it uses machine learning.” Exactly the point – cast a wide net. Include any usage of third-party AI services as well (e.g., if you use an AI SaaS for, say, resume screening). Because AI can pop up anywhere (marketing using an AI image generator, ops using an optimization algorithm, etc.), consider a quick survey or attestation process: ask teams “Do you use any AI or automated decision tools? If yes, list them.” It helps flush out shadow AI projects. This inventory is your foundation.
Classify by risk and impact
Now map each inventoried system into a risk category as defined by the Act, and also consider your own impact lens:
- High-risk candidates (per the Act): These generally include AI used in areas like safety-critical systems (transport, medical devices), employment and HR (like hiring algorithms), essential private/public services (credit scoring, housing applications), law enforcement and migration control, and any biometric identification or categorization of people. Mark systems that squarely or even possibly fall in these buckets. For example, an AI used in a recruitment tool or an AI that recommends loan approvals are high-risk by law. If unsure, lean towards caution; you can consult legal on borderline cases.
- Medium-risk / important (internal lens): Some AI might not be “high-risk” by the law’s definition but are still important for you to handle carefully. These could be decision support tools where a human is in the loop for sensitive areas. For example, an AI that suggests medical diagnoses to doctors (doctor still decides) might be medium risk: not fully automated, but critical outcomes. Or a supply chain AI that doesn’t affect people’s rights directly, but if it fails could cost a lot of money or reputation. Label these as medium significance.
- Lower-risk: AI systems that are assistive or don’t have major negative consequences if they go wrong. E.g., a chatbot that helps with FAQs, a recommendation engine on a shopping site (the stakes are low - maybe a bad recommendation, but not discrimination or injury), or internal analytics that managers use among other inputs. These likely fall under minimal or no regulatory requirements beyond general transparency in some cases.
Document your rationale briefly for each classification, especially for edge cases. If in doubt, flag it for legal review. Remember, classification is not purely a legal question - it’s also a management decision about how much risk you’re willing to tolerate in that AI’s performance. Better to slightly over-classify and then justify downgrading, than to be caught with a supposedly “low-risk” system that an auditor or incident proves was high-risk.
Core obligations for high-risk systems
For each system you deem high-risk, the Act imposes a set of obligations. It’s helpful to build a checklist or template that you’ll fill out per system:
- Data governance: Document the datasets used for training and testing. Where did the data come from? Did you assess its relevance, representativeness, any biases? The Act will require that you use training data that is appropriate, and that you’ve taken steps to ensure “bias monitoring” for protected attributes. Maintain documentation on data collection, processing, and any cleaning or augmentation done. Note any known data limitations. Perform and record bias or accuracy tests especially relevant to how the AI is used (e.g., error rates across different demographic groups, if applicable).
- Technical documentation: Write down, in a concise form, the essential information about the system: its intended purpose, an overview of its logic (you might not have to share the algorithm publicly, but you should be able to explain it at a high level), key performance metrics (accuracy, precision/recall, etc.), and known limitations or failure modes. Essentially, imagine an external auditor asking “Explain how this AI works and how you know it’s effective and safe.” You need a document that answers that. This doesn’t have to be extremely detailed initially, but cover the bases. Some companies adapt a model card or datasheet concept here, which is great.
- Logging and traceability: Ensure the system has the ability to log its operations - inputs, outputs, decisions made, any manual overrides, and when it was running which version. The Act will require keeping logs for high-risk AI to later investigate issues. Work with your IT team to set up log storage that is secure and meets retention requirements (the Act might say keep logs for X years, and consider other laws like GDPR on data retention). Check privacy implications - if the logs contain personal data, you need to protect and possibly anonymize them while still being useful.
- Human oversight: Define the points at which humans can intervene or override. For each high-risk AI, explicitly state: who (by role) is monitoring it in real time or periodically? Under what conditions are they expected to step in and turn it off or change its output? Do users have an easy way to appeal or contest an AI-driven decision? The Act likely requires that high-risk systems are designed so they can be effectively overseen by people. This might involve adding a simple “fail-safe” - e.g., a credit scoring AI that flags borderline cases for manual review, or an emergency off-switch for an autonomous machine. Document your oversight mechanism and provide training to those humans on their authority and how to exercise it.
- Accuracy and robustness: Set target performance metrics and tolerances for the AI and have a process for testing them periodically. For example, “This facial recognition system should have at least 98% identification accuracy in controlled conditions and no more than 1 in 10,000 false matches. We will test it on a fresh dataset from real-world conditions every 6 months.” Also test for cybersecurity - how does the AI handle malicious inputs or attempts to game it? Record the results of these tests. If any test fails or shows drift (performance degrading over time), have a plan to retrain or adjust. Robustness also means considering worst-case scenarios: what happens if the AI output is wrong - do you have mitigations so it doesn’t lead straight to disaster?
- Post-market monitoring: Once the AI is deployed, the obligations don’t stop. Set up a way to capture incidents, complaints, or near-misses. For instance, if users can report “the AI suggested something incorrect or harmful,” make sure those reports are logged and reviewed. Maintain a log of “errors/ incidents” and how they were addressed. Also monitor for “drift” - changes in data or context that could make the model less valid. A feedback loop should exist: if something goes wrong or conditions change, you update the system (either retrain the model, adjust thresholds, or even shut it down until fixed).
Add two cross-cutting elements across all high-risk compliance efforts:
- Quality management system: The Act essentially expects you to have a quality management system (QMS) in place for AI processes. This means having version control for models and data, internal approvals before deployment, and assigned roles for maintaining documentation and approvals. If a model is updated, you should have a mini “change management” process (who signs off? did you re-run tests?). You don’t have to be ISO 9000 certified necessarily, but adopt some structure: for example, require a compliance or technical lead to sign off that all these checklist items are done before a new version goes live.
- Incident response plan: Decide what counts as an incident worth escalating (e.g., a significant error that could harm someone, any data breach involving the AI, etc.), and have a plan: who gets informed (internal and potentially regulators, since the Act may require reporting serious incidents), how you investigate, and how quickly you’ll respond. This overlaps with human oversight but is more formal - an “escalation policy” for AI issues.
Completing this checklist for each high-risk system will likely be the bulk of your compliance heavy lifting. It’s resource-intensive, so prioritize: which systems need to be brought up to code first (perhaps those currently in use in the EU market, or those most likely to face scrutiny)?
General-purpose and foundation models
A special category in the Act (as of drafts) concerns general-purpose AI (GPAI) and foundation models. This is tricky because if you build or integrate large models like GPT-style, the rules are evolving. If you build or fine-tune large models:
- Prepare transparency documentation for them. This might include summaries of capabilities, known risks (e.g., it sometimes produces biased language or hallucinations), and recommended use and misuse info. Essentially, think of it like a datasheet or “model card” that you might have to publish or share with clients.
- Extra due diligence: test these models in as many scenarios as you can imagine that relate to fundamental rights or safety. For instance, could it produce hate speech? Does it leak personal data from training? The Act might require risk management here, even if you’re not deploying it directly but offering it as a service.
If you are a downstream deployer of a foundation model (i.e., you use someone else’s large model via API or open-source):
- Collect the provider’s documentation. The big players (OpenAI, Google, etc.) will likely publish info to comply - get that and keep it in your records.
- Run fit-for-purpose tests on the model for your specific use case. Don’t assume a general model works well on your specific data or context. For example, if you’re using GPT-xx to answer medical questions in your app, test a bunch of medical queries relevant to EU patients and see how it does. Check for errors or problematic responses. Essentially, you have to validate that the model, as integrated into your product, meets the quality and risk standards of the Act. If it doesn’t, you might need to put additional guardrails or fine-tune it with safety instructions.
Think of it this way: the Act doesn’t want companies shrugging “well, it’s a general AI, who knows what it’ll do.” You’re expected to know, as far as reasonable, what it can do in your context and to manage that. Some ideas for testing general models:
- Red team exercises: Come up with worst-case or tricky prompts that a bad actor or naive user might input in your context. See if the model gives dangerous or false outputs. E.g., prompt it for illegal advice or biased content. Record results and adjust either the model or usage policies accordingly.
- Hallucination checks: Have a set of questions with known answers (a ground truth set) and see how often the model is correct versus making stuff up. Define acceptable error rates and monitor them.
- Privacy and safety: Test if the model retains sensitive info. For instance, if you enter a sensitive personal detail, will it leak that detail if asked differently later? Or test if it refuses properly when it should (like giving medical diagnosis or personal data).
Document all these tests as part of your compliance documentation for that AI system.
Vendor and tool diligence
Most likely, you’re not doing everything in-house. You rely on vendors for AI components or tools (from cloud AI services to off-the-shelf AI software). The Act doesn’t let you off the hook for compliance just because you bought something - you’ll have to ensure your vendors meet requirements or that you compensate for any gaps. Standardize an AI vendor questionnaire to send out (or include in procurement):
- Ask about their training data (origin, was it lawfully obtained, any known biases).
- What evaluation have they done? Can they share results or summaries? (If they haven’t done any, that’s a flag.)
- What safety measures or risk mitigations do they have (like did they include bias mitigation, or is there a monitoring API)?
- How often do they update models and how will they inform you of changes or issues?
- Do they have an incident response if their model is found doing something harmful - will they help you fix it?
Also clarify contractual points:
- Notification of major changes: Put in contracts that the vendor must inform you if they significantly change the model or discover a flaw. Nothing worse than quietly getting a model update that behaves differently and you not knowing.
- Documentation access: You should have access to all documentation needed to prove compliance - so contract for that. If a regulator asks you about algorithmic decisions, you want to have what you need from the vendor readily.
- Liability split: Try to get clarity that if their model fails or has unfixable compliance issues, they will cooperate and possibly indemnify you for certain issues. Some vendors may resist, but even a clear support obligation is valuable.
- Incident handling: If something goes wrong (say, the AI triggers a scandal), require that the vendor assist with investigations (like giving log data, technical explanations) within a short time frame.
Mapping vendor provided capabilities versus what the Act needs is a good exercise. For example, if the Act says “users must be informed when interacting with an AI system,” and you’re using a vendor’s chatbot, ensure it has a feature to display “I am an AI” message - if not, you’ll have to implement that. In summary, treat AI suppliers like you would critical component suppliers in other regulated industries - demand transparency, quality controls, and shared responsibility.
Documentation that scales
The nightmare scenario is you create binderfuls of paperwork that immediately go out of date and no one reads. To avoid that, make documentation lightweight and maintainable:
- One-pagers per system: For each AI system, aim to maintain a one-page (okay, if needed up to two-page) summary that hits the key points: what it is, why it’s high/medium/low risk, current version, date of last compliance review, key metrics, who’s responsible, and where to find more detailed docs if needed. Think of it like an index card for the system. This is easier to update than a 100-page manual.
- Link out to detail: Store detailed artifacts separately - e.g., a data management spreadsheet, the full technical doc, test results - perhaps in a SharePoint or Confluence space for AI compliance. In your summary docs, just link to these. That way, when something changes (like you retrain on a new dataset), you update the dataset record and maybe a line in the summary, not re-write a massive PDF.
- Version control: Keep track of documentation versions and model versions. One approach is to assign each major AI system an ID or code, and then any docs/files related carry that and a date/version in the filename. For example, “HR-RecruiterAI_v1.2_modelcard_2025-06-01.docx”. It sounds pedantic, but when you’re juggling multiple systems, it helps avoid confusion.
- Regular updates: Set a review schedule (quarterly is a good starting point) to update each document. More frequently if the system changes often. The Act will likely expect you to keep docs up to date with any “material change.” A lightweight regime could be: every quarter, the product owner or AI owner spends 30 minutes reviewing the one-pager and underlying logs to refresh anything necessary.
Avoid gigantic binders that no one reads. Instead, strive for concise, accessible documentation with pointers to evidence. Regulators (and your own execs) will appreciate brevity with substance over volume with fluff.
Communicate internally and externally
Don’t keep all this compliance work in a silo. You need buy-in and understanding from your leadership, and in some cases reassurance to customers or regulators:
- Executive briefings: Periodically brief the C-suite or relevant leaders on your AI compliance status - not just as a risk thing but as a value-add. For example: “We have 5 high-risk AI systems, all of which now have risk controls in place and documentation per the EU AI Act. Two moderate issues were identified and we’ve mitigated them. We’re on track for compliance by [date].” This builds confidence internally that you’re handling it, and if trade-offs are needed (like delaying a feature to address compliance), leadership will understand why. Consider providing an “AI compliance dashboard” with simple status lights for each major system.
- Clear internal policy: Communicate a simple internal policy or guidelines for anyone developing or deploying AI. For example: “Any new use of AI must be cleared with the AI compliance team and added to the inventory. High-risk uses are subject to approval by [CTO or AI Governance Board]. All AI systems must allow human override and must log their decisions.” If you haven’t already, setting an internal AI governance committee or point person is useful. The key is to create awareness: people on different teams should know there’s a process and point of contact about AI.
- External communication: Be ready to articulate your approach to AI governance to customers, partners, or regulators. This could be a section on your website or an entry in your privacy/GDPR notices that now covers AI. Something like: “We implement the EU AI Act principles by ensuring transparency and oversight of our AI systems. For high-impact AI, we maintain documentation and human review. We are committed to trustworthy AI.” Keep it factual, not too boastful, and of course accurate. If customers or the public are affected by a high-risk AI (say, an automated loan decision tool), the law likely requires you to provide certain information to them. Be specific in those contexts: e.g., “Our loan decisions involve an algorithmic assessment. You have the right to an explanation of the decision and to contest it. We ensure a human reviews all denials.” Precision and openness here will earn trust.
- Regulator engagement: If you’re in a sector with active regulators, you might consider a proactive briefing or white paper to them about how you’re implementing AI governance. This is optional, but can score points. For example, some companies submitted their AI ethics approaches to EU bodies during consultations. Even a Q&A in an annex of a compliance report - ready to be handed over - can be good. It’s part of showing you take it seriously.
In essence, don’t let all your good compliance work be invisible. Use it to strengthen your narrative that your organization uses AI responsibly. In a world of hype and fear around AI, that can be a competitive advantage. A template for an executive readout might include:
- Summary of AI systems “in scope” with owners and risk tier (high/med/low).
- Top gaps or issues identified and how you plan to remediate them by quarter. (E.g., “need to improve bias control in system X by retraining on more diverse data by Q3”).
- Metrics like number of incidents caught or reduction in error rates after fixes.
- Decisions needed from leadership in the next month (like resources or policy decisions).
This keeps it action-oriented and shows progress.
The role of AI ethics
Beyond strict legal compliance, many leading organizations are adopting broader AI ethics principles – fairness, accountability, transparency, etc. While not required by law, these can bolster your program:
- An AI ethics board or working group (internal or with external advisors) can guide tricky cases that the law doesn’t explicitly cover. For instance, maybe an AI application is legal but raises ethical questions about job displacement or surveillance. Having a forum to discuss that and potentially set internal guidelines (even stricter than the law) can save reputational headaches later.
- Ethical frameworks often go into areas like avoiding amplifying unjust bias, ensuring explainability as much as possible, and respecting user privacy beyond the minimum. If you publicly commit to such principles, ensure your compliance work aligns. This could mean adding an ethics review in your development cycle (like a checklist: “Could this system be used in harmful ways? Did we consider diversity in data? Have we engaged stakeholders who might be impacted?”).
- Talent and brand: Being vocal about AI ethics isn’t just virtue signaling - it helps attract and retain employees who want to be proud of where they work. Customers, especially in Europe, care about these values too. So highlight, when appropriate, that not only are you following the law, you’re also going further to do the right thing. But be honest; don’t oversell and then get caught short. Ground every ethics claim in something concrete you’re doing.
The bottom line is that a culture of accountability around AI can become a competitive differentiator. It builds trust with regulators (they may scrutinize you a bit less if they see you’re earnest), with customers (who may prefer your product because they feel safer with it), and with partners.
A 90-day plan
If you’re starting now (2025) to operationalize this, here’s a rough sequence to get moving quickly:
- Weeks 1-2: Complete the AI inventory and assign owners for each system. Draft a preliminary classification (high/med/low) for each. Stand up a small cross-functional team (IT, legal, product, data science) to steer this effort. Also, draft an initial internal AI policy (even a one-pager) so everyone is on the same page going forward.
- Weeks 3-6: For each high-risk (or likely high-risk) system, do a deep dive. Fill out as much of the compliance checklist as you can. Identify gaps: maybe one system has no logging - flag that engineering work. Maybe another has training data issues - plan to fix or justify. Simultaneously, create the documentation skeletons (even if incomplete) for each system. It’s okay if at 6 weeks they’re half-filled, you’ll iterate. The goal is to know where you need work. Present findings to execs: “Here are our gaps and what we need to address them.”
- Weeks 7-10: Execute on closing gaps. Implement logging where missing, write or update user-facing notices for transparency, conduct the bias and accuracy tests, train staff on oversight procedures, etc. Also address vendor engagement in this window - send those questionnaires, start contract addendums if needed. By week 10, you want no system completely unaddressed; each either compliant or with a remediation path.
- Weeks 11-12: Run a tabletop exercise or simulation of an incident (or an audit). Essentially, test your readiness. For example, pretend a regulator did an audit on System A - gather all your docs as if submitting. See if anything’s obviously missing. Or simulate that System B had a major screw-up - walk through your incident response plan. This will reveal any last holes and also prepare the team for real scenarios. Finalize your ongoing monitoring plans (who will do quarterly reviews, etc.). By the end of 90 days, compile a brief report of everything done and remaining to leadership.
The goal for 2025 is not perfection - regulators themselves are just gearing up. It is to have a credible program that knows its AI systems, controls the biggest risks, and has a plan to continuously improve. If you get that in place, you’ll be in good shape to handle whatever specific guidance or enforcement comes.
Common gaps and how to close them
As you go through this, you’ll likely discover some recurring issues. Here are a few and strategies to handle them:
- Shadow or “rogue” AI (unsanctioned tools): You might find out months later that a team was using some AI service without telling anyone (maybe a marketing team using an AI analytics SaaS). To combat this, survey periodically and make the rounds in company meetings to ask, “Anyone using AI we haven’t discussed?” Encourage a blameless disclosure culture (“We’re not here to ban it, we just need to know so we can help support and document it”). Consider adding a section to project kick-off forms or IT procurement checklists: “Does this involve any AI or algorithmic component?” so it triggers involvement of your team. Basically, continually surface “shadow AI”.
- Unclear system boundaries: Sometimes what exactly constitutes a “system” is fuzzy. For instance, a platform with multiple AI features - do you treat each separately or as one? The Act tends to consider a “system” as a stand-alone AI product or module. It’s better to break it down into manageable parts for analysis, then also consider the aggregate. Ensure you’ve defined who (or what process) has final decision authority in any chain where AI is one component among many. Define the human’s role and where the AI’s autonomy stops. Clarify these boundaries in documentation (“AI module suggests pricing, but human manager approves final price - AI does not independently set price”).
- Marketing vs reality: Be cautious that your sales or marketing claims about AI match what you’ve documented. If marketing is saying “Our AI will revolutionize hiring with unbiased decisions!” but your own evaluation shows it does have some bias issues (which you’re mitigating), you need to rein that in. Overhyping can lead to unsourced claims. Either tie marketing statements to documented evaluations (“independent audit confirming X” as proof) or tone them down. Likewise, ensure you’re not making technical promises you can’t back up. It’s actually a compliance issue - misleading statements could be considered unfair practices.
Finally, remember that compliance is ongoing. The EU AI Act enforcement will evolve; standards and best practices will develop. Keep an eye on updates from regulatory sandboxes or guidelines from the EU’s planned AI Office. Adapt your program as needed. The effort you put in now not only keeps you out of trouble, it genuinely improves your AI systems and can be showcased as a trust point. When your competitors are scrambling late or (worse) facing penalties, you’ll be able to say to customers and regulators, “We’ve got this under control.” And that’s a strong position to be in as trustworthy AI becomes a market expectation.
VSCode Visible Files
src/content/posts/eu-ai-act-field-guide-2025.mdx
VSCode Open Tabs
src/content/posts/us-privacy-2025-brief.mdx src/content/posts/ira-industrial-policy-operators-2025.mdx src/content/posts/market-brief.mdx src/content/posts/crisis-qa-blueprint-2025.mdx src/content/posts/first-briefing.mdx src/content/posts/media-interviews-high-stakes-2025.mdx src/content/posts/monitoring-without-crowdtangle-2025.mdx src/content/posts/stakeholder-mapping-deep-dive-2025.mdx src/content/posts/evidence-playbook-2025.mdx src/content/posts/eu-ai-act-field-guide-2025.mdx
Current Time
8/9/2025, 12:06:46 AM (America/New_York, UTC-4:00)
Context Window Usage
365,140 / 1,048.576K tokens used (35%)
Current Mode
ACT MODE