NEWS 11 min read

EU AI Act: What Developers Actually Need to Know

The EU AI Act is now enforceable and most developers have no idea what it requires. Here's the no-BS breakdown of what you must do, what you can ignore, and the fines.

By EgoistAI ·
EU AI Act: What Developers Actually Need to Know

What Does the EU AI Act Require from Developers?

The EU AI Act is the world’s first comprehensive AI law, and as of February 2025, its prohibited practices provisions are enforceable — with the full framework phasing in through August 2026. If you build, deploy, or distribute AI systems that touch EU citizens, you need to comply or face fines up to 35 million euros or 7% of global annual turnover (whichever is higher). The core framework: AI systems are classified into four risk tiers — unacceptable, high, limited, and minimal — with obligations scaling based on the tier. Most indie developers and startups building chatbots, content generators, or coding tools fall into the “limited risk” category, which mainly requires transparency obligations. But if you’re building anything involving biometrics, hiring, credit scoring, or critical infrastructure, you’re in high-risk territory and the compliance burden is serious.

How Does the EU AI Act Risk Classification Work?

This is the backbone of the entire regulation. Everything flows from which risk tier your AI system falls into.

Unacceptable Risk (Banned)

These AI applications are outright prohibited in the EU:

  • Social scoring: Systems that rank people based on social behavior or personality traits
  • Manipulative AI: Subliminal techniques that distort behavior and cause harm
  • Emotion recognition in workplaces and schools (with narrow exceptions)
  • Untargeted facial recognition scraping: Building face databases by scraping images from the internet or CCTV (yes, Clearview AI, this means you)
  • Predictive policing: AI systems that predict criminal behavior based solely on profiling

High Risk

These require full compliance — documentation, human oversight, data governance, the works:

  • AI in hiring and recruitment
  • Credit scoring and financial assessments
  • AI used in education for grading or admissions
  • Medical devices that use AI
  • AI in law enforcement (non-banned applications)
  • Critical infrastructure management (energy, water, transport)
  • Biometric identification (permitted uses)

Limited Risk

Transparency obligations only. This is where most developer-built AI lives:

  • Chatbots: Must disclose that users are interacting with AI
  • AI-generated content: Must be labeled as AI-generated (text, images, audio, video)
  • Deepfakes: Must be clearly labeled
  • Emotion recognition systems (non-prohibited uses): Must inform users

Minimal Risk

No specific obligations. Free to build:

  • AI-powered spam filters
  • AI in video games
  • Recommendation algorithms (with some exceptions)
  • Internal productivity tools

What Are the Compliance Requirements for High-Risk AI?

If your system falls into the high-risk category, here’s what you’re actually required to do:

RequirementWhat It MeansPractical Impact
Risk Management SystemDocument risks throughout the AI lifecycleOngoing risk assessments, not just at launch
Data GovernanceTraining data must be relevant, representative, and error-freeYou need data documentation and quality checks
Technical DocumentationFull system documentation before market placementArchitecture docs, training methodology, performance metrics
Record-KeepingAutomatic logging of system operationsAudit trails for all AI decisions
TransparencyClear instructions for deployers on proper useUser manuals, limitation disclosures
Human OversightHumans must be able to override AI decisionsKill switches, human-in-the-loop mechanisms
Accuracy & RobustnessSystems must perform consistently and resist manipulationTesting against adversarial attacks, bias testing
CybersecurityAppropriate security measuresStandard security practices, nothing exotic
Conformity AssessmentPre-market evaluationSelf-assessment for most; third-party audit for biometrics
CE MarkingCompliance marking before EU market entryAdministrative step after conformity assessment

That’s a lot. And it’s designed to be a lot. The EU wants to make you think twice before deploying AI in high-stakes domains without proper safeguards.

How Does the AI Act Affect General-Purpose AI Models Like GPT and Claude?

Here’s where it gets interesting for the broader AI ecosystem. The Act created a special category: General-Purpose AI (GPAI) models.

If you’re building on top of GPT, Claude, Gemini, or Llama, the obligations split between you (the deployer/developer) and the model provider:

GPAI providers (OpenAI, Anthropic, Google, Meta) must:

  • Provide technical documentation
  • Share information with downstream developers for compliance
  • Comply with EU copyright law
  • Publish a summary of training data

GPAI with “systemic risk” (models trained with >10^25 FLOPs) must also:

  • Conduct model evaluations including adversarial testing
  • Assess and mitigate systemic risks
  • Report serious incidents
  • Ensure cybersecurity protections

What this means for you as a developer building on these APIs: The model provider handles most of the GPAI obligations. Your job is to handle the obligations specific to your application — transparency labels, risk assessment for your use case, and ensuring your deployment doesn’t turn a limited-risk model into a high-risk application.

What Are the Actual Fines for Non-Compliance?

The EU doesn’t mess around with penalties:

ViolationMaximum Fine
Prohibited AI practices35M euros or 7% global turnover
High-risk non-compliance15M euros or 3% global turnover
Providing incorrect information7.5M euros or 1.5% global turnover

For SMEs and startups, the fines are capped at the lower of the two amounts. So a bootstrapped startup won’t face a 35M euro fine — but the percentage-based calculation can still hurt.

The enforcement mechanism: Each EU member state designates a national supervisory authority. There’s also an EU AI Office at the European Commission level for GPAI oversight. Enforcement started with prohibited practices in February 2025, with full enforcement of high-risk requirements rolling in through August 2026.

What’s the Timeline for Compliance?

This is the part most developers miss — the Act doesn’t hit all at once:

  • February 2, 2025: Prohibited practices enforceable. AI literacy obligations active.
  • August 2, 2025: GPAI model obligations kick in. Governance structures must be established.
  • August 2, 2026: Full enforcement of high-risk AI system requirements.
  • August 2, 2027: Extended deadline for high-risk AI embedded in regulated products (medical devices, vehicles, etc.).

We’re currently between the second and third milestones. If you’re building high-risk AI, you have until August 2026 to get compliant. That sounds like a lot of time. It isn’t.

How Do I Know If My AI Project Needs to Comply?

Run through this checklist:

  1. Does your AI system interact with EU citizens? If yes, the Act applies regardless of where your company is based.
  2. Is your use case on the prohibited list? If yes, stop. You can’t build it.
  3. Is your use case on the high-risk list? Check Annex III of the Act for the full list. If yes, start your compliance process now.
  4. Does your system generate content, interact with users, or detect emotions? You have transparency obligations.
  5. Are you using a GPAI model (GPT, Claude, Gemini)? Your model provider handles most GPAI obligations. You handle deployment-level compliance.

Practical Compliance Steps for Most Developers

If you’re building a typical AI application — chatbot, content generator, coding assistant, recommendation engine — here’s your minimum viable compliance:

  1. Add AI disclosure: Tell users they’re interacting with AI. A simple “Powered by AI” notice works.
  2. Label AI-generated content: If your app generates text, images, or audio, label it.
  3. Document your system: Keep records of what model you use, what data you feed it, and what it does.
  4. Conduct a basic risk assessment: Write down the potential harms and how you mitigate them. This doesn’t need to be a 200-page report.
  5. Stay updated: The EU AI Office is publishing guidelines and codes of practice throughout 2025-2026.

How Does the EU AI Act Compare to US and Chinese AI Regulation?

AspectEU AI ActUS (Executive Order + State Laws)China AI Regulations
ApproachComprehensive legislationFragmented (federal EO + state laws)Sector-specific rules
Risk classificationFour-tier systemNo unified frameworkActivity-based
EnforcementDedicated AI authoritiesVaries by agency/stateCyberspace Administration
PenaltiesUp to 7% global turnoverVaries by stateAdministrative + criminal
GPAI rulesYes (specific provisions)No specific frameworkYes (algorithm registration)
TransparencyMandatory for limited+ riskVoluntary (mostly)Mandatory for deepfakes
Effective datePhased 2025-2027Ongoing, fragmentedVarious dates
Extraterritorial reachYesLimitedLimited

The EU Act is the most comprehensive and the one most likely to set global standards — the “Brussels Effect” in action. If you comply with the EU AI Act, you’re likely compliant (or close to it) everywhere else.

Should Startups and Indie Developers Be Worried?

Short answer: probably not, unless you’re building in high-risk domains.

The Act includes several SME-friendly provisions:

  • Regulatory sandboxes: EU member states must establish sandbox environments where startups can test AI systems under regulatory supervision before full deployment.
  • Reduced fees: Lower conformity assessment fees for small companies.
  • Proportional fines: Penalties capped at the lower amount for SMEs.
  • Clear guidance: The EU AI Office is publishing codes of practice specifically for smaller organizations.

If you’re an indie developer building a chatbot, a content tool, or an AI-powered SaaS, your main obligation is transparency. Slap on an AI disclosure, label your generated content, and keep basic documentation. That’s manageable.

Where it gets tricky: if you’re building AI tools used for hiring, education, or finance — even indirectly. A “resume optimizer” powered by AI could be classified as high-risk because it influences hiring decisions. Think carefully about how your tool is used, not just how you intend it to be used.

For developers managing multiple AI service subscriptions to test compliance across different models, GamsGo can help reduce costs by offering discounted access to premium AI tools.

FAQ

Does the EU AI Act apply to open-source AI?

Partially. Open-source models released under permissive licenses are largely exempt from GPAI provider obligations, UNLESS they pose systemic risk (>10^25 FLOPs training compute). The exemption doesn’t apply to high-risk deployments — if you deploy an open-source model in a high-risk use case, you still have full compliance obligations.

Do I need to comply if my company is outside the EU?

Yes, if your AI system is placed on the EU market or its output is used in the EU. The Act has extraterritorial reach, similar to GDPR. If EU citizens use your product, you need to comply.

What counts as “AI-generated content” that needs labeling?

Text, images, audio, and video produced by AI systems. This includes chatbot responses, AI-generated images, synthetic voice, and deepfakes. The labeling must be machine-readable where technically feasible.

Can I use ChatGPT/Claude in my product and stay compliant?

Yes. The GPAI provider (OpenAI, Anthropic) handles model-level obligations. You handle deployment-level obligations — transparency, risk assessment for your specific use case, and ensuring your application doesn’t create prohibited or high-risk scenarios without compliance.

What is an “AI literacy” obligation?

Article 4 requires providers and deployers to ensure their staff have sufficient AI literacy — understanding of how AI works, its limitations, and its risks. This doesn’t mean everyone needs a PhD in machine learning. Basic training is sufficient.

How is the EU AI Act enforced?

Each EU member state designates national competent authorities for enforcement. The EU AI Office (under the European Commission) oversees GPAI and systemic risk models. Enforcement actions can include fines, market withdrawal orders, and mandatory corrective measures.

Share this article

> Want more like this?

Get the best AI insights delivered weekly.

> Related Articles

Tags

eu ai actai regulationcompliancedevelopersai policyeurope2026

> Stay in the loop

Weekly AI tools & insights.