Skip to content

EU AI Act Risk Classification

The EU AI Act requires classifying AI systems into four risk levels: Unacceptable, High-Risk, Limited-Risk, and Minimal-Risk. Each level has different compliance requirements. TruthVouch automatically classifies based on system characteristics, then lets you override if needed.

AI systems inventory with automatic risk classification

The Four Risk Levels

Unacceptable Risk (Prohibited)

Systems that pose unacceptable risk and cannot be deployed in the EU.

Prohibited Practices:

  • Social scoring by public authorities
  • Subliminal manipulation
  • Exploitation of vulnerabilities (targeting children or disabled)
  • Real-time biometric ID in public spaces (with limited exceptions)

TruthVouch Action: If system classified as Unacceptable, DO NOT DEPLOY. Must redesign to remove prohibited practices.

High-Risk

Systems that require full compliance — risk assessments, documentation, testing, audit trails, human oversight.

Triggers (any of these):

TriggerWhyExample
Autonomous decisions affecting legal rightsDecisions without human appealLoan denial, hiring rejection, benefits eligibility
Biometric systemsPrivacy & identification riskFacial recognition, fingerprint, iris scan, voice recognition
Critical infrastructureSafety-critical systemsPower grid, water treatment, transportation control
Education & trainingFairness & opportunity riskStudent grading, course recommendations, school admissions
EmploymentDiscrimination & fairness riskResume screening, performance evaluation, scheduling
Pricing & financial servicesEconomic harm riskInsurance quotes, credit decisions, lending
Large-scale personal data + automated decisionsScale of impactProfiling 10K+ people, automated targeting
Real-time biometric ID (with exceptions)Privacy & identificationLaw enforcement real-time facial recognition (under strict conditions)
Content moderation at scaleRights impactFlagging illegal content affecting 100K+ users

Compliance Requirements (Annex III & IV):

  1. Risk assessment document
  2. Technical documentation (Annex IV)
  3. Bias and fairness testing
  4. Audit trail implementation
  5. Human oversight processes
  6. Post-deployment monitoring
  7. Incident reporting procedure

Limited-Risk

Systems that require transparency disclosures to users.

Triggers:

  • AI that interacts with humans or makes decisions affecting them
  • Not autonomous (human review involved)
  • Not high-risk data or scope

Examples:

  • Chatbots (must disclose AI-generated)
  • Content recommendation
  • Content moderation (human-in-the-loop)
  • Language generation (disclosure required)

Compliance Requirements:

  1. Transparent disclosure to users that AI is involved
  2. Clear explanation of what AI does
  3. User can opt for human alternative (if available)

Minimal-Risk

Systems with no requirements — low risk to individuals or society.

Examples:

  • Spell checkers
  • Grammar correction
  • Search result ranking (non-personalized)
  • Email spam filters
  • Syntax highlighters
  • Accessibility features

Auto-Classification Algorithm

Compliance AI follows EU AI Act Annex I logic:

IF prohibited_practice THEN Unacceptable
ELSE IF (autonomous_decision AND legal_effect) THEN High-Risk
ELSE IF (biometric_system) THEN High-Risk
ELSE IF (critical_infrastructure) THEN High-Risk
ELSE IF (personal_decisions + scale > threshold) THEN High-Risk
ELSE IF (no_autonomous AND transparency_disclosed) THEN Limited-Risk
ELSE IF (minimal_risk_indicator) THEN Minimal-Risk
ELSE Limited-Risk (default)

Factors considered:

  1. System type (LLM, biometric, recommender, etc.)
  2. Decision scope (autonomous vs. assisted vs. informational)
  3. Data sensitivity (special categories, scale)
  4. Jurisdiction (EU trigger)
  5. User/subject population (children, vulnerable groups)
  6. Legal/economic impact (binding decisions, financial)

How to Override Classification

If you disagree with auto-classification:

  1. Go to Registry > [System Name] > Risk Classification
  2. Click Override
  3. Select new risk level
  4. Explain reasoning in comments
  5. Save

Example override:

  • Auto-classified: High-Risk (loan approval recommendation)
  • Your override: Limited-Risk
  • Reason: “Bank review all loans autonomously; AI is informational only. All decisions made by human loan officer. No autonomous decisions.”

Compliance AI logs all overrides for audit trail.

Risk Level Impact on Compliance

RequirementUnacceptableHigh-RiskLimited-RiskMinimal-Risk
Deployment in EUProhibitedAllowed with complianceAllowed with disclosureAllowed
Risk AssessmentN/ARequiredOptionalNo
Technical DocumentationN/ARequired (Annex IV)NoNo
Bias TestingN/ARequiredRecommendedNo
Audit TrailN/ARequiredRecommendedNo
Human OversightN/ARequiredRecommendedNo
Post-Deployment MonitoringN/ARequiredRecommendedNo
User NotificationN/AYes (“high-risk AI”)Yes (“AI-generated”)No
Incident ReportingN/AWithin 72 hours (Article 73)NoNo

Risk Classification Workflow

  1. New system created → Auto-classified based on profile
  2. Classification displayed → System details show risk level with reasoning
  3. Team reviews → Confirm or override classification
  4. Scans adjust requirements → High-risk scans include all requirements; limited-risk only check transparency
  5. Annual review → Re-classify if system changes (scope, data, deployment)

Common Classification Examples

Example 1: Customer Service Chatbot

System: AI chatbot answers customer questions on website

  • Auto-inputs: LLM, customer data, real-time interaction, no autonomous decisions
  • Auto-classification: Limited-Risk
  • Reasoning: User knows it’s AI (chatbot interface), human support available, no binding decisions, no special data
  • Requirements: Disclose “This is an AI chatbot” in UI, offer “speak with human” option

Example 2: Hiring Resume Screener

System: ML model filters resumes, recommends candidates to humans

  • Auto-inputs: Decision-making, employment impact, training data could have bias
  • Auto-classification: High-Risk
  • Reasoning: Employment decisions under EU AI Act Annex I; significant impact on individuals; potential for discrimination
  • Requirements: Full compliance (risk assessment, bias testing, audit trail, human review, monitoring)

Example 3: Loan Approval System (Autonomous)

System: AI autonomously approves/denies loans under €50K

  • Auto-inputs: Autonomous decision, legal effect (binding), financial impact, potential discrimination
  • Auto-classification: Unacceptable or High-Risk (EU AI Act doesn’t auto-approve loans, so High-Risk)
  • Reasoning: Article 4 high-risk criteria triggered
  • Requirements: Cannot be deployed in EU unless redesigned (add human review for all decisions)

Example 4: Email Spam Filter

System: ML model classifies emails as spam or legitimate

  • Auto-inputs: No personal decisions, automated utility, no impact on individuals
  • Auto-classification: Minimal-Risk
  • Reasoning: No legal/economic impact, not a decision system, widely accepted utility
  • Requirements: None

Annual Re-Classification

Systems should be re-classified if they change:

  1. Scope of deployment — Now serving EU users (changes jurisdiction)
  2. System capabilities — Now autonomous (was assisted before)
  3. Data processing — Now processes health data (was not before)
  4. User population — Now used in hiring (was internal only)

Compliance AI flags systems needing re-classification annually.

Regulatory Transition

August 2, 2025: Prohibited practices banned (immediate) August 2, 2026: Full EU AI Act compliance required

If your system is High-Risk and in the EU, target August 2026 for full compliance. Start now.

Next Steps

  • View your system’s classification: Go to Registry > [System Name]
  • Override if needed: Click “Risk Classification” tab
  • Understand requirements: See EU AI Act Framework
  • Prepare for compliance: Run a scan using Running Scans