EthiCompass

EU AI Act Compliance

The EU AI Act Is in Force.
Can You Prove Compliance?

Most organizations deploying AI systems cannot produce the evidence regulators now require. EthiCompass maps every AI system to EU AI Act Articles 9–15 using a peer-reviewed, 7-dimension framework — with an immutable audit trail that proves compliance, not just claims it.

Explore the 7 Dimensions →

The Cost of “We Think
We’re Compliant”

The EU AI Act applies to every organization deploying high-risk AI systems in the European Union. Enforcement is not future tense — it is happening now.

Regulators are not asking whether you have an AI governance policy. They are asking you to prove that every AI system has been evaluated, that risks have been documented, and that compliance evidence is auditable.

“We’re working on it” is not defensible. “Here is our immutable audit trail” is.

Up to 7%

of global annual revenue — maximum EU AI Act fine

$2.3M

average cost of a single AI compliance incident

73%

of enterprises deploying AI have no formal compliance framework

Art. 9–15

the specific requirements your AI systems must demonstrably meet

What Regulators Expect

Articles 9–15: What You Need to Prove

The EU AI Act defines specific obligations for high-risk AI systems. For each article, we show what the regulation requires and what evidence you need to produce.

Art. 9

Risk Management

What It Requires

A continuous, iterative risk management process throughout the AI system lifecycle. Risks must be identified, evaluated, and mitigated with documented evidence.

What a Regulator Expects

A risk register with quantified risk scores, documented mitigation actions, and proof of ongoing monitoring — not a one-time assessment.

Art. 10

Data Governance

What It Requires

Training, validation, and testing data must meet quality criteria. Data practices must prevent bias and ensure representative datasets across protected groups.

What a Regulator Expects

Demographic parity metrics, bias testing results, and documentation that data governance practices are enforced — not just described.

Art. 11

Technical Documentation

What It Requires

Comprehensive technical documentation that demonstrates compliance before the AI system is placed on the market or put into service. Documentation must be kept up to date.

What a Regulator Expects

Model cards, data sheets, methodology descriptions, and a complete record of compliance evaluations — maintained continuously, not created retroactively when asked.

Art. 12

Record-Keeping

What It Requires

AI systems must have automatic logging capabilities that ensure traceability of the system's functioning throughout its lifecycle. Logs must be retained for an appropriate period.

What a Regulator Expects

An immutable, tamper-proof audit trail that captures every evaluation, every flag, and every decision — with retention periods that exceed the system's operational lifetime.

Art. 13

Transparency

What It Requires

AI systems must be sufficiently transparent to enable users to interpret the system's output and use it appropriately.

What a Regulator Expects

Explainability scores for every AI decision, human-readable justifications for flags and recommendations, and documented evidence that transparency mechanisms are operational.

Art. 14

Human Oversight

What It Requires

AI systems must be designed to allow effective oversight by natural persons during the period in which the system is in use, including the ability to override or reverse automated decisions.

What a Regulator Expects

A dashboard showing human intervention rates, override logs, and evidence that oversight mechanisms are used — not just available.

Art. 15

Accuracy, Robustness & Cybersecurity

What It Requires

AI systems must achieve consistent performance levels, be resilient to errors and adversarial attacks, and meet cybersecurity standards appropriate to the risk level.

What a Regulator Expects

Critical error rates, adversarial resilience test results, performance drift monitoring, and evidence that robustness is continuously measured — not tested once and assumed.

Built for the EU AI Act.
Not Retrofitted.

Our 7-dimension framework maps natively to Articles 9–15. Each dimension produces the specific evidence regulators expect — scored, traceable, and stored in an immutable audit trail.

Article

Art. 9 — Risk Management

Dimension

7. Regulatory Compliance

Proves

Continuous risk scoring across your entire AI portfolio, with quantified risk levels and documented mitigation

Article

Art. 10 — Data Governance

Dimension

1. Discrimination & Fairness

Proves

Demographic parity ratios, bias disparity indices, and evidence that AI outputs are fair across protected groups

Article

Art. 11 — Technical Documentation

Dimension

Platform: Audit Trail

Proves

Complete, automatically generated compliance documentation maintained continuously alongside every evaluation

Article

Art. 12 — Record-Keeping

Dimension

Platform: Immutable Audit Trail

Proves

Cryptographically signed records with 7+ year retention — every evaluation, every flag, every decision

Article

Art. 13 — Transparency

Dimension

3. Explainability & Transparency

Proves

Explainability coverage index, human-readable justifications for every flag, full traceability from score to criterion

Article

Art. 14 — Human Oversight

Dimension

Platform: Human Review Layer

Proves

Human intervention rates, override logs, and escalation records for low-confidence recommendations

Article

Art. 15 — Accuracy & Robustness

Dimension

5. Factuality + 6. Robustness

Proves

Critical error rates, adversarial success rates, performance drift monitoring — all measured continuously

Additional coverage: Toxicity & Harmful Language and Privacy & Data Protection provide compliance evidence beyond Articles 9–15, covering GDPR alignment, PII exposure monitoring, and harmful content prevention.

Peer-Reviewed Methodology

When a Regulator Asks How You Evaluate Compliance, What Do You Show Them?

Most AI governance platforms ask you to trust their proprietary compliance engine. When a regulator asks how it works, you point to a vendor’s marketing page.

EthiCompass is different. Our 7-dimension framework was developed by PhD researchers in AI ethics, bias detection, and regulatory compliance. It is validated through peer-reviewed publications across three domains.

When a regulator asks how you evaluate AI compliance, you point to published science. That is the difference between a vendor opinion and defensible evidence.

Explore the 7-Dimension Framework →

Two Ways to Prove
EU AI Act Compliance

Whether you need a compliance baseline or continuous monitoring, EthiCompass delivers the evidence regulators expect.

OneCheck

Your EU AI Act Compliance Baseline

A comprehensive audit of your AI systems in 3 weeks.

  • Every AI system scored across all 7 dimensions
  • EU AI Act gap analysis mapped to Articles 9–15
  • Prioritized remediation roadmap ranked by regulatory risk
  • Immutable documentation for regulators and your board

Best for: Organizations that need to understand their EU AI Act readiness before committing to a platform.

Enterprise

Full Platform

Continuous EU AI Act Compliance

The full platform for ongoing monitoring and audit-ready evidence.

  • Everything in OneCheck, plus:
  • Continuous real-time monitoring across all AI systems
  • Real-time alerting when compliance scores drift
  • Immutable audit trail with 7+ year retention
  • Customizable policy layers (enhance, never weaken)
  • Board-ready dashboards and regulator-ready documentation

Best for: Organizations deploying AI at scale that need continuous compliance assurance under the EU AI Act.

“Deployed with a Fortune 500 financial services organization managing 100+ AI systems in a regulated environment. Live in production and preventing compliance incidents.”

SOC 2 ControlsEU AI Act AlignedGDPR CompliantEncrypted End-to-End7+ Year Audit Retention

The EU AI Act Does Not Ask If You’re Working on Compliance.
It Asks You to Prove It.

7 scientifically validated dimensions. Immutable audit trails. Defensible evidence mapped to Articles 9–15. Start with an assessment or deploy continuous compliance.