Chapter V — General-purpose AI modelsArticle 51

Article 51: Classification of GPAI Models with Systemic Risk

In effect since 2 Aug 20257 min readEUR-Lex verified Apr 2026

Article 51 defines when a general-purpose AI model is classified as having systemic risk. A GPAI model has systemic risk when it has high-impact capabilities evaluated on the basis of appropriate technical tools and methodologies, including indicators and benchmarks—or when designated by a Commission decision (including upon qualified alert from the scientific panel). A GPAI model is presumed to have high-impact capabilities when the cumulative amount of computation used for its training exceeds 10^25 FLOPs (the threshold in Annex XIII), subject to Commission update via delegated acts. Classification triggers additional obligations under Article 55 and Section 2 of Annex XI.

Who does this apply to?

  • -Providers of GPAI models that meet or may meet the systemic-risk criteria (especially the compute threshold)
  • -The AI Office and the scientific panel for AI (classification and monitoring roles)
  • -Downstream providers integrating GPAI models who need to know upstream systemic-risk status
  • -Compliance teams tracking Annex XIII thresholds and Commission delegated acts

Scenarios

A new model is trained with cumulative compute exceeding 10^25 FLOPs.

Presumed to have systemic risk under Article 51(2); the provider must notify the AI Office under Article 52 and comply with Article 55 obligations.
Ref. Art. 51(2)

A model is below the compute threshold but the scientific panel issues a qualified alert citing high-impact capabilities demonstrated in real-world deployments.

The Commission may designate the model as systemic risk by decision under Article 51(1) regardless of the compute threshold.
Ref. Art. 51(1)

The Commission updates the Annex XIII threshold via delegated act after advances in training efficiency make 10^25 FLOPs less meaningful as a risk proxy.

Article 51(2) references the updated threshold; providers must re-assess against the new benchmark.
Ref. Art. 51(2) + Annex XIII

What Article 51 does (in plain terms)

Article 51 answers the question: when is a GPAI model considered to have systemic risk? Two routes:

Route 1 — High-impact capabilities (Article 51(1)): A GPAI model is classified as systemic risk when it has high-impact capabilities evaluated using appropriate technical tools and methodologies, including indicators and benchmarks. The AI Office may designate on its own initiative or after a qualified alert from the scientific panel (Article 90).

Route 2 — Compute presumption (Article 51(2)): A GPAI model is presumed to have high-impact capabilities when the cumulative amount of computation used for its training, measured in floating-point operations (FLOPs), is greater than 10^25. The Commission may update this threshold by delegated act under Article 97 based on evolving technological benchmarks.

Providers can rebut the presumption with substantiated arguments through the Article 52 procedure—the threshold is not absolute, but the burden is on the provider.

What triggers when classification happens

Once a GPAI model is classified as systemic risk (by notification or designation):

1. Article 55 additional obligations activate: model evaluation, adversarial testing (red teaming), tracking and reporting serious incidents, ensuring adequate cybersecurity protection. 2. Annex XI Section 2 documentation becomes mandatory: detailed evaluation strategies, red-teaming measures, and system architecture. 3. The model appears on the AI Office's published list of systemic-risk GPAI models. 4. Providers may need to adhere to relevant codes of practice under Article 56 or demonstrate equivalent compliance.

The 10^25 FLOPs threshold — context and limitations

The 10^25 FLOPs presumption was calibrated to frontier models at the time of legislative negotiations. It is a rebuttable presumption, not a hard classification boundary—models below the threshold can still be designated if they demonstrate high-impact capabilities through other evidence.

As training efficiency improves (distillation, architecture advances, data quality), the compute proxy may under- or over-capture risk. The Commission has delegated-act power to adjust the threshold and the Annex XIII criteria to reflect state of the art.

How Article 51 connects to the rest of the Act

  • Article 52Procedure for classification (notification, AI Office designation, rebuttal, publication).
  • Article 53Baseline GPAI obligations (apply to all GPAI providers, regardless of systemic risk).
  • Article 55Additional obligations activated by systemic-risk classification.
  • Article 56Codes of practice as a compliance pathway for systemic-risk obligations.
  • Annex XISection 2 documentation triggered by systemic-risk status.
  • Annex XIIICriteria and thresholds for systemic-risk classification (including the 10^25 FLOPs presumption).
  • Article 90Scientific panel that may trigger qualified alerts for AI Office classification.
  • Article 97Delegated acts for updating the Annex XIII threshold.
  • Article 101Penalties for GPAI providers.
  • Article 113Application dates.

Official wording: Article 51 (English)

1. A general-purpose AI model shall be classified as a general-purpose AI model with systemic risk if it meets any of the following conditions:
(a) it has high impact capabilities evaluated on the basis of appropriate technical tools and methodologies, including indicators and benchmarks;
(b) based on a decision of the Commission, ex officio or following a qualified alert from the scientific panel, it has capabilities or an impact equivalent to those set out in point (a) having regard to the criteria set out in Annex XIII.
2. A general-purpose AI model shall be presumed to have high impact capabilities pursuant to paragraph 1, point (a), when the cumulative amount of computation used for its training measured in floating point operations is greater than 10^25.
3. The Commission shall adopt delegated acts in accordance with Article 97 to amend the thresholds listed in paragraphs 1 and 2 of this Article, as well as to supplement benchmarks and indicators in light of evolving technological developments, such as algorithmic improvements or increased hardware efficiency, when necessary, for these thresholds to reflect the state of the art.

Recitals (preamble) on EUR-Lex

The recitals in the same consolidated AI Act on EUR-Lex contextualise systemic risk, high-impact capabilities, the 10^25 FLOPs calibration, the scientific panel role, and the Commission's delegated-act update mechanism. Use the official preamble on EUR-Lexdo not rely on unofficial recital lists without checking sequence and wording against the authentic text.

Compliance checklist

  • Calculate cumulative training compute (FLOPs) for each model and compare against the 10^25 threshold.
  • Monitor Commission delegated acts for threshold updates to Annex XIII.
  • Set up internal triggers to notify the AI Office without delay if the threshold is met (Article 52).
  • If below the threshold: document why systemic-risk classification does not apply (high-impact capabilities assessment).
  • If designated: immediately operationalise Article 55 obligations and Annex XI Section 2 documentation.
  • Track the AI Office's published list of systemic-risk models for upstream dependencies.

Assess your GPAI model against systemic-risk criteria—free assessment.

Start Free Assessment

Related Articles

Related annexes

  • Annex XI — GPAI technical documentation (Section 2 triggered by systemic-risk classification)
  • Annex XII — Transparency information
  • Annex XIII — Criteria for systemic-risk classification (including 10^25 FLOPs threshold)

Frequently asked questions

Is the 10^25 FLOPs threshold permanent?

No. The Commission can update it via delegated act under Article 97 based on evolving technological benchmarks and state of the art.

Can a model be systemic risk without crossing the compute threshold?

Yes. Article 51(1)(b) allows the Commission to designate based on equivalent capabilities or impact, using Annex XIII criteria, even if the model is below 10^25 FLOPs.

Do open-source models avoid systemic-risk classification?

No. Open-source status does not affect Article 51 classification. If an open-weights model meets the criteria, it is systemic risk and the Article 53(2) open-source exception falls away.

What is the scientific panel?

An independent expert body under Article 90 that can issue qualified alerts to the AI Office about GPAI models potentially meeting systemic-risk criteria, triggering the Article 52 classification procedure.

Who is responsible when we fine-tune a base model?

Fine-tuning can create a new provider relationship. Whether the fine-tuned model separately triggers Article 51 depends on the cumulative compute and capabilities of the resulting model. Map your facts against the definitions.