EU AI Act Annex XIII: Criteria For The Designation Of General-Purpose AI models With Systemic Risk Referred To In Article 51

Nov 6, 2025by Maya G

Introduction

As artificial intelligence becomes more powerful and pervasive, regulators across the globe are grappling with one pressing question — how do we ensure large AI models are safe, transparent, and accountable? The EU Artificial Intelligence Act (AI Act) provides a comprehensive answer. It introduces a tiered regulatory framework for General-Purpose AI (GPAI) models — the foundational systems that can be used across a wide range of applications. One of the Act’s most important sections, Annex XIII, outlines the criteria for designating GPAI models as having “systemic risk” under Article 51. This designation determines the level of regulatory scrutiny and compliance obligations a model provider will face.

EU AI Act Annex XIII: Criteria For The Designation Of General-Purpose AI models With Systemic Risk Referred To In Article 51

General-purpose AI models can handle diverse tasks across industries. Unlike specialized AI, they offer flexible, tailored solutions that can transform entire sectors.

  • Versatility: These models can be adapted to a broad spectrum of applications, making them suitable for diverse industries. Their ability to learn and adjust to new tasks quickly allows organizations to deploy them across multiple functions, driving efficiency and innovation. This adaptability is crucial in rapidly changing markets, where businesses must pivot swiftly to maintain a competitive edge.

  • Scalability: They can handle varying amounts of data and complexity, making them ideal for both small-scale and large-scale operations. This characteristic is particularly beneficial for startups and large corporations alike, as it ensures that the AI solutions can grow alongside the business. As data volumes increase, scalable AI models can continue to deliver high performance without the need for significant reconfiguration.

  • Interoperability: General-purpose AI models can integrate with different systems and platforms, enhancing their utility. This interoperability enables businesses to leverage existing technology investments while incorporating cutting-edge AI capabilities. By seamlessly connecting with various systems, these models facilitate a more cohesive and efficient digital ecosystem, ultimately driving better outcomes.

  • Data Privacy Concerns: The use of vast amounts of data can lead to privacy breaches if not managed properly. Ensuring robust data protection measures is vital to maintaining user trust and compliance with legal requirements. Organizations must implement stringent data governance practices to safeguard sensitive information and prevent unauthorized access.

  • Bias and Discrimination: AI models can unintentionally perpetuate biases present in their training data, leading to unfair outcomes. Addressing these biases requires continuous monitoring and refinement of AI algorithms to ensure equitable treatment for all users. Companies must prioritize diversity and inclusion when developing AI solutions to minimize the risk of discriminatory practices.

  • Security Vulnerabilities: As with any technology, AI models can be vulnerable to cyberattacks, which could have widespread implications. Protecting AI systems from malicious threats necessitates a proactive approach to cybersecurity, including regular updates and threat assessments. Organizations must invest in robust security infrastructure to safeguard their AI assets against potential breaches.

  • Economic Disruption: The automation potential of AI could lead to job displacement and economic shifts. Balancing automation with workforce development is essential to mitigate negative impacts on employment. Policymakers and businesses must collaborate to create reskilling initiatives that empower workers to thrive in an AI-driven economy.

Key Criteria For Systemic Risk Under Annex XIII

Annex XIII defines a structured set of technical and societal criteria that guide the European Commission’s assessment. Let’s explore them in detail:

1. Compute Power and Model Complexity

  • The total computational resources used to train the model, measured in FLOPs (floating-point operations), is a key indicator.

  • Models exceeding a specific compute threshold (often benchmarked around 10²⁵ FLOPs or similar magnitudes) may be automatically designated as high-impact.

  • The architecture complexity, number of parameters, and training duration also contribute to systemic risk evaluation.

 2. Autonomy And Adaptability

  • Models capable of autonomous decision-making, self-improvement, or multi-domain adaptation are considered more risky.

  • This includes large language models (LLMs) and multimodal AI systems that can perform diverse tasks beyond their original training scope.

3. Performance And Generality

  • Models that demonstrate general intelligence capabilities — i.e., those performing well across multiple benchmarks or cognitive tasks — may be classified as systemically risky.

  • High zero-shot and few-shot learning performance is another indicator of advanced general-purpose ability.

 4. Scale Of Use and Distribution

  • The number of downstream providers and users relying on the model.

  • The geographic scope and domain coverage (e.g., finance, healthcare, education, defense).

  • Models widely integrated into critical infrastructure or consumer applications carry higher systemic risk.

5. Potential For Societal or Economic Disruption

  • The model’s ability to influence human behavior, public discourse, or markets.

  • Risks related to disinformation, deepfakes, automation of sensitive decisions, or mass manipulation.

6. Capability For Generating High-Risk Outputs

  • The likelihood that a model can produce harmful, deceptive, or illegal content (e.g., misinformation, hate speech, or malicious code).

  • Models with content-generation capabilities (text, image, video, or code) are especially scrutinized.

Comparing AI Models: Choosing The Right, Responsible Solution

  • Performance: Evaluate the accuracy and efficiency of the model in executing tasks. High-performance models deliver superior results, driving business value and customer satisfaction. Organizations should prioritize models that consistently meet or exceed performance benchmarks in relevant domains.

  • Adaptability: Consider how easily the model can be adapted for different use cases. Flexible models offer the versatility needed to address diverse challenges and opportunities across industries. By selecting adaptable solutions, businesses can future-proof their AI investments against changing market dynamics.

  • Cost-Effectiveness: Analyze the costs associated with deploying and maintaining the model. Cost considerations encompass not only initial investments but also long-term operational expenses. Organizations must balance financial constraints with the potential benefits of AI to achieve sustainable growth.

  • Ethical Implications: Consider the ethical impact of using the model in real-world scenarios. Ethical AI practices prioritize fairness, accountability, and transparency, fostering trust and credibility. Companies should evaluate how their AI solutions align with societal values and contribute to positive outcomes.

What Happens When A Model Is Designated As “Systemic Risk”?

When a GPAI model is classified as systemic, the provider must comply with enhanced obligations under the AI Act, including:

  • Detailed Technical Documentation (Annex XI) – covering architecture, data, testing, and energy usage.

  • Comprehensive Transparency Information (Annex XII) – shared with downstream providers to ensure safe integration.

  • Rigorous Model Evaluation & Testing – including adversarial testing, bias analysis, and safety alignment.

  • Incident Reporting – mandatory disclosure of serious malfunctions, breaches, or misuse cases.

  • Energy & Sustainability Reporting – transparency on environmental impact during training and operation.

Conclusion

Annex XIII of the EU AI Act represents the EU’s proactive stance on managing the risks of advanced AI systems. By setting clear criteria for designating General-Purpose AI models with systemic risk, it ensures that the most powerful models are subject to rigorous oversight and accountability. As AI models grow in scale and influence, Annex XIII acts as a safety net for society — balancing innovation with protection. It ensures that AI technologies, no matter how advanced, remain aligned with European values of transparency, fairness, and human-centric progress. For developers, this isn’t just a compliance challenge — it’s an opportunity to build trust and credibility in the world’s most regulated AI market.