Article 79: Procedure at National Level for Dealing with AI Systems Presenting a Risk
Article 79
Procedure at National Level for Dealing with AI Systems Presenting a Risk
Updated on 31 July 2024 based on the version published in the Official Journal of the EU dated 12 July 2024 and entered into force on 1 August 2024.
1. AI systems presenting a risk shall be understood as a ‘product presenting a risk’ as defined in Article 3, point 19 of Regulation (EU) 2019/1020, in so far as they present risks to the health or safety, or to fundamental rights, of persons.
2. Where the market surveillance authority of a Member State has sufficient reason to consider an AI system to present a risk as referred to in paragraph 1 of this Article, it shall carry out an evaluation of the AI system concerned in respect of its compliance with all the requirements and obligations laid down in this Regulation. Particular attention shall be given to AI systems presenting a risk to vulnerable groups. Where risks to fundamental rights are identified, the market surveillance authority shall also inform and fully cooperate with the relevant national public authorities or bodies referred to in Article 77(1). The relevant operators shall cooperate as necessary with the market surveillance authority and with the other national public authorities or bodies referred to in Article 77(1).
Where, in the course of that evaluation, the market surveillance authority or, where applicable the market surveillance authority in cooperation with the national public authority referred to in Article 77(1), finds that the AI system does not comply with the requirements and obligations laid down in this Regulation, it shall without undue delay require the relevant operator to take all appropriate corrective actions to bring the AI system into compliance, to withdraw the AI system from the market, or to recall it within a period the market surveillance authority may prescribe, and in any event within the shorter of 15 working days, or as provided for in the relevant Union harmonisation legislation.
The market surveillance authority shall inform the relevant notified body accordingly. Article 18 of Regulation (EU) 2019/1020 shall apply to the measures referred to in the second subparagraph of this paragraph.
3. Where the market surveillance authority considers that the non-compliance is not restricted to its national territory, it shall inform the Commission and the other Member States without undue delay of the results of the evaluation and of the actions which it has required the operator to take.
4. The operator shall ensure that all appropriate corrective action is taken in respect of all the AI systems concerned that it has made available on the Union market.
5. Where the operator of an AI system does not take adequate corrective action within the period referred to in paragraph 2, the market surveillance authority shall take all appropriate provisional measures to prohibit or restrict the AI system’s being made available on its national market or put into service, to withdraw the product or the standalone AI system from that market or to recall it. That authority shall without undue delay notify the Commission and the other Member States of those measures.
6. The notification referred to in paragraph 5 shall include all available details, in particular the information necessary for the identification of the non-compliant AI system, the origin of the AI system and the supply chain, the nature of the non-compliance alleged and the risk involved, the nature and duration of the national measures taken and the arguments put forward by the relevant operator. In particular, the market surveillance authorities shall indicate whether the non-compliance is due to one or more of the following:
- non-compliance with the prohibition of the AI practices referred to in Article 5;
- a failure of a high-risk AI system to meet requirements set out in Chapter III, Section 2;
- shortcomings in the harmonised standards or common specifications referred to in Articles 40 and 41 conferring a presumption of conformity;
- non-compliance with Article 50.
7. The market surveillance authorities other than the market surveillance authority of the Member State initiating the procedure shall, without undue delay, inform the Commission and the other Member States of any measures adopted and of any additional information at their disposal relating to the non-compliance of the AI system concerned, and, in the event of disagreement with the notified national measure, of their objections.
8. Where, within three months of receipt of the notification referred to in paragraph 5 of this Article, no objection has been raised by either a market surveillance authority of a Member State or by the Commission in respect of a provisional measure taken by a market surveillance authority of another Member State, that measure shall be deemed justified. This shall be without prejudice to the procedural rights of the concerned operator in accordance with Article 18 of Regulation (EU) 2019/1020. The three-month period referred to in this paragraph shall be reduced to 30 days in the event of non-compliance with the prohibition of the AI practices referred to in Article 5 of this Regulation.
9. The market surveillance authorities shall ensure that appropriate restrictive measures are taken in respect of the product or the AI system concerned, such as withdrawal of the product or the AI system from their market, without undue delay.
Table of Contents
Chapter III: High-Risk AI Systems
- Section 1: Classification of AI Systems as High-Risk
- Section 2: Requirements for High-Risk AI Systems
- Article 8: Compliance with the Requirements
- Article 9: Risk Management System
- Article 10: Data and Data Governance
- Article 11: Technical Documentation
- Article 12: Record-Keeping
- Article 13: Transparency and Provision of Information to Deployers
- Article 14: Human Oversight
- Article 15: Accuracy, Robustness and Cybersecurity
- Section 3: Obligations of Providers and Deployers of High-Risk AI Systems and Other Parties
- Article 16: Obligations of Providers of High-Risk AI Systems
- Article 17: Quality Management System
- Article 18: Documentation Keeping
- Article 19: Automatically Generated Logs
- Article 20: Corrective Actions and Duty of Information
- Article 21: Cooperation with Competent Authorities
- Article 22: Authorised Representatives of Providers of High-Risk AI Systems
- Article 23: Obligations of Importers
- Article 23a
- Article 24: Obligations of Distributors
- Article 25: Responsibilities along the AI Value Chain
- Article 26: Obligations of Deployers of High-Risk AI Systems
- Article 27: Fundamental Rights Impact Assessment for High-Risk AI Systems
- Section 4: Notifying Authorities and Notified Bodies
- Article 28: Notifying Authorities
- Article 29: Application of a Conformity Assessment Body for Notification
- Article 30: Notification Procedure
- Article 31: Requirements Relating to Notified Bodies
- Article 32: Presumption of Conformity with Requirements Relating to Notified Bodies
- Article 33: Subsidiaries of Notified Bodies and Subcontracting
- Article 34: Operational Obligations of Notified Bodies
- Article 35: Identification Numbers and Lists of Notified Bodies
- Article 36: Changes to Notifications
- Article 37: Challenge to the Competence of Notified Bodies
- Article 38: Coordination of Notified Bodies
- Article 39: Conformity Assessment Bodies of Third Countries
- Section 5: Standards, Conformity Assessment, Certification, Registration
- Article 40: Harmonised Standards and Standardisation Deliverables
- Article 41: Common Specifications
- Article 42: Presumption of Conformity with Certain Requirements
- Article 43: Conformity Assessment
- Article 44: Certificates
- Article 45: Information Obligations of Notified Bodies
- Article 46: Derogation from Conformity Assessment Procedure
- Article 47: EU Declaration of Conformity
- Article 48: CE Marking
- Article 49: Registration
Chapter V: General-Purpose AI Models
Chapter VI: Measures in Support of Innovation
- Article 57: AI Regulatory Sandboxes
- Article 58: Detailed Arrangements for, and Functioning of, AI Regulatory Sandboxes
- Article 59: Further Processing of Personal Data for Developing Certain AI Systems in the Public Interest in the AI Regulatory Sandbox
- Article 60: Testing of High-Risk AI Systems in Real World Conditions Outside AI Regulatory Sandboxes
- Article 61: Informed Consent to Participate in Testing in Real World Conditions Outside AI Regulatory Sandboxes
- Article 62: Measures for Providers and Deployers, in Particular SMEs, including Start-Ups
- Article 63: Derogations for Specific Operators
Chapter VII: Governance
Chapter IX: Post-Market Monitoring, Information Sharing and Market Surveillance
- Section 1: Post-Market Monitoring
- Section 2: Sharing of Information on Serious Incidents
- Section 3: Enforcement
- Article 74: Market Surveillance and Control of AI Systems in the Union Market
- Article 75: Mutual Assistance, Market Surveillance and Control of General-Purpose AI Systems
- Article 76: Supervision of Testing in Real World Conditions by Market Surveillance Authorities
- Article 77: Powers of Authorities Protecting Fundamental Rights
- Article 78: Confidentiality
- Article 80: Procedure for Dealing with AI Systems Classified by the Provider as Non-High-Risk in Application of Annex III
- Article 81: Union Safeguard Procedure
- Article 82: Compliant AI Systems which Present a Risk
- Article 82a
- Article 83: Formal Non-Compliance
- Article 84: Union AI Testing Support Structures
- Article 79: Procedure at National Level for Dealing with AI Systems Presenting a Risk
- Section 4: Remedies
- Section 5: Supervision, Investigation, Enforcement and Monitoring in Respect of Providers of General-Purpose AI Models
- Article 88: Enforcement of the Obligations of Providers of General-Purpose AI Models
- Article 89: Monitoring Actions
- Article 90: Alerts of Systemic Risks by the Scientific Panel
- Article 91: Power to Request Documentation and Information
- Article 92: Power to Conduct Evaluations
- Article 93: Power to Request Measures
- Article 94: Procedural Rights of Economic Operators of the General-Purpose AI Model
Chapter XIII: Final Provisions
- Article 102: Amendment to Regulation (EC) No 300/2008
- Article 103: Amendment to Regulation (EU) No 167/2013
- Article 104: Amendment to Regulation (EU) No 168/2013
- Article 105: Amendment to Directive 2014/90/EU
- Article 106: Amendment to Directive (EU) 2016/797
- Article 107: Amendment to Regulation (EU) 2018/858
- Article 108: Amendment to Regulation (EU) 2018/1139
- Article 109: Amendment to Regulation (EU) 2019/2144
- Article 110: Amendment to Directive (EU) 2020/1828
- Article 111: AI Systems Already Placed on the Market or Put into Service and General-Purpose AI Models Already Placed on the Marked
- Article 112: Evaluation and Review
- Article 113: Entry into Force and Application