Zum Hauptinhalt springen
Risk level: high Title III — High-Risk AI Systems Requirements

Article 14 — Human Oversight

Human oversight 1. High-risk AI systems shall be designed and developed in such a way, including with appropriate human-machine interface tools, that they can be effectively overseen by natural persons during the period in which they are in use. 2. Human oversight shall aim to prevent or minimise the risks to health, safety or fundamental...

Article 14
Human oversight
1. High-risk AI systems shall be designed and developed in such a way, including with appropriate human-machine
interface tools, that they can be effectively overseen by natural persons during the period in which they are in use.
2. Human oversight shall aim to prevent or minimise the risks to health, safety or fundamental rights that may emerge
when a high-risk AI system is used in accordance with its intended purpose or under conditions of reasonably foreseeable
misuse, in particular where such risks persist despite the application of other requirements set out in this Section.
3. The oversight measures shall be commensurate with the risks, level of autonomy and context of use of the high-risk
AI system, and shall be ensured through either one or both of the following types of measures:
(a) measures identified and built, when technically feasible, into the high-risk AI system by the provider before it is placed
on the market or put into service;
(b) measures identified by the provider before placing the high-risk AI system on the market or putting it into service and
that are appropriate to be implemented by the deployer.
4. For the purpose of implementing paragraphs 1, 2 and 3, the high-risk AI system shall be provided to the deployer in
such a way that natural persons to whom human oversight is assigned are enabled, as appropriate and proportionate:
(a) to properly understand the relevant capacities and limitations of the high-risk AI system and be able to duly monitor its
operation, including in view of detecting and addressing anomalies, dysfunctions and unexpected performance;
(b) to remain aware of the possible tendency of automatically relying or over-relying on the output produced by a high-risk
AI system (automation bias), in particular for high-risk AI systems used to provide information or recommendations for
decisions to be taken by natural persons;
(c) to correctly interpret the high-risk AI system’s output, taking into account, for example, the interpretation tools and
methods available;
60/144 ELI: http://data.europa.eu/eli/reg/2024/1689/oj
OJ L, 12.7.2024 EN
(d) to decide, in any particular situation, not to use the high-risk AI system or to otherwise disregard, override or reverse
the output of the high-risk AI system;
(e) to intervene in the operation of the high-risk AI system or interrupt the system through a ‘stop’ button or a similar
procedure that allows the system to come to a halt in a safe state.
5. For high-risk AI systems referred to in point 1(a) of Annex III, the measures referred to in paragraph 3 of this Article
shall be such as to ensure that, in addition, no action or decision is taken by the deployer on the basis of the identification
resulting from the system unless that identification has been separately verified and confirmed by at least two natural
persons with the necessary competence, training and authority.
The requirement for a separate verification by at least two natural persons shall not apply to high-risk AI systems used for
the purposes of law enforcement, migration, border control or asylum, where Union or national law considers the
application of this requirement to be disproportionate.

Related Blog Articles

HR & Recruitment AI: The Most Common High-Risk Category Under the EU AI Act

How Annex III employment use cases trigger high-risk obligations and what HR teams should impleme...

High-Risk AI Systems: Are You Affected?

Many companies are closer to Annex III obligations than they think. Here is how to assess your ex...

FRIA Guide for High-Risk AI Deployments

What a Fundamental Rights Impact Assessment includes, when it applies, and how SMEs can run FRIA ...

Check Your Compliance

Find out if this article applies to your AI system.

Take Risk Quiz

Compliance Templates

Download ready-made templates for EU AI Act compliance.

View Templates

Machen Sie unsere kostenlose Risikobewertung

Finden Sie in 2 Minuten heraus, wo Ihr Unternehmen unter der EU-KI-Verordnung steht.

Quiz starten