EUR-Lex Access to European Union law

Back to EUR-Lex homepage

This document is an excerpt from the EUR-Lex website

Document 52021SC0084

COMMISSION STAFF WORKING DOCUMENT IMPACT ASSESSMENT Accompanying the Proposal for a Regulation of the European Parliament and of the Council LAYING DOWN HARMONISED RULES ON ARTIFICIAL INTELLIGENCE (ARTIFICIAL INTELLIGENCE ACT) AND AMENDING CERTAIN UNION LEGISLATIVE ACTS

SWD/2021/84 final

Brussels, 21.4.2021

SWD(2021) 84 final

COMMISSION STAFF WORKING DOCUMENT

IMPACT ASSESSMENT

Accompanying the

Proposal for a Regulation of the European Parliament and of the Council

LAYING DOWN HARMONISED RULES ON ARTIFICIAL INTELLIGENCE (ARTIFICIAL INTELLIGENCE ACT) AND AMENDING CERTAIN UNION LEGISLATIVE ACTS











{COM(2021) 206 final} - {SEC(2021) 167 final} - {SWD(2021) 85 final}


Table of contents

1.Introduction: technological, socio-economic, legal and political context

1.1. Technological context

1.2. Socio-economic context

1.3. Legal context

1.3.1. Relevant fundamental rights legislation

1.3.2. Relevant product safety legislation

1.3.3. Relevant liability legislation

1.4. Political context

1.5. Scope of the impact assessment

2.Problem definition

2.1. What are the problems?

2.2. What are the main problem drivers?

2.3. How will the problem evolve?

3.WHY SHOULD THE EU ACT?

3.1. Legal basis

3.2. Subsidiarity: Necessity of EU action

3.3. Subsidiarity: Added value of EU action

4.Objectives: What is to be achieved?

4.1.General objectives

4.2.Specific objectives

4.3. Objectives tree/intervention logic.

5. What are the available policy options?

5.1. What is the baseline from which options are assessed?

5.2. Option 1: EU legislative instrument setting up a voluntary labelling scheme

5.3. Option 2: A sectoral, ‘ad-hoc’ approach

5.4. Option 3: Horizontal EU legislative instrument establishing mandatory requirements for high-risk AI applications

5.5. Option 3+: Horizontal EU legislative instrument establishing mandatory requirements for high-risk AI applications + co-regulation through codes of conduct for non-high risk applications

5.6. Option 4: Horizontal EU legislative instrument establishing mandatory requirements for all AI applications, irrespective of the risk they pose

5.7. Options discarded at an early stage

6.What are the impacts of the policy options?

6.1. Economic impacts

6.1.1.Functioning of the internal market

6.1.2.Impact on uptake of AI

6.1.3.Costs and administrative burdens

6.1.4.SME test

6.1.5.Competitiveness and innovation

6.2. Costs for public authorities

6.3. Social impact

6.4. Impacts on safety

6.5. Impacts on fundamental rights

6.6. Environmental impacts

7.How do the options compare?

7.1. Criteria for comparison

7.2. Achievement of specific objectives

7.2.1. First specific objective: Ensure that AI systems placed on the market and used are safe and respect fundamental rights and Union values

7.2.2.Second specific objective: Ensure legal certainty to facilitate investment and innovation

7.2.3.Third specific objective: Enhance governance and effective enforcement of fundamental rights and safety requirements applicable to AI systems

7.2.4.Fourth specific objective: Facilitate the development of a single market for lawful, safe and trustworthy AI applications and prevent market fragmentation

7.3. Efficiency

7.4. Coherence

7.5.Proportionality

8.Preferred option

9.How will actual impacts be monitored and evaluated?

1.Introduction: technological, socio-economic, legal and political context

1 2 As part of the Commission’s overarching agenda of making Europe ready for the digital age, the Commission is undertaking considerable work on Artificial Intelligence (AI). The overall EU strategy proposed in the White Paper on AI proposes an ecosystem of excellence and trust for AI. The concept of an ecosystem of excellence in Europe refers to measures which support research, foster collaboration between Member States and increase investment into AI development and deployment. The ecosystem of trust is based on EU values and fundamental rights, and foresees robust requirements that would give citizens the confidence to embrace AI-based solutions, while encouraging businesses to develop them. The European approach for AI aims to promote Europe’s innovation capacity in the area of AI, while supporting the development and uptake of ethical and trustworthy AI across the EU economy. AI should work for people and be a force for good in society. 

The development of an ecosystem of trust is intended as a comprehensive package of measures to address problems posed by the introduction and use of AI. In accordance with the White Paper and the Commission Work Programme, the EU plans to adopt a set of three inter-related initiatives related to AI:

(1)European legal framework for AI to address fundamental rights and safety risks specific to the AI systems (Q2 2021);

(2)EU rules to address liability issues related to new technologies, including AI systems (Q4 2021-Q1 2022);

(3)Revision of sectoral safety legislation (e.g. Machinery directive, Q1 2021, General Product Safety Directive, Q2 2021).

These three initiatives would be complementary and their adoption will proceed in stages. Firstly, as entrusted by the European Council, requested by the European Parliament and supported by the results of the public consultation on the White Paper on AI, the European Commission will adopt European legal framework for AI. This legal framework should set the ground for other forthcoming initiatives by providing: (1) a definition of an AI system; (2) a definition of ‘high risk’ AI system, and (3) common rules to ensure that AI systems placed or put into service in the Union market are trustworthy. The introduction of the European legal framework for AI will be supplemented with revisions of the sectoral safety legislation and changes to the liability rules. This staged, step-by-step and complementary approach to regulate AI aims to ensure regulatory coherence throughout the Union, therefore contributing to legal certainty for developers and users of AI systems and citizens. More details on the scope of the existing safety and liability legislation are discussed in section 1.3. (legal context) and the interaction between the three initiatives are presented in section 8 (preferred option).

This impact assessment focuses on the first AI initiative, the European legal framework for AI. The purpose of this document is to assess the case for action, the objectives, and the impact of different policy options for a European framework for AI, as envisaged by the 2020 Commission work programme.

The Proposal for a European legal framework for AI and this impact assessment build on two years of analysis of evidence and involvement of stakeholders, including academics, businesses, non-governmental organisations, Member States and citizens. The preparatory work started in 2018 with the setting up of a High-Level Expert Group on AI (HLEG) which had an inclusive and broad composition of 52 well-known experts tasked to advise the Commission on the implementation of the Commission’s Strategy on Artificial Intelligence. In April 2019, the Commissioned welcomed 3 the key requirements set out in the HLEG ethics guidelines for Trustworthy AI, 4 which had been revised to take into account more than 500 submissions from stakeholders. The Assessment List for Trustworthy Artificial Intelligence (ALTAI) 5 made these requirements operational in a piloting process with over 350 organisations. The White Paper on Artificial Intelligence further developed this approach, inciting comments from more than 1250 stakeholders. As a result, the Commission published an Inception Impact Assessment that in turn attracted more than 130 comments. 6 Additional stakeholder workshops and events were also organised, the results of which support the analysis and the proposals made in this impact assessment. 7

1.1.Technological context

8 9 10 Today, AI is one of the most vibrant domains in scientific research and innovation investment around the world. Approaches and techniques differ according to fields, but overall AI is best defined as an emerging general-purpose technology: a very powerful family of computer programming techniques that can be deployed for desirable uses, as well as more harmful ones. The precise definition of AI is highly contested. In 2019, the Organisation for Economic Co-operation and Development (OECD) adopted the following definition of an AI system: ‘An AI system is a machine-based system that can, for a given set of human-defined objectives, make predictions, recommendations, or decisions influencing real or virtual environments. AI systems are designed to operate with varying levels of autonomy.’ 

11 12 13 The OECD Report on Artificial Intelligence in Society provides a further explanation on what an AI system is. An AI system, also referred to as ‘intelligent agent’, “consists of three main elements: sensors, operational logic and actuators. Sensors collect raw data from the environment, while actuators act to change the state of the environment. Sensors and actuators are either machines or humans. The key power of an AI system resides in its operational logic. For a given set of objectives and based on input data from sensors, the operational logic provides output for the actuators. These take the form of recommendations, predictions or decisions that can influence the state of the environment.”

Figure 1: A high-level conceptual view of an AI system

Source: OECD, ReportArtificial Intelligence in Society, p.23.

14 15 16 17 AI systems are typically software-based, but often also embedded in hardware-software systems. Traditionally AI systems have focused on ‘rule-based algorithms’ able to perform complex tasks by automatically executing rules encoded by their programmers. However, recent developments of AI technologies have increasingly been on so called ‘learning algorithms’. In order to successfully ‘learn’, many machine learning systems require substantial computational power and availability of large datasets (‘big data’). This is why, among other reasons, despite the development of ‘machine learning’ (ML), AI scientists continue to combine traditional rule-based algorithms and ‘new’ learning based AI techniques. As a result, the AI systems currently in use often include both rule-based and learning-based algorithms.

1.2. Socio-economic context

18 The use of AI systems leads to important breakthroughs in a number of domains. By improving prediction, optimising operations and resource allocation, and personalizing service delivery, the use of AI can support socially and environmentally beneficial outcomes and provide key competitive advantages to companies. The use of AI systems in healthcare, farming, education, infrastructure management, energy, transport and logistics, public services, security, and climate change mitigation, can help solve complex problems for the public good. Combined with robotics and the Internet of Things (IoT), AI systems are increasingly acquiring the potential to carry out complex tasks that go far beyond human capacity. 

19 20 A recent Nature article found that AI systems could enable the accomplishment of 134 targets across all the Sustainable Development Goals, including finding solutions to global climate problems, reducing poverty, improving health and the quality and access to education, and making our cities safer and greener. In the ongoing Covid-19 pandemic, AI systems are being used, for example, in the quest for vaccines, in disease detection via pattern recognition using medical imagery, in calculating probabilities of infection, or in emergency response with robots replacing humans for high-exposure tasks in hospitals. This example alone already indicates the breadth of possible benefits of AI systems. Other practical applications further show how citizens can reap a lot of benefits whern accessing improved services such as personalised telemedicine care, personalised tutoring tailored to each student, or enhanced security through applications that ensure more efficient protection against cybersecurity risks.

21 22 23 24 25 26 The successful uptake of AI technologies also has the potential to accelerate Europe’s economic growth and global competitiveness. McKinsey Global Institute estimated that by 2030 AI technologies could contribute to about 16% higher cumulative global gross domestic product (GDP) compared with 2018, or about 1.2% additional GDP growth per year. AI systems and the new business models they enable are progressively developing to at-scale deployment. Accordingly, those AI systems will increasingly impact all sectors of the economy. The International Data Corporation AI market development forecast suggests that global revenues for the AI market are expected to double and surpass USD 300 billion by as early as 2024. Many businesses in various sectors of the EU economy are already seizing these opportunities. In addition to the ICT sector, the sectors using AI most intensively are education, health, social work and manufacturing. However, Europe is home to only 3 of the top 25 AI clusters worldwide and has only a third as many AI companies per million employees as the US.

Table 1: AI technologies adopted in European businesses

AI Technologies

Currently use it

Plan to use it

Process or equipment optimisation

13%

11%

Anomaly detection

13%

7%

Process automation

12%

11%

Forecasting, price-optimisation and decision-making

10%

10%

Natural language processing

10%

8%

Autonomous machines

9%

7%

Computer vision

9%

7%

Recommendation/personalisation engines

n/a

7%

Creative and experimentation activities

7%

4%

Sentiment analysis

3%

3%

27 Source: Ipsos Survey, 2020 

28 29 The same elements and techniques that power socio-economic benefits of AI systems can also bring about risks or negative consequences for individuals or for society as a whole. For example, deployment of AI systems may be intentionally used by a developer or an operator to deceive or manipulate human choices, or altogether disable human agency, control and intermediation. This possible use of AI could have strong negative consequences for the protection of fundamental rights and for human safety. In the world of work, AI systems could also undermine the effective enforcement of labour and social rights.

30 In light of the speed of technological change and possible challenges, the EU is committed to strive for a balanced approach. European Commission President von der Leyen stated: ‘In order to release that potential we have to find our European way, balancing the flow and wide use of data while preserving high privacy, security, safety and ethical standards.’

1.3. Legal context

European Union law does not have a specific legal framework for AI. Thus, as it currently stands, EU law does not provide for a definition of an AI system, nor for horizontal rules related to the classification of risks related to AI technologies. The development and uptake of AI systems more broadly, as outlined in this section, takes place in the context of the existing body of EU law that provides non-AI specific principles and rules on protection of fundamental rights, product safety, services or liability issues.

1.3.1. Relevant fundamental rights legislation

31 The Union is founded on the values of human dignity and respect of human rights that are further specified in the EU Charter of Fundamental Rights (the Charter). The provisions of the Charter are addressed to the institutions and bodies of the Union and to the Member States only when they are implementing Union law. Some fundamental rights obligations are further provided for in EU secondary legislation, including in the field of data protection, non-discrimination and consumer protection. This body of EU secondary legislation is applicable to both public and private actors whenever they are using AI technology. 

32 33 34 35 In this context, the EU acquis on data protection is particularly relevant. The General Data Protection Regulation and the Law Enforcement Directive aim to protect the fundamental rights and freedoms of natural persons, and in particular their right to the protection of personal data, whenever their personal data are processed. This covers the processing of personal data through ‘partially or solely automated means’, including any AI system. Users that determine the purpose and means of the AI processing (‘data controllers’) have to comply with a number of data processing principles such as lawfulness, transparency, fairness, accuracy, data minimization, purpose and storage limitation, confidentiality and accountability. On the other hand, natural persons, whose personal data are processed, have a number of rights, for instance, the right to access, correction, not to be subject to solely automated decision-making with legal or similarly significant effects unless specific conditions apply. Stricter conditions also apply for the processing of sensitive data, including biometric data for identification purposes, while processing that poses high risk to natural persons’ rights and freedoms requires a data protection impact assessment.

36 37 Users of AI systems are also bound by existing equality directives. The EU equality acquis prohibits discrimination based on a number of protected grounds (such as racial and ethnic origin, religion, sex, age, disability and sexual orientation) and in specific context and sectors (for example, employment, education, social protection, access to goods and services). This existing acquis has been complemented with the new EU Accessibility Act setting requirements for the accessibility of goods and services, to become applicable as of 2025. 

38 Consumer protection law and obligations to abstain from any unfair commercial practices listed in the Unfair Commercial Practice Directive are also highly relevant for businesses using AI systems.

Furthermore, EU secondary law in the area areas of asylum, migration, judicial cooperation in criminal matters, financial services and online platforms is also relevant from a fundamental rights perspective when AI is developed and used in these specific contexts.

1.3.2. Relevant product safety legislation

39 40 41 42 43 In addition, there is a solid body of EU secondary law on product safety. The EU safety legislation aims to ensure that only safe products are placed on the Union market. The overall EU architecture on safety is based on the combination of horizontal and sectoral rules. This includes the General Product Safety Directive (GPSD) applicable to consumer products, insofar there are not more specific provisions in harmonised sector-specific safety legislation, as for example, the Machinery Directive (MD), the Medical Device Regulation (MDR) and the EU framework on the approval and market surveillance of motor vehicles (in particular Vehicle Safety Regulation). 

44 45 Reviews of both the MD and the GPSD are currently under way. Those reviews aim to respond, among other things, to the challenges of new technologies, such as IoT, robotics and AI. In addition, delegated acts are expected to be soon adopted by the Commission under the Radio Equipment Directive to enact certain new requirements on data protection and privacy, cybersecurity and harm to the network. Moreover, in the automotive sector new rules on automated vehicles, cybersecurity and software updates of vehicles will become applicable as part of the vehicle type approval and market surveillance legislation from 7 July 2022.

While the European  identifies the review of the General Product Safety Directive, the Machinery Directive and the Radio Equipment Directive as priorities, other pieces of product legislation may well be updated in the future in order to address existing gaps linked to new technologies.

46 The product safety legislation is technology-neutral and focuses on the safety of the final product as a whole. The revisions of the product safety legislation do not have the objective to regulate AI as such, but aim primarily at ensuring that the integration of AI systems into the overall product will not render a product unsafe and the compliance with the sectoral rules will not be affected.

1.3.3. Relevant liability legislation

Safety legislation sets rules to ensure that products are safe and safety risks are addressed, nevertheless, damages can still occur. For that purpose, the liability rules at national and EU level complement the safety legislation and determine which party is liable for harm, and under which conditions a victim can be compensated. A longstanding approach within the EU with regard to product legislation is based on a combination of both safety and liability rules. In practice, while being driven by different regulatory rationales and objectives, safety and liability initiatives are essential and complementary in nature.

47 48 49 50 At EU level, the Product Liability Directive (PLD) is currently the only EU legal framework that harmonizes part of national liability law, introducing a system of ‘strict’ liability without fault of the producer for physical or material damage caused by a defect in products placed on the Union market. While the PLD provides legal certainty and uniform consumer protection as a safety net applicable to all products, its rules also face increasing challenges posed by emerging technologies, including AI. As part of the upcoming revision, the Commission will be exploring several options to adapt the EU product liability rules to the digital world, for instance by adapting the definition of product and producer, or by extending the application of the strict liability regime to certain other types of damages (e.g. to privacy or personal data). The Commission will also explore options on how to address the current imbalance between consumers and producers by reversal or alleviation of the burden of proof (with access to information and presumption of defectiveness under certain circumstances), and explore the abolishment of existing timelines and threshold (€500). At national level, non-harmonised civil liability frameworks complement these Union rules by ensuring compensation for damages from products and services and by addressing different liable persons. National liability systems usually include fault-based and strict liability regimes. 

51 52 In order to ensure that victims who suffer damage to their life, health or property as a result of AI technologies have access to the same compensation as victims of other technologies, the Commission has announced possible revision of rules on liability. The main objective of the revision is to ensure that damages caused by AI systems are covered. In order to achieve this objective, together with the update of the PLD, the Commission is also considering possible new AI-specific rules harmonising certain aspects of national civil liability frameworks with regard to the liability for certain AI systems. In particular, options which are currently under evaluation include the possible setting of strict liability for AI operators, possibly combined with mandatory insurance for AI applications with a specific risk profile as well as adaptation of burden of proof concerning causation and fault for all other AI applications. 

In addition to the general review of liability rules, the Commission is examining liability challenges which are specific to certain sectors, such as health-care, and which may deserve specific considerations.

53 54 The relationship between the proposed European legal framework for AI analysed in this impact assessment and the forthcoming new rules on liability is further discussed under the preferred option in section 8. In terms of timing for the adoption of the new rules on liability, the Commission decided for a staged approach. First, the Commission will propose in Q2 2021 the AI horizontal framework (the current initiative) and then, the EU rules to address liability issues related to new technologies, including AI systems (expected Q4 2021 – Q1 2022). The future changes to the liability rules will take into account the elements of the horizontal framework with a view to designing the most effective and proportionate solutions with regard to liability. Moreover, compliance with the requirements of the AI horizontal framework will be taken into account for assessing liability of actors under future liability rules.

With regard to intermediary liability, for example when sellers place faulty products through online marketplaces, the E-Commerce Directive regulates the liability exemptions for online intermediaries. This framework is currently updated in the Commission’s proposal for a Digital Services Act.

1.3.4. Relevant legislation on services

55 The EU also has a comprehensive legal framework on services that is applicable whenever AI software is provided as a stand-alone service or integrated into other services, including in particular digital services, audio-visual media services, financial services, transport services, professional services and others. In the context of information society services, the e-Commerce Directive provides for an applicable regulatory framework laying down horizontal rules for provisions of such services in the Union. The proposed Digital Services Act includes rules on liability exemption for providers of intermediary services (i.e. mere conduit; caching; hosting), which are to date contained in the e-Commerce Directive that remains largely unchanged and fully applicable. At the same time, the Digital Services Act introduces due diligence obligations for providers of intermediary services so as to keep users safe from illegal goods, content or services and to protect their fundamental rights online. These due diligence obligations are adapted to the type and nature of the intermediary service concerned and transparency and accountability rules will apply for algorithmic systems, including those based on AI, used by online platforms.

56 57 58 In the field of financial services, the risk governance requirements under the existing legislation provide a strong regulatory and supervisory framework for assessment and management of risks. Specific rules additionally apply in relation to trading algorithms. With respect to creditworthiness assessment, European Banking Authority guidelines have been recently adopted to improve regulated financial institutions’ practices and associated governance arrangements, processes and mechanisms in relation to credit granting, management and monitoring, including when using automated models in the creditworthiness assessment and credit decision-making processes. However, no sector-specific EU guidance or rules currently apply to non-regulated entities when assessing the creditworthiness of consumers.

1.4. Political context

59 60 61 62 63 64 To address the opportunities and challenges of AI, in April 2018 the European Commission put forward a European approach to AI in its Communication “Artificial Intelligence for Europe.” In June 2018, the Commission appointed the High-Level Expert Group on Artificial Intelligence, which produced two deliverables: the Ethics guidelines for trustworthy AI and the Policy and investment recommendations for trustworthy AI. In December 2018, the Commission presented a Coordinated Plan on AI with Member States to foster the development and use of AI. 

65 66 In June 2019, in its Communication “Building Trust in Human-Centric Artificial Intelligence”, the Commission endorsed the seven key requirements for Trustworthy AI identified by the HLEG. After extensive consultation, on 17 July 2020, the HLEG published an Assessment List for Trustworthy Artificial Intelligence for self-assessment (ALTAI) which was tested by over 350 organisations.

67 68 In February 2020, the European Commission published a White Paper on AI setting out policy options for a regulatory and investment oriented approach. It was accompanied by a Commission Report on the safety and liability implications of AI. The White Paper opened a wide public consultation where more than 1 215 contributions were received from a wide variety of stakeholders, including representatives from industry, academia, public authorities, international organisations, standardisation bodies, civil society organisations and citizens. This clearly showed the great interest from stakeholders around the globe in shaping the future EU regulatory approach to AI, as assessed in this impact assessment.

The European Council and the European Parliament (EP) also repeatedly called for the Commission to take legislative action to ensure a well-functioning internal market for AI systems where both benefits and risks are adequately addressed at EU level.

69 70 71 In 2017, the European Council called for a ‘sense of urgency to address emerging trends’ including ‘issues such as artificial intelligence …, while at the same time ensuring a high level of data protection, digital rights and ethical standards’. In its 2019 Conclusions on the Coordinated Plan on the development and use of artificial intelligence Made in Europe, the Council further highlighted the importance of ensuring that European citizen's rights are fully respected and called for a review of the existing relevant legislation to make it fit for purpose for the new opportunities and challenges raised by AI. The European Council has also called for a clear determination of what should be considered as high-risk AI applications. 

72 The most recent Conclusions from 21 October 2020 further called for addressing the opacity, complexity, bias, a certain degree of unpredictability and partially autonomous behaviour of certain AI systems, to ensure their compatibility with fundamental rights and to facilitate the enforcement of legal rules.

73 74 75 In 2017, the European Parliament (EP) adopted a Resolution on Civil Law Rules on Robotics urging the Commission to analyse the impact of use of AI technologies in the main areas of EU legislative concern, including ethics, liability, standardisation and institutional coordination and oversight, and to adopt legislation where necessary. In 2019, the EP adopted a Resolution on a Comprehensive European Industrial Policy on Artificial Intelligence and Robotics. In June 2020, the EP also set up a Special Committee on Artificial Intelligence in a Digital Age (AIDA) tasked to analyse the future impact of AI systems in the digital age on the EU economy and orient future EU priorities. 

76 77 78 79 80 81 82 In October 2020, the EP adopted a number of resolutions related to AI, including on ethics, liability and copyright. EP resolutions on AI in criminal matters and AI in education, culture and the audio-visual sector are forthcoming. The EP Resolution on a Framework of Ethical Aspects of Artificial Intelligence, Robotics and Related Technologies, specifically recommends to the Commission to propose a legislative action to harness the opportunities and benefits of AI, but also to ensure protection of ethical principles. The EP resolution on internal market aspects of the Digital Services Act presents the challenges identified by the EP as regards AI-driven services. 

83 84 85 86 87 88 At international level, the ramifications in the use of AI systems and related challenges have also received significant attention. The Council of Europe started work on an international legal framework for the development, design and application of AI, based on the Council of Europe’s standards on human rights, democracy and rule of law. It has also recently issued guidelines and proposed safeguards and certain prohibitions of the use of facial recognition technology considered particularly intrusive and interfering with human rights. The OECD adopted a Council Recommendation on Artificial Intelligence. The G20 adopted human-centred AI Principles that draw on the OECD AI Principles. UNESCO is also starting to develop a global standard setting instrument on AI. Furthermore, the EU, together with many advanced economies, set up the Global Partnership on Artificial Intelligence (GPAI). As part of the EU-Japan Partnership for Sustainable Connectivity and Quality Infrastructure, concluded in September 2019, both sides reconfirmed their intention to continue promoting policies that boost innovation including in Artificial Intelligence, cloud, quantum computing and blockchain.

In addition to EU and international initiatives, many countries around the world started to consider adopting their own ethical and accountability frameworks on AI and/or automated decision-making systems. In Canada, a Directive on Automated Decision-Making came into effect on April 1, 2020 and applies to the use of automated decision systems in the public sector that “provide external services and recommendations about a particular client, or whether an application should be approved or denied.” The Directive includes an Algorithmic Impact Assessment and transparency obligations vis-à-vis persons affected by the automated decision. In 2020, the Government of New Zealand, together with the World Economic Forum, was spearheading a multi-stakeholder, policy project, structured around three focus areas: 1) obtaining of a social licence for the use of AI through an inclusive national conversation; 2) the development of in-house understanding of AI to produce well-informed policies; and 3) the effective mitigation of risks associated with AI systems to maximize their benefits. In early 2020, the United States’ government adopted overall regulatory principles. On this basis, the White House released the first-ever guidance for Federal agencies on the regulation of artificial intelligence applications in the public sector that should comply with key principles for Trustworthy AI. 89 Other countries with regulatory initiatives on AI include, for example, Singapore, Japan, Australia, the United Kingdom and China. 90 Annex 5.1 summarises the main ongoing initiatives in these third countries undertaken to address the challenges posed by AI and to harness its potential for good.

1.5. Scope of the impact assessment

This report assesses the case for an EU regulatory framework for the development and use of AI systems and examines the impact of different policy options. The use of AI for exclusive military purposes remains outside the scope of the present initiative due to its implications for the Common Foreign and Security Policy (CFSP). 91 Insofar as ‘dual use’ products and technologies 92 have AI features and can be used for both military and civil purposes, these goods will fall into the scope of the current initiative on AI.

The forthcoming initiative on liability and the ongoing revisions of sectoral safety legislation are subject to separate initiatives and remain equally outside the scope of this impact assessment, as discussed in section 1.3 above.

2. Problem definition

2.1.What are the problems?

93 The analysis of the available evidence suggests that there are six main related problems triggered by the development and use of AI systems that the current initiative aims to address.

Table 2: Main problems

MAIN PROBLEMS

STAKEHOLDERS CONCERNED

1.Use of AI poses increased risks to safety and security of citizens

Citizens, consumers and other victims

Affected businesses

2.Use of AI poses increased risk of violations of citizens’ fundamental rights and Union values

Citizens, consumers and other victims

Whole groups of the society,

Users of AI systems liable for fundamental rights violations

3.Authorities do not have powers, procedural frameworks and resources to ensure and monitor compliance of AI development and use with applicable rules

National authorities responsible for compliance with safety and fundamental rights rules

4.Legal uncertainty and complexity on how existing rules apply to AI systems dissuade businesses from developing and using AI systems

Businesses and other providers developing AI systems

Businesses and other users using AI systems

5.Mistrust in AI would slow down AI development in Europe and reduce the global competitiveness of the EU economy

Businesses and other users using AI systems

Citizens using AI systems or being affected by them

6.Fragmented measures create obstacles for cross-border AI single market and threaten Union’s digital sovereignty

Businesses developing AI, mainly SMEs affected

Users of AI system, including consumers, businesses and public authorities

Problem 1: The use of AI poses increased risks to safety and security of citizens

The overall EU architecture of safety frameworks is based on a combination of horizontal and sectoral rules. 94 This includes the horizontal GPSD and sector-specific legislation, as for example, the Machinery Directive. The EU safety legislation has significantly contributed to the high-level of safety of products put into circulation in the EU Single Market. However, it is increasingly confronted with the challenges posed by new technologies, some of which specifically relate to AI technologies. 95  

Two main reasons explain the limitations of the existing EU safety and security framework in relation to the application to AI technologies.

96 97 98 Firstly, the nature of safety risks caused by AI: The main causes of safety risks generated by complex software or algorithms are qualitatively different from risks caused by physical products. The specific characteristics of AI systems, could lead to safety risks including:

·Biases in data or models: training data can have hidden biases that can be ‘learnt’ by the AI model and reproduced in its outputs. AI algorithms can also introduce biases in their reasoning mechanisms by preferring certain characteristics of the data. For example, a medical imaging recognition device trained on data biased towards a specific segment of population may lead in certain cases to safety risks due to misdiagnosis in patients not well represented by the data.

·Edge cases: unexpected or confusing input data can result in failures due to the limited ability that AI models might exhibit to generalise well from training data. For example, image recognition systems in an autonomous vehicle could malfunction in the case of unexpected road situations, endangering passengers and pedestrians.

·Negative side effects: the realization of a task by an autonomous AI system could lead to harmful effects if the scope of action of the system is not correctly defined and does not consider the context of use and the state of the environment. For instance, an industrial robot with an AI system designed to maximize the work rate in a workshop could damage property or accidentally hurt people in situations not foreseen in its design.

99 100 AI models which are run on top of ICT infrastructure are composed of a diverse range of digital assets. Therefore, cybersecurity issues affecting this broader digital ecosystem also extend to AI systems and can result in important AI safety risks. This is particularly relevant considering the new systems enabled by AI, such as cyber-physical systems, where cybersecurity risks have direct safety implications.

101 Moreover, AI systems can also be subject to malicious attempts to exploit AI specific vulnerabilities, including: 

·Evasion – an attack when an attacker modifies input data, sometimes in an imperceptible manner, so that the AI model cannot correctly identify the input and this leads to wrong outputs. Examples include attacks to evade anti-spam filters, spoofing attacks against biometric verification systems or stickers added to stop signs to make autonomous vehicles to perceive them as speed signs.

·Data poisoning – an action aiming to modify the behaviour of the AI model by altering the training datasets, especially when the data used is scraped from the web, sourced from data exchanges or from open datasets. The ‘learning’ systems where model parameters are constantly updated using new data, are particularly sensitive to data poisoning. 102  For example, poisoning the data used to train a chatbot could make it disclose sensitive information or adopt inappropriate behaviour.

·Model extraction – attacks that aim to build a surrogate system that imitates the targeted AI model. The goal is to get access to a representation of the AI model that will allow the attacker to understand and mimic the logic of the system and possibility build more sophisticated attacks, like evasion or data poisoning or steal sensitive information from training data.

·Backdoor – refers to a typical risk in programming which is not limited to AI applications, but is more difficult to detect and avoid in the case of AI due to its opacity. 103 The presence of a so-called ‘backdoor’ 104 makes unauthorised access to a system possible.

105 AI specific risks are not or are only partly covered by the current Union safety and security legislation. While Union safety legislation covers more generally the risks stemming from software being a safety component (and usually embedded) in a hardware product, stand-alone software (except in the medical device framework) – including when used in services – or software uploaded to a hardware device after this device is placed on the market are not currently covered. Thus, services based on AI technology, such as transport services or infrastructure management, are not covered.

106 107 Moreover, even when software is covered by EU safety legislation, no specific safety or performance requirements are set for AI systems. For example, there are no AI-technology specific requirements ensuring reliability and safety over its lifecycle. The EU legal framework on cybersecurity applies to ICT products, services and processes, and therefore could also potentially cover AI technologies. However, no scheme for AI currently exists and there are no established AI cybersecurity standards of best practice for developers in the design phase, notably when it comes to ‘security by design’ for AI. Moreover, the certification schemes for cybersecurity are of a voluntary nature. 

This lack of clear safety provisions covering specific AI risks, both for AI systems being safety components of products and AI systems used in services, can be highly problematic for users and consumers of AI applications.

Secondly, the lifecycle of an AI product: Under the current legal framework, ex-ante conformity assessment procedures are mainly conceptualized for products that are ‘stable’ in time after deployment. The current safety legislation does not contain specific provisions for products that are possibly subject to evolution during their lifecycle. Yet, certain AI systems are subject to considerable change after their first deployment on the market.

Such potential and unanticipated modifications to the performance of those AI systems after their placement on the market could still occur and cause injuries and physical damage. 108 A product that is already on the market is normally required to undergo a new conformity assessment if there are substantial modifications of the product. However, the exact interpretation of the notion of “substantial modification” in an AI context needs to be clearly defined, also in light of the complexity of the AI value chain, 109 lest it should lead to legal uncertainty.

As a general conclusion, the specificities of AI applications might require the establishment of some specific safety requirements to ensure a high level of protection of users and consumers as well as to provide legal certainty for businesses, notably for use cases where the lack of proper performance or reliability of AI could have a severe impact on life and health of individuals. This is regardless of whether AI systems are a safety component of products or are used in services.

Stakeholders views: In the Public consultation on White Paper on AI, 83% of all respondents consider that the fact that AI may endanger safety’ is ‘important’ (28%) or ‘very important’ (55%). Among SMEs, 72% found safety to be an important or very important concern, whereas only 12% said it was not important or not important at all. This position was even more pronounced among large businesses, with 83% saying that safety was (very) important and only 4% finding the issue unimportant. 80% of academic and other research institutions and 88% of civil society organisations agreed that safety was a (very) important concern. Among EU citizens, 73% found safety to be an important or very important issue. Of those stakeholders who said safety was not a (very) important concern, 43% were EU citizens (which make up 35% of all respondents) and 20% were SMEs (7%).

Problem 2: Use of AI poses an increased risk of citizens’ fundamental rights and Union values violations

110 111 The use of AI can have a significant impact on virtually all fundamental rights as enshrined in the EU Charter of Fundamental Rights. AI use can be positive, promoting certain rights (e.g. the right to education, health care) and contributing to important public interests (e.g. public security, public health, protection of financial interests). It can also help reduce some adverse impacts on fundamental rights by improving the accuracy or efficiency of decision-making processes and addressing biases, delays or errors in individual human decisions. On the other hand, AI used to replace or support human decision-making or for other activities such as surveillance may also infringe upon individual’s rights. This is not a flaw of the technology per se, but the responsibility of the humans who are designing and using it and who must ensure these violations do not happen in the first place.

If breaches of fundamental rights do happen, these can also be very difficult to detect and prove, especially when the system is not transparent. This challenges the effective enforcement of the existing EU legislation aimed at safeguarding fundamental rights, as listed in section 1.3.

112 113 114 While it remains difficult to quantify the real magnitude of the risks to fundamental rights, a growing amount of evidence suggests that Union citizens might be affected in an increasingly wide range. Moreover, a growing body of case law and legal challenges to the use of AI breaching fundamental rights is also emerging across different Member States. The following sections will focus on some of the most prominent fundamental rights risks. 

2.1.1. Use of AI may violate human dignity and personal autonomy

The right to human dignity is an inviolable right that requires every individual to be treated with respect as a human being and not as a mere ‘object’ and their personal autonomy respected. Depending on the circumstances of the case and the nature of the interaction, this might be challenging if people are misled in believing that they are interacting with another person when they are actually interacting with an AI system.

115 Moreover, AI is often used to sort and classify traits and characteristics that emanate from datasets that are not based on the individual concerned. Organisations that use such data to determine individual’s status as a victim (e.g. women at risk of domestic violence) or individual’s likelihood to reoffend in case of predictive risk assessments could violate individuals’ right to human dignity since the assessment is no longer based on the personal individual situation and merits.

116 117 118 AI can also be used for manipulation that can be particularly harmful for certain users. While psychological science shows that these problems are not new, the growing manipulative capabilities of algorithms that collect and can predict very sensitive and privacy intrusive personal information can make people extremely vulnerable, easily deceived or hyper-nudged towards specific decisions that do not align with their goals or run counter to their interests. Evidence suggests that AI supported products or services (toys, personal assistants etc.) can be intentionally designed or used in ways that appeal to the subliminal perception of individuals, thus causing them to take decisions that are beyond their cognitive capacities. Even if the techniques used are not subliminal, for certain categories of vulnerable subjects, in particular children, these might have the same adverse manipulative effects if their mental infirmity, age or credulity are exploited in harmful ways.  As the AI application areas develop, these (mis)uses and risks will likely increase.

2.1.2. Use of AI may reduce privacy protection and violate the right to data protection

119 The EU has a strong and modern legal framework on data protection with the Law Enforcement Directive and the General Data Protection Regulation recently evaluated as fit for purpose. Still, the use of AI systems might challenge the effective protection of individuals since the right to private life and other fundamental rights violations can occur even if non-personal, including anonymized data is processed.

120 The arbitrary use of algorithmic tools gives unprecedented opportunities for indiscriminate or mass surveillance, profiling and scoring of citizens and significant intrusion into people’s privacy and other fundamental rights. Beyond affecting the individuals concerned, such use of technology has also an impact on society as a whole and on broader Union values such as democracy, freedom, rule of law, etc. 

121 122 123 124 125 A particularly sensitive case is the increasing use of remote biometric identification systems in publicly accessible spaces. Currently, the most advanced variety of this family of applications is facial recognition, but other varieties exist, such as gait recognition or voice recognition. Wherever such a system is in operation, the whereabouts of persons included in the reference-database can be followed, thus impacting their personal data, privacy, autonomy and dignity. Moreover, freedom of expression, association and assembly might be undermined by the use of the technology resulting in a chilling effect on democracy. On the other hand, the use of such systems has been considered by some justified in limited cases when strictly necessary and proportionate for safeguarding important public interests of public security. Public and private operators are already using such systems in Europe, but because of privacy and other fundamental rights violations, their operation has been blocked by data protection authorities in schools or in other publicly accessible spaces. Despite these serious concerns and potential legal challenges, many countries consider using biometric identification systems at a much larger scale to cope with increasing security risks. 

Apart from identification, facial, gait, iris or voice recognition technology is also used to attempt to predict individual’s characteristics (e.g. sex, race or even sexual orientation), emotions and to detect whether people are lying or telling the truth. 126  Biometrics for categorisation and emotion recognition might lead to serious infringements of peoples’ privacy and their right to the protection of personal data as well as to their manipulation. In addition, there are serious doubts as to the scientific nature and reliability of such systems. 127  

128 While EU data protection rules in principle prohibit the processing of biometric data for the purpose of uniquely identifying a natural person except under specific conditions permitted by law, the White Paper on AI opened a discussion on the specific circumstances, if any, which might justify such use, and on common safeguards.

Stakeholders views: In a recent survey, between 45% and 60% of consumers believed that AI will lead to more abuse of personal data (BEUC, Consumers see potential of artificial intelligence but raise serious concerns, 2020). 76.7 % of respondents to the White Paper on AI consider that the systems for remote biometric identification in public spaces have to be regulated in one way or another, 28.1% consider that they should never be authorized at publicly accessible spaces. Recently, 12 NGOs also started an EU-wide campaign called ‘Reclaim Your Face’ to urge EU to ban facial recognition in public spaces. The Commission has also registered a European Citizens' Initiative entitled ‘Civil society initiative for a ban on biometric mass surveillance practices'. Stakeholders views: In a recent survey, between 45% and 60% of consumers believed that AI will lead to more abuse of personal data (BEUC, Consumers see potential of artificial intelligence but raise serious concerns, 2020). 76.7 % of respondents to the White Paper on AI consider that the systems for remote biometric identification in public spaces have to be regulated in one way or another, 28.1% consider that they should never be authorized at publicly accessible spaces. Recently, 12 NGOs also started an EU-wide campaign called ‘Reclaim Your Face’ to urge EU to ban facial recognition in public spaces. The Commission has also registered a European Citizens' Initiative entitled ‘Civil society initiative for a ban on biometric mass surveillance practices'.

2.1.3. Use of AI may lead to discriminatory outcomes

129 130 131 132 Algorithmic discrimination can occur for several reasons at many stages and it is often difficult to detect and mitigate. Problems may arise due to flawed design and developers who unconsciously embed their own biases and stereotypes when making the classification choices. Users might also misinterpret the AI output in concrete situations or use it in a way that is not fit for the intended purpose. Moreover, bias causes specific concerns for AI techniques dependent on data, which might be unrepresentative, incomplete or contain historical biases that can cement existing injustices with the ‘stamp’ of what appears to be scientific and evidence-based legitimacy. Developers or users could also intentionally or unintentionally use proxies that correlate with protected characteristics under EU non-discrimination legislation such as race, sex, disability etc. Although being based on seemingly neutral criteria, this may disproportionately affect certain protected groups giving rise to indirect discrimination (e.g., using proxies such as postal codes to account for ethnicity and race). As explained in the driver section 2.2., the algorithms can also introduce themselves biases in their reasoning mechanisms by favouring certain characteristics of the data on which they have been trained. Varying levels of accuracy in the performance of AI systems may also disproportionately affect certain groups, for example facial recognition systems that detect gender well for white men, but not for black women or that do not detect as person those using wheelchairs.

The use of discriminatory AI systems notably in sectors such as employment, public administration, judiciary or law enforcement, might also violate many other fundamental rights (e.g. right to education, social security and social assistance, good administration etc.) and lead to broader societal consequences, reinforcing existing or creating new forms of structural discrimination and exclusion. 

133 134 135 136 137 138 For example, evidence suggests that in the employment sector AI is playing an increasingly key role in making hiring decisions mainly facilitated by intermediary tech service providers. This can negatively affect potential candidates in terms of discriminatory filtering at different moments of recruitment procedures or afterwards. Another problematic area is the administration of social welfare assistance, with some recent cases of suspected discriminatory profiling of unemployed people in Denmark, Poland or Austria. Financial institutions and other organisations might also use AI for assessing individual’s creditworthiness to support decisions determining the access to credit and other services such as housing. While this can increase the opportunities for some people to get access to credit on the basis of more diverse data points, there is also a risk that systems for assessing scores might unintentionally induce biases, if not properly designed and validated. In law enforcement and criminal justice, AI models trained with past data can be used to forecast trends in the development of criminality in certain geographic areas, to identify potential victims of criminal offences such as domestic violence or to assess the threats posed by individuals to commit offences based upon their criminal records and overall behaviour. In the EU, cases of these predictive policing systems exist in a number of Member States. At the borders, specific groups such as migrants and asylum seekers can also have their rights significantly affected if discriminatory AI systems are used by public authorities. 

139 Under the existing EU and national anti-discrimination law, it could be very difficult to launch a complaint as the affected person most likely do not know that an AI systems is used and even if they do, they are not aware how it functions and how its outputs are applied in practice. This makes it very difficult, if not impossible, for the persons concerned to establish the facts needed to establish prima facie discrimination, or prove it. It might also be very challenging for supervisory authorities and courts to detect and assess discrimination, in particular in cases when there is no readily available and relevant statistical evidence.

Stakeholders views: 2020 BEUC survey - Consumers see potential of artificial intelligence but raise serious concerns: In a recent consumer organization survey in nine Member States, between 37% and 51% of respondents agree or strongly agree that AI will lead to unfair discrimination based on individual characteristics or social categories. In the public consultation on the AI White Paper, 89% of all respondents found that AI leading to discriminatory outcomes is an important or very important concern. This was a (very) important concern for 76% of SMEs and only 5% found it not important (at all). Large businesses were even more concerned: 89% said discrimination was a (very) important concern. Similarly, 91% of academic and research institutions and 90% of civil society organisations thought this was (very) important concern. Meanwhile, EU citizens were less concerned, although 78% still found this to be (very) important. Of those stakeholders stating that discriminatory outcomes were not important or very important concern, EU citizens (35%), academic and research institutions (19%) and SMEs (15%) were represented the most. For academic and research institutions and SMEs, this share was significantly larger than their representation in the overall sample. Stakeholders views: 2020 BEUC survey - Consumers see potential of artificial intelligence but raise serious concerns: In a recent consumer organization survey in nine Member States, between 37% and 51% of respondents agree or strongly agree that AI will lead to unfair discrimination based on individual characteristics or social categories. In the public consultation on the AI White Paper, 89% of all respondents found that AI leading to discriminatory outcomes is an important or very important concern. This was a (very) important concern for 76% of SMEs and only 5% found it not important (at all). Large businesses were even more concerned: 89% said discrimination was a (very) important concern. Similarly, 91% of academic and research institutions and 90% of civil society organisations thought this was (very) important concern. Meanwhile, EU citizens were less concerned, although 78% still found this to be (very) important. Of those stakeholders stating that discriminatory outcomes were not important or very important concern, EU citizens (35%), academic and research institutions (19%) and SMEs (15%) were represented the most. For academic and research institutions and SMEs, this share was significantly larger than their representation in the overall sample.

2.1.4. Use of AI might violate the right to an effective remedy, fair trial and good administration

140 141 One prominent threat to the right to an effective remedy is the lack of transparency in the use and operation of AI systems. Without access given to relevant information, individuals may not be able to defend themselves and challenge any decision taken or supported by AI systems that might adversely affect them. This jeopardizes their right to be heard as well as the right to an effective remedy and fair trial. 

142 Furthermore, the use of automated decision-making in judicial proceedings might particularly affect the right of affected persons to access to court and to fair trial, if these systems are not subject to appropriate safeguards for transparency, accuracy, non-discrimination and human oversight. 

143 144 145 146 The opacity of AI could also hamper the ability of persons charged with a crime to defend themselves and challenge the evidence used against them. In the context of AI-enabled individual risk assessments increasingly used in law enforcement, singling out people without reasonable suspicion or on the basis of biased or flawed data might also threaten the presumption of innocence. Public authorities may also not be able to properly reason their individual administrative decisions which is required as part of the principle and the right to good administration.

Stakeholders views: In the public consultation on the White Paper on AI, only 2% of all respondents said that AI breaching fundamental rights is not (at all) an important concern. 85% of SMEs considered this (very) important, while none found the issue to be unimportant. Similarly, 87% large businesses found this to be (very) important– only one respondent (1%) found it not important. 93% and 94% of academic/research institutions and civil society organisations, respectively, were (very) concerned about fundamental rights breaches. EU citizens were also concerned: 83% found potential breaches of fundamental rights (very) important. Among those stakeholders who found this not to be a (very) important concern, academic and research institutions were the largest group with 33% (much higher than their 14% share of the entire sample). Stakeholders views: In the public consultation on the White Paper on AI, only 2% of all respondents said that AI breaching fundamental rights is not (at all) an important concern. 85% of SMEs considered this (very) important, while none found the issue to be unimportant. Similarly, 87% large businesses found this to be (very) important– only one respondent (1%) found it not important. 93% and 94% of academic/research institutions and civil society organisations, respectively, were (very) concerned about fundamental rights breaches. EU citizens were also concerned: 83% found potential breaches of fundamental rights (very) important. Among those stakeholders who found this not to be a (very) important concern, academic and research institutions were the largest group with 33% (much higher than their 14% share of the entire sample).

Problem 3: Competent authorities do not have powers, resources and/or procedural frameworks to ensure and monitor compliance of AI use with fundamental rights and safety rules

The specific characteristics of many AI technologies, set out in section 2.2., often make it hard to verify how outputs and decisions have been reached where AI is used. As a consequence, it may become impossible to verify compliance with existing EU law meant to guarantee safety and protect fundamental rights. For example, to determine whether a recruitment decision is justified or involved discrimination, enforcement authorities need to determine how this decision was reached. Yet, since there is no requirement for producers and users of AI systems to keep proper documentation and ensure traceability of these decision-making processes, public authorities may not be able to properly investigate, prove and sanction a breach.

The governance and enforcement mechanisms under existing sectoral legislation also suffer from shortcomings. Firstly, the use of AI systems may lead to situations where market surveillance and supervisory authorities may not be empowered to act and/or do not have the appropriate technical capabilities and expertise to inspect these systems.

Secondly, existing secondary legislation on data protection, consumer protection and non-discrimination legislation relies primarily on ex-post mechanisms for enforcement and focuses on individual remedies for ‘data subjects’ or ’consumers’. To evaluate compliance with fundamental rights, the purpose and use of an application needs to be assessed in context and it is the responsibility of every actor to comply with their legal obligations. Unless legal compliance in view of the intended purpose and context is taken into account already at the design stage, harmful AI systems might be placed on the market and violate individual fundamental rights at scale before any enforcement action is taken by competent authorities.

147 Thirdly, as set out above, the current safety legislation does not provide yet for clear and specific requirements for AI systems that are embedded in products. Outside the scope of product safety legislation, there is also no binding obligation for prior testing and validation of the systems before they are placed on the market. Moreover, after systems are placed on the market and deployed, there is no strict ex post obligation for continuous monitoring which is, however, essential given the continuous learning capabilities of certain AI system or their changing performance due to regular software updates.

148 Fourthly, the secondary legislation on fundamental rights primarily places the burden for compliance on the user and often leaves the provider of the AI system outside its scope. However, while users remain responsible for a possible breach of fundamental rights obligations, providers might be best placed to prevent and mitigate some of the risks already at an early development stage. Users are also often unable to fully understand the workings of AI applications if not provided with all the necessary information. Because of these gaps in the existing legislation, procedures by supervisory authorities may not result in useful findings.

149 Finally, given the complexity and rapid speed of AI development, competent authorities often lack the necessary resources, expertise and technological tools to effectively supervise risks posed by the use of AI systems to safety and fundamental rights. They also do not have sufficient tools for cooperation with authorities in other Member States for carrying out joint investigations, or even at national level where, for example, various sectoral legislation might intersect and lead to violations of multiple fundamental rights or to risks to both safety and fundamental rights.

Problem 4: Legal uncertainty and complexity on how to ensure compliance with rules applicable to AI systems dissuade businesses from developing and using the technology

150 Due to the specific characteristics of AI set out in section 2.2., businesses using AI technology are also facing increasing legal uncertainty and complexity on how to comply with existing legislation. Considering the various sources of risks at all different levels, organisations involved in the complex AI value chain might be unclear who exactly should ensure compliance with the different requirements under the existing legislation. 

For example, providers of AI systems might be unclear on what measures they should integrate to minimize the risks to safety and fundamental rights’ violation, while users might not be able to remedy features of the design that are inadequate for the context of application. In this context, the evolving nature of risks also pose particular problems to correctly attribute responsibilities. Providers of AI systems may have limited information with regard to the harm that AI can produce post deployment (especially if the application context has not been taken into account in the design of the system), while users may be unable to exercise due care when operating the AI system if not properly informed about its nature and provided with guidance about the required oversight and ex post control. The lack of clear distribution of obligations across the AI value chain taking into account all these specific features of the AI technology leads to significant legal uncertainty for companies, while failing to effectively minimise the increasing risks to safety and fundamental rights, as identified above.

151 Another problem is that there are no harmonized standards under EU law as to how general principles or requirements on security, non-discrimination, transparency, accuracy, human oversight should be implemented specifically as regards AI systems in the design and development stage. This results in legal uncertainty on the business side, which affects both the developer and the user of the AI system.

152 Also, there are no clear red lines when companies should not engage in the use of AI for certain particularly harmful practices beyond the practices explicitly listed in the Unfair Commercial Practice Directive. Certification for trustworthy AI product and services that are currently available on the Union market is also missing and creates uncertainties across the AI value chain.

Without a clear legal framework, start-ups and developers working in this field will not be able to attract the required investments. Similarly, without certainty on applicable rules and clear common standards on what is required for a trustworthy, safe and lawful AI, developers and providers of AI systems and other innovators are less likely to pursue developments in this field.

153 As a result, businesses and public authorities using AI technology fear becoming responsible for fundamental rights infringements, which may dissuade them from using AI. At European level, 69% of companies cite a ‘need for new laws or regulation’ as an obstacle to adoption of AI. SMEs were more likely than large companies to say that this was not a challenge or barrier, whereas small enterprises were more likely to identify this as a major barrier (31% for companies with 5 to 9 employees and 28% for those with 10 to 49 employees) compared to medium and large companies.

Figure 2: Obstacles to the use of AI (by companies)

Source: Ipsos Survey, 2020

154 At national level, a survey of 500 companies by the German Technical Inspection Association TÜV found that 42% of businesses expect legal challenges due to the use of AI, while 84% said that there was uncertainty around AI within their company. As a result, 84% wanted AI products to be clearly marked for users, and 87% were in favour of a risk-based approach to regulation.

155 Yet, a McKinsey global survey (2017) showed that companies that are committed to AI have significantly higher profit margins across sectors. These companies also expect a margin increase of up to five percentage points more than industry average in the following three years. Hence, direct costs from legal uncertainty in the market of AI development and use are also accompanied by a missed potential for business innovation. In the end, even consumers will suffer, as they will miss out beneficial products and services not developed by businesses due to the fear of legal consequences.

Problem 5: Mistrust in AI would slow down AI development in Europe and reduce the global competitiveness of the EU economy

156 157 If citizens observe that AI repeatedly endangers the safety of individuals or infringes their fundamental rights, they are unlikely to be willing to accept the use of AI technologies for themselves or by other users. In a recent survey aimed at consumers in nine EU Member States, around 56% of respondents said they had low trust in authorities to exert effective control over AI. Similarly, only 12% of the survey respondents in Sweden and Finland reported to trust private companies’ abilities to address the ethical dilemmas AI brings. 77% thought that companies that are developing new AI solutions should be bound to ethical guidelines and regulation. 

158 Faced with reluctant private and business customers, businesses will find it then more difficult to invest in and adopt AI than if customers embrace AI. As a result, demand for new and innovative AI applications will be sub-optimal. This mistrust will hamper innovation because companies will be more hesitant in offering innovative services for which they first have to establish credibility, which will be particularly challenging in a context of growing public fears about the risks of AI. That is why in a recent study, 42% of EU executives surveyed see “guidelines or regulations to safeguard ethical and transparent use” as part of a policy to promote AI to the benefit of Europe.

Similarly, a recent survey of European companies found that for 65% of respondents a lack of trust among citizens is an obstacle to the adoption of AI. 159 There is already a substantial share of companies not preparing for the AI-enabled future: 40% of the companies neither use any AI application whatsoever nor intend to do so. 160 The problem is particularly acute for SMEs that cannot rely on their brand to reassure customers that they are trustworthy. Significantly fewer large companies (34%) than SMEs (between 41% and 43% depending on company size) saw lack of trust as not an obstacle. The share of companies citing this as a major obstacle was relatively evenly distributed across company sizes (between 26% and 28%).

Figure 3: Trust in AI applications

Source: Ipsos survey 20-28 September 2018, quoted by Statista

161 162 Without a sound common framework for trustworthy AI, Europe could lose out on the beneficial impact of AI on competitiveness. Yet, the benefits of rapid adoption of AI are generally estimated as being very significant. McKinsey estimates that if Europe (EU28) on average develops and distributes AI according to its current assets and digital position relative to the world, it could add some 2.7 trillion, or 20%, to its combined economic output by 2030. According to a European Value Added Assessment, prepared by the European Parliament, a common EU framework on the ethics of AI has the potential to bring the European Union 294.9 billion in additional GDP and 4.6 million additional jobs by 2030.

Thus, a slower adoption of AI would have significant economic costs and would hamper innovation. It would also mean foregoing part of the wide array of the societal benefits that AI is poised to bring in areas such as health, transport or pollution reduction. It would thus negatively affect not just businesses and the economy, but consumers and the society as well.

Problem 6: Fragmented measures create obstacles for a cross-border AI single market and threaten the Union’s digital sovereignty

163 In the absence of a common European framework to address the risks examined before and build trust in AI technology, Member States can be expected to start taking action at a national level to deal with these specific challenges. While national legislation is within the Member States’ sovereign competences, there is a risk that diverging national approaches will lead to market fragmentation and could create obstacles especially for smaller companies to enter multiple national markets and scale up across the EU Single Market. Yet, as noted in section 1.2., AI applications are rapidly increasing in scale. Where advanced models work with billions of parameters, companies need to scale up their models to remain competitive. Since the high mobility of AI producers could lead to a race to the bottom where companies move to Member States with the lightest regulation and serve the entire EU market from there, other Member States may take measures to limit access from other Member States, leading to further market fragmentation.

164 165 That is why Member States in general support a common European approach to AI. In a recent position paper 14 Member States recognise the risk of market fragmentation and emphasise that the ‘main aim must be to create a common framework where trustworthy and human-centric AI goes hand in hand with innovation, economic growth and competitiveness’. Earlier, in its conclusions of 9 June 2020, the Council called upon the Commission ‘to put forward concrete proposals, taking existing legislation into consideration, which follow a risk-based, proportionate and, if necessary, regulatory approach for artificial intelligence.’

166 167 While waiting for a European proposal, some Member States are already considering national legislative or soft-law measures to address the risks, build trust in AI and support innovation. For example, the German Data Ethics Commission has called for a five-level risk-based system of horizontal regulation on AI that would go from no regulation for the most innocuous AI systems to a complete ban for the most dangerous ones. Denmark has just launched the prototype of a Data Ethics Seal, whilst Malta has introduced a voluntary certification system for AI. Spain is in the process of adopting a Code of Ethics and considering certification of AI products and services. Finland issued recommendations for self-regulation and the development of responsibility standards for the private sector, while Italy envisages certificates to validate and to monitor AI applications developed in an ethically sound way. Moreover, several Member States (e.g. Belgium, Sweden, Netherlands and Portugal) are considering the need for binding legislation on the legal and ethical aspects of AI.

168 169 170 In addition to this increasingly patchy national landscape, there is an ongoing proliferation of voluntary international technical standards for various aspects of ‘Trustworthy AI’ adopted by international standardisation organisations (e.g. IEEE, ISO/IEC, ETSI, ITU-T, NIST). While these standards can in principle be very helpful in ensuring safe and trustworthy AI systems, there is also a growing risk of divergence between them since they are adopted by different international organisations. Moreover, these technical standards may not be fully compliant with existing EU legislation (e.g., on data protection or non-discrimination), which creates additional liability risks and legal uncertainty for companies adhering to them. In addition, there is also a proliferation of national technical standards on AI that are adopted or being developed by a number of Member States and many other third countries around the world. This means national standards risk not being fully interoperable or compatible with each other, which will create obstacles to cross-border movement and to the scaling up of AI-driven services and products across different Member States.

The impact of this increasing fragmentation is disproportionately affecting small companies. This is because large companies, especially global ones, can spread the additional costs for operating across an increasingly fragmented single market over their larger sales, especially when they already have established a dominant position in some markets. Meanwhile, SMEs and start-ups which do not have the market power or the same resources may be deterred from entering the markets of other Member States and thus fail to profit from the single market. This problem is further exacerbated since big tech players have not only a technological advantage but also exclusive access to large and quality data necessary for the development of AI. They may try to use this information asymmetry to seek economic advantages and further harm smaller companies. These dominant tech companies may also try to free ride on political efforts aiming to increase consumer protection by ensuring that the adopted standards for AI are in line with their own business practices to the detriment of newcomers and smaller players. This risk is significantly higher when the AI market is fragmented with individual Member States taking unilateral actions.

171 172 173 All these diverging measures stand in the way of a seamless and well-functioning single market for trustworthy AI in the Union and pose particular legal barriers for SMEs and start-ups. This in turn negatively affects the global competiveness of the EU industry, both regarding AI providers and the industries using AI, giving advantage to companies from third countries that are already dominant on the global market. Beyond the purely market dimension, there is a growing risk that the ‘digital sovereignty’ of the Union and the Member States might be threatened since such AI-driven products and services from foreign companies might not completely comply with Union values and/or legislation or they might even pose security risks and make the European infrastructure more vulnerable. As stated by the President of the Commission von der Leyen, to ensure ‘tech sovereignty’, the EU should strengthen its capability to make its own choices, based on its own values, respecting its own rules. Aside from strengthening the EU internal market, such tech sovereignty will also facilitate the development and the leverage of Union’s tools and regulatory power to shape global rules and standards on AI.

2.2. What are the main problem drivers?

174 175 The uptake of AI systems has a strong potential to bring benefits, economic growth and enhance EU innovation and global competitiveness. However, as mentioned in the section above, in certain cases, the use of AI systems can also create problems for businesses and national authorities as well as new risks to safety and fundamental rights of individuals. The key cause explaining the analysed problems are the specific characteristics of AI systems which make them qualitatively different from previous technological advancements. Table 3 below explains what each characteristic means and why they can create problems for fundamental rights and safety. 

Table 3: Five specific characteristics of AI and Problem Drivers

CHARACTERISTICS

EXPLANATION (simplified)

WHY IS IT A PROBLEM?/ DRIVERS

Opacity/ (lack of transparency)

Limited ability of human mind to understand how certain AI systems operate

A lack of transparency (opacity)) in AI (due to complexity or how the algorithm was realized in code or how the application is realised) makes it difficult to monitor, identify and prove possible breaches of laws, including legal provisions that protect fundamental rights.

Complexity

Multiplicity of different components and processes of an AI system and their interlinks

The complexity of AI makes it difficult to monitor, identify and prove possible breaches of laws, including legal provisions that protect fundamental rights.

Continuous adaptation and unpredictability

Functional ability of some AI systems to continuously ‘learn’ and ‘adapt’ as they operate, sometimes leading to unpredictable outcomes.

Some AI systems change and evolve over time and may even change their own behaviour in unforeseen ways. This can give rise to new risks that are not adequately addressed by the existing legislation.

Autonomous behaviour

Functional ability of some AI systems to generate outputs such as ‘decisions’ with limited or no human intervention

The autonomous behaviour of AI systems can affect safety because of the functional ability of an AI system to perform a task with minimum or no direct human intervention.

Data

Functional dependence on data and the quality of data

The dependence of AI systems on data and their quality, the AI’s ‘ability’ to infer correlations from data input and learn from data, including proxies, can reinforce systemic biases and errors, and exacerbate discriminatory and adverse results.

176 177 178 179 Certain AI systems may include only some of the characteristics. However, as a rule, the more specific characteristics a given AI system has, the higher the probability that it becomes a ‘black box’. The term ‘black box’ reflects the limited ability of even most advanced experts to monitor an AI system. This stands in considerable difference with the ability to monitor other technologies. The OECD report explains this ‘black box’ effect with the example of neural networks as follows: “Neural networks iterate on the data they are trained on. They find complex, multi-variable probabilistic correlations that become part of the model that they build. However, they do not indicate how data would interrelate. The data are far too complex for the human mind to understand.” ‘Black box’ AI systems present new challenges for public policy compared to traditional and other technologies.

180 These specific characteristics of AI systems may create new (1) safety and security and (2) fundamental rights risks and accelerate the probability or intensity of the existing risks, as well as (3) make it hard for enforcement authorities to verify compliance with and enforce the existing rules. This set of problems leads in turn to (4) legal uncertainty for companies, (5) potentially slower uptake of AI technologies, due to the lack of trust, by businesses and citizens as well as (6) regulatory responses by Member States to mitigate possible externalities. Consequently, problems triggered by specific characteristics of AI may lead to safety risks and breaches of fundamental rights and challenge the effective application of and compliance with the EU legal framework for the protection of fundamental rights and safety.

Table 4: Problem Tree

181 Rapid developments and uptake of AI systems increase this challenge. The Ipsos 2019 survey of European business indicates that 42% of enterprises currently use at least one AI technology, a quarter of them use at least two types, and 18% have plans to adopt AI technologies in the next two years. Moreover, the intensity of use of AI technology by business is expected to grow in the next two years. This data suggests that in the next two years, it is likely that more than half of all EU businesses will be using AI systems. Thus, AI systems already affect business and consumers in the EU on a large scale.

182 183 184 According to the European Commission Better Regulation Guidelines and accompanying Toolbox, a public policy intervention may be justified, among other grounds, when regulations fail or when protection and fulfilment of fundamental rights afforded to citizens of the Union provide grounds for intervention. Several factors may cause regulatory failures, including, when existing public intervention becomes “out of date as the world evolves.” Protection and fulfilment of fundamental rights afforded to citizens of the Union may also provide important reasons for policy intervention ‘because even a perfectly competitive and efficient economy can produce outcomes that are unacceptable in terms of equity’.

2.3. How will the problem evolve?

Given the increasing public awareness of the potential for violation of safety and fundamental rights that AI has, it is likely that the proliferation of ethics principles will continue. Companies would adopt these principle unilaterally in an effort to reassure their potential customers. However, such non-binding ethical principles cannot build the necessary trust as they cannot be enforced by affected parties and no external party or regulator is actually empowered to check whether these principles are duly respected by the companies and the public authorities developing and using AI systems. Moreover, the multiplicity of commitments would require consumers to spend an extraordinary amount of time understanding which commitments apply to which application.

As a consequence, and given the significant commercial opportunities offered by AI solutions, ‘untrustworthy’ AI solutions could ensue, with a likely backlash against AI technology as a whole by citizens and businesses. If and when this happens, European citizens will lose out on the benefits of AI and European companies will be placed at a significant disadvantage compared to their overseas competitors with a dynamic home market.

Over time, technological developments in the fields of algorithmic transparency, accountability and fairness could improve the situation, but progress and impact will be uncertain and uneven across Europe. On the contrary, as AI develops, it can be implemented in more and more situations and sectors, so that the problems identified above apply to an ever-growing share of citizens’ life.

It cannot be excluded that over the long run and after a sufficient number of incidents, consumers will prefer companies with a proven track record of trustworthy AI. However, apart from the damage done in the meantime, this would have the consequence of favouring large companies, who can rely on their brand image, over SMEs who will face increasing legal barriers to enter the market.

3.WHY SHOULD THE EU ACT?

3.1. Legal basis

The initiative constitutes a core part of the EU single market strategy given that artificial intelligence has already found its way into a vast majority of services and products and will only continue to do so in the future. EU action on the basis of Article 114 of the Treaty on the Functioning of the European Union can be taken for the purposes of the approximation of the provisions laid down by law, regulation or administrative action in the Member States when it has as its object the establishment and functioning of the internal market. The measures must be intended to improve the conditions for the establishment and functioning of the internal market and must genuinely have that object, actually contributing to the elimination of obstacles to the free movement of goods or services, or to the removal of distortions of competition.

185 Article 114 TFEU may be used as a legal basis to prevent the occurrence of these obstacles resulting from diverging national laws and approaches how to address the legal uncertainties and gaps in the existing legal frameworks applicable to AI. The present initiative aims to improve the functioning of the internal market by setting harmonized rules on the development, placing on the Union market and the use of products and services making use of the AI technology or provided as stand-alone AI applications. Some Member States are already considering national rules to ensure AI is safe and is developed and used in compliance with fundamental rights obligations. This will likely lead to a further fragmentation of the internal market and increasing legal uncertainty for providers and users on how existing and new rules will apply to AI systems.

186 Furthermore, the Court of Justice has recognised that applying heterogeneous technical requirements could be valid grounds to trigger Article 114 TFEU. The new initiative will aim to address that problem by proposing harmonised technical standards for the implementation of common requirements applicable to the design and development of certain AI systems before they are placed on the market. The initiative will also address the situation after AI systems have been placed on the market by harmonising the way in which ex-post controls are conducted.

187 Based on the above, Article 114 TFEU is the applicable legal basis for the present initiative. In addition, considering that this Regulation contains certain specific rules, unrelated to the functioning of the internal market, restricting the use of AI systems for ‘real-time’ remote biometric identification by the law enforcement authorities of the Member States, which necessarily limits the processing of biometric data by those authorities, it is appropriate to base this Regulation, in as far as those specific rules are concerned, on Article 16 of the Treaty.

3.2. Subsidiarity: Necessity of EU action

The intrinsic nature of AI which often relies on large and varied datasets and which might be embedded in any product or service circulating freely within the internal market mean that the objectives of the initiative cannot effectively be achieved by Member States alone. An emerging patchy framework of potentially divergent national rules will hamper the seamless provision of AI systems across the EU and is ineffective in ensuring the safety and protection of fundamental rights and Union values across the different Member States. Such an approach is unable to solve the problems of ineffective enforcement and governance mechanisms and will not create common conditions for building trust in the technology across all Member States. National approaches in addressing the problems will only create additional legal uncertainty, legal barriers and will slow market uptake of AI even further. Companies could be prevented from seamlessly expanding into other Member States, depriving consumers and other users of the benefits of their services and products and affecting negatively the competitiveness of European companies and the economy.

3.3. Subsidiarity: Added value of EU action

The objectives of the initiative can be better achieved at Union level so as to avoid a further fragmentation of the Single Market into potentially contradictory national frameworks preventing the free circulation of goods and services embedding AI. In their positions to the White paper on AI all Member States support a coordinated action at EU level to prevent the risk of fragmentation and create the necessary conditions for a single market of safe, lawful and trustworthy AI in Europe. A solid European regulatory framework for trustworthy AI will also ensure a level playing field and protect all European citizens, while strengthening Europe’s competitiveness and industrial basis in AI. 188 A common EU legislative action on AI could boost the internal market and has great potential to provide European industry with a competitive edge at the global scene and economies of scale that cannot be achieved by individual Member States alone. Setting up the governance structures and mechanisms for a coordinated European approach to AI across all sectors and Member States will enhance safety and the respect of fundamental rights, while allowing businesses, public authorities and users of AI systems to capitalise on the scale of the internal market and use safe and trustworthy AI products and services. Only common action at EU level can also protect Union’s tech sovereignty and leverage its tools and regulatory powers to shape global rules and standards.

4.Objectives: What is to be achieved?

The problems analysed in section 2 above are complex and cannot be fully addressed by any single policy intervention. This is why the Commission proposes to address emerging problems related specifically to AI systems gradually. The objectives of this initiative are defined accordingly.

Table 5: General/Specific objectives

1.

2.

3.

4.

4.1.General objectives

The general objective of the intervention is to ensure the proper functioning of the single market by creating the conditions for the development and use of trustworthy artificial intelligence in the Union.

4.2.Specific objectives

The specific objectives of this initiative are as follows:

·set requirements specific to AI systems and obligations on all value chain participants in order to ensure that AI systems placed on the market and used are safe and respect the existing law on fundamental rights and Union values;

·ensure legal certainty to facilitate investment and innovation in AI by making it clear what essential requirements, obligations, as well as conformity and compliance procedures must be followed to place, or use an AI system in the Union market;

·enhance governance and effective enforcement of the existing law on fundamental rights and safety requirements applicable to AI systems by providing new powers, resources and clear rules for relevant authorities on conformity assessment and ex post monitoring procedures and the division of governance and supervision tasks between national and EU levels;

·facilitate the development of a single market for lawful, safe and trustworthy AI applications and prevent market fragmentation by taking EU action to set minimum requirement for AI systems to be placed and used in the Union market in compliance with the existing law on fundamental rights and safety.

4.2.1.Ensure that AI systems placed on the market and used are safe and respect the existing law on fundamental rights and Union values

189 Safeguarding the safety and fundamental rights of EU citizens is a cornerstone of European values. However, the emergence of AI creates new challenges regarding safety and protection of fundamental rights and hinder the enforcement of these rights, due to the specific features of this technology (see section 2.2.). However, the same rights and rules that apply in the analogue world should also be respected when AI systems are used. The first specific objective of the initiative is, therefore, to ensure that AI systems that are developed, placed on the market and/or used in the Union are safe and respect the existing law on fundamental rights and Union values by setting requirements specific to AI systems and obligations on all value chain participants. The ongoing review of the sectoral safety legislation will pursue a similar objective to ensure safety of products embedding AI technology, but focusing on the overall safety of the whole product and the safe integration of the AI system into the product.

4.2.2. Ensure legal certainty to facilitate investment and innovation in AI

190 Due to the specific characteristics of AI, businesses using AI technology are also facing increasing legal uncertainty and complexity on how to comply with existing legislation. As long as the challenges and risks to safety and fundamental rights have not been addressed, companies must also calculate the risk that legislation or other requirements will be introduced, without knowing what this will imply for their business models. Such legal uncertainty is detrimental for investment and especially for innovation. The second objective is therefore to promote investment and innovation by creating single market-wide legal certainty and an appropriate framework that stimulates innovation by making it clear what essential requirements, obligations as well as conformity and compliance procedures must be followed to place or use an AI system in the Union market. The complementary initiative on liability rules will also aim to increase legal certainty in the use of AI technology, but by ensuring a high-level of protection of victims who have suffered harms caused by certain AI systems. 

4.2.3.Enhance governance and effective enforcement of the existing law on fundamental rights and safety requirements applicable to AI systems

The technological features of AI such as opacity and autonomous behaviour might cause violations of safety rules and the existing law on fundamental rights that may not even be noticed by the concerned person, and that, even when they are noticed, are often difficult to prove. Existing competent authorities might also face difficulties in auditing the compliance of certain AI systems with the existing legislation due to the specific technological features of AI. They might also lack powers to intervene against actors who are outside their jurisdiction or lack sufficient resources and a mechanism for cooperation and joint investigations with other competent authorities. The enforcement and governance system needs to be adapted to these new challenges so as to ensure that possible breaches can be effectively detected and sanctioned by enforcement authorities and those affected. The third objective is therefore to improve the governance mechanism and effective enforcement of the existing law on fundamental rights and safety requirements applicable to AI by providing new powers, resources and clear rules for relevant authorities on conformity assessment and ex post monitoring procedures and the division of governance and supervision tasks between national and EU levels.

4.2.4. Facilitate the development of a single market for lawful, safe and trustworthy AI applications and prevent market fragmentation

The safety and fundamental rights risks posed by AI may lead citizens and consumers to mistrust this technology, inciting in turn Member States to address these problems with national measures which may create obstacles to cross border sales, especially for SMEs. The fourth objective is, hence, to foster trustworthy AI, which will reduce the incentives for national and potentially mutually incompatible legislations and will remove legal barriers and obstacles to cross border movement of products and services embedding the AI technology by taking EU action to set minimum requirements for AI systems to be placed and used in the Union market in compliance with the existing law on fundamental rights and safety. The complementary initiative on liability rules would also aim to increase trust in the AI technology, but by ensuring a high-level of protection of victims who have suffered harms caused by certain AI systems.

4.3. Objectives tree/intervention logic.

Figure 4: Intervention logic

The specific characteristics of certain AI systems (opacity, complexity, autonomous behaviour, unpredictability and data dependency) may create (1) safety and security and (2) fundamental rights risks and (3) make it hard for enforcement authorities to verify compliance with and enforce the existing rules. This set of problems in turn leads to other problems causing (4) legal uncertainty for companies, (5) potentially slower uptake of AI technologies, due to the lack of trust, by businesses and citizens as well as (6) unilateral regulatory responses by Member States to mitigate possible externalities.

Firstly, current EU law does not effectively ensure protection for safety and fundamental rights risks specific to AI systems, as shown in the problem definition (Problems 1 and 2). Particularly, risks caused by opacity, complexity, continuous adaptation, autonomous behaviour and the data dependency of AI systems (drivers) are not fully covered by the the existing law. Accordingly, this initiative sets out the specific objective to set requirements specific to AI systems and obligations on all value chain participants in order to ensure that AI systems placed or used in the Union market are safe and respect the existing law on fundamental rights and Union values (specific objective 1).

Secondly, under current EU law, competent authorities do not have sufficient powers, resources and/or procedural frameworks in place to effectively ensure and monitor compliance of AI systems with fundamental rights and safety legislation (problem 3). The specific characteristics of AI systems (drivers) often make it hard to verify how outputs/decisions have been reached where AI is used. As a consequence, it may become impossible to verify compliance with existing EU law meant to guarantee safety and protection of fundamental rights. Competent authorities also do not have sufficient powers and resources to effectively inspect and monitor these systems. To address these problems, the initiative sets the objective to enhance governance and effective enforcement of the existing law on fundamental rights and safety requirements applicable to AI systems by providing new powers, resources and clear rules for relevant authorities on conformity assessment and ex post monitoring procedures and the division of governance and supervision tasks between national and EU levels (specific objective 3). 

Thirdly, current EU legislation does provide certain requirements related to safety and protection of fundamental rights that apply to new technologies, including AI systems. However, those requirements are not specific to AI systems, they lack legal certainty or standards on how to be implemented and are not consistently imposed on different actors across the value chain. Considering the specific characteristics of AI (drivers), providers and users do not have clarity as to how existing obligations should be applied to AI systems for these systems to be considered safe, trustworthy and in compliance with the existing law on fundamental rights (problem 4). Furthermore, the lack of clear distribution of obligations across the AI value chain also contributes to problems 1 and 2. To address those problems, the initiative sets the objective to clarify what essential requirements, obligations, as well as conformity and compliance procedures actors must follow to place, or use an AI system in the Union market (specific objective 2).

Finally, in the absence of an EU legislation on AI that addresses the new specific risks to safety and fundamental rights, businesses and citizens distrust the technology (problem 5), while Member States’ unilateral action to address that problem risks to create obstacles for a cross-border AI single market and threatens the Union’s digital sovereignty (problem 6). To address these problems, the proposed initiative has the objective to facilitate the development of a single market for lawful, safe and trustworthy AI applications and prevent market fragmentation by taking EU action to set minimum requirements for AI systems to be placed and used in the Union market in compliance with the existing law on fundamental rights and safety (specific objective 4).

5.What are the available policy options?

The analysed policy options are based on the following main dimensions: a) The nature of the EU legal act (no EU intervention/ EU act with voluntary obligations/ EU sectoral legislation/ horizontal EU act); b) Definition of AI system (voluntary/ ad hoc for specific sectors/ one horizontal definition); c) Scope and content of requirements and obligations (voluntary/ ad hoc depending on the specific sector/ risk-based/ all risks covered); d) Enforcement and compliance mechanism (voluntary/ ex ante or ex post only/ ex ante and ex post); e) Governance mechanism (national, national and EU, EU only).

The policy options, summarized in Table 6 below, represent the spectrum of policy options based on the dimensions outlined above.

Table 6: Summary of the analysed policy options

Option 1

EU Voluntary labelling scheme

Option 2

Ad hoc sectoral approach

Option 3

Horizontal risk- based act on AI

Option 3+

Codes of conduct

Option 4

Horizontal act for all AI

Nature of act

An EU act establishing a voluntary labelling scheme

Ad hoc sectoral acts (revision or new) 

A single binding horizontal act on AI

Option 3 + code of conducts

A single binding horizontal act on AI

Scope/ Definition of AI

One definition of AI, however applicable only on a voluntary basis

Each sector can adopt a definition of AI and determine the riskiness of the AI systems covered

One horizontally applicable AI definition and methodology for determination of high- risk (risk-based)

Option 3 + industry-led codes of conduct for non-high-risk AI

One horizontal AI definition, but no methodology/or gradation (all risks covered)

Requirements

Applicable only for voluntarily labelled AI systems.

Applicable only for sector specific AI systems with possible additional safeguards/ limitations for specific AI use cases per sector

Risk-based horizontal requirements for prohibited and high risk AI systems

+ min. information requirements for certain other AI systems

Option 3 + industry-led codes of conduct for non-high-risk AI

For all AI systems irrespective of the level of the risk

Obligations

Only for providers who adopt voluntary scheme and no obligations for users of certified AI systems

Sector specific obligations for providers and users depending on the use case

Horizontal obligations for providers and users of high-risk AI systems

Option 3 + commitment to comply with codes of conduct for non-high-risk AI

Same as Option 3, but applicable to all AI (irrespective of risk)

Ex ante enforcement

Self-assessment and an ex ante check by national competent authorities responsible for monitoring compliance with the EU voluntary label

Depends on the enforcement system under the relevant sectoral acts.

Conformity assessment for providers of high-risk systems (3rd party for AI in a product and other systems based on internal checks) + registration in an EU database. 

Option 3 + self- assessment for compliance with codes of conduct for non- high-risk AI

Same as Option 3, but applicable to all AI (irrespective of risk)

Ex post enforcement

Monitoring by authorities responsible for EU voluntary label

Monitoring by competent authorities under the relevant sectoral acts

Monitoring of high-risk systems by market surveillance authorities

Option 3 + unfair commercial practice in case of non-compliance with codes

Same as Option 3, but applicable to all AI (irrespective of risk)

Governance

National competent authorities designated by Member States responsible for the EU label + a light EU cooperation mechanism

Depends on the sectoral acts at national and EU level; no platform for cooperation between various competent authorities.

At the national level but reinforced with cooperation between Member States authorities and with the EU level (AI Board)

Option 3 + without EU approval of the codes of conduct

Same as Option 3, but applicable to all AI (irrespective of risk)

5.1. What is the baseline from which options are assessed?

191 192 Under the baseline scenario, there would be no specific legislation at European level comprehensively addressing the issues related to AI discussed above. Ongoing revisions of other existing legislations, such as the review of the Machinery Directive 2006/42/EC and the General Product Safety Directive 2001/95/EC would continue. Both directives are technology neutral and their review will address aspects that are related to new digital technologies, not only specific to AI systems. In other areas, in particular with regard to use of automated tools, including AI, by online platforms, the rules proposed in the Digital Services Act and the Digital Markets Act (once adopted) would establish a governance system to address risks as they emerge and ensure a sufficient user-facing transparency and public accountability in the use of these systems. 

193 In parallel to these revisions, the EU would also promote industry-led initiatives for AI in an effort to advance ‘soft law’ but would not establish any framework for such voluntary codes. Currently, an increasingly large number of AI principles and ethical codes has already been developed by industry actors and other organisations. In the Union, the HLEG developed a set of Ethics Guidelines for Trustworthy AI with an assessment list aimed at providing practical guidance on how to implement each of the key requirements for AI. The ‘soft law’ approach could build upon existing initiatives and consist of reporting on the voluntary compliance with such initiatives based on self-reporting (without any involvement of public supervisory authorities or other accredited organisations); encouraging industry-led coordination on a single set of AI principles; awareness raising among AI systems developers and users around the existence and utility of existing initiatives; monitoring and encouraging the development of voluntary standards that could be based on the non-binding HLEG ethical guidelines.

In the absence of a regulatory initiative on AI, the risks identified in section 2 would remain unaddressed. EU legislation on the protection of fundamental rights and safety would remain relevant and applicable to a large number of emerging AI applications. However, increased violations of fundamental rights and a higher exposure to safety risks including problems with enforcement of existing EU law may grow as AI continues to develop.

194 195 196 197 198 In the baseline scenario, there is also a large offer of forecasts of the AI market which all assume an unopposed development of AI and significant growth with projections for the EU market in 2025 between €32 billion and €66 billion. However, by not considering the possibility of backlashes, the forecasts may prove over-optimistic in the absence of regulation. As an example of such a backlash, in March 2020 one major forecaster predicted a compound annual growth rate for the facial recognition market of 14.5% from 2020 of 2027. Yet in June 2020, following claims that facial recognition systems were of discriminatory nature, one market leader (IBM) stopped developing and selling these systems, while two other major players (Amazon and Microsoft) decided to suspend their sales to a major customer (the law enforcement sector). Similar developments cannot be excluded in other AI use cases, especially where claims of discrimination have already led to pressure from public opinion (e.g. recruitment software, sentencing support or financial services). Similarly, the use of CT scans for COVID diagnosis has not been rolled out as quickly as possible due to the reluctance of hospitals to use uncertified technologies.

Consequently, the lack of any decisive policy action by the EU could lead to increased fragmentation due to interventions at Member State level, as public opinion would put pressure on politicians and law-makers to address the concerns described above. As a result of national approaches the single market for AI products and services would be further fragmented with different standards and requirements that will create obstacles to cross border movement. This would reduce the competitiveness of European businesses and endanger Europe’s digital autonomy.

Stakeholders views: In the public consultation on the White Paper on AI, 16% of SMEs saw current legislation as sufficient to address concerns related to AI. 37% saw gaps in current legislation, and 40% the need for new legislation. Among large businesses too, a majority of respondents said that current legislation was insufficient. Academic and research institutions overwhelmingly came to the conclusion that current legislation was not sufficient. Only 2% said otherwise, while 48% saw a need for new legislation and 35% saw gaps in existing legislation. Almost no civil society organisation deemed current legislation sufficient. Among those stakeholders claiming that current legislation was sufficient, EU citizens, large companies (both 25%), and business associations (22%) were the largest groups. For large companies and business associations, this was around double their share in the overall sample of respondents. For SMEs (13%), this was also true.

5.2. Option 1: EU legislative instrument setting up a voluntary labelling scheme

Under this option, an EU legislative instrument would establish an EU voluntary labelling scheme to enable providers of AI applications certify their AI systems’ compliance with certain requirements for trustworthy AI and obtain an EU-wide label. While participation in the scheme would be voluntary, the instrument would envisage an appropriate enforcement and governance system to ensure that providers who subscribe comply with the requirements and take appropriate measures to monitor risks even after the AI system is placed on the market. Given the voluntary character of the initiative aimed at the AI system’s certification, the instrument would not impose certification obligations on users of labelled AI systems since these would be impractical and not voluntary in nature.

Table 6.1. Summary Option 1: EU Voluntary labelling scheme

Nature of act

An EU act establishes a voluntary labelling scheme, which becomes binding once adhered to

Scope

OECD definition of AI; adherence possible irrespective of the level of risk, but certain risk differentiation amongst the certified AI systems also possible

Content

Requirements for labelled AI systems: data, transparency and provision of information, traceability and documentation, accuracy, robustness and human oversight (to be ensured by providers who choose to label their AI system)

Obligations

Obligations for providers (who voluntarily agree to comply) for quality management, risk management and ex post monitoring

No obligations for users of certified AI systems (impractical given the voluntary character of the label aimed at certification of specific AI systems) 

Ex ante enforcement

Self-assessment and ex ante check by national competent authorities responsible for monitoring compliance with the EU voluntary label

Ex post enforcement

Ex post monitoring by national competent authorities responsible for monitoring compliance with the EU voluntary label

Governance

National competent authorities designated by Member States as responsible for the EU label + a light EU cooperation mechanism

5.2.1. Scope of the EU voluntary labelling scheme

Given the voluntary nature of the EU voluntary labelling scheme, this would be applicable regardless of the level of risk of the AI system, but certain risk differentiation amongst the certified AI systems could also be envisaged. The instrument would build on the internationally recognized OECD definition of AI, 199 because it is technology neutral and future proof. This policy choice is justified because a technology-specific definition 200 could cause market distortions between different technologies and quickly become outdated. The OECD definition is also considered sufficiently broad and flexible to encompass all problematic uses as identified in the problem section. Last but not least, it would allow a consensus with the international partners and third countries and ensure that the proposed scheme would be compatible with the AI frameworks adopted by the EU’s major trade partners.

Stakeholders views: In the Public Consultation on the White Paper on AI there were some disagreements between stakeholder groups regarding the exact definition of AI, proposed as comprising ‘data’ and ‘algorithms’. At least 11% of large companies and 10% of SMEs found this definition too broad. Only 2% of large companies and no SMEs said it was too narrow. This likely reflects concerns about too many AI systems falling under potential future requirements, thus creating an additional burden on companies. On the other hand, the civil society organisations tended to find it too narrow. Furthermore, at least 11% of large companies and 5% of SMEs said that the definition was unclear and would need to be refined.

5.2.2. Requirements for Trustworthy AI envisaged in the EU voluntary labelling scheme

The voluntary labelling scheme would impose certain requirements for Trustworthy AI which would aim to address the main sources of risks to safety and fundamental rights during the development and pre-market phase and provide assurance that the AI system has been properly tested and validated by the provider for its compliance with existing legislation.

These requirements for trustworthy AI would be limited to the necessary minimum to address the problems and include the following 5 requirements, identified in the White Paper: a) data governance and data quality; b) traceability and documentation; c) algorithmic transparency and provision of information; d) human oversight, and e) accuracy, robustness and security.

Figure 5: Requirements for Trustworthy AI systems

The 5 requirements above are the result of two years of preparatory work and derived from the Ethics Guidelines of the HLEG, 201 piloted by more than 350 organisations. 202 They are also largely consistent with other international recommendations and principles 203 which would ensure that the proposed EU framework for AI would be compatible with those adopted by the EU’s international trade partners. The EP also proposed similar requirements, 204 but it was decided not to include some of the EP proposals or the HLEG principles as requirements. This is because they were considered either too vague, too difficult to operationalize, 205 or already covered by other legislation. 206  

207 The proposed minimum requirements are already state-of-the-art for many diligent business operators and they would ensure minimum degree of algorithmic transparency and accountability in the development of AI systems. Requiring that AI systems are developed with high quality datasets that reflect the specific European context of application and intended purpose would also help to ensure that these systems are reliable, safe and minimize the risk of discrimination once deployed by users. These requirements have also been largely supported by stakeholders in the consultation on the White Paper on AI. 

Figure 6: Stakeholder consultation results on the requirements for AI

Source: Public Consultation on the White Paper on Artificial Intelligence

To prove compliance with the requirements outlined above, providers who choose to subscribe to the voluntary labelling scheme would also have to establish within their organisation an appropriate quality management and risk management system, including with prior testing and validation procedures to detect and prevent unintended consequences and minimize the risks to safety and fundamental rights. 208  To take into account the complexity and the possibility for continuous adaptation of certain AI systems and the evolving risks, providers would also have to monitor the market ex post and take any corrective action, as appropriate (problems 1, 2 and 3).

5.2.3. Enforcement and governance of the EU voluntary labelling scheme

While participation in the labelling scheme would be voluntary, providers who choose to participate would have to comply with these requirements (in addition to existing EU legislation) to be able to display a quality ‘Trustworthy AI’ label. The label would serve as an indication to the market that the labelled AI application is trustworthy, thus addressing partially the mistrust problem for those certified AI applications (problem 5).

The scheme would be enforced through ex ante self-assessment 209 and ex-post monitoring by competent authorities designated by the Member States. This is justified by the need to improve governance and enforceability of the requirements specific to AI systems (problem 3) as well as for practical reasons. On the one hand, competent authorities would first have to register the commitment of the provider to comply with the AI requirements and check if this is indeed the case. On the other hand, ex post supervision would be necessary to ensure that compliance is an ongoing process even after the system has been placed on the market. Where relevant, this would also aim to address the evolving performance of the AI system due to its continuous adaptation or software updates.

The instrument would also establish appropriate sanctions for providers participating in the scheme that have claimed compliance with the requirements, but are found to be non-compliant. The sanctions would be imposed following an investigation with a final decision issued by the competent national authority responsible for the scheme at national level. 210  

A light mechanism for EU cooperation is also envisaged with a network of national competent authorities who would meet regularly to exchange information and ensure uniform application of the scheme. An alternative would be not to envisage any cooperation at EU level, but this would compromise the European character and uniform application of the European voluntary labelling scheme.

Despite some inherent limitations of the voluntary labelling scheme in ensuring legal certainty and achieving the development of a truly single market for trustworthy AI (problems 4 and 6), this option would still have some positive effects to increase trust and address AI challenges by means of a more gradual regulatory approach, so it should not be excluded a priori from the detailed assessment.    

5.3. Option 2: A sectoral, ‘ad-hoc’ approach

This option would tackle specific risks generated by certain AI applications through ad-hoc legislation or through the revision of existing legislation on a case by case basis. 211 There would be no coordinated approach on how AI is regulated across sectors and no horizontal requirements or obligations. The sector specific acts adopted under this option would include sector specific requirements and obligations for providers and users of certain risky AI applications (e.g. remote biometric identification, deep fakes, AI used in recruitments, prohibition of certain AI uses etc.). Their content would depend on the specific use case and would be enforced through different enforcement mechanisms without a common regulatory framework or platform for cooperation between national competent authorities. The development and use of all other AI systems would remain unrestricted subject to the existing legislation.

Table 6.2. Summary Option 2: Ad hoc sectoral approach

Nature of act

Case-by-case binding sectoral acts (review of existing legislation or ad hoc new acts)

Scope

Different sectoral acts could adopt different definitions of AI that might be inconsistent. Each sectoral act will determine the risky AI applications that should be regulated.

Content

Sector specific requirements for AI systems (could be similar to Option 1, but adapted to sectoral acts)

+

Additional safeguards for specific AI use cases:

- Prohibition of certain harmful AI practices

- Additional safeguards for permitted use of remote biometric identification (RBI) systems, deep fakes, chatbots.

Obligations

a. Sector specific obligations for providers (could be similar to Option 1, but adapted to ad hoc sectoral acts)

b. Sector specific obligations for users depending on the use case (e.g. human oversight, transparency in specific cases etc.)

Ex ante enforcement

Would depend on the enforcement system under the relevant sectoral acts

For use of remote biometric identification (RBI) systems at publicly accessible spaces (when permitted): prior authorisation required by public authorities

Ex post enforcement

Ex post monitoring by competent authorities under the relevant sectoral acts

Governance

Would depend on the existing structures in the sectoral acts at national and EU level; no platform for cooperation between various competent authorities.

5.3.1. Scope of the ad-hoc sectoral acts

It would be for each ad-hoc piece of legislation to determine what constitutes risky AI applications that require regulatory intervention. These different acts might also adopt different definitions of AI which would increase legal uncertainty and create inconsistencies across sectors, thus failing to address effectively problems 4 and 6.

To address problems 1 and 2, the ad hoc approach would target both risks to fundamental rights and safety and cover the following sectoral initiatives:

·With regard to AI which are safety components of products covered by new approach or old-approach safety legislations, this option would entail the review of those legislations so as to include dedicated requirements and obligations addressing safety and security risks and, to the extent appropriate, fundamental right risks related to the AI safety components of those products which are considered high-risk. 212  

·With regard to other AI with mainly fundamental rights implications, the sectoral approach would exclude integration of the new specific requirements for AI in the data protection legislation, because it is designed as a technology-neutral legislation covering automated personal data processing in general (i.e. automated and non-automated processing). This means that each specific AI use case posing high risks to fundamental rights would have to be regulated through new ad-hoc initiatives or integrated into existing sectoral legislation to the extent that such exist. 213

5.3.2. Ad hoc sector specific AI requirements and obligations for providers and users

The content of the ad hoc initiatives outlined above would include: a) ad hoc sector specific requirements and obligations for providers and users of certain risky AI systems; b) additional safeguards for the permitted use of remote biometric identification systems in publicly accessible places; and c) certain prohibited harmful AI practices.

a)Ad hoc sector specific AI requirements and obligations for providers and users

Firstly, the ad hoc sectoral acts would envisage AI requirements and obligations for providers similar to those in Option 1, but specifically tailored to each use case. This means they may be different from one use case to another, which would allow consideration of the specific context and the sectoral legislation at place. This approach would also encompass the full AI value chain with some obligations placed on users, as appropriate for each use case. Examples include obligations for users to exercise certain human oversight, prevent and manage residual risks, keep certain documentation, inform people when communicating with an AI system, in case the latter believe they are interacting with a human, 214 and label deep fakes, if not used for legitimate purposes so as to prevent the risk of manipulation. 215  

However, this ad hoc approach would also lead to sectoral market fragmentation and increase the risk of inconsistency between the new requirements and obligations, in particular where multiple legal frameworks apply to the same AI system. All these potential inconsistencies could further increase legal uncertainty and market fragmentation (problems 4 and 6). The high number of pieces of legislation concerned would also make the timelines of the relevant initiatives unclear and potentially very long with the mistrust in AI further growing (problem 5).

b)Additional safeguards for the permitted use of remote biometric identification systems in publicly accessible places

One very specific and urgent case that requires regulatory intervention is the use of remote biometric identification systems in publicly accessible spaces. 216  EU data protection rules prohibit in principle the processing of biometric data for the purpose of uniquely identifying a natural person, except under specific conditions. In addition, a dedicated ad hoc instrument would prohibit certain uses of remote biometric identification systems in publicly accessible spaces given their unacceptable adverse impact on fundamental rights, while other uses of such systems would be considered high-risk because they pose significant risks to fundamental rights and freedoms of individuals or whole groups thereof.  217

For these high-risk uses, in order to balance risks and benefits, in addition to the requirements and obligations placed on the provider before the system is placed on the market (as per option 1), the ad hoc instrument would also impose additional safeguards and restrictions on the use of such system. In particular, uses of remote facial recognition systems in publicly accessible places that are not prohibited would require submission of a data protection impact assessment by the user to the competent data protection authority to which the data protection authority may object within a defined period. Additional safeguards would ensure that such use for legitimate security purposes is limited only to competent authorities. It would have to comply with strict procedural and substantive conditions justifying the necessity and proportionality of the interference in relation to the people who might be included in the watchlist, the triggering events and circumstances that would allow the use and strict limitations on the permitted geographical and temporal scope. All these additional safeguards and limitations would be on the top of the existing data protection legislation that would continue to apply by default.

An alternative policy option requested by some civil society organisations is to prohibit entirely the use of these systems in publicly accessible spaces, which would however prevent their use in duly justified limited cases for security purposes (e.g., in the presence of an imminent and foreseeable threat of terrorism or for identifying offenders of a certain numerus clausus of serious crimes when there is a clear evidence that they are likely to occur in a specific place at a given time). Another option would be not to impose any further restrictions on the use of remote biometric identification in publicly accessible places and apply only the requirements for Trustworthy AI (as per option 1). However, this policy choice was also discarded as it would not effectively address the high risks to fundamental rights posed by these systems and the current potential for their arbitrary abuse without an effective oversight mechanism and limitations on the permitted use (problems 2 and 3).

Stakeholders views: In the public consultation on the White Paper on AI, 28% of respondents supported a general ban of this technology in public spaces, while another 29.2% required a specific EU guideline or legislation before such systems may be used in public spaces. 15% agreed with allowing remote biometric identification systems in public spaces only in certain cases and under conditions and another 4.5% asked for further requirements (on top of the 6 requirements for high-risk applications proposed in the white paper) to regulate such conditions. Only, 6.2% of respondents did not think that any further guidelines or regulations are needed. Business and industry were more likely to have no opinion on the use of remote biometric identification, or to have a slightly more permissive stance: 30.2% declared to have no opinion on this issue, 23.7% would allow biometric identification systems in public spaces only in certain cases and under conditions, while 22.4% argued for a specific EU guideline or legislation before such systems may be used in public spaces. On the other hand, civil society was more likely to call for bans (29.5%) or specific EU guidelines/legislation (36.2%). 55.4% of the respondent citizens were most likely to call for a ban, while academia (39%) were more supportive of specific EU guidelines/legislation.

The Commission has also registered a European Citizens' Initiative entitled ‘Civil society initiative for a ban on biometric mass surveillance practices'.

c)Prohibition of certain harmful AI practices

Finally, to increase legal certainty and set clear red lines when AI cannot be used (problems 2 and 4), the ad-hoc approach would also introduce dedicated legislation to prohibit certain other particularly harmful AI practices that go against the EU values of democracy, freedom and human dignity, and violate fundamental rights, including privacy and consumer protection. 218 Alternatively, these could be integrated into relevant existing laws once reviewed. 219  

Evidence and analysis in the problem definition suggest that exiting legislation does not provide sufficient protection and there is a need for the prohibitions of i) certain manipulative and exploitative AI systems, and ii) general purpose social scoring:

I.AI systems that manipulate humans through subliminal techniques beyond their consciousness or exploit vulnerabilities of specific vulnerable groups such as children in order to materially distort their behaviour in a manner that is likely to cause these people psychological or physical harm. As described in the problem section, this prohibition is justified by the increasing power of algorithms to subliminally influence human choices and important decisions interfering with human agency and the principle of personal autonomy. This prohibition is consistent with a number of recommendations of the Council of Europe 220 and UNICEF. 221  

II.AI systems used for general purpose social scoring of natural persons done by public authorities defined as large scale evaluation or classification of the trustworthiness of natural persons based on their social behaviour in multiple contexts and/or known or predicted personality characteristics that lead to detrimental treatment in areas unrelated to the context in which the information was collected, including by restricting individual’s fundamental rights or limiting their access to essential public services. This prohibition is justified because such mass scale citizens’ scoring would unduly restrict individuals’ fundamental rights and be contrary to the values of democracy, freedom and the principles that all people should be treated as equals before the law and with dignity. A similar prohibition of social scoring was requested by more than 60 NGOs and also recommended by the HLEG. 222 The EP has also recently requested a prohibition of intrusive citizens’ mass scale social scoring in one of its reports on AI. 223

Other manipulative and exploitative practices enabled by algorithms that are usually identified as harmful (e.g., exploitative profiling and micro-targeting of voters and consumers) were considered as potential candidates for prohibition but discarded, since these problems have been specifically examined and targeted by the recent proposal for the Digital Services Act. 224 To a large extent, they are also already addressed by existing Union legislation on data protection and consumer protection that impose obligations for transparency, informed consent/opt out and prohibit unfair commercial practices, thus guaranteeing the free will and choice of people when AI systems are used. Furthermore, other prohibitions requested by NGOs (e.g., in relation to predictive policing, use of AI for allocation of social security benefits, in border and migration control and AI-enabled individualised risk assessments in the criminal law) 225 were also considered, but eventually discarded. That is because the new requirements for trustworthy AI proposed by the sectoral ad hoc acts under this option would aim to address these problematic use cases and ensure that the AI systems used in these sensitive contexts are sufficiently transparent, non-discriminatory, accurate and subjected to meaningful human oversight without the need to prohibit outright the use of AI in these contexts that could also be beneficial, if subjected to appropriate safeguards. 226

5.3.3. Enforcement and governance of the ad hoc sectoral acts

The enforcement mechanism would depend on each sector and legislative act, and could vary according to the specific class of applications. For example, in the context of safety legislation, enforcement of the new requirements and obligations would be ensured through the existing ex ante conformity assessment procedures and the ex post market surveillance and monitoring system. The enforcement of the prohibited practices would take place through ex post monitoring by the competent data protection or consumer protection authorities. The new rules on remote biometric identification systems would be enforced by the data protection authorities responsible for the authorisation of their use subject to the conditions and limitations in the new instrument and the existing data protection legislation.

As regards the governance system, the sectoral national authorities under each framework would be responsible for supervising the compliance with the new requirements and obligations. There would be no platform for cooperation, nor possibilities for joint investigations between various competent authorities responsible for the implementation of the ad hoc sectoral legislation applicable to AI. The cooperation at EU level would be limited to the existing mechanisms and structures under each sectoral act. Competent authorities would still be provided with new powers, resources and procedures to enforce the sectoral rules applicable to certain specific AI systems, which could partially address problem 3.

5.4. Option 3: Horizontal EU legislative instrument establishing mandatory requirements for high-risk AI applications

Option 3 would envisage a horizontal EU legislative instrument applicable to all AI systems placed on the market or used in the Union following a proportionate risk-based approach. The horizontal instrument would establish a single definition of AI (section 5.4.1) and harmonised horizontal requirements and obligations to address in a proportionate, risk-based manner and limited to the strictly necessary the risks to safety and fundamental rights specific to AI (section 5.4.2). A common system for enforcement and governance of the new rules would also be established applicable across the various sectors (section 5.4.3) complemented with specific measures to support innovation in AI (section 5.4.4).

The rules would be uniform and harmonised across all Member States which would address the problems of legal uncertainty and market fragmentation and help citizens and companies build trust in the AI technology placed on the Union market (problems 4 to 6).

Table 6.3. Summary of Option 3: Horizontal risk-based act on AI

Nature of act

A single binding horizontal act following a risk-based approach

Scope

OECD definition of AI (reference point also for other sectoral acts); clear methodology and criteria how to determine what constitutes a high-risk AI system

Content

Risk-based approach:

a. Prohibited AI practices and additional safeguards for the permitted use of remote biometric identification systems in publicly accessible spaces (as per Option 2)

b. Horizontal requirements as per Option1, but binding for high-risk AI and operationalized through harmonised standards

c. Minimal transparency for non-high-risk AI (inform when using chatbots and deep fakes as per Option 2)

+Measures to support innovation (sandboxes etc.)

Obligations

Binding horizontal obligations for all actors across the value chain:

a. Providers of high-risk AI systems as per Option 1 + conformity (re-) assessment, reporting of risks/breaches etc.

b. Users of high-risk AI systems (human oversight, monitoring, minimal documentation)

Ex ante enforcement

Providers:

a. Third party conformity assessment for high-risk AI in products (under sectoral safety legislation)

b. Mainly ex ante assessment through internal checks for other high-risk AI systems + registration in a EU database

Users: Prior authorisation for use of Remote biometric identification in publicly accessible spaces (as per Option 2)

Ex post enforcement

Ex post monitoring by market surveillance authorities designated by Member States

Governance

Governance at national level with a possibility for joint investigations between different competent authorities + cooperation at EU level within an AI Board

5.4.1. A single definition of AI

227 Like option 1, the horizontal instrument would build on the internationally recognized OECD definition of an AI system, because it is technology neutral and future proof. To provide legal certainty, the broad definition may be complemented with a list of specific approaches and techniques that can be used for the development of AI systems with some flexibility to change the list to respond to future technological developments. The definition of AI in the horizontal act would act as a reference point for other sectoral legislation and would ensure consistency across the various legislative frameworks applicable to AI, thus enhancing legal certainty for operators and reducing the risk of sectoral market fragmentation (problems 4 and 6).

5.4.2. Risk-based approach with clear and proportionate obligations across the AI value chain

The horizontal instrument would follow a risk-based approach where AI applications would be regulated only where strictly necessary to address the risks and with the minimum necessary regulatory burden placed on operators. The risk-based approach would have the following elements: a) prohibited AI practices and additional safeguards for the permitted use of remote biometric recognition systems in publicly accessible spaces; b) a consistent methodology for identification of high-risk AI systems; c) horizontal requirements for high-risk AI systems and clear and proportionate obligations for providers and users of these systems; d) minimal transparency requirements for certain low-risk AI systems.

a)Prohibited AI practices and additional safeguards for the permitted use of remote biometric recognition systems in publicly accessible spaces;

Firstly, the instrument would prohibit some harmful AI practices with a view to increasing legal certainty and setting clear red-lines when AI cannot be used (problems 2 and 4). These would include the same practices as envisaged in Option 2 (e.g., manipulative and exploitative AI and general-purpose social scoring of citizens). This option would also integrate the same prohibitions of certain uses of remote biometric identification systems in publicly accessible spaces and additional safeguards of the use of such systems when permitted as per Option 2.

b)A consistent methodology for identification of high-risk AI systems

228 Secondly, the instrument would introduce clear horizontal rules for a number of ‘high-risk’ AI use cases with demonstrated high-risks for safety and/or fundamental rights (problems 1, 2 and 4). The list of applications considered ‘high-risk’ would be identified on the basis of common criteria and a risk assessment methodology specified in the legal act as follows:

1.AI systems that are safety components of products would be high-risk if the product or device in question undergoes third-party conformity assessment pursuant to the relevant new approach or old approach safety legislation. 229 - 230  

2.For all other AI systems, 231 it would be assessed whether the AI system and its intended use generates a high-risk to the health and safety and/or the fundamental rights and freedom of persons on the basis of a number of criteria that would be defined in the legal proposal. 232 These criteria are objective and non-discriminatory since they treat similar AI systems similarly, regardless of the origin of the AI system (EU or non-EU). They also focus on the probability and severity of the harms to the health and safety and/or the fundamental rights, taking into account the specific characteristics of AI systems of opacity, complexity etc.

233 234 Although evidence for individual legal challenges and breaches of fundamental rights is growing, robust and representative evidence for harms inflicted by the use of AI is scarce due to lack of data and mechanisms to monitor AI as a set of emerging technology. To address these limitations, the initial assessment for the level of risk of the AI systems is based on the risk assessment methodology above and on several other sources listed in Annex 5.4.

235 236 237 Based on the evidence in the problem definition, the sources and the methodology outlined above Annex 5.3 includes the list of all sectoral product safety legislation that would be affected by the new initiative (new and old approach) and explains how the AI horizontal regulation would interplay with existing product safety legislation. For other AI systems that are mainly having fundamental rights implications, a large pool of AI use cases has been screened by applying the criteria above with Annex 5.4 identifying the initial list of high-risk AI systems proposed to be annexed to the horizontal instrument. This classification of high-risk AI systems is largely consistent with the position of the EP with certain exceptions. 

Some flexibility would be provided to ensure that the list of high-risk AI systems is future proof and can respond to technological and market developments by empowering the Commission within preliminary circumscribed limits to amend the list of specific use cases through delegated acts. 238 Any change to the list of high-risk AI use cases would be based on the solid methodology described above, supporting evidence and expert advice. 239 To ensure legal certainty, future amendments would also require impact assessment following broad stakeholder consultation and there would always be a sufficient transitional period for adaptation before any amendments become binding for operators.

In contrast to the risk-based approach presented above, an alternative could be to place the assessment of the risk as a burden on the provider of the AI system and foresee in the legislation only general criteria for the risk assessment. This approach could make the risk assessment more dynamic and capture high-risk use cases that the initial assessment proposed by the Commission may miss. This option has been, however, discarded because the economic operators would face significant legal uncertainty and higher burden and costs for understanding whether the new rules would apply in their case.

Stakeholders views: During the public consultation on the White Paper on AI, the limitation of requirements to high-risk AI applications was supported by 42.5% while 30.6% doubted such limitation. A majority (51%) of SMEs favoured limiting new compulsory requirements to high-risk applications, 21% opposed this. With regard to large businesses, a clear majority also favoured such an approach as well as the academic/research institutions. The stance of most civil society organisations differed from this view: more organisations opposed rather than supported this approach. At the same time, several organisations advocated fundamental or human rights impact assessments and cautioned against creating loopholes, for example regarding data protection, for low-risk applications. Of those stakeholders opposing the idea of limiting new requirements to high-risk AI applications, almost half were EU citizens (45%), with civil society and academic and research institutions being the second-largest groups (18% and 15%, respectively). For all these groups, this was higher than their share in the composition of the overall sample.

c)Horizontal requirements for high-risk AI systems and obligations on providers and users

The instrument would define horizontal mandatory requirements for high-risk AI systems that would have to be fulfilled for any high-risk AI system to be permitted on the Union market or otherwise put into service. The same requirements would apply regardless of whether the high-risk AI system is a safety component of a product or a stand-alone application with mainly fundamental rights implications (systems covered by both Annex 5.3. and Annex 5.4).

The requirements would be the same as in option 1 (incl. data, algorithmic transparency, traceability and documentation etc.), but operationalized by means of voluntary technical harmonized standards. In line with the principles of the New Legislative Framework, these standards would provide a legal presumption of conformity with requirements and constitute an important means to facilitate providers in reaching and demonstrating legal compliance. 240 The standards would improve consistency in the application of the requirements as compared to the baseline and ensure compatibility with Union values and applicable legislation, thus contributing to all 4 specific objectives. The reliance on harmonised standards would also allow the horizontal legal framework to remain sufficiently agile to cope with technological progress. While the legal framework would contain only high-level requirements setting the objectives and the expected outcomes, technological solutions for implementation would be left to more flexible market-driven standards that are updated on a regular basis to reflect technological progress and state-of-the-art. The governance mechanism of the European standardisation organisations who are usually mandated to produce the relevant harmonised standards would also ensure full consistency with ongoing and future standardisation activities at international level. 241

Furthermore, the instrument would place clear, proportionate and predictable horizontal obligations on providers of ‘high-risk’ AI systems placing such system on the Union market as well as on users. 242  Considering the various sources of risks and the AI specific features, responsibility would be attributed for taking reasonable measures necessary as the minimum to ensure safety and respect of existing legislation protecting fundamental rights throughout the whole AI lifecycle (specific objectives 1, 2, 3 and 4).

Figure 7: Obligations for providers of high-risk AI systems

Figure 8: Obligations for users of high-risk AI systems

These clear and predictable requirements for high-risk AI systems and obligations placed on all AI value chain participants are mostly common practice for diligent market participants and would ensure minimum degree of algorithmic transparency and accountability in the development and use of AI systems. Without creating new rights, these rules would help to ensure that reasonable and proportionate measures are taken to avoid and mitigate the risks to fundamental rights and safety specific to AI systems and ensure that the same rules and rights in the analogue world apply when high-risk AI systems are used (problems 1 and 2). The requirements for algorithmic transparency and accountability, and trustworthy AI would be enforceable and effectively complied with (problem 3) and businesses and other operators would also have legal certainty on who does what and what are the good practice and state-of-the-art technical standards to demonstrate compliance with the legal obligations (problem 4). These harmonised rules across all sectors would also help to increase the trust of citizens and users that AI use is safe, trustworthy and lawful (problem 5) and prevent unilateral Member States actions that risk to fragment the market and to impose even higher regulatory burdens on operators developing or using AI systems (problem 6).

d)Minimal transparency obligations for non-high-risk AI systems

For all other non-high risk AI systems, the instrument would not impose any obligations or restrictions except for some minimal transparency obligations in two specific cases where people might be deceived (problem 2) which are not effectively addressed by existing legislation 243 . This would include:

·Obligation to inform people when interacting with an AI system (chatbot) in cases where individuals might believe that they are interacting with another human being;

·Label deep fakes except when these are used for legitimate purposes such as to exercise freedom of expression and subject to appropriate safeguards for third parties’ rights.

These minimal transparency obligations would apply irrespective of whether the AI system is embedded in products or not. All other non-high-risk AI systems would be shielded from potentially diverging national regulations which would stimulate the creation of a single market for trustworthy AI and prevent the risk of market fragmentation for this substantial category of non-high-risk AI systems (problems 4, 5 and 6).

5.4.3. Enforcement of the horizontal instrument on AI

For the enforcement of the horizontal instrument, there are three options: a) ex post system; b) ex ante system; or c) a combination of ex ante and ex post enforcement.

a)Ex post enforcement of the horizontal instrument

Firstly, enforcement could rely exclusively on an ex-post system for market surveillance and supervision to be established by national competent authorities designated by the Member States. 244 Their task would be to control the market and investigate compliance with the obligations and requirements for all high-risk AI systems already placed on the market. Market surveillance authorities would have all powers under Regulation (EU) 2019/1020 on market surveillance, including inter alia powers to:

·follow up on complaints for risks and non-compliance;

·make on-site and remote inspections and audits of the AI systems;

·request documentations, technical specifications and other relevant information from all operators across the value chain, including access to source code and relevant data;

·request remedial actions from all concerned operators to eliminate the risks, or where the non-compliance or the risk persists - prohibit or order its withdrawal, recall from the market or immediate suspension of its use;

·impose sanctions for non-compliance with the obligations with proportionate, dissuasive and effective penalties; 245  

·benefit from the EU central registration database established by the framework as well as from the EU RAPEX system for the exchange of information among authorities on risky products.

Member States would have to ensure that all national competent authorities are provided with sufficient financial and human resources, expertise and competencies in the fields of AI, including fundamental rights and safety risks related to AI to effectively fulfil their tasks under the new instrument. The minimal transparency obligations for low-risk AI and the prohibited AI practices would also be enforced ex post. In order to avoid duplications, for high-risk AI systems which are safety components of products, covered by sectoral safety legislations, the ex-post enforcement of the horizontal instrument would rely on existing market surveillance authorities designated under those legislations (see more details in Annex 5.3).

The governance system would also enable cooperation between market surveillance authorities and other competent authorities supervising enforcement of existing Union and Member State legislation (e.g., equality bodies, data protection) as well as with authorities from other Member States. The mechanism for cooperation would also include new opportunities for exchange of information and joint investigations at national level as well as in cross border cases. All these new powers and resources for market surveillance authorities and mechanisms for cooperation would aim to ensure effective enforcement of the new rules and the existing legislation on safety and fundamental rights (problem 3).

b)Ex ante enforcement of the horizontal instrument

246 Secondly, ex ante conformity assessment procedures could be made mandatory for high-risk AI systems in consistency with the procedures already established under the existing New-Legislative Framework (NLF) product safety legislation. After the provider has done the relevant conformity assessment, it should register stand-alone AI system with mainly fundamental rights implications in an EU database that would be managed by the Commission. This would allow competent authorities, users and other people to verify if the high-risk AI system complies with the new requirements and to ensure enhanced oversight by the public authorities and the society over these systems (problems 3 to 5).

The ex-ante verification (through internal checks or with the involvement of a third-party) could be split according to the type of risks and level of interplay with existing EU legislation on product safety.

Figure 9: Types of ex ante conformity assessment procedures

247 In any of the cases above, recurring re-assessments of the conformity would be needed in case of substantial modifications to the AI systems (changes which go beyond what is pre-determined by the provider in its technical documentation and checked at the moment of the ex-ante conformity assessment). In order to operationalise this approach for continuously learning AI systems and keep the administrative burden to a minimum, the instrument would clarify that: 1) a detailed description of pre-determined algorithm changes and changes in performance of the AI systems during their lifecycle (with information for solutions envisaged to ensure continuous compliance), should be part of the technical documentation and evaluated in the context of the ex-ante conformity assessment; 2) changes that have not been pre-determined at the moment of the initial conformity assessment and are not part of the documentation would require a new conformity assessment.

248 Harmonised standards to be adopted by the EU standardisation organisations would play a key role in facilitating the demonstration of compliance with the ex-ante conformity assessment obligations. For remote biometric identification systems or where foreseen by sectoral product safety legislation, providers could replace the third-party conformity assessment with an ex-ante conformity assessment through internal checks, provided that harmonised standards exist and they have complied with those standards. Work on AI standardisation is already ongoing, many standards, notably of foundational nature, have already been produced and the preparation of many others is ongoing. The assumption of the Commission is that a large set of relevant harmonised standards could be available within 3-4 years from now that would coincide with the timing needed for the legislative adoption of the proposal and the transitional period envisaged before the legislation becomes applicable to operators.

c)Combination of an ex ante and ex post system of enforcement

249 As a third option, the ex-ante enforcement could be combined with an ex-post system for market surveillance and supervision as described above. Since this option most effectively addressed problems 1, 2 and 3, the combination between ex post and ex ante enforcement system has been chosen and the other alternatives discarded.

Figure 10: Compliance and enforcement system for high-risk AI systems

d)Alternative policy choices for the system of enforcement and obligations

Four alternative policy choices for the ex-ante obligations and assessment have also been considered: i) distinction between safety and fundamental rights compliance; ii) ex-ante conformity assessment through internal checks, or third party conformity assessment for all high-risk AI systems; iii) registration of all high-risk AI systems in the EU database or no registration at all, or iv) additional fundamental rights/algorithmic impact assessment.

I)Distinction between safety and fundamental rights compliance

A first alternative approach could be to apply an NLF-type ex-ante conformity assessment only to AI systems with safety implications, while for AI systems posing fundamental rights risks (Annex 5.4.) the instrument could envisage documentation and information requirements for providers and more extensive risk-management obligations for users. This approach was however discarded. Given the importance of the design and development of the AI system to ensure its trustworthiness and compliance with both safety and fundamental rights, it is appropriate to place responsibilities for assessment on the providers. This is because users are already bound by the fundamental rights legislation in place, but there are gaps in the material scope of the existing legislation as regards the obligations of producers as identified in problem 3 of this impact assessment.

II)Ex-ante conformity assessment through internal checks or third party conformity assessment for all high-risk AI systems

A second alternative would be to apply ex-ante conformity assessment through internal checks for all high-risk AI systems or to foresee third-party involvement for all high-risk AI systems. On the one hand, a comprehensive ex-ante conformity assessment through internal checks, combined with a strong ex-post enforcement, could be an effective and reasonable solution given the early phase of the regulatory intervention and the fact the AI sector is very innovative and expertise for auditing is just about to be accumulated. The assessment through internal checks would require a full, effective and properly documented ex ante compliance with all requirements of the regulation and compliance with a robust quality and risk management systems and post-market monitoring. Equipped with adequate resources and new powers, market surveillance authorities would also ensure ex officio enforcement of the new rules through systematic ex-post investigations and checks; request remedial actions or withdrawal of the risky or non-compliant AI systems from the market and/or impose financial sanctions. On the other hand, high-risk AI systems that are safety components of products are, by definition, already subject to the third party conformity assessment foreseen for the relevant product under sectoral product safety legislation (see more details in Annex 5.3), so the new horizontal initiative should not disrupt but rather integrate into that system.

In conclusion, the combination of the two alternatives reflects regulatory and practical considerations and results in an appropriate mix of enforcement tools to deal respectively with safety and fundamental right risks. 250

III)Require registration of all high-risk AI systems in the EU database or no registration at all

As to the registration obligation applicable only to stand-alone AI systems with mainly fundamental rights implications (Annex 5.4), an alternative policy choice would be to require registration of any high-risk AI system, including systems that are safety components of products or devices. However, this option was discarded because this latter category of AI systems might already be subject to registration according to the existing product safety legislation (e.g. medical device database) and duplication of databases should be avoided. Furthermore, in the scenario where sectoral safety legislation does not establish a registration obligation for the products, the registration in a central database of high-risk AI systems that are components of products would prove to be of limited value for the public and the market surveillance authorities given that the product as a whole is not subject to central registration obligations.

A second alternative would be not to require registration even for the high-risk AI systems with fundamental rights implications, but this policy choice was also discarded. The reason is that without such a public database the specific objectives of the initiative would be compromised, particularly in relation to increasing public trust and the enforceability of the existing law on fundamental rights (problems 3 and 5). Keeping the registration obligation for these systems with fundamental rights implications is thus justified given the need for increased transparency and public oversight over these systems. 251  

IV)Require an additional fundamental rights/algorithmic impact assessment

Another alternative for high-risk AI systems with fundamental rights implications would be to require a fundamental rights impact assessments/algorithmic impact assessments as implemented in Canada and the U.S. and recommended by some stakeholders, the Council of Europe 252 and the Fundamental Rights Agency. 253 However, this was also discarded, because users of high-risk AI systems would normally be obliged to do a Data Protection Impact Assessment (DPIA) that already aims to protect a range of fundamental rights of natural persons and which could be interpreted broadly, so new regulatory obligation was considered unnecessary.

Stakeholders views: During the public consultation on the White Paper on AI, 62% of respondents supported a combination of ex-post and ex-ante market surveillance systems. 3% of respondents support only ex-post market surveillance. 28% supported third party conformity assessment of high-risk applications, while 21% of respondents support ex-ante self-assessment. While all groups of stakeholders had the combination of ex-post and ex-ante market surveillance systems as their top choice, industry and business respondents preferred ex-ante self-assessment to external conformity assessment as their second best choice.

5.4.4. Governance of the horizontal instrument on AI

The governance system would include enforcement at national level with a cooperation mechanism established at EU level.

At national level, Member States would designate competent authorities responsible for the enforcement and supervision of the new rules and the ex post market surveillance. As explained in details above, they should be provided with competences to fulfil their tasks in an effective manner, ensuring that they have adequate funding, technical and human resource capacities and mechanisms to cooperate given that a single AI system may fall within the sectoral competences of a number of regulators or some AI systems may be currently not supervised at all. The new reporting obligations for providers to inform competent authorities in case of incidents and breaches of fundamental rights obligations of which they have become aware would also significantly improve the effective enforcement of the rules (problem 3).

At EU level, coordination would be ensured through a mechanism for cross-border investigations and consistency in implementation across Member States and the establishment of a dedicated EU body (e.g. EU Board on AI) 254 responsible for providing uniform guidance on the new rules. 255 The establishment of an AI Board is justified by the need to ensure a smooth, effective and uniform implementation of the future AI legislation across the whole EU. Without any governance mechanism at EU level, Member States could interpret and apply very differently the new rules and would not have a forum to reach consistency and cooperate. This would fail to enhance governance and effective enforcement of fundamental rights and safety requirements applicable to AI systems (problem 3). Eventually, the divergent and ineffective application of the new rules would also lead to mistrust and lower level of protection, legal uncertainty and market fragmentation that would also endanger specific objectives 1, 3 and 4. Since there is no other body at EU level that encompasses the full range of competences to regulate AI across all different sectors in relation to both fundamental rights and safety, establishing a new EU body is justified.

5.4.5. Additional measures to support innovation

In line with specific objective 3, Option 3 would also envisage additional measures to support innovation including: a) AI regulatory sandboxing scheme and b) other measures to reduce the regulatory burden and support SMEs and start-ups.

a)AI regulatory sandboxing scheme

The horizontal instrument would provide the possibility to create AI regulatory sandboxes by one or more competent authorities from Member States at national or EU level. The objective would be to enable experimentation and testing of innovative AI technologies, products or services for a limited time before their placement on the market and pursuant to a specific testing plan under the direct supervision by competent authorities ensuring that appropriate safeguards are in place. 256  Through direct supervision and guidance by competent authorities, participating providers would be assisted in their efforts to reach legal compliance with the new rules, benefitting from increased legal certainty on how the rules should apply to their concrete AI project (problem 4). This would be without prejudice to the powers of other supervisory authorities who are not associated to the sandboxing scheme.

The instrument would set clear limits for the experimentation. No derogations or exemptions from the applicable legislation would be granted, taking into account the high risks to safety and fundamental rights and the need to ensure appropriate safeguards. 257 Still, the competent authorities would have certain flexibility in applying the rules within the limits of the law and within their discretionary powers when implementing the legal requirements to the concrete AI project in the sandbox. 258 Any significant safety risks or adverse impact on fundamental rights identified during the testing of such systems should result in immediate rectification and, failing that, in the suspension of the system until such rectifications can take place. 259  

The regulatory sandboxes would foster innovation and increase legal certainty for companies and other innovators giving them a quicker access to the market, while minimising the risks for safety and fundamental rights and fostering effective compliance with the legislation through authoritative guidance given by competent authorities (problems 1, 2, 3 and 4). They would also provide regulators with new tools for supervision and hands-on experience to detect early on emerging risks and problems or possible need for adaptations to the applicable legal framework or the harmonised technical standards (problem 3). Evidence from the sandboxes would also help national authorities identify new high-risk AI use cases that would further inform the regular reviews by the Commission of the list of high-risk AI systems to amend it, as appropriate.

b)Other measures to reduce the regulatory burden and support SMEs and start-ups

To further reduce the regulatory burden on SMEs and start-ups, the national competent authorities could envisage additional measures such as provision of priority access to the AI regulatory sandboxes, specific awareness raising activities tailored to the needs of the SMEs and start-ups etc.

Notified bodies should also take into account the specific interests and needs of SMEs and start-ups when setting the fees for conformity assessment and reduce them proportionately.

As part of the implementing measures, Digital Innovation Hubs and Testing Experimentation Facilities established under the Digital Europe Programme could also provide support as appropriate. This could be achieved by providing, for example, by relevant training to providers on the new requirements and upon request, providing relevant technical and scientific support as well as testing facilities to providers and notified bodies in order to support them in the context of the conformity assessment procedures.

Stakeholders views: Stakeholders suggested different measures targeted at fostering innovation in the public consultation on the White Paper. For example, out of the 408 position papers that were submitted, at least 19 discussed establishing regulatory sandboxes as one potential pathway to better allow for experimentation and innovation under the new regulatory framework. Generally, at least 19 submissions cautioned against creating regulatory burdens that are too heavy for companies and at least 12 submissions highlighted the benefits of AI as a factor to be taken into account when contemplating new regulation, which might create obstacles in reaping those benefits. At least 12 Member States supported regulatory sandboxes in their national strategies. The Council Conclusion on regulatory sandboxes (13026/20) also highlight that regulatory sandboxes are increasingly used in different sectors, can provide an opportunity for advancing regulation through proactive regulatory learning and support innovation and growth of all businesses, especially SMEs.

5.5. Option 3+: Horizontal EU legislative instrument establishing mandatory requirements for high-risk AI applications + voluntary codes of conduct for non-high risk applications

Option 3+ would combine mandatory requirements and obligations for high-risk AI applications as under option 3 with voluntary codes of conduct for non-high risk AI.

Table 6.4. Summary of Option 3+

Nature of act

Option 3 + code of conducts non high-risk AI

Scope

Option 3 + voluntary codes of conduct non-high-risk AI

Content

Option 3 + industry-led codes of conduct for non-high-risk AI

Obligations

Option 3 + commitment to comply with codes of conduct for non-high-risk AI

Ex ante enforcement

Option 3 + self- assessment for compliance with codes of conduct for non- high-risk AI

Ex post enforcement

Option 3 + unfair commercial practice in case of non-compliance with codes

Governance

Option 3 + without EU approval of the codes of conduct

Under this option, the Commission would encourage industry associations and other representative organisations to adopt voluntary codes of conduct so as to allow providers of all non-high-risk applications to voluntarily comply with similar requirements and obligations for trustworthy AI. These codes could build on the existing self-regulation initiatives described in the baseline scenario and adapt the mandatory requirements for Trustworthy AI to the lower risk of the AI system. It is important to note that the obligatory minimal transparency obligations for non-high-risk AI systems under option 3 would continue to apply simultaneously with the voluntary codes of conduct.

The proposed system of voluntary codes of conduct would be light for companies to subscribe and not include a ‘label’ or a certification of AI systems. Combination with a voluntary labelling scheme for low-risk AI was discarded as an option because it could be still too complex and costly for SMEs to comply with. Furthermore, a separate label for trustworthy AI may create confusion with the CE label that high-risk AI systems would obtain under option 3. Under a voluntary labelling scheme, it would also be very complex and lengthy to create standards suitable for a potentially very high number of non-high-risk AI systems. Last but not least, such a voluntary labelling scheme for non-high-risk AI also received mixed reactions in the stakeholder consultation.

 Stakeholders views: During the public consultation on the White Paper on AI, 50.5% of respondents found voluntary labelling useful or very useful for non-high-risk application, while another 34% of respondents did not agree with that approach. Public authorities, industry and business and private citizens were more likely to agree, while non-governmental organisations were divided. The Council conclusions of 9 June 2020 specifically called upon the Commission to include a ‘voluntary labelling scheme that boosts trust and safeguards security and safety’. In a recent non-paper, representatives from ministries of 14 Member States call for a ‘voluntary European labelling scheme’ and in its recent resolution the European Parliament also envisaged it for non-high risk AI systems.Stakeholders views: During the public consultation on the White Paper on AI, 50.5% of respondents found voluntary labelling useful or very useful for non-high-risk application, while another 34% of respondents did not agree with that approach. Public authorities, industry and business and private citizens were more likely to agree, while non-governmental organisations were divided. The Council conclusions of 9 June 2020 specifically called upon the Commission to include a ‘voluntary labelling scheme that boosts trust and safeguards security and safety’. In a recent non-paper, representatives from ministries of 14 Member States call for a ‘voluntary European labelling scheme’ and in its recent resolution the European Parliament also envisaged it for non-high risk AI systems.

To reap the benefits of a voluntary framework for non-high risk AI, Option 3+ proposes instead that the Commission would encourage the providers of non-high risk AI systems to subscribe to and implement codes of conduct for Trustworthy AI developed by industry and other representative associations in Member States.

These industry-led codes of conduct for trustworthy AI could integrate and operationalise the main principles and requirements as envisaged under Options 1 and 3. The codes of conduct may also include other elements for Trustworthy AI that have not been included in the requirements and the compliance procedures under option 1 and 3 (e.g., proposed by HLEG, EP or Council of Europe in relation to diversity, accessibility, environmental and societal-well-being, fundamental rights or ethical impact assessments etc.). Before providers could give publicity to their adherence to a code of conduct, they should undergo the self-assessment procedure established by the code of conduct to confirm compliance with its terms and conditions. False or misleading claims that a company is complying with a code of conduct should be considered unfair commercial practices.

The Commission would not play any active role in the approval or the enforcement of these codes and they would remain entirely voluntary. As part of the review clause of the horizontal instrument, the Commission would evaluate the proposed scheme for codes of conduct for non-high risk AI and, building on the experience and the results, propose any necessary amendments.

5.6. Option 4: Horizontal EU legislative instrument establishing mandatory requirements for all AI applications, irrespective of the risk they pose

Under this option, the same requirements and obligations as the ones for option 3 would be imposed on providers and users of AI systems, but this would be applicable for all AI systems irrespective of the risk they pose (high or low).

Table 6.5. Summary of Option 4: Horizontal act for all AI systems

Nature of act

A single binding horizontal act, applicable to all AI

Scope

OECD definition of AI; applicable to all AI systems without differentiation between the level of risk

Content

Same as Option 3, but applicable to all AI systems (irrespective of risk)

Obligations

Same as Option 3, but applicable to all AI systems (irrespective of risk)

Ex ante enforcement

Same as Option 3, but applicable to all AI systems (irrespective of risk)

Ex post enforcement

Same as Option 3, but applicable to all AI systems (irrespective of risk)

Governance

Same as Option 3, but applicable to all AI systems (irrespective of risk)

5.7. Options discarded at an early stage

No options were discarded from the outset. However, in analysing specific policy options certain policy choices were made (i.e. sub-options within the main option. This selection of sub-options is summarized in Table 7 below.

Table 7: Summary of selected and discarded sub-options

POLICY OPTIONS

SELECTED SUB-OPTION

DISCARDED ALTERNATIVE SUB-OPTIONS

Relevant for Option 1, 3, 3+ and 4

OECD Definition of AI (technology neutral) – pp. 40 and 48

Technology-specific definition (e.g. Machine learning)

Relevant for all Options

5 requirements (proposed in the White Paper on AI) – p. 40

Other discarded requirements:

·Environmental and societal well-being

·Social responsibility and gender equality

·Privacy

·Effective redress

Relevant for Option 2, 3, 3+ and 4

Prohibitions of certain use of remote biometric identification in public spaces + additional safeguards and limitations for the permitted use (p. 45)

Other discarded alternatives:

·Complete prohibition of remote biometric identification systems in publicly accessible spaces

·Application of the requirements for trustworthy AI (as per option 1) without additional restrictions on the use

Relevant for Option 2, 3, 3+ and 4

Prohibition of other harmful AI practices:

·Manipulative and exploitative AI

·General purpose social scoring

(p. 46)

Complete prohibition of other AI uses:

·Other manipulative and exploitative AI uses (e.g. profiling and micro-targeting of voters, consumers etc.)

·Predictive policing

·AI used for allocation of social security benefits

·AI used in border and migration control

·Individualised risk assessments in the criminal law context

Relevant for Option 3 and 3+

List of high-risk AI systems identified by the legislator (pp.49-50)

Each provider is obliged to assess if its AI system is high-risk or not on the basis of criteria defined by the legislator

Relevant for Option 3 and 3+

AI systems included in the list of high-risk AI (Annex 5.4)

A larger pool of AI use cases has been screened and discarded (drawing from EP proposals, ISO report, stakeholder consultation and additional research)

Relevant for Option 3 and 3+

Transparency requirements for non-high-risk AI in relation to chatbots and labelling of deep fakes

Other AI uses such as use of automated decision affecting people, profiling and micro-targeting of individuals

Relevant for Option 1, 3, 3+ and 4

Ex ante and ex post enforcement (p.42 and pp.53-55)

Only ex ante or only ex post enforcement

Relevant for Option 3, 3+ and 4

Ex ante conformity assessment (split between assessment though internal checks and third party conformity assessment) + registration in an EU database of high-risk AI systems with fundamental rights implications (p. 54)

Discarded alternatives:

·Distinguish between safety and fundamental rights ex ante assessments

·Ex ante assessment through internal checks for all high-risk AI systems or third party conformity assessment for all high-risk AI systems

·Registration in the EU database of all high-risk AI systems or no database at all

·Additional fundamental rights/algorithmic impact assessment

Relevant for Option 3, 3+ and 4

Option1: Governance system with national competent authorities + light mechanism for EU cooperation

(p. 42)

Option 3, 3+ and 4: Governance system with national competent authorities + European AI Board (p. 57)

No cooperation at EU level

Relevant for Option 3+

Option 3 + voluntary codes of conduct for non-high-risk AI

(pp. 59-60)

Option 3 + voluntary labelling for non-high-risk AI



6.What are the impacts of the policy options?

The policy options were evaluated against the following economic and societal impacts, with a particular focus on impacts on fundamental rights.

6.1. Economic impacts

6.1.1. Functioning of the internal market

The impact on the internal market depends on how effective the regulatory framework is in preventing the emergence of obstacles and fragmentation by mutually contradicting national initiatives addressing the problems set out in section 2.1.2.4.    

Option 1 would have a limited impact on the perceived risks that AI may pose to safety and fundamental rights. A labelling scheme would give information to businesses that would wish to deploy AI and consumers when purchasing or using AI applications, thus redirecting some demand from non-labelled products to labelled products. However, the extent of this shift - and hence the incentive for both AI suppliers to adopt the voluntary label - is uncertain. Therefore, at least in some Member States, public opinion is expected to continue to put pressure towards a legislative solution, possibly leading to at least partial fragmentation.

Option 2 would address the risk of fragmentation for those classes of applications for which specific legislation is introduced. Since these are likely to be the ones where concerns have become most obvious and most urgent, it is possible that Member States will refrain from additional legislation. Where they see a need for supplementary action, they could bring it to the attention of the EU to make further EU-wide proposals. However, some Member States may consider that a horizontal approach is also needed and pursue such an approach at national level.

Options 3 and 3+ effectively address the risks set out in section 2.2. in all application areas which are classified sensitive or ‘high-risk, with option 3 in addition also ensuring a European approach for low-risk applications. 260 Hence, no Member State will have an incentive to introduce additional legislation. Where Member States would wish to classify an additional class of applications as high-risk, they have at their disposal a mechanism (the possibility to amend the list in the future) to include this class into the regulatory framework. Only if a Member State wishes to include an additional class of applications, but fails to convince the other Member States, could there be a potential risk for unilateral action. However, since the most risky application fall within the scope of the regulatory framework, it is unlikely that Member States would take such a step for a class of applications at the margin of riskiness.

Option 4 addresses the risks created by AI in all possible applications. Thus, Member States are unlikely to take unilateral action.

6.1.2.Impact on uptake of AI

Currently, in the European Union the share of companies use AI at a scale is 16% lower than in the US, where growth continues. 261 There is thus ample scope to accelerate uptake of AI in the EU. Faster uptake by companies would yield significant economic benefits. As an example, by 2030, companies rapidly adopting AI are predicted to gain about 122% in economic value (economic output minus AI-related investment and transition costs). In contrast, companies only slowly adapting AI could lose around 23% of cash flow compared with today. 262  

The regulatory framework can enhance the uptake of AI in two ways. On the one hand, by increasing users’ trust it will lead to a corresponding increase in the demand by AI using companies. On the other hand, by increasing legal certainty it will make it easier for AI suppliers to develop new attractive products which users and consumers appreciate and purchase.

Option 1 can increase users’ trust for those AI systems that have obtained the label. However, it is uncertain how many applications will apply, and hence the increase in user’s trust remains uncertain. Also, users will have more trust when they can rely on legal requirements, which they can enforce in courts if need be, than if they have to rely on voluntary commitments. Regarding legal certainty, option 1 does provide this neither to AI suppliers nor to AI using companies, since it has no direct legal effect.

Option 2 would enhance users’ trust in those types of AI applications to which regulations apply. However, regulation, whether new or amended existing legislation, would only occur once concerns have emerged, and may thus be delayed. Moreover, it would provide AI suppliers and AI using companies with legal certainty only regarding these particular classes of applications and mightlead to inconsistencies in the requirements imposed by sectorial legislations, hampering uptake.

Option 3 would enhance users’ trust towards the high-risk cases, which are those where trust is most needed. Hence, its positive effect on uptake would be precisely targeted. Moreover, it would not allow a negative reputation to build up in the first place, but ensure a positive standing from the outset. Option 3+ would in addition allow further trust building by AI suppliers and AI using companies and individuals where they see fit. Option 4 would have the same effect on trust for the high-risk cases but would in addition increase trust for many applications where this would have marginal effect. For options 3, 3+ and 4 legal certainty would rise, enabling AI suppliers to bring new products more easily to market.

6.1.3.Compliance costs and administrative burdens 263

Stakeholders views: With regard to costs arising due to regulation, more than three quarters of companies did not explicitly mention such costs. However, at least 14% of SMEs and 13% of large companies addressed compliance costs as a potential burden resulting from new legislation in their position papers. Further at least 10% and 9%, respectively, also mentioned additional administrative burdens tied to new regulation in this context.

264 The costs are calculated relative to the baseline scenario not taking into account potential national legislation. However, if the Commission does not take action, Member States would be likely to legislate against the risks of artificial intelligence. This could lead to similar or even higher costs if undertakings were to comply with distinct and potential mutually incompatible national requirements. 

265 The estimates in this section are taken from Chapter 4 “Assessment of the compliance costs generated by the proposed regulation on Artificial Intelligence” of the Study to Support an Impact Assessment of Regulatory Requirements for Artificial Intelligence in Europe”.  

The costs estimations are based on a Standard Cost Model, assessing required time (built on a reference table from Normenkontrollrat (2018)) and evaluating costs by the reference hourly wage rate indicated by Eurostat for the Information and Communication sector (Sector J in the NACE rev 2 classification). The cost estimation is built upon time expenditures of activities induced by the selected requirements for an average AI unit of an average firm in order to arrive at meaningful estimates. In practice, AI systems are very diverse, ranging from very cheap to very expensive systems.

This methodology is regularly used to estimate the costs of regulatory intervention. It assumes that businesses need to adopt measures to comply with every requirement set out. Thus, it represents the theoretical maximum of costs. For companies that already fulfil certain specific requirements, the corresponding cost for these specific requirements would in practice be zero, e.g. if they already ensure accuracy and robustness of their system, the costs of this requirement would be zero.

Option 1 would create comparable compliance costs to those of the regulatory approach (see option 3), assuming the voluntary labelling contains similar requirements. The administrative burden per AI application would likely be lower, as the documentation required without conformity assessment will be lighter. The aggregate costs would then depend on the uptake of the labelling, i.e. what share of AI applications would apply to obtain the label and can therefore range from 0 to a theoretical maximum of around €3 billion (see calculations in option 3). Presumably, only those applications would apply to obtain the label that would benefit from user trust or those that process personal data (thus excluding industrial applications), so it would be less than 100%. At the same time, there are applications that would benefit from user trust but are not high-risk applications as in option 3. Hence, the share would be above the share of option 3. However, this estimate depends on the success of the label, i.e. a general recognition of the label. Note that companies would only accept the administrative burden if they considered the costs lower than the benefits. 

For option 2, the compliance costs and administrative burden would depend on the specific subject regulated and are thus impossible to estimate at this point in time. Since only one class of applications will fall in the scope of each regulation, the share of total AI applications covered will be much smaller even than the share of high-risk applications. It is quite possible that the requirements may be more stringent, since these will be the most controversial applications. In any case, a business developing or using several of these classes of applications (e.g. remote biometric identification in publicly accessible spaces and deep fakes), would not be able to exploit synergies in the certification processes.

In option 3 there would be five sets of requirements, concerning data, documentation and traceability, provision of information and transparency, human oversight and robustness and accuracy. As a first step, it is necessary to identify the maximum costs of the measures necessary to fulfil each of these requirements and adding them up gives total compliance costs per AI application (Table 8). However, economic operators would already take a certain number of measures even without explicit public intervention. In particular, they would have to still ensure that their product actually works, i.e. robustness and accuracy. This cost would therefore only arise for companies not following standard business procedures (Table 8a). For the other requirements, operators would also take some measures by themselves, which would however not be sufficient to comply with the legal obligations. Here, they may build additional measures on top of the existing in order to achieve full compliance. In a second step, it is therefore necessary to estimate which share of these costs would be additional expenditure due to regulatory requirements. In addition, it should be noted that human oversight represents overwhelmingly an operating cost which arises to the user, if at all (depending on the use case)(Table 8b).

Table 8: Maximum fixed compliance costs and administrative burden for AI suppliers

Compliance costs regarding data

€2 763

Administrative burden regarding documentation and traceability

€4 390

Administrative burden regarding provision of information

€3 627

Table 8a: Additional costs for companies not following state-of-the-art business procedures

Compliance costs regarding robustness and accuracy

€10 733

Table 8b: Operating costs for AI users

Compliance costs regarding human oversight

€7 764

In option 3, the theoretical maximum compliance costs and administrative burden of algorithmic transparency and accountability per AI application development (the sum of the three sets of requirements for AI suppliers) amount to around €10 000 for companies following standard business procedures.

In accounting for the share of costs which correspond to a normal state-of-the-art business operation (“business-as-usual factor”), one can expect a steep learning curve, as companies will integrate the actions they take to fulfil the requirements with the actions they take for business purposes (for instance add a testing for non-discrimination during the regular testing of an AI application). As a result the adjusted maximum costs taking into account the business-as-usual can be estimated at around two thirds of the theoretical costs, i.e. on average at around €6 000 - €7 000 by 2025.

Since the average development cost of an AI system is assumed in the Standard Cost Model to be around €170 000 for the purposes of the cost calculation, this would amount to roughly 4-5%. These costs are for the software alone; when AI is embedded with hardware, the overall costs increase significantly as the project becomes much more expensive and the AI compliance costs as a share of total costs become correspondingly smaller.

It is useful to relate the estimated costs of this option with compliance costs and administrative burdens of other recent initiatives. 266 Although these costs are not strictly comparable (AI costs are per product, GDPR costs are for the first year, VAT costs are per year), they nevertheless give an idea of the order of magnitude. For example, regarding GDPR, a recent study 267 found that 40% of SMEs spent more than €10 000 on GDPR compliance in the first year, including 16% that spent more than €50 000. Another report 268 found that an average organization spent 2,000-4,000 hours in meetings alone preparing for GDPR. Another benchmark are VAT registration costs, which amount to between €2500 and €4000 annually per Member State, resulting in €80-90 000 for access to the entire EU market.

With estimates for AI investment in the EU by 2025 in the range of €30 billion to €65 billion, 4-5% of the upper estimate of €65 billion translate into a maximum estimate of aggregate compliance costs for all AI applications of about €3 billion in 2025. However, since option 3 only covers high-risk applications, one has to estimate the share of AI applications which would fall under the scope of the obligations, and adjust the costs accordingly. At this stage, it is not possible to estimate the costs precisely, since the legislator has not yet decided the list of high-risk applications. Nevertheless, given that in this option high-risk applications are based on exceptional circumstances, one could estimate that no more than 5% to 15% of all applications should be concerned by the requirements. Hence the corrected maximum aggregate compliance costs for high-risk AI applications would be no more than 5% to 15% of the maximum for all applications, i.e. €100 million to €500 million.

However, in practice the compliance costs and administrative burden for high-risk applications are likely to be lower than estimated. That is because the business-as-usual factor mentioned above has been calculated for an average AI application. For high-risk applications, companies would in any case have to take above-average precautions. Indeed, faced with sceptical or hostile parts of public opinion, companies will have to pay attention to issues like data representativeness regardless of legal obligations. As a result, the additional costs generated by the legislation would in practice be smaller than the estimated maximum.

For AI users, the costs for documentation would be negligible, since it will mostly rely on in-built functions such as use logs that the providers have installed. In addition, there would be the annual cost for the time spent on ensuring human oversight where this is appropriate, depending on the use case. This can be estimated at €5000 – €8000 per year (around 0.1 FTE).

In option 3+ the additional aggregate costs would depend on how many companies submit their applications to a code of conduct; if the requirements of the code of conduct were the same as for high-risk applications, the maximum aggregate compliance costs and administrative burden of option 3+ would lie between €100 million to €500 million and again a theoretical maximum of €3 billion. However, it is likely that the codes of conduct will have fewer requirements, since they cover less risky applications. Aggregate compliance costs are thus likely to be lower.

In option 4, since all AI applications have to comply with the obligations, 4-5% per AI application with an upper estimate of €65 billion would correspond to a maximum estimate of aggregate compliance costs for of about €3 billion in 2025.

Verification costs

In addition to meeting the requirements, costs may accrue due to the need to demonstrate that the requirements have been met.

For option 1, ex-ante conformity assessment through internal checks would be combined with ex-post monitoring by the competent authorities. The internal checks would be integrated into the development process. No external verification costs would therefore accrue.

Under option 2, rules for verification would be laid down in the specific legislative acts and are likely to be different from one use case to another. Thus, one cannot estimate them at this stage.

In option 3, for AI systems that are safety components of products under the new legislative approach, the requirements of the new framework would be assessed as part of the already existing conformity assessments which these products undergo. For remote biometric identification systems in publicly accessible places, a new ex-ante third party conformity assessment procedure would be created. Provided that harmonised standards exist and the providers have applied those standards, they could replace the third-party conformity assessment with an ex-ante conformity assessment through internal checks applying the same criteria. All other high-risk applications would equally be assessed via ex-ante conformity assessments through internal checks applying the same criteria.

Third-party conformity assessment for AI applications comes in two elements: the assessment of a quality management system the provider would have to implement, which is already a common feature in product legislation, and the assessment of the technical characteristics of the individual AI system itself (so-called EU technical documentation assessment).

Companies supplying products that are third-party conformity assessed already have a quality management system in place. Companies supplying remote biometric identification systems in publicly accessible places, which is a very controversially discussed topic, can equally be presumed to have a quality management system, since no customer would want to risk their reputation by using such a system that hasn’t been properly quality controlled. After adapting to the AI requirements, the quality management system has to be audited by the notified body and be proven compliant with the standards and the regulation. The initial audit costs between €1 000 and €2 000 per day, and the amount of days will depend on the number of employees. The audits need to be re-audited yearly, which will take less time and a correspondingly smaller costs. These costs could be further reduced when companies make use of existing standards as described above. 269 Moreover, the Regulation foresees that Notified Bodies, when setting their fees, shall take into account the needs of SMEs.

In addition, for each individual product the notified body will have to review documentation meant to prove that the product complies with the AI regulation to ensure that it is indeed compliant with the requirements. Such a review is expected to take between one and two and a half days. This amounts to a range of €3,000-7,500 for the notified body to monitor compliance with the documentation requirements.

With an assumed average cost for an AI system of €170 000, this amounts to between 2% and 5%. Applied to a maximum investment volume of €65 billion, aggregate costs would be between €1 billion and €3 billion if all AI systems were thus tested. However, only 5% to 15% of all AI applications are estimated to constitute a high risk, and only a subset of these (AI systems that are safety components of products and remote biometric identifications systems in publicly accessible places) would be subject to third-party conformity assessment. Hence, taking 5% as a reasonable estimate, aggregate costs would amount to around €100 million.

Analogue to the discussion above for compliance costs, when AI is embedded the total development costs are much higher than the software alone and the share of AI verification costs is correspondingly smaller. The share of 2% to 5% of total development costs would thus only apply to non-embedded AI applications that nevertheless need to undergo new third-party conformity assessment, i.e. remote biometric identification in publicly accessible spaces.

Again, to put these figures into perspective, the costs of other recent initiatives help. For example, for the Cybersecurity Act, in France the Certification Sécuritaire de Premier Niveau (CSPN) costs about €25,000 – €35,000 while in the Netherlands the Baseline Security Product Assessment (BSPA) costs on average €40,000 270 . Similarly, the conformity assessment for a laptop is estimated at around € 25 000. 271 The average cost of a conformity assessment under the Machinery Directive is € 275 000. 272

In option 3+, additional aggregate verification costs would consist in random checks of companies having introduced a code of conduct, financed by fees from participating companies, if the code of conduct foresees such random checks. This would amount to a fraction of the costs of verification of the high-risk application. Total aggregate costs would thus lie slightly above option 3 (around € 100 million). 

Under option 4, the assessment costs for each application would be identical, but all AI applications would be covered, resulting in aggregate costs between €1 billion and €3 billion.

Table 9: Overview: Estimated maximum aggregate compliance costs and administrative burden by 2025

COMPLIANCE + ADMIN COSTS

VERIFICATION COSTS

Option 1

Between €0 and €3 billion (all voluntary)

€0

Option 2

n/a

n/a

Option 3

Between €100 million and €500 million

Around €100 million

Option 3+

Between €100 / €500 million and €3 billion (voluntary above €100 / €500 million)

Slightly above €100 million

Option 4

Around €3 billion in 2025

Between €1 billion and €3 billion

Notabene: does not include the compliance costs for human oversight, which accrue to the user, not the AI supplier

6.1.4.SME test 

Under option 1, SMEs would only sign up to the voluntary labelling scheme if the benefits in terms of credibility outweigh the costs. Thus, proportionality is ensured by design. Option 2 limits the requirements to specific well-defined cases if and when problems arise or can be anticipated. Each ad-hoc regulation will thus only concern a small share of SMEs. SMEs working on several classes of applications subject to ad-hoc regulation, would, however, have to comply with multiple specific sets of requirement, increasing administrative burden.

Regarding SMEs, the approach proposed in options 3 and 3+, precisely targeting only a narrow set of well-defined high-risk AI applications and imposing only algorithmic transparency and accountability requirements, keeps costs to a minimum and ensures that the burden is no more than proportionate to the risk. For example, users’ record-keeping will be done automatically through system logs, which providers will be required to make available. By establishing clear requirements and procedures to follow at horizontal level, it also keeps administrative overhead as low as possible.

The vast majority of SMEs would not be affected at all, since obligations would be introduced only for high-risk applications. These non-affected SMEs would benefit from additional legal certainty, since they could be sure that their applications are not considered high-risk and will therefore not be subject to additional compliance costs or administrative burdens. The AI supplying SMEs concerned would however have to bear the limited costs just as large companies. Indeed, due to the high scalability of digital technologies, small and medium enterprises can have an enormous reach, potentially impacting millions of citizens despite their small size. Thus, when it comes to high risk applications, excluding SMEs from the application of the regulatory framework could seriously undermine the objective of increasing trust. However, they would benefit from a single set of requirements, streamlining compliance across applications. Under option 3+ SMEs that are not covered by the scope could invest in additional trust by adopting a code of conduct, if they see an economic advantage in doing so.

As with all regulations, the AI supplying SMEs concerned would in principle be more affected than large companies for several reasons. Firstly, in so far as large companies produce more AI applications, they can distribute the one-off costs of familiarising themselves (including legal advice if necessary) over more applications and would also experience a faster learning curve. Nevertheless, most of the additional fixed compliance costs generated by the legislation occur for every new application and thus do not provide economies of scope to the larger companies. Secondly, in so far as their applications find more customers they can distribute the fixed costs of regulation (such as the testing for non-discrimination effects) over more customers (economies of scale). However, many AI applications are bespoke developments for specific customers where this will not be possible, since the fixed costs may have to be incurred again (e.g. training data is likely to be different for customised applications). Thirdly, SMEs financial capacity to absorb additional burdens is much more limited. SMEs produce an average annual value added of €174 000, going as low as €69 000 for micro-enterprises (less than ten employees), compared to €71.6 million for large enterprises. 273 This compares with estimated compliance costs of €6 000 - €7 000 for those SMEs that would develop or deploy high-risk AI applications.

SMEs are also expected to benefit significantly from a regulatory framework. Firstly, small and therefore generally not well-known companies will benefit more from a higher overall level of trust in AI applications than large established companies, who already have large bases of established and trusting customers (e.g. an increase in trust in AI is less likely to benefit large platform operators or well-known e-Commerce companies, whose reputation depends on many other factors). This applies especially to the companies using AI in the business-to-consumer market, but also to the AI suppliers in the business-to-business market, where customers will value the reassurance that they are not exposed to legal risks from the application they are purchasing or licensing. Secondly, legal uncertainty is a bigger problem for SMEs than for large companies with their own legal department. Thirdly, for small enterprises seamless market access to all Member States is more important than for large companies, which are better able to handle different regulatory requirements. SMEs also lack the scale to recoup the costs of adapting to another set of regulatory requirements by sufficiently large sales in other Member States. As a result, SMEs profit more from the avoidance of market fragmentation. Thus, the legislation would reduce the existing disadvantages of SMEs on the markets for high-risk AI applications.

Options 3 and 3+ also foresee to implement regulatory sandboxes allowing for the testing of innovative solutions under the oversight of the public authorities. These sandboxes would allow proportionate application of the rules to the SMEs as permitted in the existing legislation and thus allow a space for experimentation under the new rules and the existing legal framework. This will support the SMEs in reaching compliance in the pre-market phase that will ultimately facilitate their entry into the market. The regulatory oversight shall give guidance to providers how to minimize the associated risks and allow competent authorities to exercise their margin of discretion and flexibility as permitted by the applicable rules. Before an AI system can be placed on the market or put into service into a ‘live’ environment, the provider should ensure compliance with the applicable standards and rules for safety and fundamental rights and complete the applicable conformity assessment procedure. Direct guidance from the competent authorities will minimise the legal risk of non-compliance and thus reduce the compliance costs for SMEs participating in the sandboxing scheme, for example by reducing the need for legal or technical advice from third parties. Moreover, it will allow SMEs to bring their products and services to market faster.

SMEs, like any other AI provider, will also be able to rely on harmonized standards that will guide them in the implementation of the new requirements based on standardized good practices and procedures. This would alleviate the SMEs from the burden of developing these standards and good practices on their own and help them to build trust in their products and services, which is key not only for consumers, but also businesses customers across the value chain.

As a result, the foreseen regulatory requirements would not create a barrier to market entry for SMEs. One should also recall that notified bodies are bound to take the size of the company into account when setting their fees, so that SMEs will have lower costs than large companies for conformity assessment.

Option 4 would lead to SMEs being exposed to the regulatory costs when developing or using any AI application, no matter whether the application poses risks or not, or whether consumer trust is an important sales factor for this application. Despite the limited costs, it would thus expose SMEs as well as large companies to disproportionate expenditures. Regulatory sandboxes analogue to options 3 and 3+ could be foreseen, but in order to have a similar effect, there would have to be many more of them, since many more applications would be in the scope of the regulatory framework.

In addition, all options would envisage measures to support SMEs, including through the AI resources and services made available by the AI-on-demand platform 274 and through the provision of model compliance programmes and guidance and support through the Digital Innovation Hubs and the Testing and Experimentation Facilities. 275

The combined effect of the regulation on those SMEs providing AI will depend on how effective the support measures are in offsetting cost increases generated by the new legal requirements. The additional fixed compliance costs per AI system have been estimated at €6 000 - €7 000 for an average high-risk AI system of € 170 000, with another €3 000 to € 7 500 for conformity assessment (see section 6.1.3), while the monetary value of the support measures can not be determined with accuracy.

Access to free advice in the framework of regulatory sandboxes will be especially valuable to SMEs, at the beginning in particular, since they not only save on legal fees, but also receive guidance, reducing legal uncertainty to a minimum. Nevertheless, familiarisation with the requirements only accounts for a small part of the compliance costs. Access to the experimentation facilities of the Digital Innovation Hubs (DIH) and Testing and Experimentation Facilities (TEF) can be very valuable for SMEs thanks to their free services, although this may vary across sectors. For some sectors with little hardware requirements, cost savings will be smaller. For others, testing will require considerable physical infrastructure and free access to testing facilities is thus more beneficial. Contrary to large companies, SMEs cannot amortise costs for their own facilities over a large number of products. Note that cost savings provided by access to DIHs and TEFs may reduce both costs that are due to the regulation and costs which are not linked to the regulatory requirements. Finally, reduced costs for conformity assessment can partially compensate disadvantages SMEs may face due to the smaller scale of their operations. Moreover, by providing a focal point, regulatory sandboxes, DIHs and TEFs also facilitate partnering with complementary enterprises, knowledge acquisition and links to investors. For example, in a financial sector sandbox, a recent study found that entry into the sandbox was followed by an average increase in capital raised of 15% over the following two years. 276

As a result, with the support measures the cost of regulatory requirements to SMEs are smaller than without such measures, but the costs are not completely offset. Whether the additional costs can at the margin discourage some SMEs from entering into certain markets for high-risk AI applications will depend on the competitive environment for the specific application and its technical specificities.

6.1.5.Competitiveness and innovation

To date the European AI market currently accounts for roughly a fifth of the total world market and is growing fast. It is thus highly attractive not just for European firms but also for competitors from third countries.

277 Table 10: AI investment estimates (€ million) from 2020 to 2025

AI INVESTMENTS

2020

2021

2022

2023

2024

2025

Global (Grand View)

48 804

70 231

101 064

145 433

209 283

301 163

EU (Grand View)

10 737

15 451

22 234

31 995

46 042

66 256

Global (Allied Market)

15 788

24 566

38 224

59 476

92 545

144 000

EU (Allied Market)

3 473

5 404

8 409

13 085

20 360

31 680

Source: Contractor’ interpolation based on Allied Market Research, Grand View Research, and Tractica

The international competition is particularly tough because companies develop or adapt AI application in-house only to a smaller extent, and to a larger extent purchase them from external providers, either as a ready-to-use system or via hired external contractors. Thus, AI providers companies from outside the EU find it relatively easy to win market share, and as a result supply chains are often international.

Figure 11: Most common AI sourcing strategies for enterprises

Notabene: Applies to enterprises using at least one or two technologies.

Source: European enterprise survey on the use of technologies based on artificial intelligence,

European Commission 2020 (Company survey across 30 European countries, N= 9640)

Against this background, the impact of the options on competitiveness and innovation is crucial. In principle, the impact of a regulatory framework on innovation, competitiveness and investment depends on two contradicting factors. On the one hand, the additional compliance costs and administrative burdens (see section 6.1.3.) make AI projects more expensive and hence less attractive for companies and investors. From an economic point of view, whether the obligations are imposed on the user or on the developer is irrelevant, since any costs the developer has to bear will eventually be passed on to the user. On the other hand, the positive impact on uptake (see section 6.1.2.) is likely to increase demand even faster, and hence make projects more attractive for companies and investors. The overall impact will depend on the balance of these two factors.

Under Option 1, companies will only undergo the additional costs if they consider that the increased uptake of their products and services will outweigh the additional costs. It will thus not negatively affect innovation and thus the competitiveness of European providers of AI applications.

Under Option 2, only a small number of specific applications would have to undergo the additional costs. A positive effect on uptake is possible, but less likely for revisions of existing legislation than for ad-hoc legislation addressing a specific issue, since there would be no publicity effect. Innovation would become more expensive only for the specific applications regulated. Where regulation already exists, e.g. for many products, the impact will be lower, since companies are already equipped to deal with requirements.

However, under Option 3 increased costs and increased uptake through higher trust would be limited to a small subset of particularly high-risk applications. It is possible that AI providers would therefore focus investment on applications that do not fall in the scope of the regulatory framework, since the additional costs of the requirements would make innovations in non-covered AI applications relatively more attractive. Option 3+ would have similar effects, insofar as applications outside the scope would not be obliged to undergo the additional costs. Option 4 would see no such shift of supply but would see a much larger overall increase in cost, thus dampening innovation across all AI applications.

For options 3, 3+ and 4, there is no reason why investment into the development of ‘high risk’ use cases of AI would move to third countries in order to serve the European market, because for AI suppliers the requirement are identical on all markets. On the EU market foreign competitors would have to fulfil the same requirements; on third country markets, EU companies would not be obliged to fulfil the criteria (if they sell only to these markets). However, there is a theoretical risk that certain high-risk applications could not be sold profitably on the EU market or that the additional costs for users would make them unprofitable in use. For example, at the margin a recruitment software could be profitable for the provider if sold without respecting the requirements, but not if the provider would have to prevent it from discriminating against women. In those cases the choice implicit in the regulation would be that the respect of the fundamental right in question (in this case: non-discrimination) prevails over the loss of economic activity. Nevertheless, given the size of the EU market, which in itself accounts for 20% of the world market, it is very unlikely that the limited additional costs of algorithmic transparency and accountability would really prevent the introduction of this technology to the European market.

In addition, for Options 2, 3, 3+ and 4 there will be in addition the positive effect of legal certainty. By fulfilling the requirements, both AI providers and users can be certain that their application is lawful and do not have to worry about possible negative consequences. Hence, they will be more likely to invest in AI and to innovate using AI, thus becoming more competitive.

6.2. Costs for public authorities

278 Under options 1, 3, 3+ and 4, in-house conformity assessment as well as third-party conformity assessment would be funded by the companies (through fees for the third party mechanism). However, Member States would have to designate a supervisory authority in charge of implementing the legislative requirements and/or the voluntary labelling scheme, including market monitoring. Their supervisory function could build on existing arrangements, for example regarding conformity assessment bodies or market monitoring, but would require sufficient technological expertise. Depending on the pre-existing structure in each Member States, this could amount to 1 to 25 Full Time Equivalent (FTE) per Member State. The resource requirement would be fairly similar, whether ex-ante enforcement takes place or not. If it does, there is more work to supervise the notified bodies and/or the ex-ante conformity assessment through internal checks of the companies. If it doesn’t, there will be more incidents to deal with.

Options 1, 3, 3+ and 4 would benefit from a European coordination body to exchange best practices and to pool resources. Such a coordination body would mainly work via regular meetings of national competent authorities, assisted by secretarial support at EU level. This could amount to 10 FTE at EU level. As an additional option, the board could be supported by external expertise, e.g. in the form of a group of experts; the expertise would have to paid when needed and the cost would depend on the extent to which such expertise would be required. In addition, the EU would have to fund the database of high-risk AI applications with impacts mainly for fundamental rights. The additional costs at EU level should be more than offset by the reduction in expertise needed at national level, especially regarding the selection of applications for regulation and the gathering of a solid evidence base to support such a selection, which would be carried out at European level. Indeed, one of the key reasons why a European coordination mechanism is needed is that the pooling of resources is more efficient than a purely national build-up of expertise. Which costs option 2 would cause to public authorities would depend on the specific legislations. It is thus impossible to estimate at this stage. For the ad-hoc modifications of existing legislation with existing enforcement and supervisory structures, the costs to public authorities would be incremental, and European coordination could rely on existing structures as well. For ad-hoc legislation on new issues, e.g. remote biometric identification in publicly accessible spaces, Member States would have to build up new enforcement and supervisory structures with a more significant cost, including the building up of European coordination structures where they do not exist yet.

6.3. Social impact

279 280 281 All options, by increasing trust and hence uptake of AI applications, will lead to additional labour market impacts of AI. Generally, increasing uptake of AI applications is considered to cause a loss of some jobs, to create some others, and to transform many more, with the net balance uncertain. The increase in the use of AI and its facilitation also has important implications for skills, both in terms of requiring high-level AI skills and expertise and in ensuring that people can effectively use and interact with AI systems across the breadth of applications. The High-Level Group High-Level Expert Group on the Impact of the Digital Transformation on EU Labour Markets and the Special Report requested by former Commission president Juncker have recently analysed these effects in depth or the Commission. Increasing uptake of AI will therefore reinforce these effects, and the stronger one option increases uptake, the stronger this effect will be.

By setting requirements for training data in high-risk applications, options 3 and 3+ would contribute to reducing involuntary discrimination by AI systems, for example used in recruiting and career management, thus improving the situation of disadvantaged groups and leading to greater social cohesion. Option 4 would have the same impact on a larger set of applications; however, since the additional applications are not high risk, the marginal impact of reducing discrimination is less significant. Option 2 would only have this effect where the classes of applications that was subject to ad-hoc regulation was prone to unfair discrimination. Similarly, option 1 would only have this effect for the applications obtaining the label and only in so far as these applications were high risk and prone to unfair discrimination.

Given the effect of AI applications to enable efficiencies, expand and improve service delivery across sectors, advancing the uptake of AI will also accelerate the development of socially beneficial applications, such as in relation to education, culture or youth. For example, by enabling new forms of personalised education, AI could improve education overall, and in particular for individuals that do not learn well under a one-size-fits-all approach. Similarly, by enabling new forms of collaboration, new insights and new tools, it allows young people to engage in creative activities. It could also be used to improve accessibility and provide support to persons with disabilities for example through innovative assistive technologies.

The potential for health improvement by AI applications in terms of better prevention, better diagnosis and better treatment, is widely recognised. Here, option 3 would address the most pertinent applications. However, since trust is so important in this sector, it would be very beneficial to give other AI applications as well the opportunity to prove their trustworthiness, even if they are not strictly high-risk. Option 3+ would therefore be highly relevant. The benefits of option 1 would be limited in this field of applications, since voluntary commitments do not yield the same level of confidence. Option 2 would well address the issue of AI applications for health, since the health sector already has a well-developed regulatory system.

6.4. Impacts on safety

All options aim to fill gaps in relation to the specific safety and security risks posed by AI-embedded in products in order to minimize the risks of death, injury and material damages.

While option 2 primarily concerns amendments to existing legislation for AI embedded in products but no new regulations for AI in services or as a stand-alone application, options 3, 3+ and 4 extend the scope of the horizontal framework to AI used in services and decision-making processes (for example software used for automatically managing critical utilities with severe safety consequences).

Compared to option 2, benefits of options 3 and 4 are generated by several factors. First of all, the risks to safety from the introduction of AI applications would decrease since a larger scope of AI systems posing risks to safety would be subject to AI-specific requirements. In particular, these requirements would concern AI components that are integrated into both products and services. With regard to the AI safety components of products already covered by option 2, option 3, 3+ and 4 would have greater benefits in terms of legal certainty, consistency and harmonised implementation of requirements aimed at tackling risks which are inherent to AI systems. This is because options 3, 3+ and 4 will avoid sectoral approach to tackling AI risks and regulate them in a harmonised and consistent manner. A horizontal instrument under options 3, 3+ and 4 would also provide harmonized requirements for managing the evolving nature of risks which will help to ensure that products are continuously safe during their lifecycle. This would be of particular value to AI providers and users who often operate in several sectors.

Moreover, under option 3, 3+ and 4, the process of development and adoption of harmonised standards on AI systems would be significantly streamlined, with the production of a consistent and comprehensive set of horizontal and vertical standards in the field. This would very much support providers of AI and manufacturers of AI-driven products in demonstrating their compliance with relevant rules. In addition, the integration of the requirements for AI embedded in products into conformity assessment procedures foreseen under sectoral legislation minimises the burden on sector-specific providers and, more generally, sector-specific operators.

While option 3 will impose new safety requirements only for high-risk AI systems, the positive safety-related benefits for society under option 4 are expected to be higher since all AI systems will have to comply with the new requirements for safety, security, accuracy and robustness and be accordingly tested and validated before being placed on the market. Option 3+ is fundamentally the same as option 3 in terms of binding legal requirements, while it introduces a system of codes of conduct for companies supplying or using low-risk AI. This voluntary system could be however a tool to push market operators to engage in ensuring a higher safety baseline for their products even if they are low-risk.

6.5. Impacts on fundamental rights

Strengthening the respect of EU fundamental rights and effective enforcement of the existing legislation is one of the main objectives of the initiative.

All options will have some positive effects on the fundamental rights protection, although their extent will largely depend on the intensity of the regulatory intervention. While option 1 voluntary labelling may marginally facilitate compliance with fundamental rights legislation by setting common requirements for trustworthy AI, these positive effects will be only for providers of AI systems who voluntarily decide to subscribe to the scheme. By contrast, binding requirements under options 2 to 4 will significantly strengthen the respect of fundamental rights for the AI systems covered under the different options.

A sectoral ‘ad-hoc’ approach under option 2 will provide legal certainty and fill certain gaps in or complement the existing non-discrimination, data protection and consumer protection legal frameworks, thus addressing risks to specific rights, covered by these frameworks. However, option 2 might lead to delays, inconsistencies and will be limited to the scope of application of each sectoral legislation.

A horizontal framework under options 3 to 4 will ensure consistency and address cross-cutting issues of key importance for the effective protection of the fundamental rights. Such a horizontal instrument will establish common requirements for trustworthy AI applicable across all sectors and will prohibit certain AI practices considered as contravening the EU values. Options 3 to 4 will also impose specific requirements relating to the quality of data, documentation and traceability, provision of information and transparency, human oversight, robustness and accuracy of the AI systems which are expected to mitigate the risks to fundamental rights and significantly improve the effective enforcement of all existing legislation. Users will also be better informed about the risks, capabilities and limitations of the AI systems, which will place them in a better position to take the necessary preventive and mitigating measures to reduce the residual risks.

An ex ante mechanism for compliance with these requirements and obligations will ensure that providers of AI systems take measures to minimize the risks to the fundamental rights by design since otherwise they will not be allowed to place their AI systems on the Union market. Conformity assessment through independent third party notified bodies would be more effective than ex ante conformity assessment through internal checks as an enforcement mechanism in this respect to ensure the effective protection of the fundamental rights. In particular, documentation and transparency requirements will be important to ensure that fundamental rights can be duly enforced before judicial or other administrative authorities. In addition, the ex post market surveillance and supervision by competent authorities should ensure that any violation of fundamental rights can be investigated and sanctioned in a proportionate, effective and dissuasive manner. Authorities will also have stronger powers for inspection and joint investigations. The obligations placed on providers to report to the competent national authorities serious breaches of obligation under Union and Member State law intended to protect fundamental rights will further improve the detection and sanctioning of these infringements.

The positive effect on the fundamental rights will be different depending on whether option 3, 3+ or 4 is chosen. While option 4 envisages horizontal regulatory intervention for all AI systems irrespective of the risk, option 3 targets only systems posing ‘high risks’ that require regulatory action because of their expected severity and high risks for the fundamental rights and safety. Given the larger scope of applications to all AI systems, option 4 might therefore lead to better protection of all fundamental rights examined in the problem definition section. However, the regulatory burden placed on so many economic operators and users and the impact on their freedom to conduct a business can actually prevent the development of many low-risk AI applications that can benefit fundamental rights (for instance AI used for bias detection, detection of security threats etc.).

Option 3+, which combines option 3 with codes of conduct for non-high risk, might be thus most suitable to achieve an optimal level of protection of all fundamental rights. This is expected to enhance the trust in the AI technology and stimulate its uptake, which can be very beneficial for the promotion of a whole range of political, social and economic rights, while minimizing the risks and addressing the problems identified in section 2.

In addition to these overall positive benefits expected for all fundamental rights, certain fundamental rights are likely to be specifically affected by the intervention. These are analysed in Annex 5.5.

6.6. Environmental impacts

The environmental impact depends on how effective the regulatory framework is in increasing trust and hence uptake, balanced against the resources needed for compliance and against the positive effects from increased uptake.

The environmental impact of option 1 would depend on how widespread the adoption of the label would be. If the adoption were to be sufficiently large to create a public perception that AI development has become more trustworthy than previously, it would increase uptake and hence energy and resource consumption, to be balanced by efficiency gains obtained through AI applications.

The environmental impact of option 2 would vary with the specific problem addressed. However, it would not create widespread trust in AI as such, but only in the class of applications regulated. Thus, it would reduce energy and resource consumption by limiting certain applications, but increase adoption of this particular class of applications. Given the horizontal usability of AI, the impact of regulating a single class of applications would be negligible from a society-wide point of view.

In options 3 and 3+, the direct environmental impacts that can be expected from the proposed measures are very limited. On the one hand, this options prevents the development of applications on the black list, and it limits the deployment of remote biometric identification systems in publicly accessible spaces. All of this will reduce energy and resource consumption and correspondingly CO2 output.

On the other hand, the requirements do impose some additional activities with regard to testing and record-keeping. However, while machine learning is energy-intensive and rapidly becoming more so, the vast majority of the energy consumption occurs during the training phase. A significant increase in energy consumption would only take place if retraining were to be necessary on a large scale. However, whilst this may occur initially, developers will quickly learn how to make sure that their systems avoid retraining, given the enormous and rapidly increasing costs associated.

The indirect environmental impacts are more significant. On one hand, by increasing trust the measures will increase uptake and hence development and thus use of resources. It should be pointed out that this effect will not be limited to high-risk applications only – through cross-fertilization between different AI applications and re-use of building blocks, the increase in trust will also foster development in lower or no risk applications. On the other hand, many of the AI applications will be beneficial to the environment because of their superior efficiency compared to traditional (digital or analogue) technology. AI systems used in process optimisation by definition make processes more efficient and hence less wasteful, e.g. reducing the amounts of fertilizers and pesticides needed, decreasing the water consumption at equal output, etc.). AI systems supporting improved vehicle automation and traffic management contribute to the shift towards cooperative, connected and automated mobility, which in turn can support more efficient and multi-modal transport, lowering energy use and related emissions.

In addition, it is also possible to purposefully direct AI applications to improve the environment. For example, they can help pollution control and modelling the impact of climate change mitigation or adaptation measures. Finally, AI applications will minimise resource usage and energy consumption if policies encourage them to do so. Technical solutions include more efficient cooling systems, heat reuse, the use of renewable energy to supply data centres, and the construction of these data centres in regions with a cold climate. In the context of the Coordinated Plan on Artificial Intelligence with Member States, the Commission will consider options to encourage and promote AI solutions that have a neutral/positive impact on climate change and environment. This will also reduce potential environmental impacts of the present initiative.

In option 4 the direct impacts would be very similar to those in option 3. The only difference is that more testing would take place, and hence consume more energy. The indirect impacts would be identical, except that the increase in uptake could be higher if some applications which require trust but are not considered ‘high-risk’ are more readily accepted by citizens.

7.How do the options compare?

7.1. Criteria for comparison

The following criteria are used in assessing how the options would potentially perform, compared to the baseline:

·Effectiveness in achieving the specific objectives:

-ensure that AI systems placed on the market and used are safe and respect fundamental rights and Union values;

-ensure legal certainty to facilitate investment and innovation;

-enhance governance and effective enforcement of fundamental rights and safety requirements applicable to AI;

-facilitate the development of a single market for lawful, safe and trustworthy AI applications and prevent market fragmentation.

·Efficiency: cost-benefit ratio of each policy options in achieving the specific objectives;

·Coherence with other policy objectives and initiatives;

·Proportionality: whether the options go beyond what is a necessary intervention at EU level in achieving the objectives.



Table 11: Summary of the comparison of options against the four criteria

Notabene: table annotations should only be read in vertical; in the table, for options 3, 3+ and 4 it is assumed that ex-ante third party conformity assessments are mandatory for AI systems that are safety components of products and for remote biometric identification in publicly accessible spaces; “0” means same as baseline, “+” means partially better than baseline, “++” means better than baseline, “+++” means much better than baseline

7.2. Achievement of specific objectives

7.2.1. First specific objective: Ensure that AI systems placed on the market and used are safe and respect the existing law on fundamental rights and Union values

Option 1 would limit the risks for individuals regarding applications that have obtained the label, since companies would face sanctions if they claimed the label but did not actually respect the associated obligations. There would be a shift of demand to applications with the label, depending on how much attention consumers paid to this label. There is a chance that the label would become so widespread that it could set a standard that all market participants are forced to meet, but this is by no means certain. As a result, there is no guarantee that all or even most of high-risk applications would apply for the label and individuals would remain exposed to the risks identified earlier. Hence, option 1 would not be more effective than the baseline in achieving this objective.

Option 2 would effectively limit the risks for individuals, but only for those cases where action has been taken, assuming that the ad-hoc legislations will appropriately define the obligations for AI applications. Indeed, since the obligations can be precisely tailored to each use case, it will probably limit risks for the cases that are covered better than a horizontal framework. However, this effectiveness will only apply to the issues addressed in the separate legislations, leaving individuals unprotected against potential risks by other AI applications. Such an ad-hoc approach will also not be able to distribute obligations across the full AI value chain and will be limited to the material and personal scope of application of each sectorial legislation, which is likely to be more often for safety reasons than for fundamental rights. Option 2 is hence very effective for a number of cases but not comprehensive, and overall thus only be partially more effective than the baseline in achieving this objective.

Option 3 would effectively limit the risks to individuals for all applications that have been selected because the combination of the likelihood of violations and impact of such violations means that they constitute a high risk. By setting a comprehensive set of requirements and effective ex ante conformity assessment procedures, it makes violations for these applications much less likely before they are placed on the market. In addition, all providers of high-risk AI systems will have to establish and implement robust quality and risk management systems as well as post-market monitoring strategy that will provide efficient post-market supervision by providers and quick remedial action for any emerging risks. An effective ex-post market surveillance control will be also carried out by national competent authorities having adequate financial and technical resources. Moreover, additional AI applications could be added as the need arises. Hence, option 3 is more effective than the baseline in achieving this objective.

Option 3+ would have the same legal effectiveness as option 3, but in addition allow companies that produce or use applications that have not been selected as high risk to nevertheless fulfil the obligations. Since risks to individuals in reality are not binary – either low or high – but follow a continuous graduation from zero to extremely high, providing such an incentive especially to applications which are at the edge of high risk but are not covered by the legal requirements could significantly further reduce the overall risk of violation. Thus, option 3+ would be more effective than the baseline in achieving this objective.

Option 4 would very effectively limit the risks by setting the same requirements as option 3, but for all AI applications. It would thus cover the high-risk applications of option 3, the applications at the edge of high risk that make the codes of conduct of option 3+ worthwhile, and all other applications as well, including many applications where there are no or only very low risks. Individuals would be comprehensively protected, and as a result, option 4 would be much more effective than the baseline effective in achieving this objective.

7.2.2.Second specific objective: Ensure legal certainty to facilitate investment and innovation in AI

Option 1 could not foster investment and innovation by providing legal certainty to AI developers and users. While the existence of the voluntary label and the compliance with the associated requirements could function as an indication that the company intends to follow recommended practices, from a legal point of view there would be only a small change compared to the baseline. Uncertainty regarding the application of EU fundamental rights and safety rules specifically to AI would remain, and the ensuing risk would continue to discourage investment. Thus, option 1 would not be more effective than the baseline in achieving objective 2.

Option 2 would improve investment and innovation conditions by providing legal certainty only for applications that have been regulated. Thus, option 2 would only be partially more effective than the baseline in achieving objective 2.

Option 3 would improve conditions for investment and innovation by providing legal certainty to AI developers and users. They would know exactly which AI applications across all Member States are considered to constitute a high risk, which requirements would apply to these applications and which procedures they have to undertake in order to prove their compliance with the legislation, in particular where ex-ante conformity assessments (third-party or through internal checks) are part of the enforcement system. Option 3 would thus be more effective than the baseline in achieving objective 2.

Given the rapid technological evolution, legal certainty would nevertheless not be absolute, since regulatory changes cannot be excluded over time, but only be minimised as far as possible. When proposing changes, European policy-makers would be supported in their analysis by a group of experts and by national administrations, which can draw on evidence from their respective monitoring systems.

Option 3+ would provide the same legal certainty as option 3. The additional code of conduct scheme would, as in option 1, function as an indication that the company is willing to take appropriate measures, but would not assure legal certainty to those participating. However, since unlike option 1 the applications covered by the codes of conduct would be medium to low risk applications, the need for legal certainty is arguably smaller than for those applications which are covered by the high-risk requirements. Option 3+ would thus be more effective than the baseline in achieving objective 2.

Option 4 would provide the same legal certainty as option 3, but for all AI applications. However, this increased legal certainty would come at the price of increased legal complexity for applications where there is no reason for such complications, since they do not constitute a high risk. It would thus simply be more effective than the baseline in achieving objective 2.

7.2.3.Third specific objective: Enhance governance and effective enforcement of the existing law on fundamental rights and safety requirements applicable to AI systems

Option 1 would moderately improve enforcement for those applications that have obtained the label. There would be specific monitoring by the issuer of the label, which could include audits; it is even possible that the label would require ex-ante verification. However, the limited coverage would preclude these improvements from being an overall enhancement of enforcement. Since the label would coexist with a series of national legislative frameworks, governance would be more complicated than in the baseline scenario. Hence, option 1 would not be more effective than the baseline in achieving objective 3.

Option 2 would presumably improve effective enforcement and governance for regulated applications, according to the specifications laid down in the relevant sectorial legislation. However, since these may very well differ from one area to the next, overall enforcement and governance of requirements related to AI applications may become more complicated, especially for applications that could fall into several regulated categories simultaneously. As a result, option 2 would only partially be more effective than the baseline in achieving objective 3.

Options 3, 3+ and 4 would all improve enforcement. For all three options, there would be the requirements to carry out ex-ante verification, either in the form of third party ex-ante conformity assessment (the integration of AI concerns into existing third party conformity assessments, and remote biometric identification in publicly accessible spaces) or in form of ex ante assessment through internal checks (mainly services). Compared to the baseline, this is a clear improvement in enforcement. Ex-post enforcement would also be considerable strengthened because of the documentation and testing requirements that will allow assessing the legal compliance of the use of an AI system. Moreover, in all of these options competent national authorities from different sectors would benefit from enhanced competences, funding and expertise and would be able to cooperate in joint investigations at national and cross border level. In addition, a European coordination mechanism is foreseen to ensure a coherent and efficient implementation throughout the single market. Of course, option 3 and the mandatory part of option 3+ would cover only high-risk of AI applications and would thus be more effective than the baseline in achieving objective 3, while Option 4 would cover all AI applications and would thus be much more effective than the baseline in achieving objective 3.

7.2.4.Fourth specific objective: Facilitate the development of a single market for lawful, safe and trustworthy AI applications and prevent market fragmentation

Option 1 would provide a certain improvement to the baseline, by establishing a common set of requirements across the single market and allowing companies to signal their adherence, thus allowing users to choose these applications. Consumers and businesses could therefore reasonably be confident that they purchase a lawful, trustworthy and safe product if it has obtained the label, no matter from which member state it originates. The single market would be facilitated only for those applications that have obtained the label. For all other applications, the baseline would continue to apply. There is also the real possibility that individual Member States esteem that the voluntary label does not sufficiently achieve objective 1 and therefore take legislative action, leading to fragmentation of the single market. Consequently, option 1 would only be partially more effective than the baseline in achieving objective 4.

Option 2 would provide a clear improvement to the baseline, which would however be limited to those products and services for which ad-hoc legislation (including amendments to existing legislations) is introduced. For those products, consumers and businesses could be certain that the products and services they use are lawful, safe and trustworthy, and no Member States would be likely to legislate with respect to those products. This effect would come into being for each class of applications only after the ad-hoc legislation has been adopted. However, for products not covered by ad-hoc legislation, there would be no positive effect on consumer trust, and there is a real possibility of fragmentation of the single market along national borders, even assuming that the highest risk applications are those for which ad-hoc legislation would be agreed. Therefore, option 2 would only be partially more effective than the baseline in achieving objective 4.

282 Option 3 would provide a clear improvement to the baseline. On the one hand, for those cases that are covered, consumers and businesses can rely on the European framework to guarantee that the AI applications are lawful, trustworthy and safe coming from any Member States. On the other side, they can consider that those applications not covered by the legislation do not, in principle, constitute a high risk. Moreover, Member States are likely to refrain from legislation that would fragment the single market for low risk products, since they have agreed on a list of high-risk applications and since there is a mechanism to amend this list. As a result, option 3 would be more effective than the baseline in achieving objective 4.

Option 3+ would also provide a clear improvement to the baseline. It would have all the same effects of option 3, and in addition afford businesses the opportunity to signal their adherence to lawful, trustworthy and safe AI for those applications not considered high risk. .While it is uncertain how many non-high-risk applications would be developed in accordance with codes of conduct, the total increase in trust by business and consumers is at the minimum equivalent to option 3 and can be legitimately expected to be significantly higher. Option 3+ would thus be much more effective than the baseline in achieving objective 4.

Option 4 would create a comprehensive increase in trust by businesses and consumers, since they will know for all applications that providers had to fulfil the legal obligations. Moreover, since all risks will be covered, there is no risk of additional national legislation that could lead to fragmentation. However, one must also concede that the increase in costs for all AI applications (see discussion on proportionality below), including when there is no countervailing benefit because they do not extensively rely on user trust (e.g. industrial applications) can have the effect of fewer AI applications being offered, thus leading to a smaller market than otherwise. Option 4 thus only effectively achieves objective 4.

7.3. Efficiency

The costs of option 1 for AI providers and users would be similar for each AI application to the costs of option 3, if the requirements are identical, and if the enforcement mechanism is similar. On an aggregate level, the costs could be higher or lower than option 3, depending on how many companies introduce a code of conduct. However, the costs will be targeted in a less precise way because some costs AI applications that do not really need additional trust will incur them, and some applications that should undergo the requirements according to the risk-based approach will not do so. On the other hand, participation is voluntary and therefore left to the business decisions of companies. Hence, one can argue that it has no net cost – if the benefits did not outweigh the costs, companies would not participate. In that sense, option 1 would be cost effective. However, public administrations would still have to bear the costs to supervise the system, which could in principle cover all AI applications. Nevertheless, there would be no costs to policy-makers to determine high-risk applications, since any application can apply for the voluntary label.

Option 2 has overall low aggregate costs for AI providers and users, since it will only address specific problems and may often be implemented during regular revisions of existing regulations. However, the costs for each application can be significant, and the multiplicity of specific regulations may make compliance with them unnecessarily complicated. Nevertheless, it can be assumed that significant costs would only be occurred if the benefits were worth the effort. Public administrations would only incur costs in specific areas, where – in case of amending existing regulations - competent authorities would already be established. The costs of determining high risk applications would correspond to the choice of applications to be regulated. Overall, it can be assumed that option 2 is cost effective.

The costs of option 3 mainly consist in the burden on AI providers and users, which is in turn composed of the compliance costs and verification costs. While the costs for covered systems are moderate, the overall aggregate cost remains low due to the precise targeting of a small number of high-risk applications only. A limitation of third-party conformity assessments to AI systems that are safety components of products and remote biometric identification in publicly accessible spaces further limits the expenditure to the most relevant cases only. Moreover, the requirements are unified across applications, allowing for inexpensive and reusable compliance procedures. These costs are compensated by a strong positive impact on those applications where it is most needed. There would also be costs for public administrations that have to ensure enforcement, but they too would be limited, since the monitoring would only cover the applications classified as “high risk”. For policy-makers there would be the additional costs of determining, based on solid evidence, what applications should be classified as high risk, which would however be small compared to overall compliance costs. The existence of an evidence base from the monitoring systems established by national competent authorities would help in minimising the risk that AI producers exploit their information advantages to misrepresent risks without prohibitive costs. Option 3 would thus be cost effective.

Regarding option 3+, for the mandatory part, the precise targeting ensures cost effectiveness. For the codes of conduct, the voluntary character ensures cost effectiveness. Overall, Option 3+ can be considered cost effective.

Option 4 has by far the highest aggregate costs for AI providers and users, since the costs per applications are the same, but the number of applications is far greater. These vastly increased costs are compensated only to little extent by an increased trust, since most of the additionally covered application do not rely on trust. Moreover, public administrations would have to monitor and enforce the system for all AI application, which would be significantly more resource-intensive than option 3. Thus, despite the fact that there would be no costs to policy-makers to determine high-risk applications, since all applications are covered, option 4 would not be cost effective.

7.4. Coherence

All options are fully coherent with the existing legislation on safety and fundamental rights. Options 1, 3, 3+ and 4 would promote or impose obligations to facilitate the implementation of existing legislation, and to address issues that existing legislation does not cover. Options 3, 3+ and 4 would make use of existing conformity assessment procedures wherever available. Option 2 would specifically cover applications where problems have arisen or are likely to arise that are not addressed by existing legislation.

283 All options are consistent with the separate initiative on liability, which, among others, aims to address the problems outlined in the Report on the safety and liability implications of Artificial Intelligence, the Internet of Things and robotics. All options are equally coherent with the digital single market policy, by attempting to prevent the rising of barriers to cross-border commerce through the emergence of national and incompatible regulatory frameworks attempting to address the challenges raised by AI.

Options 3, 3+ and 4 are equally fully coherent with the overall strategy set out in Shaping Europe's digital future, which especially articulates a vision of “a European society powered by digital solutions that are strongly rooted in our common values”, and with the European data strategy, which argues that the “vision stems from European values and fundamental rights and the conviction that the human being is and should remain at the centre.” Building on these visions, both strategies attempt to accelerate the digital transformation of the European economy. Promoting legal certainty for the use of AI and ensuring it is trustworthy clearly contributes to this endeavour.

Option 1 has the same objective as the other initiatives but falls short in implementing these visions, since its non-binding character cannot guarantee the widespread respect of European values when it comes to AI applications. It is thus only partially coherent with European policy. Option 2 can only implement the respect of European values with regard to a subset of AI applications. It is thus equally only partially coherent with European policy

7.5Proportionality

Options 1, 2, 3 and 3+ impose procedures that are proportional to the objectives pursued. Option 1 creates burdens only for companies who have voluntarily decided to so. Option 2 would only impose burdens when a concrete problem has arisen or can be foreseen, and only for the purpose of addressing this problem.

Option 3 only imposes burdens on a small number of specifically selected high-risk applications and only sets requirements that are the minimum necessary to mitigate the risks, safeguard the single market, provide legal certainty and improve governance. Only very limited transparency obligations are imposed where needed to inform affected parties that an AI system is used and provide them with the necessary information to enable them to exercise their right to an effective remedy. For high-risk systems, the requirements relating to data, documentation and traceability, provision of information and transparency, human oversight, accuracy and robustness, are strictly necessary to mitigate the risks to fundamental rights and safety posed by AI and uncovered by other frameworks. A limitation of third-party conformity assessments to AI systems that are safety components of products and remote biometric identification in publicly accessible spaces also contributes to this precise targeting. Harmonized standards and supporting guidance and compliance tools will aim to help providers and users to comply with the requirements and minimize their costs. The costs incurred by operators are proportionate to the objectives achieved and the economic benefits that operators can expect from this initiative.

Option 3+ would have the same precise targeting plus allowing companies to follow voluntarily certain requirements for non-high-risk applications. Option 4, on the other hand, imposes burdens across all AI applications, whether justified by the risks each application poses or not. The aggregate economic cost for AI providers and AI users is therefore much higher, with no or only small additional benefits. It is thus disproportionate.

8.Preferred option

As a result from the comparison of the options, the preferred option is option 3+, a regulatory framework for high-risk AI applications with the possibility for all non-high-risk AI applications to follow a code of conduct. This option would: 1) provide a legal definition of AI, 2) establish a definition of a high-risk AI system, and 3) set up the system of minimum requirements that high-risk AI systems must meet in order to be placed on or used on the EU market. The requirements would concern data, documentation and traceability, provision of information and transparency, human oversight and robustness and accuracy and would be mandatory for high-risk AI applications. Companies who introduce codes of conduct for other non-high-risk AI systems would do so voluntarily and these systems would be in principle shielded from unilateral Member States regulations.

Compliance would be verified through ex-ante conformity assessments and ex-post supervision and market surveillance. Ex-ante conformity assessments would be applicable to providers of all high-risk AI systems. Every high-risk AI system will be certified for a specific intended purpose(s) so that its performance can be verified in concreto. If the purpose or the system’s functionality are substantially changed by the user or a third party, they will have the same obligations as the provider in case the changed system qualifies as high-risk.

284 285 Regarding high-risk AI systems which are safety components of products, the regulatory framework will integrate the enforcement of the new requirements into the existing sectoral safety legislation so as to minimise additional burdens. This integration will take place following an appropriate transitional period before the new AI requirements become binding for operators under the sectoral legislation. The mechanism of integration and extent of legal applicability of the horizontal instrument will depend on the nature and structure of the sectoral instruments in question. In particular:

·Regarding high-risk AI systems covered by NLF legislation, 286 existing NLF conformity assessment systems would be applicable for checking the compliance of the AI system with the new requirements. The application of the horizontal framework would not affect the logic, methodology or general structure of conformity assessment under the relevant NLF product safety legislation (see Annex 5.3. - e.g. under the new Medical Device Regulation, the requirements of the horizontal AI framework would be applicable within the frame of the overall risk-benefit consideration which is at the heart of the assessment under that legislation). Obligations of economic operators and ex-post enforcement provisions (as described later in this text) of the horizontal framework will also apply to the extent they are not already covered under the sectoral product safety law.

·Regarding high-risk AI systems covered by relevant Old Approach legislation 287 (e.g. aviation, cars), applicability of the horizontal framework will be limited to the ex-ante essential requirements (e.g. human oversight, transparency) for high-risk AI systems, which will have to be taken into account when amending those acts or when adopting relevant implementing or delegated legislation under those acts. 288  

289 For other high-risk AI systems, the conformity assessment could be done by the provider of the system based on ex ante assessment through internal checks. However, biometric remote identification in publicly accessible spaces would have to undergo an ex-ante third party conformity assessment, because of the particularly high-risks to breaches of fundamental rights.

290 In addition to ex-ante conformity assessments, there would also be an ex-post system for market surveillance and supervision by national competent authorities designated by the Member States. In order to facilitate cross-border cooperation, a European coordination mechanism would be established which would function primarily via regular meetings between competent national authorities with some secretarial support at EU level. The EU body would be supported by an expert group to monitor technological developments and risks and provide evidence-based advice on the need for revision and updating of the high-risk use cases in public consultation of relevant stakeholders and concerned parties. This “Board on AI” will work in close cooperation with the European Data Protection Board, the EU networks on market surveillance authorities and any other relevant structures at EU level.

This option would best meet the objectives of the intervention. By requiring a restricted yet effective set of actions from AI developers and users, it would limit the risks of violation of fundamental rights and safety of EU citizens, but would do so in targeting the requirements only to applications where there is a high risk that such violations would happen. As a result, it would keep compliance costs to a minimum, thus avoiding an unnecessary slowing of uptake due to higher prices. In order to address possible disadvantages for SMEs, it would among others provide for regulatory sandboxes and access to testing facilities. Due to the establishment of the requirements and the corresponding enforcement mechanisms, citizens could develop trust in AI, companies would gain in legal certainty, and Member States would see no reason to take unilateral action that could fragment the single market. As a result of higher demand due to better trust, higher offers due to legal certainty, and the absence of obstacles to cross-border movement of AI systems, the single market for AI would be likely to flourish. The European Union would continue to develop a fast-growing AI ecosystem of innovative services and products embedding AI technology or stand-along AI applications, resulting in increased digital autonomy. As indicated in the introduction, the AI horizontal framework outlined in this preferred option will be accompanied by review of certain sectoral product safety legislation and new rules on AI liability.

With regard to review of safety legislation, as indicated in Section 1.3.2, review of some NLF sector-specific-legislation is ongoing in order to address challenges linked to new technologies. While relevant NLF product legislation would not cover aspects that are under the scope of the horizontal legislative instrument on AI for high-risk applications, the manufacturer would still have to demonstrate that the incorporation of a high-risk AI system covered by those NLF legislations into the product ensures the safety of the product as a whole in accordance with that NLF product legislation. In this respect, for example, the reviewed Machinery Directive 2006/42/EC could contain some requirements with regard to the safe integration of AI systems into the product (which are not under the scope of the horizontal framework). In order to increase legal clarity, any relevant NLF product legislation which is reviewed (e.g. Machinery Directive 2006/42/EC) would cross reference the AI horizontal framework, as appropriate. On the other hand, the General Product Safety Directive (GPSD) is also being reviewed to tackle emerging risks arising from new technologies. In line with its nature (see Section 1.3.2), the reviewed GPSD will be applicable, insofar as there are not more specific provisions in harmonised sector-specific safety legislation (including the future AI horizontal framework). Therefore, we can conclude that all revisions of safety legislation will complement and not overlap the future AI horizontal framework.

Concerning liability, a longstanding EU approach with regard to product legislation is based on adequate combination of both safety and liability rules. This includes EU harmonised safety rules ensuring a high level of protection and the removal of barriers within the EU single market, and effective liability rules to provide for compensation where accidents nonetheless happen. For this reason, the Commission considers that only a combination of the AI horizontal framework with future liability rules can fully address the problems listed in this impact assessment specifically in terms of specific objectives 2 and 4 (legal certainty and single market for trustworthy AI). In fact, while the AI initiative shaped in this preferred option is an ex ante risk minimisation instrument to avoid and minimise the risk of harm caused by AI, the new rules on liability would be an ex post compensation instrument when such harm has occurred. Effective liability rules will also provide an additional incentive to comply with the due diligence obligations laid down in the AI horizontal initiative, thus reinforcing the effectiveness and intended benefits of the proposed initiative.

291 292 293 In terms of timing for the adoption, the Commission has decided at political level that in order to provide clarity, consistency and certainty for businesses and citizens the forthcoming initiatives related to AI, as proposed in the White Paper on AI, will be adopted in stages. First, the Commission will propose the AI horizontal legal framework (Q2 2021) which will set the definition for artificial intelligence, a solid risk methodology to define high-risk AI, certain requirements for AI systems and certain obligations for the key operators across the value chain (providers and users). Second, the liability framework (expected Q4 2021/Q1 2022) will be proposed, possibly comprising a review of the Product Liability Directive and harmonising targeted elements of civil liability currently under national law. The future changes to the liability rules will take into account the elements of the horizontal framework with a view to designing the most effective and proportionate solutions with regard to liability for damages/harm caused by AI systems as well as ensuring effective compensation of victims. The AI horizontal framework and the liability framework will complement one another: while the requirements of the horizontal framework mainly aim to protect against risks to fundamental rights and safety from an ex-ante perspective, effective liability rules primarily take care of damage caused by AI from an ex-post angle, ensuring compensation should the risks materialise. Moreover, compliance with the requirements of the AI horizontal framework will be taken into account for assessing liability of actors under future liability rules. 

Table 12: Forthcoming EU AI initiatives

AI INITIATIVE

MAIN ELEMENTS (SCOPE) WITH REGARD TO AI SYSTEMS

Horizontal legislation on AI

(current proposal)

-Sets a definition for “artificial intelligence”

-Sets risk assessment methodology and defines high-risk AI systems

-Sets certain minimum requirements for high risk AI systems (e.g. minimum transparency of algorithm, documentation, data quality)

-Sets legal obligations with regard to the conduct of key economic operators (providers and users)

-Sets a governance system at national and EU level for the effective enforcement of these rules

294 New and adapted liability rules (under reflection - expected Q4 2021-Q1 2022)

-Makes targeted adaptations to liability rules, to ensure that victims can claim compensation for damage caused by AI systems

-May introduce possible adaptations to the existing EU product liability rules (based on strict liability), including notions of product, producer, defect as well as the defences and claim thresholds.

-May propose possible harmonisation of certain elements of national liability systems (strict and fault-based)

-May provide possible specific considerations for certain sectors (e.g. healthcare)

-All possible changes will take into account foundational concepts (e.g. the definition of AI) and legal obligations with regard to the conduct of key economic operators set by the AI horizontal framework.

Sectoral safety legislation revisions

-The revisions will complement, but not overlap with the horizontal AI framework

-May set certain requirements to ensure that integration of the AI systems into the product is safe and the overall product performance is not compromised

9.How will actual impacts be monitored and evaluated?

Providing for a robust monitoring and evaluation mechanism is crucial to evaluate how far the regulatory framework succeeds in achieving its objectives. One could consider it a success if AI systems based on the proposed regulatory approach would be appreciated by consumers and businesses, with the European Union and Member States developing together a new culture of algorithmic transparency and accountability without stifling innovation. As a result, AI made in the EU incorporating a trust-based approach would become the world reference standard. AI made in Europe would be characterised by the absence of violations of fundamental rights and incidents physically harming humans due to AI.

The Commission will be in charge of monitoring the effects of the preferred policy option. For the purpose of monitoring, it will establish a system for registering stand-alone AI applications with implications mainly for fundamental rights in a public EU-wide database. This would also enable competent authorities, users and other interested people to verify if the high-risk AI system complies with the new requirements and exercise enhanced oversight over these AI applications posing increased rights to fundamental rights (Annex 5.4.). To feed this database, AI suppliers will be obliged to provide meaningful information about the system and the conformity assessment carried out.

Moreover, AI providers will be obliged to inform national competent authorities about serious incidents or AI performances which constitute a breach of fundamental rights obligations as soon as they become aware of them, as well as any recalls or withdrawals of AI systems from the market. National competent authorities will then investigate the incidents/breaches, collect all the necessary information and regularly transmit it with adequate metadata to the EU board on AI, broken down by fields of applications (e.g. recruitment, biometric recognition etc.) and calculated a) in absolute terms, b) as share of applications deployed and c) as share of citizens concerned.

. The Commission will complement this information on applied high-risk AI use cases by a comprehensive analysis of the overall market for artificial intelligence. To do so, it will measure AI uptake in regular surveys (a baseline survey has been carried out by the Commission in Spring 2020), and use data from national competent authorities, Eurostat, the Joint Research Center (through AI Watch) and the OECD. It will pay particular attention to the international compatibility of the data collections, so that data becomes comparable between Member States and other advanced economies. The joint OECD/EU AI observatory is a first step on the way to achieve this.

The following list of indicators is provisional and non-exhaustive.

Table 13: Indicators for monitoring and evaluation

OBJECTIVE

INDICATOR

SOURCE

AI systems are safe and respect EU fundamental rights and values

(negative indicators)

Number of serious incidents or AI performances which constitute a serious incident or a breach of fundamental rights obligations (semi-annual) by fields of applications and calculated a) in absolute terms, b) as share of applications deployed and c) as share of citizens concerned

National competent authorities; European Data Protection Board

Facilitate investment and innovation

(positive indicators)

Total AI investment in the EU (annual)

Total AI investment by Member State (annual)

Share of companies using AI (annual)

Share of SMEs using AI (annual)

Projects approved through regulatory sandboxes and placed on the market (annual)

Number of SMEs consulting on AI in Digital Innovations Hubs and Testing and Experimentation Facilities

Commission services and AI Watch; National competent authorities

Improve governance and enforcement mechanisms

(negative indicators)

Number of recalls or withdrawals of AI systems from the market (semi-annual); by fields of applications and calculated a) in absolute terms, b) as share of applications deployed and c) as share of citizens concerned

National competent authorities

Facilitate single market

Level of trust in artificial intelligence (annual) (positive indicator)

Number of national legislations that would fragment the single market (biannual) (negative indicator)

Commission services and AI Watch

Note: For a positive indicators, a higher value represents a better outcome. For a negative indicator a lower value represents a better outcome.

Taking into account these indicators and complementing with additional ad-hoc sources as well as qualitative evidence, the Commission will publish a report evaluating and reviewing the framework five years following the date on which it becomes applicable.



Glossary 295

Acquis

The EU's 'acquis' is the body of common rights and obligations that are binding on all EU countries, as EU Members. Source: EUR-Lex glossary

AI

Artificial Intelligence

Artificial intelligence (AI) system

An AI system is a machine-based system that can, for a given set of human-defined objectives, generate output such as content, predictions, recommendations, or decisions influencing real or virtual environments. AI systems are designed to operate with varying levels of autonomy. Source: based on OECD AI principles

Algorithm

Finite suite of formal rules (logical operations, instructions) allowing to obtain a result from input elements. This suite can be the object of an automated execution process and rely on models designed through machine learning. Source: Council of Europe AI Glossary

ALTAI

Assessment List for Trustworthy Artificial Intelligence, developed by the EU’s High-Level Expert Group on Artificial Intelligence.

Autonomous systems

ICT-based systems which have a high degree of automation and can for instance perceive their environment, translate this perception into meaningful actions and then execute these actions without human supervision.

Algorithmic bias

AI bias or (or algorithmic) bias describes systematic and repeatable errors in a computer system that create unfair outcomes, such as favouring one arbitrary group of users over others. Source: ALTAI glossary

Black-box

In the context of AI and machine learning-based systems, the black box refers to cases where it is not possible to trace back the reason for certain decisions due to the complexity of machine learning techniques and their opacity in terms of unravelling the processes through which such decisions have been reached. Source: European Commission Expert group on Ethics of connected and automated vehicles study  

Chatbot

Conversational agent that dialogues with its user (for example: empathic robots available to patients, or automated conversation services in customer relations). Source: Council of Europe Glossary

CJEU

Court of Justice of the European Union

CT

Computer Tomography

Data sovereignty

Concept that data is protected under law and the jurisdiction of the state of its origin, to guarantee data protection rights and obligations.

Data value chain

Underlying concept to describe the idea that data assets can be produced by private actors or by public authorities and exchanged on efficient markets like commodities and industrial parts (or made available for reuse as public goods) throughout the lifecycle of datasets (capture, curation, storage, search, sharing, transfer, analysis and visualization). These data are then aggregated as inputs for the production of value-added goods and services which may in turn be used as inputs in the production of other goods and services.

Deep Learning

A subset of machine learning that relies on neural networks with many layers of neurons. In so doing, deep learning employs statistics to spot underlying trends or data patterns and applies that knowledge to other layers of analysis. Source: The Brookings glossary of AI and emerging technologies

Deepfakes:

Digital images and audio that are artificially altered or manipulated by AI and/or deep learning to make someone do or say something he or she did not actually do or say. Pictures or videos can be edited to put someone in a compromising position or to have someone make a controversial statement, even though the person did not actually do or say what is shown. Increasingly, it is becoming difficult to distinguish artificially manufactured material from actual videos and images. Source: The Brookings glossary of AI and emerging technologies

DIH

Digital Innovation Hub

Distributed computing

A model where hardware and software systems contain multiple processing and/or storage elements that are connected over a network and integrated in some fashion. The purpose is to connect users, applications and resources in a transparent, open and scalable way, and provide more computing and storage capacity to users. In general terms, distributed computing refers to computing systems to provide computational operations that contribute to solving an overall computational problem.

Embedded system

Computer system with a dedicated function within a larger system, often with real-time computing constraints comprising software and hardware. It is embedded as part of a complete device often including other physical parts (e.g. electrical, mechanical, optical). Embedded systems control many devices in common use today such as airplanes, cars, elevators, medical equipment and similar.

Facial Recognition

A technology for identifying specific people based on pictures or videos. It operates by analysing features such as the structure of the face, the distance between the eyes, and the angles between a person’s eyes, nose, and mouth. It is controversial because of worries about privacy invasion, malicious applications, or abuse by government or corporate entities. In addition, there have been well-documented biases by race and gender with some facial recognition algorithms. Source: The Brookings glossary of AI and emerging technologies

GDPR

General Data Protection Regulation 2016/679

Harmonised Standard

A European standard elaborated on the basis of a request from the European Commission to a recognised European Standards Organisation to develop a standard that provides solutions for compliance with a legal provision. Compliance with harmonised standards provides a presumption of conformity with the corresponding requirements of harmonisation legislation. The use of standards remains voluntary. Within the context of some directives or regulations voluntary European standards supporting implementation of relevant legal requirements are not called ‘harmonised standards’.

HLEG

High-Level Expert Group on Artificial Intelligence.

IoT

(Internet of Things)

Dynamic global network infrastructure with self-configuring capabilities based on standard and interoperable communication protocols where physical and virtual "things" have identities, physical attributes and virtual personalities and use intelligent interfaces and are seamlessly integrated into the information network.

ISO

International Organization for Standardization

Machine learning

Machine learning makes it possible to construct a mathematical model from data, including a large number of variables that are not known in advance. The parameters are configured as you go through a learning phase, which uses training data sets to find links and classifies them. The different machine learning methods are chosen by the designers according to the nature of the tasks to be performed (grouping, decision tree). These methods are usually classified into 3 categories: human-supervised learning, unsupervised learning, and unsupervised learning by reinforcement. These 3 categories group together different methods including neural networks, deep learning etc. Source: Council of Europe Glossary

Natural language processing

Information processing based upon natural-language understanding. Source: ISO

Neural Network

Algorithmic system, whose design was originally schematically inspired by the functioning of biological neurons and which, subsequently, came close to statistical methods. The so-called formal neuron is designed as an automaton with a transfer function that transforms its inputs into outputs according to precise logical, arithmetic and symbolic rules. Assembled in a network, these formal neurons are able to quickly operate classifications and gradually learn to improve them. This type of learning has been tested by tests on games (Go, video games). It is used for robotics, automated translation, etc. Source: Council of Europe Glossary

NLF

New legislative framework

To improve the internal market for goods and strengthen the conditions for placing a wide range of products on the EU market, the new legislative framework was adopted in 2008. It is a package of measures that streamline the obligations of manufacturers, authorised representatives, importers and distributors, improve market surveillance and boost the quality of conformity assessments. It also regulates the use of CE marking and creates a toolbox of measures for use in product legislation. Source: European Commission, Internal Market, Industry, Entrepreneurship and SMEs

OECD

The Organisation for Economic Co-operation and Development.

PLD

Product Liability Directive, i.e. Council Directive 85/374/EEC of 25 July 1985 on the approximation of the laws, regulations and administrative provisions of the Member States concerning liability for defective products, OJ L 210, 7.8.1985, p. 29–33, ELI: http://data.europa.eu/eli/dir/1985/374/1999-06-04.

Self-learning AI system

Self-learning (or self-supervised learning) AI systems recognize patterns in the training data in an autonomous way, without the need for supervision. Source: ALTAI glossary

SME

Small- and Medium-sized Enterprise

An enterprise that satisfies the criteria laid down in Commission Recommendation 2003/361/EC of 6 May 2003 concerning the definition of micro, small and medium-sized enterprises (OJ L 124, 20.05.2003, p. 36): employs fewer than 250 persons, has an annual turnover not exceeding €50 million, and/or an annual balance sheet total not exceeding €43 million.

Supervised Learning

According to Science magazine, supervised learning is ‘a type of machine learning in which the algorithm compares its outputs with the correct outputs during training. Supervised learning allows machine learning and AI to improve information processing and become more accurate’. Source: The Brookings glossary of AI and emerging technologies

Training data

Samples for training used to fit a machine learning model. Source: ISO

Trustworthy

Trustworthy AI has three components: 1) it should be lawful, ensuring compliance with all applicable laws and regulations; 2) it should be ethical, demonstrating respect for, and ensure adherence to, ethical principles and values; and 3) it should be robust, both from a technical and social perspective, since, even with good intentions, AI systems can cause unintentional harm. Trustworthy AI concerns not only the trustworthiness of the AI system itself but also comprises the trustworthiness of all processes and actors that are part of the AI system’s life cycle. Source: ALTAI glossary

Use case

Use case: A use case is a specific situation in which a product or service could potentially be used. For example, self-driving cars or care robots are use cases for AI. Source: ALTAI glossary

(1)

     European Commission, White Paper on Artificial Intelligence - A European approach to excellence and trust , COM(2020) 65 final, 2020.

(2)

     See above, European Commission, White Paper on Artificial Intelligence - A European approach to excellence and trust , COM(2020) 65 final, 2020. p. 25.

(3)

     European Commission, Building Trust in Human-Centric Artificial Intelligence , COM(2019) 168.

(4)

     HLEG, Ethics Guidelines for Trustworthy AI , 2019.

(5)

     HLEG, Assessment List for Trustworthy Artificial Intelligence (ALTAI) for self-assessment , 2020.

(6)

     European Commission, Inception Impact Assessment For a Proposal for a legal act of the European Parliament and the Council laying down requirements for Artificial Intelligence .  

(7)

     For details of all the consultations that have been carried out see Annex 2.

(8)

     For the discussion on why AI can be considered as an emerging general purpose technology see for instance Agrawal, A., J. Gans and A. Goldfarb, Economic policy for artificial intelligence, NBER Working Paper No. 24690, 2018; Brynjolfsson, E., D. Rock and C. Syverson, Artificial intelligence and the modern productivity paradox: A clash of expectations and statistics, NBER Working Paper No. 24001, 2017.

(9)

     For the analysis of the available definitions and they scope see e.g. JRC, Defining Artificial Intelligence, Towards an operational definition and taxonomy of artificial intelligence , 2020. As well as the forthcoming update to this JRC Technical Report. The forthcoming update provides a qualitative analysis of 37 AI policy and institutional reports, 23 relevant research publications and 3 market reports, from the beginning of AI in 1955 until today.

(10)

     OECD, Recommendation of the Council on Artificial Intelligence , 2019 .

(11)

     OECD, Artificial Intelligence in Society , 2019, p. 23.

(12)

See above, OECD, Artificial Intelligence in Society , 2019.

(13)

     See above, OECD, Artificial Intelligence in Society , 2019.

(14)

     Rule-based algorithms are well-suited to execute applications and tasks that require high reliability and robustness. They can be used for complex simulations and can be adapted by adding new information to the system that is then processed with the help of the established rules.

(15)

     Rule-based algorithms are well-suited to execute applications and tasks that require high reliability and robustness.

(16)

One of the best known subfields of AI technology where algorithms ‘learn’ from data is ‘machine learning’ (ML) that predicts certain features, also called ‘outputs’, based on a so called ‘input’. ‘Learning’ takes place when the ML algorithm progressively improves its performance on the given task.

(17)

     For now, the majority of AI systems are rule-based.

(18)

     AI is a technology, thus it cannot be directly compared or equated with human intelligence. However, to explain how AI systems achieve ‘artificial intelligence’ the parallel to humans is telling. The AI ‘brain’ is increasingly acquiring the potential to carry our complex tasks which require a ‘body’ (sensors, actuators) and a nervous system (embedded AI). This combination of ‘brain’ and ‘body’ connected through ‘a nervous system’ allows AI systems to perform tasks such as exploring space, or the bottom of the oceans. For a graphical overview, see Figure 1.

(19)

     Vinuesa, R. et al., ‘The role of artificial intelligence in achieving the Sustainable Development Goals’, Nature communications 11(1), 2020, pp. 1-10.

(20)

     OECD,  Using artificial intelligence to help combat COVID-19 , 2020.

(21)

     According to McKinsey, the cumulative additional GDP contribution of new digital technologies could amount to €2.2 trillion in the EU by 2030, a 14.1% increase from 2017, McKinsey, S haping the Digital Transformation in Europe , 2020). PwC comes to an almost identical forecast increase at global level, amounting to USD 15.7 trillion, PwC, Sizing the prize: What’s the real value of AI for your business and how can you capitalise? , 2017.

(22)

     For a comparison, the introduction of steam engines in the 1800s boosted labour productivity by 0.3% a year and spread of IT during the 2000s by 0.6% a year (ITU/McKinsey, Assessing the Economic Impact of Artificial Intelligence, 2018).

(23)

     IDC, IDC Forecasts Strong 12.3% Growth for AI Market in 2020 Amidst Challenging Circumstances , 2020. 

(24)

     OECD, Artificial Intelligence in Society , 2019.

(25)

     European Commission, Ipsos Report, European enterprise survey on the use of technologies based on artificial intelligence , 2020.

(26)

     McKinsey, How nine digital frontrunner can lead on AI in Europe , 2020.

(27)

     European Commission, Ipsos Survey, European enterprise survey on the use of technologies based on artificial intelligence , 2020. (Company survey across 30 European countries, N= 9640).

(28)

     For detailed review of various human rights related risks see e.g. Horizon 2020 funded SIENNA project , Rodrigues, R, Siemaszko, K and Warso. Z, D4.2: Analysis of the legal and human rights requirements for AI and robotics in and outside the EU (Version V2.0). Zenodo, 2019. The researchers in this project identified the following main concerns related to fundamental rights and AI systems: lack of algorithmic transparency / transparency in automated decision-making; unfairness, bias, discrimination and lack of contestability; intellectual property issues; issues related to AI vulnerabilities in cybersecurity; issues related to impacts on the workplace and workers; privacy and data protection issues; and liability issues related to damage caused by AI systems and applications. See also, JRC Report, Artificial Intelligence: A European Perspective , 2018.

(29)

     For the discussion see Problem 2 below.

(30)

     Ursula von der Leyen, Political Guidelines for the Next European Commission 2019-2024 , 2019, p. 13.

(31)

     For a comprehensive overview of applicable EU primary and secondary legislation see SIENNA project, ibid.

(32)

      Regulation (EU) 2016/679 of the European Parliament and of the Council of 27 April 2016 on the protection of natural persons with regard to the processing of personal data and on the free movement of such data, and repealing Directive 95/46/EC (General Data Protection Regulation).

(33)

      Directive (EU) 2016/680 of the European Parliament and of the Council of 27 April 2016 on the protection of natural persons with regard to the processing of personal data by competent authorities for the purposes of the prevention, investigation, detection or prosecution of criminal offences or the execution of criminal penalties, and on the free movement of such data, and repealing Council Framework Decision 2008/977/JHA.

(34)

     This is a rather broad term, encompassing in principle any AI and automated decision-making systems.

(35)

     For an overview on how GDPR applies to AI, see e.g. Spanish Data Protection Agency, RGPD compliance of processings that embed Artificial Intelligence: An introduction , 2020. See also European Data Protection Supervisor Opinion on the European Commission’s White Paper on Artificial Intelligence – A European approach to excellence and trust , 2020.

(36)

     E.g. Council Directive 2000/43/EC of 29 June 2000 implementing the principle of equal treatment between persons irrespective of racial or ethnic origin;  Council Directive 2000/78/EC of 27 November 2000 establishing a general framework for equal treatment in employment and occupation; Directive 2006/54/EC of the European Parliament and of the Council of 5 July 2006 on the implementation of the principle of equal opportunities and equal treatment of men and women in matters of employment and occupation (recast); Council Directive 2004/113/EC of 13 December 2004 implementing the principle of equal treatment between men and women in the access to and supply of goods and services.

(37)

      Directive (EU) 2019/882 of the European Parliament and of the Council of 17 April 2019 on the accessibility requirements for products and services, OJ L 151, 7.6.2019, pp. 70–115.

(38)

      Directive 2005/29/EC of the European Parliament and of the Council of 11 May 2005 concerning unfair business-to-consumer commercial practices in the internal market and amending Council Directive 84/450/EEC, Directives 97/7/EC, 98/27/EC and 2002/65/EC of the European Parliament and of the Council and Regulation (EC) No 2006/2004 of the European Parliament and of the Council (‘Unfair Commercial Practices Directive’).

(39)

     In the context of EU sector specific safety legislation, so-called old and new approaches are traditionally distinguished. The ‘Old Approach’ refers to the very initial phase of EU regulation on products, whose main feature was the inclusion of detailed technical requirements in the body of the legislation. Certain sectors such as food or transport are still being regulated on the basis of ‘old approach’ legislations with detailed product requirements for reasons of public policy or because of their reliance on international traditions and/or agreements which cannot be changed unilaterally. The so-called ‘New Approach’ was developed in 1985, whose main objective was to restrict the content of legislation to ‘essential (high-level) requirements’ leaving the technical details to European harmonised standards. On the basis of the New Approach, the New Legislative Framework (NLF) was then developed in 2008, introducing harmonised elements for conformity assessment, accreditation of conformity assessment bodies and market surveillance. Today more than 20 sectors are regulated at EU level based on the NLF approach, e.g. medical devices, toys, radio-equipment or electrical appliances.

(40)

      Directive 2001/95/EC of the European Parliament and of the Council of 3 December 2001 on general product safety.

(41)

      Directive 2006/42/EC of the European Parliament and of the Council of 17 May 2006 on machinery, and amending Directive 95/16/EC (recast).

(42)

     New rules on automated vehicles, cybersecurity and software updates of vehicles will become applicable as part of the vehicle type approval and market surveillance legislation as from 7 July 2022, providing notably for obligations for the manufacturer to perform an exhaustive risk assessment (including risks linked to the use of AI) and to put in place appropriate risk mitigations, as well as to implement a comprehensive risk management system during the lifecycle of the product.

(43)

     Regulation (EU) 2019/2144 of the European Parliament and of the Council of 27 November 2019 on type-approval requirements for motor vehicles and their trailers, and systems, components and separate technical units intended for such vehicles, as regards their general safety and the protection of vehicle occupants and vulnerable road users, amending Regulation (EU) 2018/858 of the European Parliament and of the Council (OJ L 325, 16.12.2019), p. 1.

(44)

     The Commission intends to adopt proposals in the first quarter of 2021 for the revision of the Machinery Directive and second quarter of 2021 for the revision of the General Product Safety Directive.

(45)

      Directive 2014/53/EU of the European Parliament and of the Council of 16 April 2014 on the harmonisation of the laws of the Member States relating to the making available on the market of radio equipment and repealing Directive 1999/5/EC.

(46)

Safety legislation assesses a broad spectrum of risks and ensures that the overall interplay between different types and elements of risks does not render a product or service as a whole unsafe. These measures will also facilitate the uptake and increase certainty, by ensuring that the integration of new technologies in the product does not endanger the overall safety of a product or service. More detailed explanation about the interaction between the AI initiative examined in this impact assessment and the sectoral product legislation can be found in Annex 5.3.

(47)

      Council Directive 85/374/EEC of 25 July 1985 on the approximation of the laws, regulations and administrative provisions of the Member States concerning liability for defective products. See in particular Article 6(1), listing out the relevant circumstances under which a product is considered defective, i.e. when not providing the ‘safety which a person is entitled to expect’; see also whereas 6 and 8 of this Directive.

(48)

To obtain compensation, the injured party shall prove in court three elements: defect, damage, causal link between the two. The PLD is technology-neutral by nature.

(49)

It is unclear whether the PLD still provides the intended legal certainty and consumer protection when it comes to AI systems and the review of the directive will aim to address that problem. Software, artificial intelligence and other digital components play an increasingly important role in the safety and functioning of many products, but are not expressly covered by the PLD. The PLD also lacks clear rules in relation to changes, updates or refurbishments of products, plus it is not always clear who the producer is where products are adapted or combined with services. Finally, establishing proof of defect, harm and causation is in many cases excessively difficult for consumers, who are at a disadvantage in terms of technical information about the product, especially in relation to complex products such as AI. See Evaluation SWD(2018)157 final of Council Directive 85/374/EEC of 25 July 1985 on the approximation of the laws, regulations and administrative provisions of the Member States concerning liability for defective products, accompanying Report COM(2018) 246 from the Commission to the European Parliament, the Council and the European Economic and Social Committee on the application of the Directive; See also Report COM(202064 on the safety and liability implications of Artificial Intelligence, the Internet of Things and robotics.

(50)

     For the overview of national liability regimes applicable to the AI technologies, see e.g. European Parliamentary Research Service, Civil liability regime for artificial intelligence: European Added Value Assessment , 2020.

(51)

      See introduction above for references.

(52)

     For additional details, see the European Commission , 2020 and the Report on Liability for Artificial Intelligence and other emerging technologies .

(53)

See section 8 for a more detailed analysis.

(54)

The relevant recital provision to this extent would be included in the proposed horizontal framework initiative.

(55)

     European Commission, Digital Services Act – deepening the internal market and clarifying responsibilities for digital services , 2020.

(56)

      Commission Delegated Regulation (EU) 2017/589 of 19 July 2016 supplementing Directive 2014/65/EU of the European Parliament and of the Council with regard to regulatory technical standards specifying the organisational requirements of investment firms engaged in algorithmic trading.

(57)

European Banking Authority, Guidelines on loan origination and monitoring , 2020.

(58)

     See, also mapping of national approaches in relation to creditworthiness assessment under Directive 2008/48/EC of the European Parliament and of the Council of 23 April 2008 on credit agreements for consumers and repealing Council Directive 87/102/EEC. As set out in the recently adopted digital finance strategy, the Commission will invite the European Supervisory Authorities and the European Central Bank to develop guidance on the use of AI applications in 2021.

(59)

     European Commission, Artificial Intelligence for Europe , COM(2018) 327 final, 2018.

(60)

     European Commission, High-Level Expert Group on Artificial Intelligence , 2020.

(61)

     High-Level Expert Group on Artificial Intelligence, Ethics Guidelines for Trustworthy AI , 2019.

(62)

     High-Level Expert Group on Artificial Intelligence, Policy and investment recommendations for trustworthy AI , 2019.

(63)

     European Commission, Coordinated Plan on Artificial Intelligence , 2018.

(64)

     The Plan builds on a Declaration Cooperation on Artificial Intelligence , signed by EU Member States and Norway in April 2018. An AI Member States group has been regularly meeting since 2018, discussing among other things the ethical and regulatory aspects of AI. A review of the Coordinated plan is foreseen in 2021.

(65)

     European Commission, Communication from the Commission to the European Parliament, the Council, the European Economic and Social Committee and the Committee of the Regions, Building Trust in Human Centric Artificial Intelligence , COM(2019)168 final, 2019.

(66)

     High-Level Expert Group on Artificial Intelligence, Assessment List for Trustworthy Artificial Intelligence (ALTAI) for self-assessment , 2020.

(67)

     European Commission, White Paper on Artificial Intelligence - A European approach to excellence and trust , COM(2020) 65 final, 2020.

(68)

     European Commission, Staff Working Document on Liability for emerging digital technologies , SWD 2018/137 final.

(69)

     European Council, European Council meeting (19 October 2017) – Conclusion  EUCO 14/17, 2017, p. 8.

(70)

     Council of the European Union, Artificial intelligence b) Conclusions on the coordinated plan on artificial intelligence-Adoption  6177/19, 2019.

(71)

     European Council, Special meeting of the European Council (1and 2 October 2020) – Conclusions  EUCO 13/20, 2020.

(72)

     Council of the European Union, Presidency conclusions - The Charter of Fundamental Rights in the context of Artificial Intelligence and Digital Change , 11481/20, 2020.

(73)

     European Parliament resolution of 16 February 2017 on Civil Law Rules on Robotics, 2015/2103(INL) .

(74)

     European Parliament resolution of 12 February 2019 on a comprehensive European industrial policy on artificial intelligence and robotics, 2018/2088(INI) .

(75)

     European Parliament decision of 18 June 2020 on setting up a special committee on artificial intelligence in a digital age, and defining its responsibilities, numerical strength and term of office, 2020/2684(RSO) .

(76)

     European Parliament resolution of 20 October 2020 on a framework of ethical aspects of artificial intelligence, robotics and related technologies, 2020/2012(INL) .

(77)

     European Parliament resolution of 20 October 2020 on a civil liability regime for artificial intelligence, 2020/2014(INL).

(78)

     European Parliament resolution of 20 October 2020 on intellectual property rights for the development of artificial intelligence technologies, 2020/2015(INI).

(79)

     European Parliament Draft Report, Artificial intelligence in criminal law and its use by the police and judicial authorities in criminal matters,  2020/2016(INI) .

(80)

     European Parliament Draft Report, Artificial intelligence in education, culture and the audiovisual sector, 2020/2017(INI) .

(81)

     More details of the EP proposals are presented in section 5 when various policy options are discussed.

(82)

European Parliament resolution of 20 October 2020 on the Digital Services Act: Improving the functioning of the Single Market (2020/2018(INL)).

(83)

     See for an overview: Fundamental Rights Agency, AI Policy Initiatives 2016-2020 , 2020; or Council of Europe, Artificial Intelligence , 2020.

(84)

Consultative Committee of The Convention for the Protection of Individuals with regard to Automatic Processing of Personal Data Convention 108 Guidelines on Facial Recognition , 28 January 2021, T-PD(2020)03rev4.

(85)

     OECD, Recommendation of the Council on Artificial Intelligence ,  OECD/LEGAL/0449, 2019. 

(86)

     G20, G20 Ministerial Statement on Trade and Digital Economy , 2019.

(87)

     UNESCO, Artificial intelligence with human values for sustainable development , 2020.

(88)

     The EU is one of the founding members, alongside Australia, Canada, France, Germany, India, Italy, Japan, Mexico, New Zealand, the Republic of Korea, Singapore, Slovenia, the United Kingdom, the United States of America. The aim of this initiative is to bring together leading experts from industry, civil society, governments, and academia to bridge the gap between theory and practice on AI by supporting cutting-edge research and applied activities on AI-related priorities.

(89)

See the most recent Executive Order on Promoting the Use of Trustworthy Artificial Intelligence in the Federal Government from 3 December 2020, stipulates that when designing, developing, acquiring, and using AI in the Federal Government, agencies shall adhere to the following Principles: (a) Lawful and respectful of our Nation’s values; (b) Purposeful and performance-driven; (c) Accurate, reliable, and effective; (d) Safe, secure, and resilient; (e) Understandable; (f) Responsible and traceable; (g) Regularly monitored; (h) Transparent; (i) Accountable.

(90)

     Fjeld, J., N. Achten, H. Hilligoss, A. Nagy, and M. Srikumar, Principled Artificial Intelligence: Mapping Consensus in Ethical and Rights-Based Approaches to Principles for AI. Berkman Klein Center Research Publication No. 2020-1, 2020.

(91)

     The Common Foreign and Security Policy is regulated under Title V of the Treaty on European Union, which would be applicable for the use of AI for such exclusive military purposes.

(92)

     Modernized rules for the export control of such dual use products and technologies were agreed by the EU co-legislators in November based on the Commission’s Proposal for a Regulation setting up a Union regime for the control of exports, transfer, brokering, technical assistance and transit of dual-use items (recast) , COM(2016) 616 final. 2016/0295 (COD).

(93)

     The analysed evidence includes results of the public consultation on White Paper on AI, responses to the inception impact assessment, stakeholder consultations carried out within the framework of this impact assessment, European Parliament resolution of 20 October 2020 with recommendations to the Commission on a framework of ethical aspects of artificial intelligence, robotics and related technologies, European Council Presidency Conclusion of 21 October 2020, ongoing work of international organizations, as well as secondary literature.

(94)

For more details about the existing product safety legislation see section legal context 1.3.2.

(95)

     In this respect the Commission’s White Paper on AI was accompanied by Commission Report on safety and liability implications of AI, the Internet of Things and Robotics , 2020. Both the White Paper and the Report point to the combination of revision of existing EU safety legislation and the horizontal framework on AI to address those risks as explained in the section on the legal context 1.3.2. and 1.3.3.

(96)

     Those risks very much pertain to the quality and reliability of information which results from the output of a computing operation. Qualitatively different means that the nature (the cause/ driver) of safety risks generated by complex software or algorithms are different from risks caused by physical products.

(97)

     For explanation of AI Characteristics please see section 2.2. ‘Drivers’ below and Annex 5.2.: Five specific characteristics of AI.

(98)

This refers primarily to the ill-designed systems, see Russell, Stuart J., Peter Norvig. Artificial Intelligence: A Modern Approach. Pearson, 4th ed., 2020.

(99)

     The “AI Cybersecurity Challenges: Threat landscape for Artificial Intelligence” report published by ENISA with the support to the Ad-Hoc Working Group of Artificial Intelligence Cybersecurity presents a mapping of the AI cybersecurity ecosystem and its Threat Landscape, highlighting the importance of cybersecurity for secure and trustworthy AI.

(100)

     ENISA, JRC, Cybersecurity challenges in the uptake of Artificial Intelligence in Autonomous Driving, 2021.

(101)

     These vulnerabilities and risks in addition to their implications on safety, could also lead in certain scenarios to the situations that could have significant negative impacts on fundamental rights and Union values (see Problem 2), in particular when AI systems are used to make decisions based on personal data.

(102)

The use of ‘Evasion’, ‘data poisoning’ attacks or task misspecification may also have an objective to misdirect reinforcement learning behaviour.

(103)

     For the definition of a term, please see Annex 5.2.: Five specific characteristics of AI.

(104)

     A backdoor usually refers to any method by which authorized and unauthorized users are able to get around normal security measures and gain high level user access on a computer system, network, or software application.

(105)

     The ongoing review of certain sectorial legislations is considering these aspects (e.g. Machinery Directive and General Product Safety Directive).

(106)

      Regulation (EU) 2019/881 of the European Parliament and of the Council of 17 April 2019 on ENISA (the European Union Agency for Cybersecurity) and on information and communications technology cybersecurity certification and repealing Regulation (EU) No 526/2013 (Cybersecurity Act). The cybersecurity framework allows the development of dedicated certification schemes. Each scheme establishes and lists the relevant standards, however, any certification scheme established under this Regulation is of a voluntary nature.

(107)

     However, product specific legislation already exists in some sector: e.g. in the automotive sector new rules on automated vehicles, cybersecurity and software updates of vehicles will become applicable as part of the vehicle type approval and market surveillance legislation as from 7 July 2022, providing notably for obligations for the manufacturer to perform an exhaustive risk assessment (including risks linked to the use of AI) and to put in place appropriate risk mitigations, as well as to implement a comprehensive risk management system during the lifecycle of the product.    

(108)

     The current legal framework, does not provide conditions when self-learning AI should undergo a new conformity assessment.

(109)

     For example, developers, installation/operation/maintenance service providers at the point of use, actors responsible for the operation and maintenance of networks/platforms.

(110)

To this end, the fundamental rights of all persons concerned must be looked at and all remedies and safeguards applicable to an AI systems considered. The potential positive or adverse impact on the society as a whole and on general public interests such as public security, fight against crime, good administration, public health, protection of public finances should also be taken into account.

(111)

EU Fundamental Rights Agency, Getting the future right – Artificial intelligence and fundamental rights , 2020. Raso, F. et al., Artificial Intelligence & Human Rights: Opportunities & Risks , Berkman Klein Center for Internet & Society Research Publication, 2018.

(112)

     Reports and case studies published among others by research and civil society organisations such as AlgorithmWatch and Bertelsmann Stiftung, Automating Society – Taking Stock of Automated Decision-Making in the EU , 2019. AlgorithmWatch and Bertelsmann Stiftung, Report Automating Society , 2020. EDRi, Use cases: Impermissible AI and fundamental rights breaches , 2020.

(113)

     See e.g. Decision 216/2017 of the National Non-Discrimination and Equality Tribunal of Finland of 21 March 2017. The Joint Council for the Welfare of Immigrants, 2020.   The decision of the Hague District Court regarding the use of the SyRi scheme by the Dutch authorities etc.

(114)

 Broader considerations and analysis of the impact on all fundamental rights can be found in the study supporting this impact assessment as well as studies on AI and human rights, for example, commissioned by the EU Fundamental Rights Agency and the Council of Europe .

(115)

For example the VioGén protocol  in Spain which includes an algorithm that evaluates the risk that victims of domestic violence are going to be attacked again by their partners or ex-partners. See AlgorithmWatch and Bertelsmann Stiftung, Report Automating Society , 2020.

(116)

     See Council of Europe, Declaration by the Committee of Ministers on the manipulative capabilities of algorithmic processes, February 2019;

(117)

U.S. White House Office of Science and Technology Policy, Request for Information on the Future of Artificial Intelligence , September 1, 2016; Maurice E. Stucke Ariel Ezrachi, ‘ The Subtle Ways Your Digital Assistant Might Manipulate You ’, Wired, 2016; Judith Shulevitz, ‘Alexa, Should We Trust You?, The Atlantic, November 2018.

(118)

Anna-Lisa Vollmer, Children conform, adults resist: A robot group induced peer pressure on normative social conformity, Science Robotics, Vol. 3, Issue 21, 15 Aug 2018; Hasse, A., Cortesi, S. Lombana Bermudez, A. and Gasser, U. (2019). 'Youth and Artificial Intelligence: Where We Stand', Berkman Klein Center for Internet & Society at Harvard University; UNICEF, 'Safeguarding Girls and Boys: When Chatbots Answer Their Private Questions' , 6 August 2020.

(119)

     See Communication from the Commission, Data protection as a pillar of citizens’ empowerment and the EU’s approach to the digital transition - two years of application of the General Data Protection Regulation  COM(2020) 264 final, 2020.

(120)

  EDPS Opinion on the European Commission’s White Paper on Artificial Intelligence – A European approach to excellence and trust , 2020.

(121)

     AlgorithmWatch and Bertelsmann Stiftung, Report Automating Society , 2019 and 2020.

(122)

In their submissions to the public consultation on the White paper, some countries (e.g. France, Finland, the Check republic, Denmark) submit that the use of remote biometric identification systems in public spaces might be justified for important public security reasons under strict legal conditions and safeguards.

(123)

     For example, the Italian ministry of interior plans to employ the SARI facial recognition in Italy Cameras with facial recognition technology have also been used in a train station in Madrid or in the bus terminal. See also EU Fundamental Rights Agency, Facial recognition technology: fundamental rights considerations in the context of law enforcement , 2019.

(124)

     See e.g. EDPB, Facial recognition in school renders Sweden’s first GDPR fine, 2019. Politico, French privacy watchdog says facial recognition trial in high schools is illegal , 2019. In the UK, the Court of Appeal found that the facial recognition programme used by the South Wales police was unlawful and that ‘[i]t is not clear who can be placed on the watch list, nor is it clear that there are any criteria for determining where [the facial recognition technology] can be deployed’ (UK, Court of Appeal, R (Bridges) v. CC South Wales, EWCA Civ 1058, 11 August 2020).

(125)

     Germany put on hold its plans to use facial recognition at 134 railway stations and 14 airports, while France plans to establish a legal framework permitting video surveillance systems to be embedded with facial recognition (Stolton, S., ‘ After Clearview AI scandal, Commission ‘in close contact’ with EU data authorities ’, Euroactiv, 2020). In 2019, the Hellenic Police signed a €4 million contract with Intracom Telecom for a smart policing project (Homo Digitalis, The Greek DPA investigates the Greek Police , 2020). Italy also considers using facial recognition in all football stadiums (Chiusi, F., In Italy, an appetite for face recognition in football stadiums , 2020).

(126)

     This was researched at selected EU external borders (Greece, Hungary and Latvia) in the framework of the Integrated Portable Control System (iBorderCtrl) project, which integrates facial recognition and other technologies to detect if a person is saying the truth.

(127)

     Vincent, J., AI ‘emotion recognition’ can’t be trusted , The Verge, 2019.

(128)

     See Article 9(2) of the GDPR and Article 10 of the Law Enforcement Directive .

(129)

     Fundamental Rights Agency, #BigData: Discrimination in data-supported decision-making , 2018, p. 3. Despite the new risks posed by AI to the right to non-discrimination, the FRA report also highlights that human-decision-making is similarly prone to bias and if AI systems are properly designed and used, they offer opportunities to limit discriminatory treatment based on biased human decisions.

(130)

     Bakke, E., ‘Predictive policing: The argument for public transparency’, New York University Annual Survey of American Law, 2018, pp. 139-140.

(131)

     Postal codes were used, for example, in the Amsterdam risk assessment tool ProKid (now discontinued) to assess the risk of recidivism – future criminality – of children and young people, even if postal codes are often proxies for ethic origin as ruled by the CJEU, Case C-83/14.

(132)

     The Gender Shades project evaluates the accuracy of AI powered gender classification products.

(133)

     Research suggests that hiring platforms such as PeopleStrong or TribePad, HireVue, Pymetrics and Applied use such kind of algorithmic tools for supporting recruitment decisions, see Sánchez-Monedero et al., What does It mean to ‘solve’ the problem of discrimination in hiring? Social, technical and legal perspectives from the UK on automated hiring systems, 2020. See also VZBV, Artificial Intelligence: Trust Is Good, Control Is Better , 2018.

(134)

     Algorithms used to serve ads were found to generally prefer men over women for high-paying jobs. See Upturn, An Examination of Hiring Algorithms, Equity, Bias , 2018.

(135)

     For example, the Dutch SyRI system used to identify the risk of abusing the social welfare state, was recently found by the court to be intransparent and unduly interfering with the rights to private life of all affected persons. AI systems for social welfare also exist in Finland, Germany, Estonia and other countries, see AlgorithmWatch and Bertelsmann Stiftung, 2020.

(136)

     For example, in Germany the leading company for scoring individuals, SCHUFA run an AI system that was found by researchers and civil society to suffer from various anomalies in the data. In 2018, the Finnish National Non-Discrimination and Equality Tribunal prohibited a financial company, specialising in credits, from using certain statistical methods in credit scoring decisions. For more cases see also AlgorithmWatch and Bertelsmann Stiftung, 2019 and 2020.

(137)

E.g. the Dutch ProKid (now discontinued) and the e-Criminality Awareness System; Precobs, Krim and SKALA in Germany; KeyCrime and eSecurity in Italy, Pred-Crime in Spain. For more cases see also AlgorithmWatch and Bertelsmann Stiftung, 2019 and 2020.

(138)

For example, the UK Home Office stopped using an algorithm streaming visa applicants , because of allegations for unlawful discrimination against people of certain nationalities. Also in the UK, a speech recognition system used to detect fraud among those sitting English language exams in order to fulfil student visa requirements, reportedly resulted in the wrongful deportation of up to 7,000 people. The Algorithm Watch and Bertelsmann Stiftung , 2020 mentions several other AI tools and pilot projects in EU Member States where AI is used in the context of migration, border control and asylum procedures, e.g. p, 26, 85, 115 and 199.

(139)

     Wachter, S., B. Mittelstadt, and C. Russell, Why fairness cannot be automated: bridging the gap between EU non-discrimination law and AI , Oxford Internet Institute, University of Oxford, 2020.

(140)

     Fundamental Rights Agency, Artificial intelligence and fundamental rights , 2020, Ferguson A. G., Policing Predictive Policing, Washington University Law Review, 2017, pp. 1165-1167.

(141)

     Ibidem, see also Council of Europe, Algorithms and human rights , 2017, pp.11 and 24. See also a recent judgment from Italy (T.A.R., Rome, sect. III-bis, 22 mars 2017, n 3769) that ruled that the simple description of the algorithm, in terms of decision-making process steps, without the disclosure of the specific sequence of instructions contained in the algorithm, would not constitute an effective protection of the subjective right concerned.

(142)

EU Fundamental Rights Agency, Artificial intelligence and fundamental rights , 2020. See also the European Commission for the Efficiency of Justice, European ethical Charter on the use of Artificial Intelligence in judicial systems and their environment , 2018.

(143)

     See Erik van de Sandt et al. Towards Data Scientific Investigations: A Comprehensive Data Science Framework and Case Study for Investigating Organized Crime & Serving the Public Interest, November 2020.

(144)

     Meijer, A. and M. Wessels, ‘Predictive Policing: Review of Benefits and Drawbacks’, International Journal of Public Administration 42:12, 2019, p. 1032.

(145)

     The CJEU has ruled that the inclusion of a natural person in databases of potential suspects interferes with the presumption of innocence and can be proportionate ‘only if there are sufficient grounds to suspect the person concerned’ (CJEU, Peter Puskar , Case C‑73/16, para 114).

(146)

EU Fundamental Rights Agency, Artificial intelligence and fundamental rights , 2020.

(147)

     However, this is going to be covered in the new Machinery legal act being revised for AI systems/ components having safety functions.

(148)

     See Recital 78 of the GDPR which states that producers of systems are not directly bound by the data protection legislation.

(149)

     In sectors other than under the GDPR and the Law Enforcement Directive where data protection authorities from different Member States can cooperate.

(150)

     See, for example, Mahieu, R. and J. van Hoboken, Fashion-ID: Introducing a Phase-Oriented Approach to Data Protection? , European Law Blog, 2019. 

(151)

     For example, assurance and quality control, metrics and thresholds used, testing and validation procedures, good practice risk prevention and mitigation measures, data governance and quality management procedures, or disclosure obligations.

(152)

      Directive 2005/29/EC of the European Parliament and of the Council of 11 May 2005 concerning unfair business-to-consumer commercial practices in the internal market and amending Council Directive 84/450/EEC, Directives 97/7/EC, 98/27/EC and 2002/65/EC of the European Parliament and of the Council and Regulation.

(153)

European Commission, Ipsos Survey, 2020. Large companies were represented significantly less than SMEs (44% as opposed to just above 50%).

(154)

     Note, however, that 54% also thought that regulation of AI inhibits innovation (TÜV, Künstliche Intelligenz in Unternehmen , 2020).

(155)

     McKinsey Global Institute, Artificial Intelligence the next digital frontier? ,2017. Global survey to C-level executives, N=3.073

(156)

     BEUC, Artificial Intelligence: what consumers say , 2020.

(157)

     Tieto, People in the Nordics are worried about the development of AI – personal data processing a major concern , 2019. Number of respondents N=2648. 

(158)

McKinsey and DG CNECT, Shaping the digital transformation in Europe , 2020.

(159)

     European Commission, European enterprise survey on the use of technologies based on artificial intelligence , 2020.

(160)

     See above.

(161)

     McKinsey Global Institute, Notes from the AI frontier: tackling Europe’s gap in digital and AI , 2019.

(162)

     European Parliamentary Research Service, European added value assessment: European framework on ethical aspects of artificial intelligence, robotics and related technologies , 2020.

(163)

     See, for example, CJEU Judgement of 14 October 2004, Omega , Case C-36/02, ECLI:EU:C:2004:614, where the EU Court of Justice has stated that Member States can take unilateral measures to restrict the free movements of services and goods if necessary and proportionate to ensure respect of fundamental rights guaranteed by the legal order of the Member States.

(164)

  Non-paper - Innovative and trustworthy AI: two sides of the same coin , Position paper on behalf of Denmark, Belgium, the Czech Republic, Finland, France Estonia, Ireland, Latvia, Luxembourg, the Netherlands, Poland, Portugal, Spain and Sweden, 2020.

(165)

Council of the European Union, Shaping Europe's Digital Future-Council Conclusions , 8711/20, 2020.

(166)

     Datenethikkommission, Opinion of the German Data Ethics Commission , 2019.

(167)

     The AI Finland Project's ethics working group and the Ethics Challenge added emphasis on companies and self-regulation (AI Finland, ‘ Etiikkahaaste (Ethics Challenge) ’, Tekoäly on uusi sähkö. 2020).

(168)

     See, for example, ISO, ISO/IEC JTC 1/SC 42 Artificial intelligence , 2017; Oceanis, Repository , 2020.

(169)

     Christofi, A., et.al. ‘Erosion by Standardisation: Is ISO/IEC29134:2017 on Privacy Impact Assessment Up to GDPR Standard?’, in M. Tzanou (ed.), Personal Data Protection and Legal Developments in the European Union¸ IGI Global, 2020.

(170)

     StandICT, Standards Watch , 2020.  

(171)

See, for example, the Clearview AI scandal where AI technology that is based on scraping of billions of images online can enter the Union market and be used by businesses and law enforcement agencies (Pascu, L., ‘Hamburg data protection commissioner demands answers on biometric dataset from Clearview AI’, Biometric Update , 2020.

(172)

     Ursula von der Leyen, Shaping Europe's digital future: op-ed by Ursula von der Leyen President of the European Commission, 2020.

(173)

     European Council, Special meeting of the European Council (1 and 2 October 2020) – Conclusions , EUCO 13/20, 2020.

(174)

     See section 1. ‘Introduction’ of this impact assessment.

(175)

     For detailed analysis see Annex 5.2: Five specific characteristics of AI. Table 3 presents a simplified and non-technical explanation of the AI characteristics and their link to the problems. The main aim is to highlight main elements rather than provide a detail account of all elements of each characteristic, which is not possible due to space limitation.

(176)

Furthermore, some AI systems may include mitigating mechanisms to reduce negative effects of some of the five characteristics.

(177)

 OECD, Artificial Intelligence in Society , 2019, p.23.

(178)

     OECD, Artificial Intelligence in Society , 2019.

(179)

     For analysis and reference to the supporting evidence see e.g. OECD, Artificial Intelligence in Society , 2019.

(180)

     See ‘Problems’ section 2 of this impact assessment.

(181)

     According to a global survey, the number of business using AI grew 270% in the past four years and just in the last year tripled, Gartner, Gartner Survey Shows 37 Percent of Organizations Have Implemented AI in Some Form , 2019.

(182)

     European Commission, Commission Staff Working Document – Better Regulation Guidelines , SWD (2017) 350.

(183)

     See above, specifically Toolbox 14.

(184)

     See above, Toolbox 14, pp. 89-90.

(185)

 CJEU Judgment of the Court (Grand Chamber) of 3 December 2019, Czech Republic v European Parliament and Council of the European Union , Case C-482/17, paras. 35.

(186)

 CJEU Judgment of the Court (Grand Chamber) of 2 May 2006, United Kingdom of Great Britain and Northern Ireland v European Parliament and Council of the European Union , Case C-217/04, paras. 62-63.

(187)

Article 114 TFEU as a legal basis for EU action on AI was also suggested by the European Parliament in its 2017 resolution on civil law rules on robotics and in its 2020 resolution on ethical framework for AI, robotics and related technologies. See European Parliament resolution 2020/2012(INL) .

(188)

 For the analysis of the European added value of the EU action see also European Parliamentary Research Service, European added value assessment: European framework on ethical aspects of artificial intelligence, robotics and related technologies , 2020.

(189)

     For the interaction between the AI initiative and revision of the product safety legislation see section 8 (preferred option) and Annex 5.3.

(190)

For the interaction between the AI initiative and the initiative on liability see section 8 (preferred option).

(191)

The revision of the Machinery Directive will address risks emerging from new technologies and problems related to software with a safety function and placed independently on the market, human-robot collaboration, the loss of connection of a device and cyber risks, transparency of programming codes, risks related to autonomous machines and lifecycle related requirements. The revision of the General Product Safety Directive might address cybersecurity risks when affecting safety, mental health, evolving functionalities and substantive modifications of consumer products.

(192)

The proposal of the Digital Services Act, for examples, include obligations to maintain a risk management system, including annual risk assessments for determining how the design of intermediary service, including their algorithmic processes, as well as the use (and misuse) of their service contribute or amplify the most prominent societal risks posed by online platforms. It would also include an obligation to take proportionate and reasonable measures to mitigate the detected risks, and regularly subject the risk management system to an independent audit. Furthermore, enhanced transparency and reporting obligations with regard to content moderation and content amplification are proposed. Finally, proposal of the Digital Services Act also envisages user-facing transparency obligation of content recommender systems, enabling users to understand why, and influence how information is being presented to them, as well as far-reaching data access provisions for regulators and vetted researchers, and strong enforcement and sanctions powers including at EU level.

(193)

     See Fundamental Rights Agency, AI Policy Initiatives 2016-2020 , 2020; Jobin, A., M. Ienca and E. Vayena, ‘ The global landscape of AI ethics guidelines’ , Nature Machine Intelligence Volume 1, pp. 389–399, 2019.

(194)

     Grand View Research, Facial Recognition Market Size.  Industry Report, 2020.

(195)

     Hamilton I.A., Outrage over police brutality has finally convinced Amazon, Microsoft, and IBM to rule out selling facial recognition tech to law enforcement. Here’s what’s going on. Business Insider, 13/06/2020.

(196)

     Dastin J., Amazon scraps secret AI recruiting tool that showed bias against women,  Reuters, 11/11/ 2020.

(197)

     Larson J., et al., How We Analyzed the COMPAS Recidivism Algorithm , Propublica, 23/05/2016.

(198)

     Vigdor N., Apple Card Investigated After Gender Discrimination Complaints , The New York Times, 10/11/2019.

(199)

     AI will be defined in the legal act as ‘a machine-based system that can, for a given set of human-defined objectives, generate output such as predictions, recommendations, or decisions influencing real or virtual environments. AI systems are designed to operate with varying levels of autonomy’ based on the OECD definition (OECD, Recommendation of the Council on Artificial Intelligence , OECD/LEGAL/0449, 2019). To cover a broader range of ‘AI outputs’ (e.g. deep fakes and other content), the OECD definition has been slightly adapted referring to ‘AI outputs such as predictions, recommendations or decisions’.

(200)

For example, focusing only on machine learning technology.

(201)

     High-Level Expert Group on Artificial Intelligence, Ethics Guidelines for Trustworthy AI , 2019.

(202)

     They were also endorsed by the Commission in its 2019 Communication on human-centric approach to AI.

(203)

For example, the OECD AI Principles endorsed also by G20, the Council of Europe Recommendation CM/Rec(2020)1 of the Committee of Ministers to member States on the human rights impacts of algorithmic systems. April 2020, the U.S. President’s Executive Order from 3 December 2020 on Promoting the Use of Trustworthy Artificial Intelligence in the Federal Government etc.

(204)

The EP has proposed the following requirements for high-risk AI: human oversight; safety, transparency and accountability; non bias and non-discrimination; social responsibility and gender equality; environmental sustainability; privacy, and the right to seek redress. The requirements for accountability is for example operationalised through the documentation requirements and obligations, while non-discrimination through the requirements for data and data governance.

(205)

 Environmental and social well-being are aspirational principles included in the HLEG guidelines , the OECD and G20 principles , the draft UNICEF Recommendation of the ethics of AI as well as the EP position. They have not been included in this proposal, because they were considered too vague for a legal act and too difficult to operationalise. The same applies for social responsibility and gender equality proposed by the EP.

(206)

For example, the requirement for privacy is already regulated under the GDPR and the Law Enforcement Directive. The requirement for effective redress proposed by the EP would be covered by the separate liability initiative on AI (planned for Q4 2021) or under existing legislation (e.g. rights of data subjects under the GDPR to challenge a fully automated decision).

(207)

     For a detailed breakout of the views of the various stakeholder groups on these issues, see Annex 2.

(208)

Council of Europe Recommendation CM/Rec(2020)1 also states that risk-management processes should detect and prevent the detrimental use of algorithmic systems and their negative impacts. Quality assurance obligations have also been introduced in other regulatory initiatives in third countries such as the Canada’s Directive on Automated Decision-Making.

(209)

Possibly based on the ALTAI self-assessment tool.

(210)

     Sanctions will include 1) suspension of the label and 2) imposition of fines proportionate to the size of the company in case of serious and/or repeated infringements for providing misleading or inaccurate information. In case of minor and first time infringements, only recommendations and warnings can be issued for imposing possible future sanctions in case the non-compliance persists.

(211)

De facto, this is already happening in some sectors, for example, how drones are regulated under the EU Regulations 2019/947 and 2019/945 for the safe operation of drones in European skies or the specific rules for trading algorithms under the MiFID II / MiFIR financial legislation.

(212)

     Based on up-to date analysis, concerned NLF legislations would be: Directive 2006/42/EC on machinery (which is currently subject to review), Directive 2009/48/EU on toys, Directive 2013/53/EU on recreational craft, Directive 2014/33/EU of the European Parliament and of the Council of 26 February 2014 on lifts and safety components for lifts, Directive 2014/34/EU on equipment and protective systems intended for use in potentially explosive atmospheres, Directive 2014/53/EU on radio-equipment, Directive 2014/68/EU on pressure equipment, Directive 2014/90/EU on marine equipment, Regulation (EU) 2016/424 on cableway installations, Regulation (EU) 2016/425 on personal protective equipment, Regulation (EU) 2016/426 on gas appliances, Regulations (EU) 745/2017 on medical devices and Regulation (EU) 746/2017 on in-vitro diagnostic medical devices. The concerned old-approach legislation would be Regulation (EU) 2018/1139 on Civil Aviation, Regulation 858/2018 on the approval and market surveillance of motor vehicles, Regulation (EU) 2019/2144 on type-approval requirements for motor vehicles, Regulation (EU) 167/2013 on the approval and market surveillance of agricultural and forestry vehicles, Regulation (EU) 168/2013 on the approval and market surveillance of two- or three-wheel vehicles and quadricycles, Directive (EU) 2016/797 on interoperability of railway systems.

(213)

An example of such sectoral ad hoc legislation targeting specific use case of non-embedded AI with mainly fundamental rights implications could be the recent proposal of the New York City Council proposal for a regulation on automated hiring tools. In addition to employment and recruitment, Option 3 and Annex 5.4 identify other high risk use cases also for remote biometric identification systems in publicly accessible places, use of AI for determining access to educational institutions and evaluations; to evaluate the eligibility for social security benefits and services; creditworthiness, predictive policing as well as some other problematic use cases in law enforcement, judiciary, migration and asylum and border controls.

(214)

 The draft UNICEF recommendation on AI also emphasizes the need to protect the right of users to easily identify whether they are interacting with a living being, or with an AI system imitating human or animal characteristics.

(215)

     The EP similarly requested in a recent report that an obligation for labelling of deep fakes should be introduced in a legislation. This is in line with actions taken in some states in the U.S . and also considered in the UK to require labelling of deep fakes or prohibit their use in particular during election campaigns or for person’s impersonation. See also a recent report of Europol and United Nations Interregional Crime and Justice Research Institute on the Malicious Uses and Abuses of Artificial Intelligence , 2020 which identifies deep fakes as an emerging threat to public security.

(216)

See problem section 2.1.2.1.

(217)

 This is overall consistent with the EP position in its resolution on the ethics of AI that the use and gathering of biometric data by private entities for remote identification purposes in public areas, such as biometric or facial recognition, would not be allowed. Only Member States’ public authorities may carry out such activities under strict conditions, notably regarding its scope and duration. The Council of Europe has also proposed certain prohibitions and safeguards in the use of facial recognition technology, see Consultative Committee of The Convention for the Protection of Individuals with regard to Automatic Processing of Personal Data Convention 108 Guidelines on Facial Recognition , 28 January 2021, T-PD(2020)03rev4.

(218)

     Prohibition of certain particularly harmful AI practices has been requested by more than 60 NGOs who sent an open letter to the Commission.

(219)

     The prohibition of the manipulative practice could possibly be integrated into the Unfair Commercial Practice Directive, while the prohibitions of the general purpose social scoring of citizens could be possibly included in the General Data Protection Regulation.

(220)

     Council of Europe, Declaration on the manipulative capabilities of algorithmic processes, 13 February 2019, Recommendation CM/Rec(2020)1 on the human rights impacts of algorithmic systems also states that experimentation designed to produce deceptive or exploitative effects should be explicitly prohibited. With regard to children who are vulnerable group, this prohibition is also consistent with Recommendation CM/Rec(2018)7 on Guidelines to respect, protect and fulfil the rights of the child in the digital environment that advocates for a precautionary approach and for taking measures to prevent risks and practices adversely affecting the physical, emotional and psychological well-being of a child and to protect their rights in the digital environment.

(221)

UNICEF, Policy guidance on AI for children, September , 2020.

(222)

HLEG Policy and Investment Recommendations For Trustworthy AI, June 2019. For a similar social credit system introduced in China, see Chinese State Council Notice concerning Issuance of the Planning Outline for the Construction of a Social Credit System (2014-2020) GF No. (2014)21. Systems where people are socially scored with discriminatory or disproportionately adverse treatment have also been put in place in some Member States such as the Gladaxe system in Denmark or the SyRI system in the Netherlands.

(223)

See EP report on artificial intelligence: questions of interpretation and application of international law in so far as the EU is affected in the areas of civil and military uses and of state authority outside the scope of criminal justice (2020/2013(INI)) .

(224)

Proposal for a REGULATION OF THE EUROPEAN PARLIAMENT AND OF THE COUNCIL on a Single Market For Digital Services (Digital Services Act) and amending Directive 2000/31/EC COM/2020/825 final.

(225)

See an open letter to the Commission from EDRi and more than 60 other NGOs.

(226)

The use of these AI systems in the public sector could also be beneficial if subjected to appropriate safeguards as it would help public authorities to be more effective in the allocation of scarce public resources, thus potentially improving the access to these services and even reducing discrimination in individual human decisions that might also be biased.

(227)

     The EP has called for an instrument equally applying to AI, but also covering robotics and related technologies. ‘Related technologies’ is defined by EP as ‘software to control with a partial or full degree of autonomy a physical or virtual process, technologies capable of detecting biometric, genetic or other data, and technologies that copy or otherwise make use of human traits’ would be covered by the OECD definition to the extent that this concerns technologies that enable software to control with a partial or full degree of autonomy a physical or virtual process. ‘Technologies capable of detecting biometric, genetic or other data, and technologies that copy or otherwise make use of human traits’ are covered only to the extent that they use AI systems as defined by OECD. The rest is excluded, because any technology that is able to detect and process ‘other data’ would qualify as AI, which is considered excessively broad and beyond AI. While cognitive robotics would be included in the list of AI approaches and techniques, other robots are not related to the same AI characteristics as described in section 2.2. and do not pose specific fundamental rights risks, so these are already sufficiently covered by the existing product safety legislation.

(228)

     The definition of high-risk used in the context of a horizontal framework may be different from the notion of high-risk used in sectoral legislation because of different context of the respective legislations. The qualification of an AI system as high-risk under the AI horizontal instrument does not necessarily mean that the system should be qualified as high-risk under other sectoral acts.

(229)

This is irrespective of whether the safety components are placed on the market independently from the product or not.

(230)

NLF product legislation may also cover some AI systems which are to be considered products by themselves (e.g., AI devices under the Medical Device Regulations or AI safety components placed independently on the market which are machinery by themselves under the Machinery Directive).

(231)

This can include standalone AI systems not covered by sectoral product safety legislation (e.g., recruitment AI system) or AI systems being safety components of products which are not covered by sectoral product safety legislation under point 1 and which are regulated only by the General Product Safety Directive. An initial list of high-risk AI systems covered by this point is detailed in Annex 5.4.

(232)

These criteria include: a) the extent to which an AI system has been used or is about to be used; b) the extent to which an AI system has caused any of the harms referred to above or has given rise to significant concerns around their materialization; c) the extent of the adverse impact of the harm; d) the potential of the AI system to scale and adversely impact a plurality of persons or entire groups of persons; e) the possibility that an AI system may generate more than one of the harms referred to above; f) the extent to which potentially adversely impacted persons are dependent on the outcome produced by an AI system, for instance their ability to opt-out of the use of such an AI system; g) the extent to which potentially adversely impacted persons are in a vulnerable position vis-à-vis the user of an AI system; h) the extent to which the outcome produced by an AI system is reversible; i) the availability and effectiveness of legal remedies; j) the extent to which existing Union legislation is able to prevent or substantially minimize the risks potentially produced by an AI system.

(233)

     See problem section 2.1.2.

(234)

     As an additional criteria, it could be envisaged that broader sectors are identified to select the high-risk AI use cases, as proposed in the White Paper and by the EP. The EP report proposes to base the risk assessment on exhaustive and cumulative high-risk sectors and of high-risk uses or purposes. Risky sectors would comprise employment, education, healthcare, transport, energy, public sector, defence and security, finance, banking and insurance. The Commission considers that broad sectors are not really helpful to identify specific high risk use cases. Applications may be low-risk even in high risk sectors (i.e. document management in the justice sector) or high-risk in sectors which are classified as low risk. On the other hand, more specific fields of AI applications could be envisaged to circumscribe the possible change in the use cases as another alternative.

(235)

Final Draft of ISO/IEC TR 24030 identifies a list of 132 AI Use Cases that have been screened as a starting point by applying the risk assessment criteria and the methodology specified above. Other sources of use cases have been also considered such as those identified as high-risk in the EP report, in the public consultation on the White paper and based on other sources presented in Annex 5.4.

(236)

     See more details in Annex 5.4 on how the methodology has been applied and what are the identified high-risk use cases.

(237)

     The EP has identified as high-risk uses the following: recruitment, grading and assessment of students, allocation of public funds, granting loans, trading, brokering, taxation, medical treatments and procedures, electoral processes and political campaigns, some public sector decisions that have a significant and direct impact on the rights and obligations of natural or legal persons, automated driving, traffic management, autonomous military systems, energy production and distribution, waste management and emissions control. The proposed list by the Commission largely overlaps, it is summarised in Annex 5.4. It does not include algorithmic trading because this is regulated extensively by the Commission Delegated Regulation (EU) 2017/589. Use of AI for exclusive military purposes is considered outside the scope of this initiative given the implications for the Common Foreign and Security Policy regulated under Title V of TEU. Electoral processes and political campaigns are considered covered by the proposal for a Digital Services Act and the proposal for e-Privacy regulation. Brokering, taxation and emission controls were considered sufficiently covered by existing legislation and there is no sufficient evidence for harms caused by AI, but it could not be excluded that these might be included at a later stage with future amendments.

(238)

The EP has also proposed amendments to the list of high risk uses cases via Commission’s delegated acts.

(239)

An expert group would support the work of the European AI Board and would regularly review the need for amendment of the list of high-risk AI systems based on evidence and expert assessment.

(240)

The AI legislation would be built as a New Legislative Framework (NLF) type legislation that is implemented through harmonised technical standards.    The European Standardisation Organisations (CEN/CENELEC and ETSI) will adopt these standards on the basis of a mandate issued by the Commission and submit them to the Commission for possible publication in the Official Journal.

(241)

While CEN/CENELEC and ETSI, as European Standardisation Organisations, are the addressees of Commission’s Standardisation Requests in accordance with Regulation 2012/1025/EU, the Vienna Agreement signed between CEN and ISO in 1991 recognizes the primacy of international standards and aims at the simultaneous recognition of standards at international and European level by means of improved exchange of information and mutual representation at meetings. This usually ensures the full coordination between international and European process for standardisation. Moreover, other important international standardisation organisations, IEEE, and CEN/CENELEC have recently engaged in upscaling their level of collaboration and mutual cooperation.

(242)

     Except where users use the high-risk AI system in the course of a personal (non-business) or transient activity e.g. travellers from third countries could use for example their own self-driving car and do not comply with the new obligations, while they are in Europe.

(243)

Other use cases involving the use of AI that merit transparency requirements have also been considered (e.g., when a person is subject to solely automated decisions or micro-targeted), but these were discarded. This is because relevant transparency obligations already exist in data protection legislation (Articles 13 and 14 of the GDPR), in consumer protection law as well as in the proposals for the e-Privacy Regulation COM/2017/010 final - 2017/03 (COD) and the proposal for the Digital Services Act (COM/2020/825 final).

(244)

To ensure consistency in the implementation of the new AI instrument and existing sectoral legislation, Member States shall entrust market surveillance activities for those AI systems to the national competent authorities already designated under relevant sectoral Union legislation, where applicable (e.g. sectoral product safety, financial service).

(245)

Thresholds and criteria for assessment would be defined in the legal act to ensure effective and uniform enforcement of the new rules across all Member States. Fines would be in particular imposed for supplying incorrect, incomplete or false information and non-compliance with the obligations for ex ante conformity assessment and post market monitoring, failure to cooperate with the competent authorities etc.

(246)

See footnote 231 and Annex 5.4.

(247)

This approach is fundamentally in line with the idea of “pre-determined change control plan” developed and proposed by the US Food and Drug Administration (FDA) in the field of AI-based software in the medical field in a discussion paper produced in 2019. The effectiveness of the described approach is further reinforced by the fact that an obligation for post-market monitoring is set for providers of high-risk AI system, obliging them to collect data about the performance of their systems on a continuous basis after deployment and monitor it.

(248)

More details on the interaction between the ex-ante enforcement system envisaged in the horizontal act and its interplay with product safety legislation can be found in section 8 of this impact assessment and Annex 5.3.

(249)

This choice is also consistent with the EP position which envisages ex ante certification and ex post institutional control for compliance.

(250)

Nevertheless, the conformity assessment rules of many existing relevant sectorial legislations would allow providers of high-risk AI systems that are safety components of products to carry out a conformity assessment through internal checks if they have applied harmonised standards.

(251)

This would also contribute to the principle of societal well-being endorsed by the OECD and the HLEG and follows the recommendation of the Council of Europe CM/Rec(2020)1 to increase transparency and oversight for AI systems having significant fundamental rights implications.

(252)

 Council of Europe, Recommendation CM/Rec(2020)1 on the human rights impacts of algorithmic systems. 2020.

(253)

European Agency for Fundamental Rights, Getting The Future Right – Artificial Intelligence and Fundamental Rights , 2020.

(254)

     The AI Board would be an independent EU ‘body’ established under the new instrument. Its status would be similar to the European Data Protection Board.

(255)

     This has also been requested by the European Parliament resolution of 20 October 2020 ( 2020/2012(INL) ).

(256)

See for a similar definition Council Conclusions on regulatory sandboxes and European Commission, TOOL #21. Research & Innovation, Better Regulation Toolbox; European Commission; 6783/20 (COM (2020)103).

(257)

See also Council Conclusions on regulatory sandboxes which emphasize the need to always respect and foster the precautionary principle and ensure existing levels of protection are respected. While under certain regulatory sandboxes there is a possibility to provide complete derogations or exemptions from the existing rules, this is not considered appropriate in this context given the high risks to safety and fundamental rights. Similar approach has also been followed by other competent authorities establishing sandboxes in the financial sector where the sandbox is rather used as a tool to apply flexibility permitted by law and help reach compliance in an area of legal uncertainty instead of disapplication of already existing Union legislation. See in this sense ESMA, EBA and EIOPA Report FinTech: Regulatory sandboxes and innovation hubs , 2018.

(258)

For example, how to determine when the AI specific application is sufficiently accurate, robust or transparent for its intended purpose, whether the established risk management system and quality management systems are proportionate, in particular for SMEs and start-ups, etc.

(259)

See in this sense also Council of Europe Recommendation CM/Rec(2020)1 of the Committee of Ministers to member States on the human rights impacts of algorithmic systems (Adopted by the Committee of Ministers on 8 April 2020 at the 1373rd meeting of the Ministers’ Deputies).

(260)

     With regard to so-called ‘old approach’ products like cars the introduction of specific requirements for AI will require changes in sectorial legislation. Such modifications should follow the principles of the horizontal legislation. As these sectorial legislations are regularly updated, a timely insertion of the specific AI requirements can be expected. Therefore, as well in this area, MS should not be in need to legislate unilaterally.

(261)

     McKinsey Global Institute, Notes from the AI frontier: tackling Europe’s gap in digital and AI , 2019.

(262)

     McKinsey Global Institute, Notes from the AI Frontier: Modeling the Impact of AI on the World Economy , 2018.

(263)

Administrative burdens mean the costs borne by businesses, citizens, civil society organizations and public authorities as a result of administrative activities performed to comply with information obligations included in legal rules; compliance costs the investments and expenses that are faced by businesses and citizens in order to comply with substantive obligations or requirements contained in a legal rule. (Better Regulation tool 58)

(264)

For the estimates related to the European Added Value see e.g. European Parliamentary Research Service, European added value assessment: European framework on ethical aspects of artificial intelligence, robotics and related technologies, 2020. This analysis suggest that a common EU framework on ethics (as compared to fragmented national actions) has the potential to bring the European Union €294.9 billion in additional GDP and 4.6 million additional jobs by 2030.

(265)

  ISBN 978-92-76-36220-3

(266)

Given the assumption of the cost estimation that an average AI application costs €170 000 to develop, it is reasonable to assume that most SMEs only produce one or maximum two AI applications per year

(267)

     GDPR.EU, Millions of small businesses aren’t GDPR compliant, our survey finds . Information website, 2019.

(268)

Datagrail, The Cost of Continuous Compliance, Benchmarking the Ongoing Operational Impact of GDPR and CCPA , 2020.

(269)

     Such as ISO 9001/2015 ‘general), IEC13485 (medical devices), ISO/IEC 25010 (software).

(270)

     European Commission, Impact Assessment Accompanying the Cybersecurity Act, SWD(2017) 500 final

(271)

     Centre for Strategy & Evaluation Services (CSES): Evaluation of Internal Market Legislation for Industrial Products

(272)

     ResearchGate, Calculating average cost per company of annual conformity assessment activities . 

(273)

     Estimates for 2018 produced by DIW Econ, based on 2008-2016 figures from the Structural Business Statistics Database (Eurostat, Structural business statistics overview , 2020).

(274)

      https://www.ai4eu.eu

(275)

     Established under the Digital Europe Programme (currently under negotiation).

(276)

Inside the Regulatory Sandbox: Effects on Fintech Funding

(277)

     Assuming a constant European share of the global AI market at 22%, based on its share in the AI software market in 2019 (Statista, Revenues from the artificial intelligence software market worldwide from 2018 to 2025, by region , 2019).

(278)

     As a comparison, Data Protection Authorities in small Member States usually have between 20 and 60 staff, in big Member States between 150 and 250 (Germany is the outlier with 700; Brave, Europe’s Governments are failing the GDPR , 2020).

(279)

     OECD, Recommendation of the Council on Artificial Intelligence , OECD/LEGAL/0449, 2019.

(280)

     High-Level Expert Group on The Impact of the Digital Transformation on EU Labour Markets, Final Report with Recommendation , 2018.

(281)

     Servoz, M. AI – the future of work? Work of the future !, 2019.

(282)

For high risk, they have already agreed and cannot adopt national measures that are contrary to the uniform rules agreed within the European horizontal instrument.

(283)

     European Commission, Report from the Commission to the European Parliament, the Council and the European Economic and Social Committee, Report on the safety and liability implications of Artificial Intelligence, the Internet of Things and robotics , COM/2020/64 final, 2020.

(284)

     See footnotes 229 and 300 for additional details.

(285)

     An overview of the impact and applicability of the horizontal framework to high-risk AI systems is provided in Annex 5.3.

(286)

     Based on up-to-date analysis, the concerned NLF legislations would be: Directive 2006/42/EC on machinery (which is currently subject to review), Directive 2009/48/EU on toys, Directive 2013/53/EU on recreational craft, Directive 2014/33/EU on lifts and safety components for lifts, Directive 2014/34/EU on equipment and protective systems intended for use in potentially explosive atmospheres, Directive 2014/53/EU on radio-equipment, Directive 2014/68/EU on pressure equipment, Regulation (EU) 2016/424 on cableway installations, Regulation (EU) 2016/425 on personal protective equipment, Regulation (EU) 2016/426 on gas appliances, Regulations (EU) 745/2017 on medical devices and Regulation (EU) 746/2017 on in-vitro diagnostic medical devices. 

(287)

     Based on up-to-date analysis, the concerned old-approach legislation would be Regulation (EU) 2018/1139 on Civil Aviation, Regulation 858/2018 on the approval and market surveillance of motor vehicles, Regulation (EU) 2019/2144 on type-approval requirements for motor vehicles, Regulation (EU) 167/2013 on the approval and market surveillance of agricultural and forestry vehicles, Regulation (EU) 168/2013 on the approval and market surveillance of two- or three-wheel vehicles and quadricycles, Directive (EU) 2016/797 on interoperability of railway systems. Given the mandatory character of international standardization, Directive 2014/90/EU on marine equipment (which is a peculiar NLF-type legislation) will be treated in the same way as old-approach legislation.

(288)

The direct or indirect applicability of requirements will fundamentally depend on the legal structure of the relevant old-approach legislation, and notably on the mandatory application of international standardisation. Where application of international standardisation is mandatory, requirements for the high-risk AI systems in the horizontal framework on AI will not directly apply but will have to be taken into account in the context of future Commission’s activities in the concerned sectors.

(289)

See footnote 231 and Annex 5.4.

(290)

     For consistency purposes and in order to leverage on existing EU legislation and tools in the market surveillance domain, the provisions of the Market Surveillance Regulation 2019/1020 would apply, meaning the RAPEX system established by the General Product Safety Directive 2001/95/EC would be used for the exchange of relevant information with regard to measures taken by Member States again non-compliant AI systems.

(291)

The new liability rules on AI are currently under reflection (see section 1.3. for more details on the issues at stake).

(292)

The discussion on the requirements in the horizontal framework that relate to safety of a system and protection of fundamental rights ‘ex ante’ and ‘ex post’ placement on the market (that are both covered in the proposed horizontal framework) is discussed in another part of the text. For example, to ensure ex-post enforcement of requirements provided in the horizontal regulation, as discussed in the sections on enforcement, the proposal includes appropriate investigations by competent authorities with powers to request remedial action and impose sanctions.

(293)

The relevant recital provision to this extent would be included in the proposed horizontal framework initiative.

(294)

     As indicated in Section 1.3.3., one of the elements under reflection is the possible Revision of the Product Liability Directive. The Product Liability Directive is a technology-neutral directive applicable to all products. If and when reviewed, it would also apply to high-risk AI systems covered under the AI horizontal framework.

(295)

 If not indicated otherwise, Source of the definitions: DG CNECT Glossary . 

Top

Brussels, 21.4.2021

SWD(2021) 84 final

COMMISSION STAFF WORKING DOCUMENT

IMPACT ASSESSMENT

ANNEXES

Accompanying the

Proposal for a Regulation of the European Parliament and of the Council

LAYING DOWN HARMONISED RULES ON ARTIFICIAL INTELLIGENCE (ARTIFICIAL INTELLIGENCE ACT) AND AMENDING CERTAIN UNION LEGISLATIVE ACTS

{COM(2021) 206 final} - {SEC(2021) 167 final} - {SWD(2021) 85 final}


Annexes: Table of Contents

1.    ANNEX 1: PROCEDURAL INFORMATION

1.1.    Lead DG, Decide Planning/CWP references

1.2.    Organisation and timing

1.3.    Opinion of the RSB and responses    

1.4.    Evidence, sources and quality

2.    ANNEX 2: STAKEHOLDER CONSULTATION

2.1.    The public consultation on the White Paper on Artificial Intelligence

2.2.    Analysis of the results of the feedback on the inception impact assessment

2.3.    Stakeholder outreach

2.3.1.    Event on the White Paper with larger public

2.3.2.    Technical consultations

2.3.3.    Outreach and awareness raising events in Member States and International outreach

2.3.4.    European AI Alliance platform

3.    ANNEX 3: WHO IS AFFECTED AND HOW?

3.1.    Practical implications of the initiative    

3.1.1.    Economic operators/business        

3.1.2.    Conformity assessment, standardisation and other public bodies

3.1.3.    Individuals/citizens

3.1.4.    Researchers

3.2.    Summary of costs and benefits

4.    ANNEX 4: ANALYTICAL METHODS

5.    ANNEX 5: OTHER ANNEXES

5.1.    Ethical and Accountability Frameworks on AI introduced in Third Countries

5.2.    Five specific characteristics of AI

5.3.        Interaction between the AI initiative and product safety legislation

5.4.    List of high-risk AI systems (not covered by sectorial product legislation)

5.5.    Analyses of impacts on fundamental rights specifically impacted by the intervention

1.Annex 1: Procedural information

1.1. Lead DG, Decide Planning/CWP references

Lead DG: Directorate-General for Communications Networks Content and Technology (CNECT).

Decide: PLAN/2020/7453.

CWP: Adjusted Commission Work Programme 2020 COM(2020) 440 final: Follow-up to the White Paper on Artificial Intelligence, including on safety, liability, fundamental rights and data (legislative, incl. impact assessment, Article 114 TFEU, Q1 2021).

1.2. Organisation and timing

The initiative constitutes a core part of the single market given that artificial intelligence (AI) has already found its way into a vast majority of services and products and will only continue to do so in the future. It is based on Article 114 TFEU since it aims to improve the functioning of the internal market by setting harmonized rules on the development, placing on the Union market and the use of AI systems embedded in products and services or provided as stand-alone AI applications.

The impact assessment process started with opening of a public consultation on the AI White Paper 1 on 19 February 2020, open until 14 June 2020. The inception impact assessments was published for stakeholder comments on 23 July 2020, open for comments until 10 September 2020. For details on the consultation process, see Annex 2.

The inter-service group (ISG) met on 10 November 2020 before submission of the Staff Working Document to the Regulatory Scrutiny Board (18 November 2020). The ISG consists of representatives of the Secretariat-General, and the Directorates-General CNECT, JUST, GROW, LS, HOME, SANTE, FISMA, AGRI, JRC, DEFIS, TRADE, ENV, ENER, EMPL, EAC, MOVE, RTD, TAXUD, MARE, EEAS, ECFIN and CLIMA.

A meeting with the Regulatory Scrutiny Board was held on 16 December 2020. The Regulatory Scrutiny Board issued a negative opinion on 18 December 2020. The inter-service group met again on 18 January before re-submission of the Staff Working Document to the Regulatory Scrutiny Board (22 February 2021).

Based on the Board's recommendations of 18 December, the Impact Assessment has been revised in accordance with the following points.

1.3.Opinion of the RSB and responses

The Impact Assessment report was reviewed by the Regulatory Scrutiny Board. Based on the Board’s recommendations, the Impact Assessment has been revised to take into account the following comments:



First submission to the Regulatory Scrutiny Board

Comments of the RSB

How and where comments have been addressed

(B) Summary of findings

(1)The report is not sufficiently clear on how this initiative will interact with other AI initiatives, in particular with the liability initiative.

The report has been substantially reworked, especially in the introduction, sections 1.3, 4.2 and 8, to better explain how this initiative interacts with other AI initiatives such as the safety revisions and the AI liability initiative, emphasizing the complementarity between the three initiatives and their different scopes.

Regarding links with the liability initiative, the AI horizontal initiative is an ex ante risk minimisation instrument including a system of continuous oversight to avoid and minimise the risk of harm caused by AI, whilst the initiative on liability rules would be an ex post compensation instrument when such harm has occurred (Sections 1.3.3 and 8).

Concerning the product safety revisions, these aim primarily at ensuring that the integration of AI systems into the overall product will not render a product unsafe and compliance with the sectoral rules is not affected. On the other hand, the AI legislation will set a single definition of AI, a risk assessment methodology and impose minimum requirements specific to the high-risk AI system to address both safety and fundamental rights risks (Section 1.3.2. and 8).

Section 8 and Annex 5.3 explain in detail how the AI horizontal initiative will work in practice for sectoral safety legislation (old and new approach). Annex 5.3 also lists all pieces of sectoral product legislation that will be affected by the horizontal AI initiative.

(2) The report does not discuss the precise content of the options. The options are not sufficiently linked to the identified problems. The report does not present a complete set of options and does not explain why it discards some.

The report has been substantially re-worked to explain in detail the content of all policy options and their linkages to the problem identified in the impact assessment.

The report now sets out in detail the five requirements for AI systems and how they are linked to the problems and the drivers (opacity, autonomy, data dependency etc.) (Policy Option 1).

The prohibited practices are now clearly explained and justified with links to the problems and relevant justifications and recommendations for their prohibition (Policy Option 2). Option 2 also lists all the sectoral initiatives that would have to be undertaken and their content, including an ad hoc specific initiative that would further restrict the use of remote biometric identification systems at public spaces.

The risk assessment methodology has been explained with the precise criteria defined (Policy Option 3). All high-risk AI use cases (not covered by product sectoral legislation) are listed and justified by applying the methodology in a new Annex 5.4 supported with evidence. Annex 5.3. explains, on the other hand, the methodology for high-risk AI covered by sectoral product safety legislation and lists the relevant acts that would be affected by the new horizontal initative.

The compliance procedures and obligations for providers have been further explained for Policy Options 1, 2 and 3, linking them to the problems the AI regulatory initiative aims to solve. The same has been done for obligations of users for Options 2 and 3.

Measures to support innovation are further explained in Option 3 (e.g. sandboxes, DIHs) and how they will operate and help to address the problems.

Option 3+ has been reworked and now explains in detail the possibility for codes of conduct as a voluntary mechanism for non-high risk AI applications.

Details on the enforcement and governance system at national and EU level have been given for all policy options.

For each of these different issues, alternative policy choices/sub-otions have been considered and explanations given why they have been discarded. A new table 7 summarises the selected and discarded policy sub-options.

(3) The report does not show clearly how big the relative costs are for those AI categories that will be regulated by this initiative. Even with the foreseen mitigating measures, it is not sufficiently clear if these (fixed) costs could create prohibitive barriers for SMEs to be active in this market.

Section 6.1.3 has been reworked in order to put costs in relation to regulated AI applications. The report now provides a perspective on the level of costs by estimating costs of other regulatory requirements and explains why there is hardly any risk of depriving EU of certain technological innovations. It also distinguishes one-off and running costs and analyses which activities companies would have to undertake even without regulatory intervention.

The role of regulatory sandboxes for SMEs has been clarified, with guidance from competent authorities to facilitate compliance and reduce costs (Section 5.4.).

(C) What to improve

(1) The content of the report needs to be completed and reworked. The narrative should be improved and streamlined, by focusing on the most relevant key information and analysis.

The content of the report has been streamlined and focuses now more on the most relevant key information, such as how this initiative will interact with other AI initiatives, how the options are designed and what is their precise content, how the high-risk cases are selected. The context and the objectives of the proposal (especially Section 4.2.4.) have been detailed. The policy options have been further completed and better linked to the identified problems (Section 5). The report now presents a complete set of options and explains why it discards some.

(2) The report should clearly explain the interaction between this horizontal regulatory initiative, the liability initiative and the revision of sectoral legislation. It should present which part of the problems will be addressed by other initiatives, and why. In particular, it should clarify and justify the policy choices on the relative roles of the regulatory and liability initiatives.

The interaction between the AI horizontal initiative, the liability initiative and sectoral product safety revisions has been further explained and analysed (Introduction and Section 1.3.).

Section 4.2. explains which parts of the problems will be addressed by the horizontal AI initiative and which parts by the liability and the sectoral product safety revisions. Policy choices on the relative roles of the regulatory and liability initiatives are clarified in Section 8.

Annex 5.3. explains in detail how the AI horizontal initiative will work in practice for sectoral safety legislation (old and new approach) and lists all acts that will be affected by the horizontal AI initiative.

(3) In the presentation of the options, the report focusses mainly on the legal form, but it does not sufficiently elaborate on the content. The report should present a more complete set of options, including options that were considered but discarded. Regarding the preferred option, the report should give a firm justification on what basis it selects the four prohibited practices. There should be a clear definition and substantiation of the definition and list of high-risk systems. The same applies to the list of obligations. The report should indicate how high risks can be reliably identified, given the problem drivers of complexity, continuous adaptation and unpredictability. It should consider possible alternative options for the prohibited practices, high-risk systems, and obligations. These are choices that policy makers need to be informed about as a basis for their decisions.

The report now describes in detail the content of all policy options and clearly links them to the problem identified in the impact assessment. For each of the key dimensions linked to the content and the enforcement and governance system, it presents alternative policy choices and explains why it discards some.

Two new tables are added: Table 6 Summary of the content of all Policy Options and Table 7 Summary of selected sub-option and discarded alternative sub-options. To improve readability, summary tables of the content of each policy option have also been added.

Alternatives for the proposed mandatory AI requirements are discarded in Policy Option 1 (e.g. social and environmental well-being, accessibility, proposed by EP), but could be addressed via voluntary codes of conduct (Option 3+). 

Option 3 explains now in detail the risk assessment methodology for classification of a system as high-risk distinguishing between AI systems as safety components of products and other high-risk AI systems (stand-alone). For the second category, the methodology with the concrete criteria for assessment has been explained in detail and applied in Annex 5.4. More details are given how the high-risk cases are selected and on what evidence basis, starting from a larger pool of 132 ISO use cases and other possible applications (Annex 5.4.). In option 3, the report explains also that the methodology focusing on the severity and likelihood of harms that is appropriate to address the problem drivers of complexity, continuous adaptation and unpredictability. Alternative ways of how the risk assessment could be done are also discarded – e.g. burden placed on the provider for the risk assessment (Policy Option 3).

Alternative prohibited practices are also considered, such as the complete prohibition of remote biometric identification systems and other cases requested by civil society organisation (Policy Option 2).

Alternative ways of the proposed compliance procedure and obligations for providers and users are also analysed and discarded (Policy Option 3).

(4) The report should be clearer on the scale of the (fixed) costs for regulated applications. It should better analyse the effects of high costs on market development and composition. The report should expand on the costs for public authorities, tasked to establish evolving lists of risk rated AI products. It should explain how a changing list of high-risk products is compatible with the objective of legal certainty. The analysis should consider whether the level of costs affects the optimal balance with the liability framework. It should reflect on whether costs could be prohibitive for SMEs to enter certain markets. Regarding competiveness, the report should assess the risk that certain high-risk AI applications will be developed outside of Europe. The report should take into account experiences and lessons learnt from third countries (US, China, South Korea), for instance with regard to legal certainty, trust, higher uptake, data availability and liability aspects.

Section 6.1.3. has been reworked in order to put costs in relation to regulated AI applications. The report now also provides a perspective on the level of costs by estimating costs of other regulatory requirements. Section 6.1.4. has been strengthened to assess the impact on SMEs, and support measures for SMEs have been spelt out.

Section 6.1.5. now discards specifically the possibility that certain high-risk AI applications will only be available outside of Europe as a result of the regulatory proposal. A new annex with an overview of development in third countries has been added (Annex 5.1.).

Section 5.4.2.c) explains how a changing list of high-risk AI systems is compatible with the objective of legal certainty. The powers of the Commission would be preliminarily circumscribed by the legislator within certain limits. Any change to the list of high-risk AI use cases would also be based on the solid methodology defined in the legislation, supporting evidence and expert advice. To ensure legal certainty, future amendments would also require impact assessment following broad stakeholder consultation and there would always be a sufficient transitional period for adaptation before any amendments become binding for operators.

In presenting the proposed content of the various policy options (Section 5.), the report also takes into account experiences and lessons learnt from third countries.

(5) The report should explain the concept of reliable testing of innovative solutions and outline the limits of experimenting in the case of AI. It should clarify how regulatory sandboxes can alleviate burden on SMEs, given the autonomous dynamics of AI.

Section 5.4. has been detailed and now outlines better the limits of experimenting with AI technologies (Policy Option 3). The role of regulatory sandboxes in the mitigation of burden on SMEs has been better clarified, since options 3 and 3+ foresee implementation of regulatory sandboxes allowing for the testing of innovative solutions under the oversight of the public authorities in order to alleviate the burden on SMEs (Section 6.1.4.). It has been clarified that no exemption will be granted, and that benefits to SMEs will come from lower costs for specialist legal and procedural advice and from faster market entry.

(6) The report should better use the results of the stakeholder consultation. It should better reflect the views of different stakeholder groups, including SMEs and relevant minority views, and discuss them in a more balanced way throughout the report.

The report has been reworked and completed with additional breakdowns of stakeholder views based on the public consultation on the White Paper on AI, for instance on the various problems identified in the impact assessment, on the need for regulation, on sandboxes, on costs and administrative burdens, on the limitation of requirements to high-risk applications, on the definition of AI, on the use of remote biometric identification systems in public spaces.

(7) The report should make clear what success would look like. The report should elaborate on monitoring arrangements and specify indicators for monitoring and evaluation.

The report has been further elaborated and detailed on monitoring and evaluation (Section 9). Success has been defined two-fold: 1) Absence of violation of safety and fundamental rights of individuals; 2) Rapid uptake of AI based on widespread trust. Thus, AI made in EU would become a world reference.

Additional details on the reporting systems and the indicators have been provided: AI providers would be obliged to report safety incidents and breaches of fundamental rights obligations when brought to their attention; competent authorities would monitor and investigate incidents; the Commission would maintain a publicly accessible database of high-risk AI systems with mainly fundamental rights implications; and it will also monitor uptake of AI and market developments.

Indicators for monitoring and evaluation are specified.

Second submission to the Regulatory Scrutiny Board

(1) The report should explain the methodology and sources for its cost calculations in the relevant annex. It should include a detailed discussion of where and why the presented costs deviate from the supporting study. The report should better discuss the combined effect of the foreseen support measures for SMEs (lower fees for conformity assessments, advice, priority access to regulatory sandboxes) and the (fixed) costs, including for new market entrants.

Annex 4 has been expanded to provide more details on the methodology extracted from the support study. An explanation on where and why assumptions and figures differ from the support study was provided.

A new section has been added as the end of 6.1.4 setting out how the support measures provide benefits to SMEs and how far this counteracts the costs generated by the regulation.

1.4. Evidence, sources and quality

To ensure a high level of coherence and comparability of analysis for all potential policy approaches, an external study was procured to feed into the impact assessment. It reviewed available evidence of fundamental rights or safety-related risks created by AI applications, as well as assessed the costs of compliance with the potential requirements outlined in the AI White Paper. The study also reviewed evidence of potential compliance costs based on the review of literature or other countries and analysed results of the public consultation launched by the White Paper. The estimation of the costs of compliance can be found in Annex 4 of this impact assessment.

In order to gather more evidence following the consultation on the AI White Paper, the Commission organised in July, September and November 2020 five (closed) expert webinars on (1) Requirements for high-risk AI, (2) Standardisation, (3) Conformity assessment and (4) Biometric identification systems (5).

On 9 October 2020, the Commission organised the Second European AI Alliance Assembly, with the participation of over 1 900 viewers. Featuring Commissioner Thierry Breton, representatives of the German Presidency of the European Council, Members of the European Parliament as well as other high-level participants, the event focused on the European initiative to build an Ecosystem of Excellence and of Trust in Artificial Intelligence. 2 The sessions included plenaries as well as parallel workshops and breakout sessions. Viewers were able to interact and ask questions to the panellists.

Furthermore, the Commission held a broad stakeholder consultation on the White Paper. There were numerous meetings with companies, business associations, civil society, academia, Member States and third countries’ representatives. In addition, Commission representatives participated in more than fifty (online) conferences and roundtables, organised by Member States, civil society, business associations, EU representations and delegations and others.

In addition, the IA takes into account the analysis and the work that contributed to the Ethical Guidelines adopted by the high-Level Expert Group on AI (HLEG AI) and the results of the testing of the Assessment List of the HLEG AI. The guidelines are based on the analysis of more than 500 submissions from stakeholders. The Assessment List of the HLEG AI, adopted in the second half of 2019, where more than 350 organisation participated.

Finally, to further support evidence based analysis, the Commission has conducted extensive literature review, covering academic books, journals and well as a wide spectrum of policy studies and reports, including by non-governmental organisations. They have been quoted in the main body of the Impact Assessment.

2.Annex 2: Stakeholder consultation

In line with the Better Regulation Guidelines, 3 the stakeholders were widely consulted as part of the impact assessment process.

2.1.The public consultation on the White Paper on Artificial Intelligence

The main instrument was the public consultation on the White Paper on Artificial Intelligence that ran from 19 February to 14 June 2020. The questionnaire of the consultation was divided in three sections:

·Section 1 referred to the specific actions, proposed in the White Paper’s Chapter 4 for the building of an ecosystem of excellence that can support the development and uptake of AI across the EU economy and public administration;

·Section 2 referred to a series of options for a regulatory framework for AI, set up in the White Paper’s Chapter 5;

·Section 3 referred to the Report on the safety and liability aspects of AI. 4

The summary below only address the questions relating to Sections 2 and 3 of the public consultation, where the regulatory framework is discussed.

The consultation targeted interested stakeholders from the public and private sectors, including governments, local authorities, commercial and non-commercial organisations, experts, academics and citizens. Contributions arrived from all over the world, including the EU’s 27 Member States and countries such as India, China, Japan, Syria, Iraq, Brazil, Mexico, Canada, the US and the UK.

The public consultation included a set of closed questions that allowed respondents to select one or more options from a list of answers. In addition to the given options, respondents could provide free text answers to each question of the questionnaire or insert position papers with more detailed feedback. 

In total, 1 215 contributions were received, of which 352 were from companies or business organisations/associations, 406 from citizens (92% EU citizens), 152 on behalf of academic/research institutions, and 73 from public authorities. Civil society’s voices were represented by 160 respondents (among which 9 consumers’ organisations, 129 non-governmental organisations and 22 trade unions), 72 respondents contributed as ‘others’.

Out of 352 business and industry representatives 222 were individual companies/businesses, while 130 came from business associations, 41.5% of which were micro, small and medium-sized enterprises. The rest were business associations. Overall, 84% of business and industry replies came from the EU-27. Depending on the question, between 81 and 598 of the respondents used the free text option to insert comments.

Over 450 position papers were submitted through the EU Survey website, either in addition to questionnaire answers (over 400) or as stand-alone contributions (over 50). This brings the overall number of contributions to the consultation to over 1 250. Among the position papers, 72 came from non-governmental organisations (NGOs), 60 from business associations, 53 from large companies, 49 from academia, 24 from EU citizens, 21 from small and medium enterprises (SMEs), 19 from public authorities, 8 from trade unions, 6 from non-EU citizens, 2 from consumer organisations, with 94 not specified.

Main concerns

In the online survey, the overwhelming majority of participants (95%) responded to the section on the regulatory options for AI. Out of the concerns suggested in the White Paper, 90% and 87% of respondents found the possibility of AI breaching fundamental rights and the use of AI that may lead to discriminatory outcomes, respectively, as the most important ones. The possibility that AI endangers safety or takes actions that cannot be explained were also considered as (very) important by respectively 82% and 78% of respondents. Concerns over AI’s possible lack of accuracy (70%) and lack of compensations following harm caused by AI (68%) follow.

The most reoccurring out of 390 free text answers received for this question, highlighted the benefits of AI, to express the need of a balanced regulatory approach and the avoidance of ‘overregulation’ (48 comments). However, other comments add to the concerns related to AI. According to those, future regulation should pay attention to issues such as the transparency of decisions made by AI (32), the attribution of accountability for those decisions (13) as well as ensuring the capacity of human beings to making their own choices without being influenced by algorithms (human agency / human in the loop) (19). A number of non-governmental organisations underlined the need for a democratic oversight (11) while aspects such as equality (11), data quality (7), labour rights (5), safety (4) and others 5 were mentioned as well.

The importance of fundamental rights and other ethical issues was also underlined by many position papers. 42 position papers, 6 of which are arguing in favour of human rights impact assessments, mentioned the issue as one of their top three topics. Fundamental rights issues were mostly emphasized by NGOs (16), 5 of which were in- favour of introducing a human rights / fundamental rights impact assessment for AI. In addition, many respondents brought up other ethical issues such as discrimination and bias (21), the importance of societal impacts (18), data protection (15), civil society involvement (9) and human oversight (7).

What kind of legislation

In the relevant online survey question 6 , 42% of respondents found the introduction of a new regulatory framework on AI as the best way to address the concerns listed in the previous paragraph. Among the main arguments used by participants in 226 free text answers was that current legislation might have gaps when it comes to addressing issues related to AI and therefore a specific AI legislation is needed (47 comments). According to other comments such legislation should come along with appropriate research and gap analysis processes (39).

Other free text answers, however, highlighted that this process should take place with caution in order to avoid overregulation and the creation of regulatory burdens (24). 33% of participants to the online questionnaire thought that the gaps identified, could be addressed through the adaptation of current legislation, in a way that new provisions do not overlap with existing ones. Standardisation (17 comments) or the provision of guidelines (14 comments) were some alternative solutions mentioned in the free text answers 7 while others mentioned that there should be a regular review of existing legislation, accounting for technological change (2 comments). On the same topic, only 3% of participants in the online survey thought that current legislation is sufficient, while the rest declared to have other opinions (18%) or no opinion at all (4%).

Mandatory requirements

The vast majority of online respondents seemed to overwhelmingly agree with compulsory requirements introduced by the White Paper in the case of high-risk applications. Clear liability and safety rules were supported by 91% of respondents and were followed by information on the nature and purpose of an AI system (89%), robustness, and accuracy of AI systems (89%). Human oversight (85%), quality of training datasets (84%) and the keeping of records and data (83%) followed.

Figure 1: Agreement to introduce compulsory requirements in the case of high-risk applications (in %)

Source: online survey, multiple-choice questions

In the 221 free text answers received on this topic, 35 referred to other documents and standards (e.g. German Data Ethics Commission – mentioned in 6 comments) while 33 of them called for criteria that are more detailed and definitions that would allow the limitation of requirements to high-risk applications only. However, other comments did not support a simple distinction between ‘high’ and ‘low’ risk AI. Some partly coordinated responses (16) were in favour of an impact assessment on fundamental/human rights impact assessment while others supported that all AI applications should be regulated as such (16) or based on use cases (13). Like for the question above, comments repeated that requirements should be proportionate (8) and avoid overregulation or any kind of unnecessary burdens for companies (6). 8

In the position papers, the requirements were often not the main topics. When they were one of the major issues, the majority in favour of legislation was somewhat smaller. While many position papers did not mention regulatory requirements in their top three topics (54%), 23% generally agreed with the White Paper's approach to regulatory requirements for high-risk AI, while 12% generally disagreed. Some stakeholders also expressed other opinions (12%).    

Among the 12% of stakeholders who expressed another opinion (47 in total), some argued that no new AI requirements were needed (7), while others asked for additional requirements (e.g. on intellectual property or AI design) to be considered (7). Other comments highlighted that the requirements must not stifle innovation (6), or that they needed to be more clearly defined (3).

‘Human oversight’ was the most mentioned requirement (109 mentions), followed by ‘training data’ (97), ‘data and record keeping’ (94), ‘information provision’ (78) and ‘robustness and accuracy’ (66).

Many business associations (73%) and large companies (59%) took a stance on regulatory requirements, while the other stakeholder types, including SMEs, did not take a stance on the issue as often. In addition, business stakeholders tended to broadly agree with the Commission on the issue as presented on the AI White Paper (31%). Those who expressed other opinions mainly highlighted that new rules/requirements were not needed (3.7%), or that requirements should be proportionate (2.2%).

Only 39% of academic stakeholders mentioned regulatory requirements (19). When they did, they tended to be in favour of them (22%) or they expressed other opinions (10%). The positioning of NGOs was similar: while only 38% mentioned the regulatory requirements, those who did were also mostly in favour of them (21%).

High-risk applications

Concerning the scope of this new possible legislation, participants where asked on whether it should be limited to high-risk applications only. While 42.5% of online questionnaire respondents agreed that the introduction of new compulsory requirements should only be limited to high-risk AI applications, another 30.6% doubted such limitation. The remaining 20.5% had other opinions and 6.3% had no opinion at all. It is interesting to note that respondents from industry and business were more likely to agree with limiting new compulsory requirements to high-risk applications with a percentage of 54.6%.

However, several online respondents did not appear to have a clear opinion regarding what high-risk means: although 59% of respondents supported the definition of high-risk provided by the White Paper 9 , only 449 out of 1215 (37% of consultation participants) responded to this question.

In the 59 free text answers received, 10 found the definition provided in the White Paper unclear and asked for more details/criteria to be provided. Other comments found problematic the clause according to which ‘there may also be exceptional instances where, due to the risks at stake, the use of AI applications for certain purposes is to be considered as high-risk’ (7) while some suggested additional criteria for the definition of ‘high-risk’ (4). Coordinated responses (4) support existing (sectorial) definitions of ‘high and low risk’ while others (4) suggested the identification of high-risk application instead of sectors. For others, the classification of entire sectors as 'high-risk' could bring disadvantages and hamper innovation (3). Other comments focus on the importance of ‘legal certainty’ (5) which could be reduced by overly frequent reviews of high-risk sectors (3) 10 .

Consultation participants were also asked to indicate AI applications or uses which according to them can be considered as high-risk. The table below lists the top answers received:

Table 1: Other AI Applications and uses that can be considered as “high-risk” according to free text answers

TOP AI APPLICATIONS AND USES CONSIDERED AS “HIGH-RISK”

MIN. NO. OF MENTIONS

Applications related to autonomous weapons / defense sector

41

Remote biometric identification (e.g. facial recognition)

34

Applications in critical infrastructure (e.g. electricity, water supply, nuclear)

28

Reference to other documents/standards

25

Applications related to health

22

Applications in HR and employment

21

Applications analysing/manipulating human behaviour

18

Applications in predictive policing

18

Applications enabling mass surveillance

15

Applications used in political communication / disinformation

12

Applications related to security, law enforcement

12

The definition of ‘high-risk’ seemed to be the most important point of for stakeholders submitting position papers as well. A large number of papers commented that this definition was unclear or needed improvement (74 out of 408 stakeholders). Many believed that the simple distinction between high and low risk was too simplified and some proposed to introduce more levels of risk. Some believed that the definition was too broad, while others believed that it was too narrow.

In this context, some stakeholders proposed alternative approaches to defining 'high-risk' with more risk levels: some position papers (at least 6) suggested following a gradual approach with five risk levels, as proposed by the German Data Ethics Commission to create a differentiated scheme of risks. Other stakeholders (at least 5) suggested the adoption of risk matrixes, which combine the intensity of potential harm with the level of human implication/control in the AI decision. The probability of harm was another criterion for risk, repeatedly mentioned by stakeholders. 

Similarly, many position papers addressed the two-step approach proposed in the White Paper to determining ‘high-risk’ AI. At least 19 position papers considered the approach inadequate, at least 5 argued against the sectoral approach and many others put forth a diverse set of suggestions and criticism.

One notable suggestion for the risk assessment approach was to take into account all subjects affected by the AI application: multiple stakeholders argued that not only individual risks, but also collective risks should be considered, as there were also risks affecting society as a whole (e.g. with regards to democracy, environment, human rights). The impression that the definition of ‘high-risk’ needs to be clarified was shared by all stakeholder types.

The two-step risk assessment approach received most comments from business stakeholders. At least 5 business associations and large companies argued against the sectoral approach to determining high-risk and were supportive of a contextual assessment. On the contrary, two out of the three SMEs that mentioned the risk assessment approach expressly, supported the sectoral approach.

Remote biometric identification in public spaces

Online questionnaire respondents were concerned about the public use of such systems with 28% of them supporting a general ban of this technology in public spaces, while another 29.2% required a specific EU guideline or legislation before such systems may be used in public spaces. 15% agreed with allowing remote biometric identification systems in public spaces only in certain cases and under conditions and another 4.5% asked for further requirements (on top of the 6 requirements for high-risk applications proposed in the white paper) to regulate such conditions. Only, 6.2% of respondents did not think that any further guidelines or regulations are needed. 17.1% declared to have no opinion.

In the 257 free text answers received, participants mainly referred to the concerns that the use of biometric identification brings. According to 38 comments, remote biometric identification in public spaces endangers fundamental rights in general while other comments supported such concerns by referring to other documents, standards (30) or even the GDPR (20). 15 comments referred to the risk of mass surveillance and imbalances of power that the use of such technology may bring, 13 others referred to privacy while 10 more mentioned that biometric identification endangers the freedom of assembly/expression. However, there were also 13 comments referring to possible benefits coming from the use of this technology while 13 more mentioned that the use of remote biometric identification in public spaces should be allowed for specific purposes only, e.g. security, criminal or justice matters. Other 8 comments stressed that there are uses of facial recognition that do not pose 'high risks' or endanger fundamental rights and the introduction of guidelines could be beneficial for the correct use of the technology (8). The management of such systems by qualified staff could, according to 7 more comments, guaranty human oversight in its use. 11

Among the position papers, a part of the stakeholders specifically mentioned remote biometric identification in public spaces (96) as one of their top three topics. Of these, a few argued for a ban of remote biometric identification in public spaces (19), and 7 respondents for a moratorium. A few more were in favour of conditioning its use to tight regulation and adequate safeguards in public spaces (19). Almost half of the stakeholders who positioned themselves in favour of a ban of biometric identification in public spaces were NGOs. This contrasts with the 34 business stakeholders who mentioned biometric identification, among which only one was in favour of a ban. A moratorium for remote biometric identification in public spaces was also mentioned by academic stakeholders: four research institutions were in favour of a moratorium of biometric identification until clear and safe guidelines were issued by the EU.

Enforcement and voluntary labelling

To make sure that AI is trustworthy, secure and in respect of European values, the White Paper suggested conformity assessment mechanisms for high-risk applications. The public consultation proposed several options to ensure that AI is trustworthy, secure and in respect of European values. 62% of online survey respondents supported a combination of ex-post and ex-ante market surveillance systems. 3% of respondents supported only ex-post market surveillance. 28% supported external conformity assessment of high-risk applications. 21% of respondents supported ex-ante self-assessment.

To the options above, respondents added further ones through 118 free text answers. Among those, 19 suggested an (ex-ante) assessment of fundamental rights while 14 comments were in favour of self-assessment and 11 more suggested that independent external bodies/experts should ensure assessment. There were also 8 comments supporting that existing assessment processes are sufficient while 8 others where against ex-ante assessment as that might be a burden for innovation.

Voluntary labelling systems could be used for AI applications that are not considered of high-risk. 50.5% of online respondents found it useful or very useful, while another 34% did not agree with the usefulness of such system. 15.5% of respondents declared that they did not have an opinion on the matter.

Still, in 301 free text answers, 24 comments appeared to be generally in favour of voluntary labelling, 6 more supported self-assessment and 46 more made reference to other documents and existing international standards that could be used as an example for such practices (e.g. the AI HLEG’s assessment list, the energy efficiency label 12 or the EEE EPPC and IEEE-SA). According to 6 comments, labelling systems need to be clear and simple while 18 comments stressed that clearer definitions and details are needed. Other 16 comments called for the involvement of stakeholders in development of labelling systems or (7 more comments) the appointment of an independent body should be responsible for the voluntary labelling. The importance of enforcement and control of voluntary labelling was stressed by 12 more comments and a harmonised EU-wide approach was suggested by 5 others. Moreover, 8 comments mentioned that systems need to be flexible to adapt to technological changes.

However, 27 replies seemed to be sceptical towards voluntary labelling systems in general and 25 more towards self-labelling/self-regulation in particular. Some of these comments mentioned that such systems may be used according to the interest of companies, according to 16 more, it is likely that such systems favour bigger players who can afford it while 23 more stressed it imposes costs that can hamper innovation for smaller ones. Moreover, 12 comments mentioned the issue of labelling for 'low risk' categories, which can create a false sense of risks, others 7 comments mentioned that the distinction among low and high risk is too simplified while 5 more said that they can create a false sense of security. 13  

52 position papers addressed the proposed voluntary labelling scheme as one of their top three topics. 21 of them were sceptical of labelling, either because they believed that it would impose regulatory burdens (especially for SMEs) or because they were sceptical of its effectiveness. Some stakeholders argued that such a scheme was likely to confuse consumers instead of building trust. On the other hand, 8 position papers were explicitly in favour, and many other stakeholders provided a diverse set of comments.

The voluntary labelling scheme received most comments through position papers submitted by business stakeholders: most of business associations (11) and SMEs (3) were sceptical of the idea, due to the costs it could impose on them or a suspected lack of effectiveness. The position of large companies mentioning voluntary labelling was quite the opposite: most tended to be in favour of it (4).

Safety and liability implications of AI, IoT and robotics

The overall objective of the safety and liability legal frameworks is to ensure that all products and services, including those integrating emerging digital technologies, operate safely, reliably and consistently, and that damage that has already occurred is remedied efficiently.

60.7% of online respondents supported a revision of the existing Product Liability directive to cover particular risks engendered by certain AI applications. 63 % of respondents supported that national liability rules should also be adapted for all AI applications (47 %) or specific AI applications (16 %) to better ensure a proper compensation in case of damage, and a fair allocation of liability. Amongst those businesses that took a position on this question (i.e. excluding ‘no opinion’ responses), there is equally clear support for such adaptations, especially amongst SMEs (81 %).

Among the particular AI related risks to be covered, online respondents prioritised cyber risks with 78% and personal security risks with 77%. Mental health risks followed with 48% of respondents flagging them, and then risks related to the loss of connectivity, flagged by 40% of respondents. Moreover, 70% of participants supported that the safety legislative framework should consider a risk assessment procedure for products subject to important changes during their lifetime.

In 163 free text answers, 23 respondents added to the risks those of discrimination/manipulation, which according to 9 others can be caused by profiling practices or automated decision making (5 comments), while 14 more (mainly NGOs) focused on the particular discrimination risk linked to online advertisement. This can also be related to another set of comments (14 in total) according to which, such risks may cause differentiated pricing, financial detriments, filter bubbles or interference in political processes (other 2 comments mentioning the risks of disinformation can be relevant here as well). Risks to personal data (11 comments), or those deriving from cyber-attacks (7 comments), risks for people with disabilities (10 comments) as well as general health risks (8 comments) were among other risks mentioned. 14 For the specific risks deriving from cyber security and connectivity loss in the automotive sector, a coordinated response of four carmakers, noted that other regulations tackle them already.

In the 173 free text answers regarding the risk assessment procedures of the safety and liability framework, as pointed by 11 comments ‘AI systems change over time’. Therefore, 16 comments mention that risk assessments need to be repeated in case of changes (after placement on the market). To the same regard, 13 comments pointed that clearer definitions of e.g. ‘important changes’ should be given during that process and 11 others that risk assessment should only be required in case of a significant change to a product (partly coordinated response).

According to 12 comments, assessment procedures could build up on the existing GDPR Impact Assessment or even involve GDPR and data protection officers (coordinated response of 10 stakeholders). 15

52 position papers addressed issues of liability as one of their top three topics, most of them providing a diverse set of comments. 8 believed that existing rules were probably sufficient and 6 were sceptical of a strict liability scheme. Those who were sceptical often argued that a strict liability scheme was likely to stifle investment and innovation, and that soft measures like codes of conduct or guidance documents were more advisable. At the same time, other contributions to the public consultation from the entire range of stakeholders expressed support for a risk-based approach also with respect to liability for AI, and suggested that not only the producer, but also other parties should be liable. Representatives of consumer interests stressed the need for a reversal of the burden of proof.

When it comes to liability, some business associations and large companies thought that existing rules were probably already sufficient (7) or they were sceptical of strict liability rules and possible regulatory burdens (5). Almost none of the other stakeholder types shared this position. A few businesses submitted position papers in favour of harmonising liability rules for AI.

Other issues raised in the position papers

The position papers submitted also raised some issues that were not part of the questionnaire.

How to define artificial intelligence? (position papers only)

As the White Paper does not contain its own explicit definition of AI, this analysis of the position papers took the definition of the HLEG on AI as a reference point. The HLEG definition of AI includes systems that use symbolic rules or machine learning, but it does not explicitly include simpler Automatic Decision Making (ADM) systems.

Position papers were analysed to determine whether and why stakeholders shared or did not share this definition or have other interesting comments on the definition of AI.

The majority of position papers made no mention of the definition of AI (up to 70%, or 286 out of 408) among their top three topics. A majority of 15.7% had a different definition than the one suggested by the HLEG (64). 9.3% found the definition was too broad (37), out of which 2.7% said that AI should only include machine learning (11). Stakeholders highlighted that a too broad definition risks leading to overregulation and legal uncertainty, and was not specific enough to AI. Another 6.6% believed that the definition was too narrow (27), with 3.7% saying that it should also include automated decision-making systems (15). Stakeholders highlighted that the definition needed to be future proof: if it was too narrow, it risks disregarding future aspects of next-generation AI.

2.7% of stakeholders agreed with the AI HLEG definition of AI (11) but 5.4% of position papers stated that the AI HLEG’s definition is unclear and needs to be refined (22). To improve the definition, stakeholders propose, for example: to clarify to what extent the definition covers traditional software; to distinguish between different types of AI; or to look at existing AI definitions made by public and private organisations. Finally, 2.2% of stakeholders provided their own definition of AI (9). 

The majority of business stakeholders believed that the AI HLEG’s definition was too broad. This trend was strongest for business associations. On the contrary, the majority of academic and NGO stakeholders believed that the HLEG's definition is too narrow.

At least 24 business stakeholders believed that the definition was too broad, while only 5 believed that it was too narrow and only 4 agreed with it. Business stakeholders were also relatively numerous in saying that the definition is unclear or needs to be refined (at least 11). The majority of academic and NGO stakeholders believed that the AI HLEG's definition was too narrow (6 and 8) and only 1 academic and 4 NGO stakeholders believed that the definition was too broad.

Costs - What costs could AI regulation create? (Position papers only.)

Costs imposed by new regulations are always a contentious topic. Some see costs imposed by regulation as an unnecessary burden to competitiveness and innovation; others see costs as a necessary by-product of making organisations comply with political, economic or ethical objectives.

In order to better understand stakeholder's perspective on the costs of AI regulation, position papers were analysed for mentions of two main types of costs: (1) compliance costs, generally defined as any operational or capital expense faced by a company to comply with a regulatory requirement; and (2) administrative burdens, a subset of compliance costs, covering 'red tape' such as obligations to provide or store information.

84% of stakeholders do not explicitly mention costs that could be imposed by a regulation on AI as one of the top three topics (344). 11% of stakeholders (46) mention compliance costs in general and 7% of stakeholders (29) (also) mention administrative burdens in particular. It must be noted that some stakeholders mentioned both types of costs.

Some stakeholders warned against the costs incurred by a mandatory conformity assessment, especially for SMEs or companies operating on international markets. Some highlighted that certain sectors were already subject to strict ex-ante conformity controls (e.g. automotive sector) and warned against the danger of legislative duplication. Several stakeholders also saw a strict liability regime as a potential regulatory burden and some noted that a stricter regime could lead to higher insurance premiums.

Some respondents also put forth other arguments related to costs, such as the potential cost saving effects of AI, the concept of 'regulatory sandboxes' as a means to reduce regulatory costs, or the environmental costs created by AI due to high energy consumption.

17% of all types of business stakeholders mentioned compliance costs and 13% (also) mentioned administrative burdens, while up to 74% of business stakeholders did not explicitly mention costs among their top three topics. Among business stakeholders, business associations are the ones that mentioned costs the most. Out of all mentions of costs from all stakeholders (75 in total), 56% came from business stakeholders (42).

Academic stakeholders also mentioned costs more often than other types of stakeholders, but also not very often overall. 13% of academic stakeholders mentioned compliance costs and 9% (also) mentioned administrative burdens, while 82% did not explicitly mention costs in their top three topics. Other stakeholders mentioned costs more rarely.

Governance - Which institutions could oversee AI governance? (Position papers only)

The institutional structure of AI governance is a key challenge for the European regulatory response to AI. Should AI governance, for example, be centralised in a new EU agency, or should it be decentralised in existing national authorities, or something in between? In order to better understand this issue, the position papers were analysed regarding their position on the European institutional governance of AI.

Most stakeholders (up to 77% or 314) did not address the institutional governance of AI.

Among the 23% of position papers who did address this issue in their top three topics, 10% of stakeholders were in favour a new EU-level institution, with 6% of stakeholders being in favour of some form of a new EU AI agency (24) and 4% in favour of a less formalised EU committee/board (15). At the same time, at least 3% of stakeholders were against establishing a new institution (14): they argued that creating an additional layer of AI-specific regulators could be counterproductive, and they advocated for a thorough review of existing regulation frameworks, e.g. lessons learned from data protection authorities dealing with GDPR, before creating a new AI-specific institution/body.

1% of stakeholders were in favour of governance through national institutions (6) and another 1% of stakeholders were in favour of governance through existing competent authorities (5) (without specifying whether these would be on the EU or national level). In addition, stakeholders also mentioned other ideas, such as the importance of cooperation between national and/or EU bodies (7); multi-stakeholder governance involving civil society and private actors (6); or sectorial governance (4).

While only 32% of academic stakeholders mention the issue in their position papers among the top three topics, they tended to be in favour of an EU AI agency (10%), but many provided a diverse set of other arguments. 24% of large companies and business associations provided a position on the issue while SMEs practically did not mention it. All business stakeholders tended to be more sceptical of formal institutionalisation: 8% of business associations and 4% of large companies are against a new institution, 5% of associations and 2% of large companies are in favour of a less formalised committee/board, and the others share other more specific positions.

Most trade unions and EU or non-EU citizens did not have a position on the issue, but if they did, the majority was in favour of an EU AI agency (25% of trade unions and 17% of EU and non-EU citizens). However, it must be noted that these percentages are very volatile due to the low number of respondents with a position on the issue.

2.2. Analysis of the results of the feedback from the inception impact assessment

The Inception Impact Assessment elicited 132 contributions from 130 different stakeholders – two organizations commented twice – from 24 countries all over the world. 89 respondents out of 130 had already answered the White Paper consultation.

Table 2: Participating Stakeholders (by type)

STAKEHOLDER TYPE

NUMBER

Business Association

55

Company/Business Organization

28

NGO

15

EU citizen

7

Academic/Research Institution

7

Other

6

Consumer Organization

5

Trade Union

4

Public Authority

3

Table 3: Participating Stakeholders
(by country)

COUNTRY

NUMBER

COUNTRY

NUMBER

Belgium

49

Finland

2

Germany

17

Hungary

2

US

11

Poland

2

Netherlands

8

Portugal

2

UK

8

Sweden

2

France

6

Bulgaria

1

Ireland

3

Czech Republic

1

Italy

3

Estonia

1

Spain

3

Japan

1

Austria

2

Lithuania

1

Denmark

2

Summary of feedback

Stakeholder mostly requested a narrow, clear and precise definition for AI. Stakeholders also highlighted that besides the clarification of the term of AI, it is important to define ‘risk’, ‘high-risk’, ‘low-risk’, ‘remote biometric identification’ and ‘harm’.

Some of the stakeholders caution the European Commission not to significantly expand the scope of future AI regulation to ADM, because if AI were to be defined as ADM, it would create regulatory obligations that hamper development.

Several stakeholders warn the European Commission to avoid duplication, conflicting obligations and overregulation. Before introducing new legislation, it would be crucial to clarify legislative gaps, to adjust the existing framework, focus on effective enforcement and adopt additional regulation only where necessary. It is essential to review EU legislation in other areas that are potentially applicable to AI and make them fit for AI. Before choosing any of the listed options, existing regulation needs to be carefully analysed and potential gaps precisely formulated.

There were many comments underlining the importance of a technology neutral and proportionate regulatory framework.

Regulatory sandboxes could be very useful and are welcomed by stakeholders, especially from the Business Association sector.

Most of the respondents are explicitly in favour of the risk-based approach. Using a risk-based framework is a better option than blanket regulation of all AI applications. The types of risks and threats should be based on a sector-by-sector and case-by-case approach. Risks also should be calculated taking into account the impact on rights and safety.

Only a few respondents agreed that there is no need for new regulation for AI technologies: option 0 “baseline”. Less than 5% of the stakeholders supported option 0.

There was a clear agreement among those stakeholders who reflected on option 1 that either per se or in combination with other options, ‘soft law’ would be the best start. Around one third of the stakeholders commented option 1 and more than 80% of them were in favour of it. Most of the supportive comments arrived from the business association (more than 75%) and company/business sector.

Option 2 ‘voluntary labelling system’ per se was not supported, since it seems to be premature, inefficient and ineffective. More than one third of the stakeholders had a view on the voluntary labelling system of which nearly 75% disagreed with option 2. It is argued mostly by the business association, company/business and NGO sectors that voluntary labelling could create heavy administrative burden and would only be useful if it is flexible, robust and clearly articulated. If the Commission would decide to introduce voluntary certification, it should be carefully addressed as it can result in a meaningless label and even increase non-compliant behaviour when there are no proper verification mechanisms.

The three sub-options 3a (legislation for specific AI applications), 3b (horizontal framework for high-risk AI applications) and 3c (horizontal framework for all AI applications) were commented on by more than 50% of the respondents. There is a majority view – more than 90% of the stakeholders who reflected on this question – that if legislation is necessary, the EU legislative instrument should be limited to ‘high-risk’ AI applications based on the feedback mostly of business associations, companies and NGOs. Legislation limited only to specific applications could leave some risky application out of the regulatory framework.

The combination of different options was a very popular choice, nearly one third of the respondents supported option 4 ‘combination of any of the options above’. Most variations included option 1 ‘soft law’. The most favoured combination with nearly 40% was option 1 ‘soft law’ with sub-option 3b ‘high-risk applications’, sometimes with sub-option 3a ‘specific applications’. Mainly business associations and companies supported this combination. Especially NGOs, EU citizens and others preferred the combination of option 2 ‘voluntary labelling’ and sub-option 3b. In small numbers, option 1, option 2 and sub-option 3b, the combinations of option 2, sub-option 3a and/or sub-option 3b were also preferred. Option 1 and sub-option 3b are often viewed favourably per se or in combination also. Sub-option 3c was not popular at all.

Among those who formulated their opinion on the enforcement models, more than 50%, especially from the business association sector were in favour of the combination of ex-ante risk self-assessment and ex-post enforcement for high-risk AI applications.

In case of an ex-ante enforcement model, there are many respondents who caution against third party ex-ante assessments and instead recommend self-assessment procedures based on clear due diligence guidance. New ex-ante conformity assessments could cause significant delays in releasing AI products has to be taken into account. Ex-ante enforcement mechanisms without any background are causing a lot of uncertainty. Ex-ante conformity assessments could be disproportionate for certain applications.

If ex-post enforcement would be chosen, it should be used with the exception of sectors where ex-ante regulation is a well-established practice. Ex-post enforcement should only be implemented in a manner that complements well against ex-ante approaches.

2.3. Stakeholder outreach

The following consultation activities (in addition to the open public consultation and the Inception Impact Assessment feedback) were organised:

2.3.1.    Event on the White Paper with larger public

In addition to the public consultations, the Commission also consulted stakeholders directly. On 9 October 2020, it organised the Second European AI Alliance Assembly with more than 1 900 participants across different stakeholder groups, where the issues addressed in the impact assessment were intensely discussed. The topical workshops held during the event on the main aspects of the AI legislative approach included biometric identification, AI and liability, requirements for Trustworthy AI, AI Conformity assessment, standards and high-risk AI applications. The AI Alliance is a multi-stakeholder forum launched in June 2018 in the framework of the European Strategy on Artificial Intelligence. During the conference, participants could interact with the different panels, which were made up of diverse stakeholders, through Sli.do. Overall, there were 647 joined participants, with 505 active participants. 338 questions were asked, and attracted over 900 likes among themselves. Over the course of the day, over 1 000 poll votes were cast.

2.3.2.    Technical consultations

The Commission organised five online workshops with experts from different stakeholder groups:

·online workshop on conformity assessment on 17 July 2020 with 26 participants from the applying industry, civil society and conformity assessment community;

·online workshop on biometrics on 3 September 2020 with 17 external participants from stakeholders such as the Fundamental Rights Agency, the World Economic Forum, the French Commission Nationale de l'Informatique et des Libertés and academia;

·online workshops on standardisation on 29 September 2020 with 27 external participants from UNESCO, OECD, Council of Europe, CEN-CENELEC, ETSI, ISO/IEC, IEEE, ITU;

·online workshop on potential requirements on 9 October 2020 with 15 external experts on AI, mainly from academia;

·online workshop on children’s rights and AI on 12 November 2020 with external experts.

·AI expert group for home affairs on surveillance technologies and data management by law enforcement on 17 December 2020.

In addition, the contractor for the analytical study organised two online workshops with experts as follows:

·online validation workshop on costs assessment on 28 September 2020 with 40 external experts;

·online validation workshop conformity assessment on 7 October 2020 with 25 external experts.

The Commission services also participated in many seminars (more than 50) and held a numerous meetings with a large variety of stakeholders from all groups.

2.3.3.Outreach and awareness raising events in Member States and International outreach

Due to the coronavirus, the planned outreach activities in Member States had to move online and were fewer than initially planned. Nevertheless, Commission services discussed the approach in meetings with large numbers of stakeholders in several Member States, including France, Germany and Italy. They also exchanged views with international bodies, in particular the Council of Europe, the G8 and G20 as well as the OECD. The EU approach was discussed in bilateral meetings with a number of third countries, for example Japan and Canada.

2.3.4.European AI Alliance platform

The Commission also used the European AI Alliance, launched in June 2018, which is a multi-stakeholder online platform intended for broad engagement with academia, industry and civil society to discuss the European AI and gather input and feedback. The European AI Alliance has more than 3 700 members representing a wide range of fields and organisations (public authorities, international organisations, consumer organisations, industry actors, consultancies, professional associations, NGOs, academia, think tanks, trade unions, and financial institutions). All Member States are represented, as well as non-EU countries.

3.Annex 3: Who is affected and how?

3.1. Practical implications of the initiative

3.1.1. Economic operators/business

This category comprises developers of AI applications, providers that put AI applications on the European market and operators/users of AI applications that constitute a particularly high risk for the safety or fundamental rights of citizens. The initiative applies to AI systems operated or used in Europe and the respective operators, independent of whether they are based in Europe or not. According to their respective role in the AI life-cycle they would all have to comply with clear and predictable obligations for taking measures with a view to preventing, mitigating and monitoring risks and ensuring safety and respect of fundamental rights throughout the whole AI lifecycle. Before placing their product on the market, providers in particular will have to ensure that the high-risk AI systems comply with essential requirements, addressing more specifically the underlying causes of risks to fundamental rights and safety (such as requirements relating to data, traceability and documentation, transparency of AI systems and information to be provided, robustness and accuracy and human oversight). They will also have to put in place appropriate quality management and risk management systems, including to identify and minimise risks and test the AI system ex ante for its compliance with the requirements and relevant Union legislation on fundamental rights (e.g. non-discrimination).

Once the system has been placed on the market, providers of high-risk AI systems would be obliged to continuously monitor, manage and mitigate any residual risks, including reporting to the competent authorities incidents and breaches of fundamental rights obligations under existing Union and Member States law.

Where feasible, the requirements and obligations will be operationalised by means of harmonized standards that may cover the process and the requirements (general or specific to the use case of the AI system). This will help providers of high-risk AI systems to reach and demonstrate compliance with the requirements and improve consistency.

In addition, for a subset of the high-risk applications (safety components of products and remote biometric identification in publicly accessible spaces), companies would have to submit their applications to third-party ex-ante conformity assessment bodies before being able to place them on the market. When harmonized standards exist and the providers apply those standards, they would not be required to undergo an ex-ante third party conformity assessment; this option would be applicable to safety components depending on the relevant sectoral safety rules for conformity assessment. For all other high-risk applications, the assessment would be carried out via an ex ante conformity assessment though internal checks.

For non-high risk AI systems, the instrument will impose minimal requirements and obligations for increased transparency in two limited cases: obligation to disclose that the human is interacting with an AI system and label deep fakes when not used for legitimate purposes.

The initiative will give rise to new compliance costs. Apart from authorisation and on-going supervisory costs, developers, providers and operators will need to implement a range of operational changes. The individual costs arising from this will largely depend on the extent to which respective AI developers, providers and operators have already implemented measures on a voluntary basis. An EU regulatory framework, however, avoids the proliferation of nationally fragmented regimes. It will thus provide AI system developers, operators and providers with the opportunity to offer services cross-border throughout the EU without incurring additional compliance costs. As the initiative pre-empts the creation of national regimes in many Member States, there can be a significant indirect cost saving in this regard for cross-border operations. Concerning AI system developers, the initiative aims to facilitate competition on a fair basis by creating a regulatory level playing field. It will also help to strengthen consumer and investor trust and should thereby generate additional revenue for AI systems developers, providers and operators.

3.1.2.Conformity assessment, standardisation and other public bodies

Standardisation bodies will be required to develop standards in the field.

Conformity assessment bodies would have to establish or adapt conformity assessment procedures for the products covered. In case of third-party conformity assessment they also would have to carry them out.

Member States would have to equip competent national authorities (e.g. market surveillance bodies etc.) adequately to supervise the enforcement of the requirements, including the supervision of the conformity assessment procedures and also the ex-post market monitoring and supervision. The ex-post system will monitor the market and investigate compliance with the obligations and requirements for all high-risk AI systems already placed on the market and used in order to effectively enforce the existing rules and sanction non-compliance.

Authorities will also have to participate in meetings as part of a coordination mechanism at EU level to provide uniform guidance about the interpretation of the new rules and consistency.

The Commission will also encourage voluntary compliance with codes of conduct developed by industry and other associations.

Supervisors will face a range of new tasks and supervisory obligations stemming from the framework. This has cost implications, both as concerns one-off investments and ongoing operational costs. Supervisors will need to invest in particular in new monitoring systems and ensure a firm enforcement of regulatory provisions. They will also need to train staff to ensure sufficient knowledge of these newly regulated markets and employ additional employees to stem the additional work. The costs for specific national authorities depends on (1) the number of AI applications monitored, and (2) the extent to which other monitoring systems are already in place.

3.1.3. Individuals/citizens

Citizens will benefit from an increased level of safety and fundamental rights protection and higher market integrity. The mandatory information and transparency requirements for high-risk AI systems and enforcement rules will enable citizens to make more informed decisions in a safer market environment. They will be better protected from possible activities that might be contrary to the EU fundamental rights or safety standards. In summary, citizens will carry lower risks, given the European regulatory approach. It can however not be excluded, that some of the compliance costs will be passed on to the citizens.

3.1.4.Researchers

There will be a boost for research, since some of the requirements (such as those related to robustness) will require continuous research and testing of products.

3.2. Summary of costs and benefits

Table 4: Overview of Benefits (total for all provisions) – Preferred Option

DESCRIPTION

AMOUNT

COMMENTS

Direct benefits

Fewer risks to safety and fundamental rights

Not quantifiable

Citizens

Higher trust and legal certainty in AI

Not directly quantifiable

Businesses

Indirect benefits

Higher uptake

Not directly quantifiable

Businesses

More beneficial applications

Not quantifiable

Citizens

Not quantifiable: impossible to calculate (e.g. economic value of avoiding fundamental rights infringements)

Not directly quantifiable: could in theory be calculated if many more data were available (or making large numbers of assumptions)

Table 5: Overview of costs – Preferred option

CITIZENS/
CONSUMERS

BUSINESSES

ADMINISTRATIONS

One-off

Recurrent

One-off

Recurrent

One-off

Recurrent

Comply with substantial require-ments

Direct costs

€ 6000 – 7000 per application

€ 5000 – 8 000 per application

Indirect costs

Verify
compliance

Direct costs

€ 3000 – 7500 per application

Indirect costs

Audit QMS

€1000 – 2000 per day, depending on complexity

Renew audit, €300 per hour, depending on complexity

Establish competent authorities

Direct costs

1-25 FTE per MS; 5 FTE at EU

Indirect costs



4.Annex 4: Analytical methods

Summary of the elements of the compliance costs and administrative burden

This annex summarises the key elements of the compliance costs and administrative burdens for enterprises, based on chapter 4 “Assessment of the compliance costs generated by the proposed regulation on Artificial Intelligence” of the Study to Support an Impact Assessment of Regulatory Requirements for Artificial Intelligence in Europe. 16  

The cost assessment achieved by the consultant relies on the Standard Cost Model, a widely known methodology to assess administrative burdens. It has been adopted by several countries around the world, including almost all EU Member States and the European Commission in its Better Regulation Toolbox.

A specific version of the model is used in this case: it features standardised tables with time estimates per administrative activity and level of complexity. The cost estimation is built on time expenditure for activities induced by the new requirements under the proposed regulation. The assessment is based on cost estimates of an average AI unit of an average firm, estimated to cost around USD 200,000 or EUR 170,000 17 .

The costs assessed here refer to two kinds of direct compliance costs:

Substantive compliance costs, which encompass those investments and expenses faced by businesses and citizens to comply with substantive obligations or requirements contained in a legal rule. These costs are calculated as a sum of capital costs, financial costs and operating costs. 

Administrative burdens are those costs borne by businesses, citizens, civil society organisations and public authorities as a result of administrative activities performed to comply with the information obligations (IOs) included in legal rules.

The approach broadly corresponds to the methodology adopted by the German government and developed with the Federal Statistical Office (Destatis). The table below shows a correspondence table used for the cost assessment in this document that allocate specific times to specific activities, differentiating each activity in terms of complexity levels.

Reference table for the assessment of compliance costs

Source: Consultant’s elaboration based on Normenkotrollrat (2018)

The translation of activities into cost estimates was obtained by using a reference hourly wage rate of EUR 32, which is the average value indicated by Eurostat for the Information and Communication sector (Sector J in the NACE rev 2 classification) 18 .

Two workshops were organised to discuss the cost estimates, one with businesses and one with accreditation bodies and standardisation organisations were invited to another workshop to discuss the team’s estimates on conformity costs.

Compliance costs regarding data

This requirement, as defined in the White Paper (pp.18-19), includes the following main activities:

Providing reasonable assurances that the use of the products or services enabled by the AI system is safe (e.g. ensuring that AI systems are trained on datasets that are sufficiently broad and representative of the European context to cover all relevant scenarios needed to avoid dangerous situations).

Take reasonable measures to ensure that the use of the AI system does not lead to outcomes entailing prohibited discrimination, e.g. obligation to use sufficiently representative datasets, especially to ensure that all relevant dimensions of gender, ethnicity and other possible grounds of prohibited discrimination are appropriately reflected.

Ensuring that privacy and personal data are adequately protected during the use of AI-enabled products and services. For issues falling within their respective scope, the GDPR and the Law Enforcement Directive regulate these matters.

Thus, the types of activities that would be triggered by this requirement include, among others:

·familiarising with the information obligation (one-off);

·assessment of data availability (this may require an internal meeting);

·risk assessment (this may require an internal meeting);

·testing for various possible risks, including safety-related and fundamental rights-related risks, to then adopt and document proportionate mitigating measures;

·anonymisation of datasets, or reliance on synthetic datasets; or implementation of data minimisation obligations;

·collecting sufficiently broad datasets to avoid discrimination.

For an average process, and a normally efficient firm, a reasonable cost estate for this activity is €2763. 19

Administrative burden regarding documents and traceability

This requirement aims to enable the verification and enforcement of compliance with existing rules. The information to be kept relates to the programming of the algorithm, the data used to train high-risk AI systems, and, in certain cases, keeping the data themselves. The White Paper (p. 19) prescribes the following actions:

Keeping accurate records of the dataset used to train and test the AI system, including a description of the main characteristics and how the dataset was selected;

Keeping the datasets themselves;

Keeping documentation on programming and training methodologies, processes and techniques used to build, test and validate the AI system;

Keeping documentation on the functioning of the validated AI system, describing its capabilities and limitations, expected accuracy/error margin, the potential ‘side effects’ and risks to safety and fundamental rights, the required human oversight procedures and any user information and installation instructions;

Make the records, documentation and, where relevant, datasets available on request, in particular for testing or inspection by competent authorities.

Ensure that confidential information is protected (e.g. trade secrets).

As a result, this obligation requires a well-trained data officer with the necessary legal knowledge to manage data and records and ensure compliance. The cost could be shared among different products and the data officer could have other functions, too. For an average process, and an efficient firm, a reasonable cost estimate per AI product for this activity is €4 390. 20

Administrative burden regarding provision of information

Beyond the record-keeping requirements, adequate information is required on the use of high-risk AI systems. According to the White Paper (p. 20), the following requirements could be considered:

Ensuring clear information is provided on the AI system’s capabilities and limitations, in particular the purpose for which it is intended, the conditions under which it can be expected to function as intended, and the expected level of accuracy in achieving the specified purpose. This information is especially important for deployers of the systems, but it may also be relevant to competent authorities and affected parties.

Making it clear to citizens when they are interacting with an AI system and not a human being.

Hence, the types of activities that would be triggered by this requirement include:

Provide information on the AI system’s characteristics, such as 

oIdentity and contact details of the provider;

oPurpose and key assumptions/inputs to the system;

oWhat the model is designed to optimise for, and the weight according to the different parameters;

oSystem capabilities and limitations;

oContext and the conditions under which the AI system can be expected to function as intended and the expected level of accuracy/margin of error, fairness, robustness and safety in achieving the intended purpose(s);

oPotential ‘side effects’ and safety/fundamental rights risks;

oSpecific conditions and instructions on how to operate the AI system, including information about the required level of human oversight.

Provide information on whether an AI system is used for interaction with humans (unless immediately apparent).

Provide information on whether the system is used as part of a decision-making process that significantly affects the person.

Design AI systems in a transparent and explainable way.

Respond to information queries to ensure sufficient post-purchase customer care. This activity was stressed by stakeholders with experience in GDPR compliance.

Given the overlaps with activities foreseen under other requirements, only the familiarisation with the specific information obligations and their compliance has been computed, rather than the cost of the underlying activities. However, it is worth noting that this requirement may also entail changes in the design of the system to enable explainability and transparency.

For an average process, and a normally efficient firm, a reasonable cost estimate for this activity is €3 627. 21

Compliance costs regarding human oversight

The White Paper acknowledges that the type and degree of human oversight may vary from one AI system to another (European Commission, 2020a, p.21). It will depend, in particular, on the intended use of the AI system and the effects of that use on affected citizens and legal entities. For instance:

Output of the AI system does not become effective unless it has been previously reviewed and validated by a human (e.g. the rejection of an application for social security benefits may be taken by a human only).

Output of the AI system becomes immediately effective, but human intervention is ensured afterwards (e.g. the rejection of an application for a credit card may be processed by an AI system, but human review must be possible afterwards).

Monitoring of the AI system while in operation and the ability to intervene in real time and deactivate (e.g. a stop button or procedure is available in a driverless car when a human determines that car operation is not safe).

In the design phase, by imposing operational constraints on the AI system (e.g. a driverless car shall stop operating in certain conditions of low visibility when sensors may become less reliable, or shall maintain a certain distance from the vehicle ahead in any given condition).

Therefore, the possible activities involved in compliance with this requirement are the following, based on the questions of the Assessment List for Trustworthy Artificial Intelligence 22 developed by the High-Level Expert Group on Artificial Intelligence:

·monitoring the operation of the AI system, including detection of anomalies, dysfunctions, and unexpected behaviour;

·ensuring timely human intervention, such as a “stop” button or procedure to safely interrupt the running of the AI system;

·conducting revisions in the design and functioning of the currently deployed AI system as well as implementing measures to prevent and mitigate automation bias on the side of the users;

·overseeing overall activities of the AI system (including its broader economic, societal, legal and ethical impact);

·implementing additional hardware/software/systems assisting staff in the above-mentioned tasks to ensure meaningful human oversight over the entire AI system life cycle;

·implementing additional hardware/software/systems to meaningfully explain to users that a decision, content, advice or outcome is the result of an algorithmic decision, and to avoid that end-users over-rely on the AI system.

This leads to a total estimate of €7 764. 23

Compliance costs regarding robustness and accuracy

According to the White Paper on Artificial Intelligence (European Commission, 2020a, p. 20), ‘AI systems must be technically robust and accurate if they are to be trustworthy. These systems, therefore, need to be developed in a responsible manner and with ex ante due and proper consideration of the risks they may generate. Their development and functioning must be such to ensure that AI systems behave reliably as intended. All reasonable measures should be taken to minimise the risk of harm.’ Accordingly, the following elements could be considered:

Requirements ensuring that the AI systems are robust and accurate, or at least correctly reflect their level of accuracy, during all lifecycle phases;

Requirements ensuring that outcomes can be reproduced;

Requirements ensuring that AI systems can adequately deal with errors or inconsistencies during all lifecycle phases;

Requirements ensuring that AI systems are resilient against overt attacks and against more subtle attempts to manipulate data or algorithms, and that mitigating measures are taken in such cases.

Compliance with this requirement entails technical and organizational measures tailored to the intended use of the AI system, to be assessed since the design phase of an AI system and throughout the moment in which the system is released on the market. It includes measures to prevent and mitigate automation bias, particularly for AI systems used to provide assistance to humans; and measures to detect and safely interrupt anomalies, dysfunctions, unexpected behaviour.

For every single AI product the following activities are envisaged:

1 | On accuracy:

·familiarising oneself with accuracy requirements;

·calculating an established accuracy metric for the task at hand;

·writing an explanation of the accuracy metric, understandable for lay people;

·procure external test datasets and calculating additional required metrics.

2 | On robustness:

·familiarising oneself with robustness requirement;

·brainstorming on possible internal limitations and external threats of the AI model;

·describing limitations of the AI system based on knowledge of the training data and algorithm;

·conducting internal tests against adversarial examples (entails possible retraining, changes to the algorithm, ‘robust learning’);

·conducting internal tests against model flaws (entails possible retraining, changes to the algorithm);

·conducting tests with external experts (e.g. workshops, audits);

·conducting robustness, safety tests in real-world conditions (controlled studies, etc.).

Moreover, additional labour is very likely be necessary to perform these tasks so that the development complies with requirements and to keep records of testing results for future conformity assessment.

For an average process, and a normally efficient firm, a reasonable cost estate for this activity is €10 733.33. 24

The business-as-usual factor

All of the above costs estimates relate to the total cost of the activities. However, economic operators would already take a certain number of measures even without explicit public intervention. To calculate this so-called business-as-usual factor, it is assumed that in the best prepared sector at most 50% of compliance costs would be reduced through existing practices. All sectors of the economy are then benchmarked with regard to their digital intensity against the best performing sector (e.g. in a sector with half the digital intensity only half as much can be accounted for business-as-usual). Next, for each sector future growth in digital intensity is forecast by extrapolating from recent years and a weighted average is calculated. As a result, the above costs are discounted by a factor of 36.67% 25 .

Instances where the data used in the impact assessment diverges from the data in the study

All the cost estimates are based on the support study. However, a few adjustments were made.

Firstly, all the figures have been rounded and where possible expressed as ranges of values. That is because the precise figures given above are the result of the mathematical modelling used in the study. However, given the assumption necessary for the calculation, the result really are only rough estimates, and indicating amounts to a single euro would signal a precision which is not backed up by the methodology. So, for example, the study’s business-as-usual factor 36.37% is used in the impact assessment as a “roughly one third” reduction.

Secondly, the compliance costs regarding robustness and accuracy have not been taken into account. Indeed, an economic operator trying to sell AI systems would anyway have to ensure that their product actually works, i.e. robustness and accuracy. This cost would therefore only arise for companies not following standard business procedures. While it is important that these requirements are included in the regulatory framework so that substandard operators need to improve their procedures, it would be misleading to include these costs for an average company. Including these costs in the overall estimate would only makes sense if one takes into account that a large share of AI providers supplies products that are either not accurate or not robust. There is no evidence to suggest that this is the case.

Note also that the compliance costs regarding human oversight have not been added with the other compliance costs into one single amount but kept separate, since it is overwhelmingly a recurring cost for AI users rather than a one-off cost for AI suppliers like the other compliance costs.

Finally, companies supplying high-risk AI systems in general already have a quality management system in place. For products, that is fundamentally because of already existing Union harmonisation legislation on product safety, which includes quality system-based conformity assessment procedures and, in some cases, also ad-hoc obligations for economic operators related to the establishment of a quality management system. Companies supplying high-risk stand-alone AI systems, such as remote biometric identification systems in publicly accessible places, which are controversially discussed topics, will equally often either already have a quality management system in place or introduce one if they want to market such a system subject to reinforced public scrutiny.,. Analogue to the reasoning above, while it is important that these requirements are included in the regulatory framework so that substandard operators need to improve their procedures, it would be misleading to include these costs for an average company.

5.ANNEX 5

5.1. ETHICAL AND ACCOUNTABILITY FRAMEWORKS ON AI INTRODUCED IN THIRD COUNTRIES

The present initiative on AI appears to be a frontrunner when it comes to proposing a comprehensive regulatory framework for AI. Governments in third countries are looking at the EU as a standard-setter (e.g. India; Japan); less eager to take action to impose regulatory constraints on AI (e.g. China); or more inclined towards sectoral approaches, rather that all-encompassing frameworks (the US). To date, no country has enacted a comprehensive regulatory framework on AI. However, a number of initiatives around the globe were taken into account in the analysis:

üThe Australian government is developing a voluntary AI Ethics framework, which includes a very broad definition of AI and eight voluntary AI Ethics principles. Guidance is developed to help businesses apply the principles in their organisations.

üIn Canada, a Directive on Automated Decision-Making came into effect on April 1, 2020 and it applies to the use by public authorities of automated decision systems that “provide external services and recommendations about a particular client, or whether an application should be approved or denied.” It includes an Algorithmic Impact Assessment and obligations to inform affected people when such systems are used.

üIn March 2019, the Japanese Cabinet Office released a document titled "Social Principles of Human-Centric AI". This document defines three basic principles: (i) Dignity, Diversity and Inclusion and Sustainability.  In July 2020, the Ministry of Economy, Trade and Industry, published a White Paper with respect to big data, the Internet of Things, AI and other digital technologies. The argument is that in order for regulations to keep up with the changes in technology and foster innovation, a new regulatory paradigm is needed.

üIn early 2020, the Personal Data Protection Commission of Singapore revised after consultation a Model AI Governance Framework, which offers detailed and readily-implementable guidance to private sector organisations to address key ethical and governance issues when deploying AI solutions.

üIn 2019, in the UK, the Office for AI published a “Guide on using artificial intelligence in the public sector” advising how the public sector can best implement AI ethically, fairly and safely. The Information Commissioner’s Office (ICO) has also published a Guidance on AI Auditing Framework, 26 providing 'best practices' during the development and deployment of AI systems for ensuring compliance with data protection laws.

üIn early 2020, the United States government adopted overall regulatory principles. On this basis the White House released the first-ever guidance for Federal agencies on the regulation of AI applications in the public sector.  Federal agencies must consider 10 principles including promoting public trust in AI, considering issues of fairness, non-discrimination, safety, and security, and assessing risks, costs, and benefits. The most recent U.S. President’s Executive Order from 3 December 2020 on Promoting the Use of Trustworthy Artificial Intelligence in the Federal Government, stipulates that when designing, developing, acquiring, and using AI in the Federal Government, agencies shall adhere to the following Principles: (a) Lawful and respectful of our Nation’s values. (b) Purposeful and performance-driven; (c) Accurate, reliable, and effective; (d) Safe, secure, and resilient; (e) Understandable; (f) Responsible and traceable; (g) Regularly monitored; (h) Transparent; (i) Accountable.

üSenate and House bills for the Algorithmic Accountability Act were proposed in the US-Congress in April 2019: they have required “impact assessments” on “high-risk” automated decision systems.” Similar bills were recently introduced by New Jersey, Washington State and New York City.

üIn February 2020, the New York City Council also proposed a bill for the use of automated employment decision tools, which requires an independent bias audit of these systems and informing job candidates that such systems have been used and that they are regulated by this act.

üStill in the Unites States, a Commercial Facial Recognition Privacy Act was proposed in March 2019. If enacted, the bill would generally prohibit organisations from using “facial recognition technology to collect facial recognition data” of end-users without providing notice and obtaining their consent.

üThe Government of New Zealand, together with the World Economic Forum, in 2020 was spearheading a multi-stakeholder policy project structured around three focus areas: 1) obtaining of a social licence for the use of AI through an inclusive national conversation; 2) the development of in-house understanding of AI to produce well-informed policies; and 3) the effective mitigation of risks associated with AI systems to maximize their benefits.

5.2. FIVE SPECIFIC CHARACTERISTICS OF AI

(1)Complexity: [multiplicity of elements that constitute an AI system and complexity of a value chain]

AI systems often have many different components and process very large amounts of data. For example, advanced AI models frequently have more than a billion parameters. These amount of parameters are not in practice understandable for humans, including for their designers and developers.

A system can be complex but still comprehensible from an ex-post perspective. For example, in the case of a rule-based system with a high number of rules, a human might not be able to say in advance what output the system would produce in a given context, but once there is an output, it can be explained based on the rules.

(2) Transparency/ Opacity: [the process by which an AI system reaches a result]

The opacity refers to the lack of transparency on the process by which AI system reaches a result. An AI system can be transparent (or conversely opaque) in three different ways: with respect to how exactly the AI system functions as a whole (functional transparency); how the algorithm was realized in code (structural transparency) and how the program actually run in a particular case, including the hardware and input data (run transparency).

Algorithms often no longer take the form of more or less easily readable code, but instead resemble a ‘black-box’. This means that while it maybe be possible to test the algorithm as to its effects, but not to understand how those effects have been achieved.

Some AI systems lack transparency because the rules followed, which lead from input to output, are not fully prescribed by a human. Rather, is some cases, the algorithm is set to learn from data in order to arrive at a pre-defined output in the most efficient way, which might not be representable by rules which a human could understand. As a result, AI systems are often opaque in a way other digital systems are not (‘the so called black box effect’). Independently from technical characteristics, a lack of transparency can also stem from systems relying on rules and functionalities that are not publicly accessible and of which a meaningful and accurate description is not publicly accessible.

The complexity and lack of transparency (opacity of AI) makes it difficult to identify and prove possible breaches of laws, including legal provisions that protect fundamental rights.

(3) Continuous adaptation: [the process by which an AI system can improve its own performance by ‘learning’ from experience] and Unpredictability: [the outcome of an AI system cannot be fully determined]

Some AI systems are not completed once put into circulation, but by their nature depend upon subsequent input, in particular on updates or upgrades. Often they need to interact with other systems or data sources in order to function properly. They therefore need to remain open by design, i.e. permit external input either via some hardware plug or through some wireless connection, and come as hybrid combinations of hardware, software, continuous software updates, and various continuous services.

“Many systems are designed to not only respond to pre-defined stimuli, but to identify and classify new ones and link them to a self-chosen corresponding reaction that has not been pre-programmed as such”. 27 Some AI systems can be used to automatically adapt or ‘learn’ while in use. In these cases, the rules being followed by the system will adapt based on the input which the system receives. This continuous adaptation will mean that the same input may produce different outputs at different times, thus rendering the system unpredictable to a certain extent.

Continuous adaptation can give rise to new risks that were not present when the system was placed on the market. These risks are not adequately addressed in the existing legislation which predominantly focuses on safety risks present at the time of placing on the market.

(4) Autonomous behaviour: [functional ability of a system to perform a task with minimum or no direct human control or supervision]

AI systems can increasingly perform tasks with less, or entirely without, direct human intervention. 28 A certain and increasing degree of autonomy (level of autonomy is a continuum) is one of the key aspects of certain AI systems. 29 This continuum ranges from systems where actions of a system are under full supervisions and control of a human to the more sophisticated AI systems that “combine environmental feedback with the system’s own analysis regarding its current situation” and thus have minimum or no human supervision in real time. This increasing degree of autonomous behaviour of some AI systems for a particular task combined with their increasing ‘ability’ to ‘interact’ with the external environment may present a particular challenge for transparency and human oversight. 30

Autonomy is not by itself a technological feature but rather the result of a design decision allowing a more or less constrained interaction between the system and the environment in pursuit of a task. The high-level objectives are defined by humans, however the underlying outputs and mechanisms to reach these objectives are not always concretely specified. The partially autonomous behaviour that developers foresee for certain AI systems is usually strongly tied to a specific context and function. Within the given context, these AI systems are designed to help reach conclusions or take decisions within pre-set boundaries without the involvement of a human operator.

Autonomy can affect the safety of the product, because certain AI systems increasingly can perform tasks with less, or entirely without, direct human intervention and in complex environments this may lead to situations where AI system may actions which have not been fully foreseen by their human designers with limited possibilities to override the AI system decision.

(5) Data

Many AI systems “increasingly depend on external information that is not pre-installed, but generated either by built-in sensors or communicated from the outside, either by regular data sources or by ad hoc suppliers. Data necessary for their proper functioning may, however, be flawed or missing altogether, be it due to communication errors or problems of the external data source, due to flaws of the internal sensors or the built-in algorithms designed to analyse, verify and process such data”. 31  

The accuracy of AI systems might be unevenly distributed in relation to different kinds of input data, depending on the data with which the system was trained. Furthermore, algorithms that are based on statistical methods produce probabilistic outputs, always containing a certain degree of error, no matter if they are designed to adapt while in use or not. Certain AI systems, due to the way and context they are exploited, present a risk of algorithmic bias as a consequence of several factors such as the considered dataset or machine learning algorithm. For instance, a machine learning algorithm may be “trained” on data to build a model that, once deployed, will process new data in a certain way (e.g. performing classification or pattern recognition). As an example, an application that is developed to recognize patterns would in the case of one common method (supervised learning) “learn” from the training data which characteristics (often called “features”) are relevant indicators for certain patterns so that the application can be used to recognize such patterns. The trained application can then be used to analyse future input data.

Both the training data and the input data (used to obtain an output) risk being discriminatory if they are unsuitable or inaccurate. For example, in recruitment contexts it is plausible that a developer only has data about accepted candidates, but no data about the would-be performance of candidates that were not hired. Besides the data, potential discrimination can also originate in the design of algorithms that are used to process the data. Relevant factors include the problem formulation, the underlying conception of a good result, potential biases in the conception of the software code, such as in the choice of input data and variables or in the benchmark for the evaluation of the outcome, which is often used to further optimise an application. There is a particular risk of biased outcomes in the case of machine learning applications. By automating at least parts of the process by which the rules are generated according to which an algorithm will produce results, it becomes possible that discriminatory rules are automatically generated. This is even likely where the data used to train a machine learning application reflects societal biases, if there is no adequate procedure to counteract these biases.

The dependence of AI systems on data and their ‘ability’ to infer correlations from data input can in certain situations can affect the values on which the EU is founded, create real health risk, disproportionately adverse or discriminatory results, reinforce systemic biases and possibly even create new ones.

5.3.INTERACTION BETWEEN THE INITIATIVE ON AI AND EXISTING SECTORAL PRODUCT SAFETY LEGISLATION

Section 1.

Existing product safety legislation does not contain specific requirement for safety and trustworthiness of AI systems. The proposed horizontal framework on AI will establish such new requirements for high-risk AI systems for certain sectoral product safety legislation (new and old approach). The acts concerned under the NLF framework and the old approach are enumerated respectively in sections A and B below.

The table below summarises how these new requirements for high-risk AI systems will be implemented and interact with existing sectoral product safety legislation.

Table 6: Overview of impact and applicability of the existing safety legislation and the AI horizontal framework to high-risk AI systems

High-risk AI / existing safety legislation

Interaction

Overall Impact

1

AI systems covered by certain sectoral safety legislation, following New Legislative Framework (NLF) 

The AI system will be high-risk if it is a safety component of a product or a device that is subject to a third party conformity assessment under the NLF legislation.

Requirements and obligations for high-risk AI systems set by the AI horizontal framework will become directly applicable and will automatically complement the existing NLF legislation.

·The new ex ante requirements for high-risk AI systems set in the AI horizontal framework will complement the existing sectoral safety requirements under NLF sectoral legislation.

·The conformity assessment procedures already existing under NLF would also apply for the checks of the new AI specific requirements.

·New obligations for providers and users will apply to the extent these are not already existing under the NLF sectoral act.

·The ex-post enforcement of the new rules for AI systems will be carried out by the same NLF market surveillance authorities responsible for the product.

2

AI systems covered by certain sectoral safety legislation, following Old Approach

(e.g. aviation, cars)

AI systems that are safety components of products under relevant old approach legislation will always be considered high-risk.

The new requirements for high-risk AI systems set by the AI horizontal framework will have to be taken into account when adopting relevant implementing or delegated legislation under those acts.

·The new ex-ante requirements for high-risk AI systems set in the AI horizontal framework will complement the existing sectoral requirements under old approach (when relevant implementing or delegated legislation under those acts will be adopted).

·The conformity assessment or authorisation procedures existing under the sectoral old approach legislation would also apply for the checks of the new AI requirements.

·The AI horizontal framework will not create any new obligations for providers and users.

·The ex-post enforcement rules of the AI horizontal framework will not apply.

A.Interaction between the proposal for AI horizontal framework and NLF safety legislation (row 1 in table above)

The proposed horizontal framework on AI will establish new requirements for high-risk AI systems that will complement the existing product safety NLF legislation. 32 An AI system will be high-risk if it is a safety component of a product or a device which undergoes a third party conformity assessment under the relevant sectoral NLF legislation. 33 A safety component of a product or device is understood as a component which provides the safety functions with regard to that specific product or device.

Based on up-to-date analysis the concerned NLF legislation that will fall under the scope of the new AI horizontal initiative include:

·Directive 2006/42/EC on machinery (which is currently subject to review);

·Directive 2009/48/EU on toys;

·Directive 2013/53/EU on recreational craft;

·Directive 2014/33/EU on lifts and safety components for lifts;

·Directive 2014/34/EU on equipment and protective systems intended for use in potentially explosive atmospheres;

·Directive 2014/53/EU on radio-equipment;

·Directive 2014/68/EU on pressure equipment;

·Regulation (EU) 2016/424 on cableway installations;

·Regulation (EU) 2016/425 on personal protective equipment

·Regulation (EU) 2016/426 on gas appliances;

·Regulations (EU) 745/2017 on medical devices;

·Regulation (EU) 746/2017 on in-vitro diagnostic medical devices.

The objective is to ensure that the new AI horizontal framework (which is in itself an NLF-type framework for the new safety requirements it creates) can be fully and smoothly integrated into the existing procedures and enforcement and governance systems established under the NLF legislation.

The new requirements for AI systems set by the AI horizontal framework would become directly applicable and be checked in the context of the conformity assessment system already existing under the relevant NLF instrument.

The Notified Bodies assessing the compliance of the provider with the new AI requirements would be the ones already designated under the relevant NLF legislation. However, the competence of the Notified Bodies in the field of AI should be assessed as part of the designation process under the relevant NLF instrument.

Obligations for certain operators in the value chain – namely manufacturers, importer, distributor, authorised representative - are generally already established in the existing NLF legislation. Obligations for economic operators (notably for providers and users) of the new AI horizontal framework apply to the extent these are not already existing under the NLF sectoral act.

With regard to market surveillance, Regulation (EU) 2019/1020 on market surveillance will apply to the AI horizontal framework. The ex-post enforcement of the new rules for high-risk AI systems will be carried out by the same NLF market surveillance authorities responsible for the product under the existing NLF legislation.

Ongoing or future reviews of NLF product legislation will not address aspects which are covered by the AI horizontal instrument. In order to increase legal clarity, any relevant NLF product legislation being reviewed (e.g. Machinery Directive 2006/42/EC subject to an ongoing review) would cross reference the AI horizontal framework, as appropriate. However, any reviewed NLF product legislation may aim to ensure that the incorporation of the AI system into the product does not compromise the safety of the product as a whole. In this respect, for example, the reviewed Machinery Directive 2006/42/EC could contain requirements for the safe integration of AI systems into the product (not covered by the AI horizontal framework).

B.Interaction between the proposal for AI horizontal framework and old-approach safety legislation (row 2 in table above)

Compared to NLF legislation, the applicability of the AI horizontal framework will be different for the old approach product safety legislation. This is because the old approach legislation follows a system of enforcement, generally based on detailed legal safety requirements (with possible integration of international standards into law) and a stronger role of public bodies in the approval system – an approach very different from the NLF logic followed by the AI horizontal initiative.

The horizontal framework on AI will establish new requirements for high-risk AI systems (e.g. transparency, documentation, data quality) that will be integrated into the existing old approach safety legislation. AI systems that are safety components of products under the old approach legislation will always be considered as high-risk AI systems. 34

Based on up-to-date analysis, the concerned old-approach legislation would be:

·Regulation (EU) 2018/1139 on Civil Aviation;

·Regulation 858/2018 on the approval and market surveillance of motor vehicles;

·Regulation (EU) 2019/2144 on type-approval requirements for motor vehicles;

·Regulation (EU) 167/2013 on the approval and market surveillance of agricultural and forestry vehicles;

·Regulation (EU) 168/2013 on the approval and market surveillance of two- or three-wheel vehicles and quadricycles;

·Directive (EU) 2016/797 on interoperability of railway systems.

·Directive 2014/90/EU on marine equipment (which is a peculiar NLF-type legislation, but given the mandatory character of international standardization in that field, will be treated in the same way as old-approach legislation).

The new requirements for high-risk AI systems set by the AI horizontal framework will have to be taken into account in the future when amending the sectoral legislation or when adopting relevant implementing or delegated acts under that sectoral safety legislation.

Existing conformity assessment/authorization procedures, obligations of economic operators, governance and ex-post enforcement under the old approach legislation will not be affected by the AI horizontal framework.

The application/relevance of the AI horizontal initiative on AI to the old approach safety legislation will be thus limited only to the new safety requirements for high-risk AI systems, when relevant implementing or delegated acts under that sectoral safety legislation will be adopted.

5.4. LIST OF HIGH RISK AI SYSTEMS (NOT COVERED BY SECTORIAL PRODUCT LEGISLATION) 

35 For AI systems that are mainly with fundamental rights implications and not covered by sectoral product safety legislation, the Commission has done the initial assessment for identifying the relevant high-risk AI systems by screening a large pool of AI use cases, covering:

·High-risk AI use cases included in the EP report 36 ; 

·A list of 132 AI use cases identified by a recent ISO report 37 and other methodologies 38 ;

·Results from the study accompanying the report, analysis by AI Watch and extensive complementary research of other sources such as analysis of case-law, academic literature and reports from international and other organisations (problem definition 2 in the impact assessment presents in short some of the most prominent use cases with significant fundamental rights implications);

·Results from the piloting of the draft HLEG ethic guidelines in which more than 350 stakeholders participated, including 50 in-depth case studies;

·Results from the public consultation on the White Paper that identify specific use cases as high-risk (or request their prohibition) and additional targeted consultations with stakeholders 39 ;

The risk assessment methodology described in the impact assessment has been applied to this large pool of use cases and the assessment of the Commission has concluded that the initial list of high-risk AI systems presented below should be annexed to Commission’s proposal of the AI horizontal instrument. Other reviewed AI use cases not included in this list have been discarded either because they do not cause harms to the health and safety and/or the fundamental rights and freedom of persons, or the probability and/or the severity of these harms has not been estimated as ‘high’ by applying the indicative criteria for risk assessment. 40  

Table 7: List of high-risk AI use cases (stand-alone) identified following application of the risk assessment methodology

HIGH-RISK USES

POTENTIAL HARMS

ESPECIALLY RELEVANT
INDICATIVE CRITERIA
*

EVIDENCE &
OTHER SOURCES

AI systems intended to be used for the remote biometric identification of persons in publicly accessible spaces

Intense interference with a broad range of fundamental rights (e.g. private life and data protection, human dignity, freedoms expression, freedom of assembly and association)

Systemic adverse impact on society at large (i.e., on democratic processes, freedom and chilling effect on civic discourse)

Already used by an increasing number of public and private actors in the EU

Potentially very severe extent of multitude of harms

High potential to scale and adversely impact a plurality of people

Vulnerability of affected people (e.g. people cannot object freely, imbalance if used by public authorities)

Indication of harm (legal challenges and decisions by courts and DPAs)

AlgorithmWatch and Bertelsmann Stiftung, Automating Society Report 2020 , 2020 (pp. 38-39, p. 104);

European Data Protection Board, Facial recognition in school renders Sweden’s first GDPR fine , 2019; 

European Data Protection Board, EDPS Opinion on the European Commission’s White Paper on Artificial Intelligence – A European approach to excellence and trust , 2020 (pp. 20-21); 

Agency for Fundamental Rights, Facial recognition technology: fundamental rights considerations in the context of law enforcement , 2019; 

Court of Appeal, United Kingdom, Decision R (Bridges) v. CC South Wales, EWCA Civ 1058 of 11 August 2020;

Buolamwini, I./ Gebru, T., Gender Shades: Intersectional Accuracy Disparities in Commercial Gender Classification , 2018;

National Institute of Standards and Technology, U.S. Department of Commerce, Face Recognition Vendor Test (FRVT) Part 3: Demographic Effects , 2019.

AI systems intended to be used to dispatch or establish priority in the dispatching of emergency first response services, including firefighters and medical aid

Injury or death of person(s), damage of property (i.e. by de-prioritising individuals in need of emergency first response services)

Potential interference with fundamental rights (e.g. human dignity, right to life, physical and mental integrity, non-discrimination)

Already used by some public authorities (firefighters, medical aid)

Potentially very severe extent of harm

High potential to scale and adversely impact a plurality of people (due to public monopoly)

Vulnerability and high dependency on such services in emergency situations

Irreversibility of harm very likely (due to physical character of the harm)

Not regulated by safety legislation

European Agency for Fundamental Rights, Getting The Future Right – Artificial Intelligence and Fundamental Rights , 2020 (pp. 34-36);

ISO A.97 System for real-time earthquake simulation with data assimilation, ISO/IEC TR 24030 - AI Use Cases 2020 (p. 101).

AI systems intended to be used as safety components in the management and operation of essential public infrastructure networks, such as roads or the supply of water, gas and electricity

Injury or death of person(s)

Potential adverse impact on the environment

Disruptions of ordinary conduct of critical economic and social activities

Potentially very severe extent of harm to people, environment and ordinary conduct of life

High potential to scale and adversely impact people and also the environment (potentially large scale due to criticality of essential public infrastructure networks)

Dependency on outcome (high degree of dependency due to potential to impact sensitive access to basic utilities)

Irreversibility of harm very likely due to the safety implications

Not regulated by safety legislation

German Data Ethics Commission, Opinion of the Data Ethics Commission , 2020.

ISO A.109 AI dispatcher (operator) of large-scale distributed energy system infrastructure, ISO/IEC TR 24030 - AI Use Cases 2020 (p. 42);

ISO A.29 Enhancing traffic management efficiency and infraction detection accuracy with AI technologies, ISO/IEC TR 24030 - AI Use Cases 2020 (pp. 103-104);

ISO A.49 AI solution for traffic signal optimization based on multi-source data fusion, ISO/IEC TR 24030 - AI Use Cases 2020 (p. 104);

ISO A.122 Open spatial dataset for developing AI algorithms based on remote sensing (satellite, drone, aerial imagery) data, ISO/IEC TR 24030 - AI Use Cases 2020 (pp. 114-115).

AI systems intended to be used for determining access or assigning individuals to educational and vocational training institutions, as well as for assessing students in educational and vocational training institutions and for assessing participants in tests commonly required for admission to educational institutions

Intense interference with a broad range of fundamental rights (e.g. non-discrimination, right to education, private life and data protection, effective remedy, rights of children)

Adverse impact on financial, educational or professional opportunities; adverse impact on access to public services;

Already used by some educational institutions

Potentially very severe extent of harm

High potential to scale and adversely impact a plurality of people (public education)

Dependency on outcome (access to education critical for professional and economic opportunities)

Insufficient remedies and protection under existing law

Indication of harm (opacity, existing legal challenges/case-law)

Tuomi, I., The use of Artificial Intelligence (AI) in education , European Parliament, 2020 (pp. 9-10);

UNESCO, Artificial Intelligence in Education: Challenges and Opportunities for Sustainable Development , 2019 (pp. 32-34);

AlgorithmWatch and Bertelsmann Stiftung, Automating Society Report 2020 , 2020 (p. 280);

Burke, L., The Death and Life of an Admissions Algorithm , Inside Higher Ed, 2020;

Department of Education Ireland, Press release on errors detected in Leaving Certificate 2020 Calculated Grades Process, 30 September 2020

ISO A.73 AI ideally matches children to daycare centers, ISO/IER TR 24030 – AI Use Cases 2020

ISO A.83 IFLYTEK intelligent marking system, ISO/IER TR 24030 – AI Use Cases 2020 (p.39).

AI systems intended to be used for recruitment – for instance in advertising vacancies, screening or filtering applications, evaluating candidates in the course of interviews or tests – as well as for making decisions on promotion and termination of work-related contractual relationships, for task allocation, monitoring or evaluating work performance and behaviour

Intense interference with a broad range of fundamental rights (e.g. workers’ rights, non-discrimination, private life and personal data, effective remedy)

Adverse impact on financial, educational or professional opportunities

Growing use in the EU

Potentially very severe effect of adverse decisions in employment context on individuals’ professional and financial opportunities and their fundamental rights

High degree of vulnerability of workers vis-à-vis (potential) employers

Insufficient remedies and protection under existing law

Indication of harm (high probability of historical biases in recruitment used as training data, opacity, case-law for unlawful use);

Datta, A. et al., Automated Experiments on Ad Privacy Settings , Proceedings on Privacy Enhancing Technologies; 2015 (pp. 92-112);

Electronic Privacy Information Center, In re HireVue , 2019;

Geiger, G., Court Rules Deliveroo Used 'Discriminatory' Algorithm , Vice, 2020; [Italy, Tribunale di Bologna, Decision of 31 December 2020, to be published .];

Sánchez-Monedero, J. et al., What does it mean to 'solve' the problem of discrimination in hiring?: social, technical and legal perspectives from the UK on automated hiring systems , 2020;

Upturn, Help Wanted: An Examination of Hiring Algorithms, Equity, and Bias , 2018;

ISO A.23 VTrain recommendation engine, ISO/IEC TR 24030 - AI Use Cases 2020 (p. 38)

AI systems intended to be used to evaluate the creditworthiness of persons or establish their credit score, with the exception of AI systems developed by small scale users for their own use

Adverse impact on economic, educational or professional opportunities;

Adverse impact on access to essential public services

Intense interference with a broad range of fundamental rights (e.g. non-discrimination, private life and personal data, effective remedy)

Growing use by credit bureaux and in the financial sector

Lack of transparency of AI based decisions making it impossible for individuals to know what type of behaviour will be relevant to assign them to their statistical group

Risk of high number of cases of indirect discrimination which are not likely to be captured by existing anti-discrimination legislation

Potentially severe harm (due to reduced access to economic opportunities when the services are provided by large scale operators, e.g. credit enabling investments and use of the score to determine access to other essential services e.g. housing, mobile services etc.)

Insufficient remedies and protection under existing law (robust financial service legislation, but assessment also done by unregulated entities; no binding specific requirements for AI)

Indication of harm (high probability of historical biases in past credit data used as training data, opacity, case law)

AlgorithmWatch, 2020, SCHUFA, a black box: OpenSCHUFA results published , 2018;

European Agency for Fundamental Rights, Getting The Future Right – Artificial Intelligence and Fundamental Rights , 2020 (pp. 71-72);  

Finland, National Non-Discrimination and Equality Tribunal, Decision 216/2017 of 21 March 2017;

European Banking Authority, Report on Big Data and Advanced Analytics , 2020 (pp. 20-21);

ISO A.27 Credit scoring using KYC data, ISO/IEC TR 24030 - AI Use Cases 2020 (pp. 43-44);

ISO A.119 Loan in 7 minutes, ISO/IEC TR 24030 - AI Use Cases 2020 (p. 46).

AI systems intended to be used by public authorities or on behalf of public authorities to evaluate the eligibility for social security benefits and services, as well as to grant, revoke, or reclaim social security benefits and services

Intense interference with a broad range of fundamental rights (e.g. right to social security and assistance, non-discrimination, private life and personal data protection, good administration, effective remedy)

Adverse impact on financial, educational or professional opportunities or on a person’s course of life; adverse impact on access to public services;

Growing use in the EU

Potentially very severe extent of harm (due to potentially crucial importance essential of social security benefits and services for individuals well-being)

High potential to scale and adversely impact a plurality of persons or groups (due to the public character of the social security benefits and services)

High degree of dependency on the outcome (due to lack of alternative for recipients) and high degree of vulnerability of recipients vis-à-vis public authorities

Indication of harm (opacity, high probability of past biased training data, challenges/case-law)

Allhutter, D. et al., AMS Algorithm on trial , Institut für Technikfolgen-Abschätzung der Österreichischen Akademie der Wissenschaften, 2020;

Netherlands, Court of The Hague, Decision C-09-550982 / HA ZA 18-388 of 5 February 2020 on Syri; Kayser-Bril, N., In a quest to optimize welfare management, Denmark built a surveillance behemoth , AlgorithmWatch, 2020;

Niklas, J., Poland: Government to scrap controversial unemployment scoring system , AlgorithmWatch, 2019;

Wills, T., Sweden: Rogue algorithm stops welfare payments for up to 70,000 unemployed , AlgorithmWatch, 2019;

European Agency for Fundamental Rights, Getting The Future Right – Artificial Intelligence and Fundamental Rights , 2020 (pp. 30-34).

Predictive policing and certain other AI systems in law enforcement, asylum, migration, border control with significant impacts on fundamental rights 

Intense interference with a broad range of fundamental rights (e.g. effective remedy and fair trial, non-discrimination, right to defence, presumption of innocence, right to liberty and security, private life and personal data, freedom of expression and assembly, human dignity, rights of vulnerable groups)

Systemic risks to rule of law, freedom and democracy

Growing use in the EU

Potentially very severe extent of harm (due to severe consequences of decisions and actions in this context)

Potential to scale at large and adversely impact a plurality of people (due to large number of individuals affected)

High degree of dependency (due inability to opt out) and high degree of vulnerability vis-à-vis law enforcement)

Limited degree of reversibility of harm

Insufficient remedies and protection under existing law

Indication of harm (high probability of historical biases in criminal data used as training data, opacity)

AlgorithmWatch, Automating Society , 2019 (pp. 37-38, 100);

Council of Europe, Algorithms and human rights , 2017, (pp. 10-11, 27-28);

European Agency for Fundamental Rights, Getting The Future Right – Artificial Intelligence and Fundamental Rights , 2020 (pp. 68-74); 

González Fuster, G., Artificial Intelligence and Law Enforcement – Impact on Fundamental Rights , European Parliament, 2020;

Gstrein, O. J. et al., Ethical, Legal and Social Challenges of Predictive Policing , Católica Law Review, 3:3, 2019 (pp. 80-81);

Oosterloo, S. & van Schie, G., The Politics and Biases of the “Crime Anticipation System” of the Dutch Police , Information, Algorithms, and Systems, 2103 (pp. 30-41);

Erik van de Sandt et al. Towards Data Scientific Investigations: A Comprehensive Data Science Framework and Case Study for Investigating Organized Crime & Serving the Public Interest, November 2020.

European Crime Prevention Network, Predictive policing, Recommendations paper, 2014.

Wright, R., Home Office told thousands of foreign students to leave UK in error , Financial Times, 2018.

Warrel, H., Home Office drops ‘biased’ visa algorithm , Financial Times, 2020;

Molnar P. and Gill L.,  Bots at the Gate: A human rights analysis of automated decision-making in Canada’s immigration and refugee system , University of Toronto, 2018.

ISO A.14 Behavioural and sentiment analytics, ISO/IEC TR 24030 - AI Use Cases 2020 (pp. 96-97).

Roxanne research project that uses AI to enhance crime investigation capabilities 

AI systems used to assist judicial decisions, unless for ancillary tasks

Intense interference with a broad range of fundamental rights (e.g. effective remedy and fair trial, non-discrimination, right to defence, presumption of innocence, right to liberty and security, human dignity as well as all rights granted by Union law that require effective judicial protection)

Systemic risk to rule of law and freedom

Increased possibilities for use by judicial authorities in the EU

Potentially very severe impact and harm for all rights dependent on effective judicial protection

High potential to scale and adversely impact a plurality of persons or groups (due to large number of affected individuals)

High degree of dependency (due to inability to opt out) and high degree of vulnerability vis-à-vis judicial authorities)

Indication of harm (high probability of historical biases in past data used as training data, opacity)

Council of Europe, Algorithms and human rights , 2017, (pp. 11-12);

European Commission for the Efficiency of Justice, European ethical Charter on the use of Artificial Intelligence in judicial systems and their environment , 2018;

U.S. Wisconsin Supreme Court Denied Writ of Certiorari of 26 June 2017, Loomis v. Wisconsin , 881 N.W.2d 749 (Wis. 2016);

Decision of the French Conseil Constitutionnel of 12 June 2018, Décision n° 2018-765 DC.Propublica,

Machine Bias: There’s software used across the country to predict future criminals, and it’s biased against blacks , 2016

5.5.: Analyses of Impacts on fundamental rights specifically impacted by the intervention

Impact on the right to human dignity

All options will require that humans should be notified of the fact that they are interacting with a machine, unless this is obvious from the circumstances or they have already been informed.

Options 2 to 4 will also prohibit certain harmful AI-driven manipulative practices interfering with personal autonomy when causing physical or psychological harms to people.

Impacts on the rights to privacy and data protection

All options will further enhance and complement the right to privacy and the right to data protection. Under options 3 to 4, providers and users of AI systems will be obliged to take mitigating measures throughout the whole AI lifecycle, irrespective of whether the AI system processes personal data or not.

All options will also require the creation of data governance standards in the training and development stage. This is expected to stimulate the use of privacy-preserving techniques for the development of data-driven machine learning models (e.g. federated learning, ‘small data’ learning etc.). New requirements relating to transparency, accuracy, robustness and human oversight would further complement the implementation of the data protection acquis by providing rules that address developers and providers who might not be directly bound by the data protection legislation. These options will harmonize and enhance technical and organisational standards on how high-level principles should be implemented (e.g. security, accuracy, transparency etc.), including in relation to high-risk AI applications.

Under options 2 to 4, AI used for some particularly harmful practices would also be prohibited such as general purpose scoring of citizens and use of AI-enabled technology that might manipulate users through specific techniques that are likely to cause physical or psychological harms.

Options 2 to 4 will also prohibit certain uses of remote biometric identification systems in publicly accessible spaces and subject the permitted uses to higher scrutiny and additional safeguards on top of those currently existing under the data protection legislation.

Impacts on the rights to equality and non-discrimination

All option will aim to address various sources of risks to the right to non-discrimination and require that sources of biases embedded in the design, training and operation of AI systems should be properly addressed and mitigated. All options except option 1 will also envisage limited testing obligation for users, taking into account the residual risk.

High quality data and high quality algorithms are essential for discrimination prevention. All options would impose requirements for documentation requirements in relation to the data and applications used and, where applicable, use of high quality data sets that should be relevant, accurate and representative for the context of application and the intended use. Obligations will also be imposed for testing and auditing for biases and adoption of appropriate bias detection and correction measures for high-risk AI system. Transparency obligations across the full AI value chain about the data used to train an algorithm (where applicable), its performance indicators and limitations will also help users to minimize the risk of unintentional bias and discrimination.

All options will also include additional requirements for accuracy and human oversight, including measures to minimize ‘automation bias’ that will help to reduce prohibited discriminatory impacts across protected groups.

Under options 2 to 4, providers and users of AI systems will be allowed to process sensitive data for the sole purpose of bias detection and mitigation and subject to appropriate safeguards. This will strike a fair balance and reconcile the right to privacy with the right to non-discrimination in compliance with the data protection legislation and the EU Charter of Fundamental Rights.

When properly designed AI systems could positively contribute to reducing bias and existing structural discrimination especially in some sectors (e.g. recruitment, police, law enforcement). For example, predictive policing might, in some contexts, lead to more equitable and non-discriminatory policing by reducing reliance on subjective human judgements.

Impact on the right to freedom of expression

Options 2 to 4 are expected to indirectly promote the right to freedom of expression insofar that increased accountability on the use of data shared by individuals could contribute to preventing the risk of a chilling effect on the right to freedom of expression.

An obligation to label deep fakes generated by means of AI could have an impact on the right to freedom of expression. That is why this obligation should not apply when the deep fakes are disseminated for legitimate purposes when authorised by law or to exercise freedom of expression or arts subject to appropriate safeguards for the rights of third parties and the public interests.

Impacts on the right to an effective remedy and fair trial and the right to good administration

The overall increased transparency and traceability of the system in the scope of all options will also enable affected parties to exercise their right to defence and right to an effective remedy in cases where their rights under Union or national law have been breached.

In addition, options 3 to 4 would require that certain AI systems used for judicial decision-making, in the law enforcement sector and in the area of asylum and migration should comply with standards relating to increased transparency, traceability and human oversight which will help to protect the right to fair trial, the right to defence and the presumption of innocence (Articles 47 and 48 of the Charter) as well as the general principles of the right to good administration. In turn, increased uptake of trustworthy AI in these sectors will contribute to improving access to legal information, possibly reducing the duration of judicial proceedings and to enhancing access to justice in general.

Finally, concerning restrictions potentially imposed by authorities, the established remedy options would always be available to providers and users of AI systems who are negatively affected by the decisions of public authorities.

Impacts on rights of special groups

All options are expected to positively affect the rights of a number of special groups. First, workers’ rights will be enhanced since recruitment tools and tools used for career management or monitoring will likely be subjected to the mandatory requirements for accuracy, non-discrimination, human oversight, transparency etc. In addition to that, workers (in a broad sense) are often the back-end operators of AI systems, so the new requirements for training and the requirements for safety and security will also support their rights to fair and just working conditions (Article 31 of the Charter).

The rights of the child (Art. 24 of the Charter) are expected to be positively affected when high-risk AI systems are affecting them (e.g. for decision-making purposes in different sectors such as social welfare, law enforcement, education etc.). Providers of the high-risk AI system should also consider children’s safety by design and take effective measures to minimize potential risks. Under option 3, this will concern only products and services considered to be ‘high-risk’, while under option 4 any product embedding AI, such as AI-driven toys, will have to comply with these requirements.

Option 2, 3, 3+ and 4 would also prohibit the design and use of AI systems with a view to distorting children’s behaviour in a manner that is likely to cause them physical or psychological harm which would also help to increase overall safety and integrity of children who are vulnerable due to their immature age and credulity.

Overall, increased use of AI applications can be very beneficial for the enhanced protection of children’s rights, for example, by detecting illegal content online and child sexual abuse, providing that it does not lead to a systematic filtering of communications, identifying missing children, providing adaptive learning systems tailored to each student’s needs and progress to name only a few examples.

Impact on the freedom to conduct a business and the freedom of science

All options will impose some restrictions on the freedom to conduct business (Article 16 of the Charter) and the freedom of art and science (Article 13 of the Charter) in order to ensure responsible innovation and use of AI. While under option 1, these restrictions will be negligent since compliance with the measures will be voluntary, options 2 to 4 envisage binding obligations that will make the restrictions more pronounced.

Under options 2, 3 and 3+, these restrictions are proportionate and limited to the minimum necessary to prevent and mitigate serious safety risks and likely infringements of fundamental rights. However, option 4 would impose requirements irrespective of the level of risk, which might lead to disproportionate restrictions to the freedom to conduct a business and the freedom of science. These restrictions are not genuinely needed to meet the policy objective and they would prevent the scientific community, businesses, consumers and the society at large from reaping the benefits of the technology when it poses low risks and does not require such an intense regulatory intervention.

Impact on intellectual property rights (Article 17(2) of the Charter)

Often economic operators seek out copyright, patent and trade secret protection to safeguard their knowledge on AI and prevent disclosure of information about the logic involved in the decision-making process, the data used for training the model etc.

The increased transparency obligations under options 2 to 4 will not disproportionately affect the right to an intellectual property since they will be limited only to the minimum necessary information for users, including the information to be included in the public EU database.

When public authorities and notified bodies are given access to source code and other confidential information, they are placed under binding confidentiality obligations.

(1)      European Commission, White Paper on Artificial Intelligence - A European approach to excellence and trust , COM(2020) 65 final, 2020.
(2)      European Commission, Second European AI Alliance Assembly , 2020.
(3)      European Commission, Commission Staff Working Document – Better Regulation Guidelines , SWD (2017) 350, 2017.
(4)

     European Commission, , Report on the safety and liability implications of Artificial Intelligence, the Internet of Things and robotics COM/2020/64 final, 2020.

(5)

     Other arguments mentioned (minimum frequency): accuracy (10) , collective harms caused by AI (5), involvement of civil society (5), manipulation (5), power asymmetries (e.g. between governments and citizens; employers and employees; costumers and large companies) (4), safety (4), legal reediness & review, environmental impact of AI (4), unemployment/employment related discrimination (3), privacy & data protection (3), compensation (2), cybersecurity (2), intentional harmful abuse from AI (2), More R&D in AI can help address concerns (2), external threats to humanity (2), intellectual property rights (2) and media pluralism (2).

(6)

     ‘Do you think that the concerns expressed above can be addressed by applicable EU legislation? If not, do you think that there should be specific new rules for AI systems?’

(7)      To that aim, some comments suggested changes in the GDPR and others supported that legislation should be technology neutral.
(8)      Further comments to this question referred to human oversight (3), the difficulty of assessment and categorisation of AI (2), the need to align the definition of high-risk with international standards (2) and continuously review them for change (2), the use of the precautionary principle in general (2) and that of GDPR for risks related to privacy (1).
(9)      An AI application should be considered high-risk where it meets the following two cumulative criteria:First, the AI application is employed in a sector where, given the characteristics of the activities typically undertaken, significant risks can be expected to occur. This first criterion ensures that the regulatory intervention is targeted on the areas where, generally speaking, risks are deemed most likely to occur. The sectors covered should be specifically and exhaustively listed in the new regulatory framework. For instance, healthcare; transport; energy and parts of the public sector. (…)Second, the AI application in the sector in question is, in addition, used in such a manner that significant risks are likely to arise. This second criterion reflects the acknowledgment that not every use of AI in the selected sectors necessarily involves significant risks. For example, whilst healthcare generally may well be a relevant sector, a flaw in the appointment scheduling system in a hospital will normally not pose risks of such significance as to justify legislative intervention. The assessment of the level of risk of a given use could be based on the impact on the affected parties. For instance, uses of AI applications that produce legal or similarly significant effects for the rights of an individual or a company; that pose risk of injury, death or significant material or immaterial damage; that produce effects that cannot reasonably be avoided by individuals or legal entities.’ (European Commission, White Paper on Artificial Intelligence - A European approach to excellence and trust , COM(2020) 65 final, 2020).
(10)      Other comments: In favour of ‘human rights impact assessments’ (2). The context of use of an AI is important for assessing its risk (2). The binary separation in high/low risk is too simplified (2). The criteria for 'high risk' do not go far enough (2). Against listing the transport sector as 'high risk' (2). The risk framework should be proportionate (2). Agree with limit to high-risk applications, but should also apply to non-AI systems (2). Reference to other documents/standards (2).
(11) Additional comments: Excessive regulation hinder innovation / imposes costs (7). Allow remote biometric identification in public spaces only if proportionate (6).More research/information is necessary (6). Allow remote biometric identification in public spaces only for specific purposes: security / criminal justice matters, only in specific cases (6). The existing framework is sufficient (6). Stakeholders should be consulted (6). Allow remote biometric identification in public spaces only under other specific conditions (5). Facial recognition may be needed for autonomous vehicles (coordinated response, car makers) (5). Legislation needs to be clear and simple (4). The definition of ‘public space’ is unclear (4). Strict rules for the storage of biometric data are important (3). Remote biometric identification in public spaces is useful for social distancing during the COVID-19 epidemic (3). Regulation should only be considered in case of consumer harm (2). Human oversight is overestimated (2). A moratorium would leave the field to other, less free countries and reduce accuracy of systems (2). Are vehicles a ‘public space’? (2). EU-level harmonisation is important (2).
(12)      European Commission, About the energy label and eco-design , 2020.
(13) Additional comments: All AI should be regulated (5). In favour of a mandatory labelling system (4). In B2B trust is created through contractual agreements (3). Standards need to be actively promoted to become effective (2). Not products/services should be labelled, but an organisation's quality of AI governance (2).
(14)      Additional comments: Risks caused by autonomous driving / autonomous systems (5). Risks linked to loss of control / choice (7). Weapons / lethal autonomous weapon systems (4). Risks for fundamental rights (3). Risks for nuclear safety (2). Significant material harm (2). Risks to intellectual property (IP) (2). Risks to employment (1).
(15) Additional comments: Recommendations on when the risk assessment should be required (8). There is no need for new AI-specific risk assessment rules (7). Existing bodies should be involved and properly equipped (4). Independent external oversight is necessary (not specified by whom) (4). Overly strict legislation can be a barrier for innovation and impose costs (4). New risk assessment procedures are not necessary (4). Trade unions should be involved (3). Long-term social impacts should be considered (3).Human oversight / final human decisions are important (3). Fundamental rights are important in the assessment (2). Legal certainty is important (2). Risk assessments are already obligatory in sectors like health care (2).
(16)   ISBN 978-92-76-36220-3
(17)
(18)

Stakeholders’ feedback suggests that EUR 32 is too low, but they are operating in more advanced economies. Given the economic differences across the EU, the EU average is a reasonable reference point here.

(19)      See support study chapter 4, section 4.2.1.
(20)      See support study chapter 4, section 4.2.2.
(21)      See support study chapter 4, section 4.2.3.
(22)      High-Level Expert Group on Artificial Intelligence, Assessment List for Trustworthy Artificial Intelligence (ALTAI) for self-assessment , 2020.
(23)      See support study chapter 4, section 4.2.4.
(24)      See support study chapter 4, section 4.2.5.
(25) See support study chapter 4, section 4.4.2.1
(26)   https://ico.org.uk/for-organisations/guide-to-data-protection/key-data-protection-themes/guidance-on-artificial-intelligence-and-data-protection/
(27)      Report from the Expert Group on Liability and New Technologies – New Technologies Formation, European Commission, 2019, p.33.
(28)      This is independent and separate from the ability of certain systems to alter the rules which they follow while in use, i.e. ‘continuous adaptation’ characteristic discussed above.
(29)      See for example, SAE International standard J3016 “Levels of Driving Automation” that defines the six levels of driving automation for road vehicles, from no automation to full automation.
(30)      For more detailed discussion of concept of autonomy see e.g. The International Committee of the Red Cross, Autonomy, artificial intelligence and robotics: Technical aspects of human control, 2019. This report cautiously explain “the perception of both autonomy and AI is constantly shifting, as advances in technology mean that some systems once considered “autonomous” and “intelligent” are now classed merely as “automated”. Importantly, there is no clear technical distinction between automated and autonomous systems, nor is there universal agreement on the meaning of these terms.”
(31)      See above, p.33.
(32)      NLF product legislation also covers some non-embedded AI systems which are considered products by themselves (e.g. devices by themselves under the Medical Device Regulations or AI safety components placed independently on the market which are “machinery by themselves under the Machinery Directive). If those non-embedded AI systems are subject to third-party conformity assessment under the relevant sectoral framework, they will be high-risk for the purpose of the AI horizontal framework.
(33) This approach is justified because the conformity assessment of any sectoral legislation already presupposes a risk assessment on the safety risks posed by the products covered by that instrument. It makes therefore sense to rely on the risk classification of a product under the relevant NLF legislation to define when an AI-driven safety component (of that product) should be considered high-risk.
(34) This is because products regulated under the old approach legislation always undergo third party conformity assessments or authorisation procedures in the legislations that will be covered by the new AI initiative.
(35)      For AI systems which are safety components of products covered by sectoral product safety legislation see Annex 5.3.
(36)      European Parliament resolution of 20 October 2020 with recommendations to the Commission on a framework of ethical aspects of artificial intelligence, robotics and related technologies, 2020/2012(INL) .
(37)      For example, classification of products as high-risk means that the AI safety component should also be treated similarly; See also Article 29 Data Protection working Party, Guidelines on Data Protection Impact Assessment (DPIA) and determining whether processing is “likely to result in a high risk” for the purposes of Regulation 2016/679.
(38) Final Draft of ISO/IEC TR 24030 - AI Use Cases. The Opinion of the German Data Ethic Commission proposing a pyramid of 5 levels of criticality of the AI systems. The Council of Europe Recommendation CM/Rec(2020)1  refers to “high risk” when the algorithmic systems is used in processes or decisions that can produce serious consequences for individuals or in situations where the lack of alternatives prompts a particularly high probability of infringement of human rights, including by introducing or amplifying distributive injustice.
(39) The Commission has also carried out targeted consultations on specific topics that have informed its assessment: 1) Principles and Requirements for trustworthy AI; 2) Biometrics, 3) Children’s rights, 4) Standardisation, 5) Conformity assessments, 6) Costs of implementation. These workshops were focusing on collecting data, evidence and complementing the public consultations on the White Paper and the Inception Impact Assessment.
(40) See Option 3 in section 5.3 of the Impact assessment. A specific assessment of the probability and severity of the harms will be done to determine if the AI system generates a high-risk to the health and safety and the fundamental rights and freedom of persons based on a set of criteria that will be defined in the legal proposal. The criteria for assessment include: a) the extent to which an AI system has been used or is about to be used; b) the extent to which an AI system has caused any of the harms referred to above or has given rise to significant concerns around their materialization; c) the extent of the adverse impact of the harm; d) the potential of the AI system to scale and adversely impact a plurality of persons or entire groups of persons; e) the possibility that an AI system may generate more than one of the harms referred to above; f) the extent to which potentially adversely impacted persons are dependent on the outcome produced by an AI system, for instance their ability to opt-out of the use of such an AI system; g) the extent to which potentially adversely impacted persons are in a vulnerable position vis-à-vis the user of an AI system; h) the extent to which the outcome produced by an AI system is reversible; i) the availability and effectiveness of legal remedies; j) the extent to which existing Union legislation is able to prevent or substantially minimize the risks potentially produced by an AI system.
Top