Wednesday, January 22, 2025

Cracking the Code: The EU Framework for Trustworthy AI

Cracking the Code:  The EU Framework for Trustworthy AI

The EU AI Act is not light bedtime reading—over one hundred and fifty pages packed with legal jargon. It covers everything from defining “input data” to laying out the documentation rules for high-risk AI systems. And, like getting lost in a dense forest, it is easy to miss the bigger picture when you are buried in the details.


Context is everything

To properly gain a more nuanced understanding of the regulatory efforts, it helps to trace through its history. Do not worry, we are not going to make you sift through six years of documents. Instead, we will break it down for you, showing how the framers of the regulations saw their work within the larger context.

Impact on Regulation

The framework for trustworthy AI released by the high level expert group has greatly influenced the EU AI Act. The legal text explicitly references the seven requirements for trustworthy AI, which although non-binding, has nevertheless informed many of the concepts and jargon that feature directly in the legislation. This is specially true of high risk AI systems when it pertains to human oversight, transparency and technical robustness obligations as is evident from some excerpts from the actual Act:

“Requirements should apply to high-risk AI systems as regards risk management, the quality and relevance of data sets used, technical documentation and record-keeping, transparency and the provision of information to deployers, human oversight, and robustness, accuracy and cybersecurity. Those requirements are necessary to effectively mitigate the risks for health, safety and fundamental rights”

“High-risk AI systems shall be designed and developed in such a way that they achieve an appropriate level of accuracy, robustness, and cybersecurity, and that they perform consistently in those respects throughout their lifecycle”

“Human oversight shall aim to prevent or minimise the risks to health, safety or fundamental rights that may emerge when a high-risk AI system is used in accordance with its intended purpose or under conditions of reasonably foreseeable misuse, in particular where such risks persist despite the application of other requirements set out in this Section”

Understanding the framework will give you a good grounding of the EU AI Act and some of its underlying principles.


An Eco-system of trust

The European Commission, between 2018 and 2020, identified two key policy objectives in order to increase the uptake of AI, address AI risks and avoid market fragmentation. These consisted of establishing common EU policies in order to establish:

  1. An ecosystem of trust in AI
  2. An ecosystem of excellence in AI

The ecosystem of trust in turn, reasoned the European Commission, required a clear EU wide legal framework for AI as well as guidelines and framework for development of trustworthy AI. We will deconstruct both today.

AI ecosystem of Trust

Legal framework for AI

The push to create a comprehensive legal framework for AI came from two main goals: building trust in AI and avoiding market fragmentation in the EU caused by different national laws.

EU Legal framework for AI

The legal framework was envisaged with three pillars in mind to mitigate the risks of AI as well as provide a redressal mechanism for any harm caused by AI:

  1. The EU AI Act which provided the rules concerning development and use of AI systems. The Act was conceived as a horizontal framework which would be largely sector agnostic.
  2. The General Safety Product Regulation, while not specific to AI, provided some sector specific guardrails. In later articles we shall see how AI systems have increased calls for certain changes in GSPR but lets keep it simple for now.
  3. An AI liability directive which is currently in the proposal stage and addresses liability issues related to AI systems.

We will deep dive into the proposed liability directive in our articles on March 6th and 13th - you can find out about it here. A primer on the EU AI Act is planned on 20th March.

Guidelines and framework for trustworthy AI

The guidelines and framework for trustworthy AI, initially released by the High Level Expert Group in 2018, is not just useful as a non-binding best practice for organizations, but as noted earlier, has had a significant influence on the legislation itself.

The framework can be divided into three logical parts:

  1. Part one defines what trustworthy AI is.
  2. Part two details the seven requirements needed to realize trustworthy AI.
  3. Part three provides concrete assessment lists to help organizations evaluate their readiness in meeting the requirements.
EU Framework for trustworthy AI

Defining trustworthy AI

Trustworthy AI is defined by the framework as AI which was lawful, ethical and robust:

  1. Lawful: Complies with all applicable laws and regulations. There are many levels of laws governing the EU including EU primary laws, secondary laws and laws of the individual member states.
  2. Ethical: Adheres to ethical principles and values.
  3. Robust: Confidence that AI systems will not cause unintentional harm and perform in a safe, secure and reliable manner.

Trustworthy AI, the framework notes, is further grounded in ethical principles which include human autonomy, prevention of harm, fairness and explicability. Explicability here is defined as a means to provide transparency and explainability in decisions and processes of the AI system.


Realizing trustworthy AI

Once trustworthy AI is defined, the framework details how to realize it. Seven key requirements are introduced which together help create trustworthy AI systems:

  • Human agency and oversight: Enabling meaningful human control. Oversight concepts are made clear such as
    • Human in the loop (HITL). Refers to the capability for human intervention in every decision cycle of the system.
    • Human on the loop (HOTL). Capability for human intervention during the design cycle of the system and monitoring the system’s operation.
    • Human in command (HIC). capability to oversee the overall activity of the AI system (including its broader economic, societal, legal and ethical impact) and the ability to decide when and how to use the system in any particular situation.
  • Technical robustness and safety: Ensuring reliability, security, and resilience. This includes resilience to cybersecurity, fallback plan and general safety, accuracy, reliability and repreducibility.
  • Privacy and data governance: Guaranteeing data protection and integrity. Privacy protection refers not only to collection in AI but also inference. For example, Digital records of human behaviour may allow AI systems to infer not only individuals’ preferences, but also their sexual orientation, age, gender, religious or political views.
  • Transparency: Enhancing traceability and communication. The elements of transparency include traceability, explainability and communication.
  • Diversity, non-discrimination, and fairness: Which consists of avoiding biases, fostering inclusivity and ensuring stakeholder participation.
  • Societal and environmental well-being: Addressing broader social and ecological impacts.
  • Accountability: Implementing mechanisms for oversight, responsibility, and redress. In particular the AI system should be auditable, report adverse impacts and address tradeoffs in a rational manner.

Assessing trustworthiness of AI systems

Finally, the framework addresses how to evaluate readiness for meeting the seven requirements essential for building a trustworthy AI system. The assessment lists around a dozen questions for each requirement which are further broken down into its constituent components as illustrated below.

AI Assessment List

The final version of the assessment list can be found here.

Opinion

The framework is useful to understand the evolving regulatory framework but has limited value at an organizational level. If you wish to understand many of the underlying concepts of the EU AI Act and its evolution, it is useful to review the framework. The assessment list also provides some guidance and can act as a starting point for practitioners who are not experienced in AI governance.

From an organizational point of view, however, the framework has limited value; and this may well be because it was not intended to be a comprehensive document for organizations but rather a simple scaffolding for regulators. The requirements and especially the assessment list offers limited guidance on how organizations can develop their AI governance frameworks or how they can manage AI risks. To this extent, organizations may find the AI Risk Management Framework by the US National Institute of Standards and Technology more useful which can be found here. Recently the European Commission and related organizations, such as, Joint Research Centre have made significant efforts to release templates and guides which are helpful to organizations - we will review these at a later date.

We will release a very detailed paper on arXiv regarding AI risk management and AI governance from an organizational perspective in the next week - so keep tuned.


References

The EU AI Act. https://eur-lex.europa.eu/legal-content/EN/TXT/?uri=CELEX%3A32024R1689&qid=1737509691447

HLEG Guidelines on trustworthy AI. https://digital-strategy.ec.europa.eu/en/library/ethics-guidelines-trustworthy-ai

HLEG AI assessment list. https://digital-strategy.ec.europa.eu/en/library/assessment-list-trustworthy-artificial-intelligence-altai-self-assessment

European Commission whitepaper on White Paper on Artificial Intelligence: a European approach to excellence and trust https://commission.europa.eu/publications/white-paper-artificial-intelligence-european-approach-excellence-and-trust_en

NIST AI Risk Management Framework. https://www.nist.gov/itl/ai-risk-management-framework