
(Images made by author with Microsoft Copilot)
On March 13, 2024, the European Parliament took a historic step by approving the Artificial Intelligence Act (AI Act). This landmark legislation marks the world’s first comprehensive regulatory framework for AI, aiming to ensure responsible development and use of this powerful technology within the European Union (EU). This blog post provides an overview of the Act, explaining its significance and key provisions.
Table of Contents
Before delving into the details, let’s first understand what the AI Act is and why it’s important.
The EU’s Artificial Intelligence Act: Importance and Overview
The AI Act is a set of harmonized rules that establish requirements for the development, deployment, and use of AI systems within the EU. It classifies AI systems into different risk categories and sets out specific obligations for each category, positioning Europe as a global leader in AI regulation.
The AI Act is significant because it provides clear guidelines for AI developers and users. This promotes responsible usage while addressing potential risks. The Act ensures safety, transparency, and accountability in AI systems, minimizing risks while fostering innovation and economic growth in the EU. By enshrining fundamental rights and values, the Act defines the future path of AI development and deployment.
Key Provisions
The core elements of the AI Act are outlined in key provisions, including:
Bans on harmful AI practices
The Act prohibits certain AI practices that pose a clear risk to fundamental rights and safety. Such prohibited practices include:
- AI systems that manipulate people’s behavior in a way that causes physical or psychological harm.
- AI systems that exploit vulnerable persons. AI systems that are used for social scoring by governments.
Classification of AI Systems
The AI Act takes a risk-based approach, classifying AI systems into the following categories:
- High-risk AI systems: These are AI systems that pose a significant risk to the health, safety, or fundamental rights of individuals. These systems are subject to strict requirements, including mandatory conformity assessment procedures and ongoing monitoring.
- Non-high-risk AI systems: These are AI systems that do not pose a significant risk to the health, safety, or fundamental rights of individuals. These systems are not subject to the same strict requirements as high-risk AI systems, but they may be subject to some requirements, such as transparency, accountability, and safety.
Transparency and accountability
The Act promotes transparency by requiring providers of high-risk AI systems to provide clear and accessible information about how their systems work. This information must be available to users before they interact with the AI system. Additionally, high-risk AI systems must bear the European conformity marking to indicate their compliance with the Act’s requirements.
Establishment of a European Artificial Intelligence Board
The independent European Artificial Intelligence Board will advise the Commission, the executive branch of the European Union, on AI policy, ethics, and risks. It will also monitor AI developments, promote responsible AI practices, and facilitate collaboration. Composed of diverse experts, the Board will be a key player in shaping the EU’s AI landscape.
Creation of an AI Office
The AI Office, established by the Commission, will support the implementation and enforcement of the AI Act. It will provide expertise and guidance to national authorities, build enforcement capacity, and foster stakeholder engagement. The Office will also monitor implementation, report findings, and contribute to the Act’s ongoing development.
Obligations for Providers and Deployers
The AI Act outlines distinct responsibilities for different parties involved in the AI ecosystem. Providers and deployers are subject to the following obligations:
- Providers of AI systems are entities that develop, design, and distribute AI systems. They are responsible for ensuring that their systems are compliant with all applicable regulations and industry standards, and that they are designed and developed in a way that minimizes risk to users. Providers must also provide clear and concise instructions for use to deployers and users, and they must be able to demonstrate that their systems have been tested and validated to ensure their safety and effectiveness.
- Deployers of AI systems are entities, excluding personal non-professional users, that use an AI system under their authority. They are responsible for using the systems in a way that is consistent with their intended purpose and that minimizes risk to users. Deployers must also ensure that their staff is properly trained on the use of the systems and that they have adequate resources to monitor and maintain the systems.
Enforcement and Penalties
Enforcement of the AI Act is primarily the responsibility of national competent authorities. These authorities are responsible for monitoring and supervising AI systems placed on the EU market to ensure compliance with the Act’s requirements. They have the power to investigate non-compliant AI systems, take enforcement actions, and cooperate with the European AI Office.
Penalties for non-compliance can be significant, including fines and other measures to prevent or mitigate the risks posed by non-compliant AI systems.
Conclusion
The EU’s Artificial Intelligence Act, expected to be finalized soon, represents a significant milestone in AI regulation. This comprehensive framework addresses AI risks, fosters innovation, and protects fundamental rights. Its impact is anticipated to reach beyond the European Union, shaping global AI development and deployment.
It’s important to note that the Act will have a varied implementation timeline once adopted. While some rules take effect within 6 months, high-risk systems have 3 years for compliance.
Learn more
For further details, you can refer to the following resources:
European Parliament, Artificial Intelligence Act: MEPs adopt landmark law, March 13, 2024.
European Parliament, Artificial Intelligence Act (text adopted), March 13, 2024.
Note: This post was researched and written with the assistance of various AI-based tools.


Leave a comment