August 1, 2024 marks an important date for businesses and digital professionals in France with the entry into force ofthe Artificial Intelligence Regulation (RIA) or AI Act.
This regulation, adopted by the European Union, aims to regulate the use of artificial intelligence (AI) in commercial, industrial and social activities, in order to guarantee safety, ethics and transparency in the use of these technologies.
In this article, we decipher these new regulations for you, share everything you need to know, and suggest the best practices to adopt to ensure your compliance.
What is RIA?
The Artificial Intelligence Regulation (AIR ) is a legislative framework drawn up by the European Union to regulate the application and development of AI. This regulation is part of a broader drive to make Europe a world leader in artificial intelligence, while protecting citizens' fundamental rights.
The RIA is based on a risk-based approach, categorizing AI systems into four levels of risk: unacceptable, high, limited, and minimal. Each category is subject to specific rules, designed to ensure a balance between technological innovation and the protection of human rights.
RIA's objective?
- Harmonization of rules for the marketing, commissioning and use of AI systems in the EU
- Offer AI systems that are safe, transparent, traceable, non-discriminatory and environmentally friendly
Why is the RIA important?
Artificial intelligence is at the heart of digital transformation. However, without adequate regulation, it can present significant risks, particularly in terms of discrimination, privacy violations and security. The RIA therefore responds to the need to establish a clear framework for the use of AI, guaranteeing that its development is carried out ethically and transparently.
This regulation is crucial for :
- Ensuring the safety of AI systems: By imposing rigorous standards for applications deemed to be high-risk.
- Protect fundamental rights: By preventing discrimination and ensuring the transparency of AI systems, particularly those used in sensitive areas such as justice, employment and public services.
- Encouraging innovation: By providing companies with a clear framework that facilitates the integration of AI into their processes while ensuring compliance
Key points to remember
- Prohibition of practices deemed unacceptable: because of their potential to violate fundamental rights (social rating, real-time biometric identification, etc.).
- Regulation of high-risk AI systems: AI systems used in critical sectors, such as healthcare, education, and justice, are subject to strict requirements in terms of transparency, risk management, and human control.
- Obligations for AI suppliers: Companies developing AI systems must ensure that their products comply with European standards, including data traceability, technical documentation and conformity assessments.
- Transparency and information: End-users need to be informed when they interact with an AI, particularly in the case of chatbots or content-generating AIs.
- Encouraging responsible innovation: The regulation introduces regulatory sandboxes to offer companies a secure, supervised environment in which to test innovative AI solutions. These sandboxes allow, among other things, the re-use of personal data, including sensitive data, when this serves a major public intereste.g improving the healthcare system).
Who checks its application?
Two types of governance :
At the European level:
- AI Office (European Commission institution)
- European AI Committee (representatives from each member state)
At the national level: France has 12 months to designate one or more supervisory authorities.
What are the penalties?
In the event of non-compliance, the RIA imposes a market withdrawal or product recall, as well as administrative fines of up to 35 million euros or 7% of annual worldwide sales.
Does the RIA replace the GDPR
NO: The GDPR concerns any processing of personal data, whereas the RIA concerns the development, marketing and deployment of AI.
Differences between RIA and GDPR
Although the RIA and the General Data Protection Regulation (GDPR ) are both European legal frameworks, they aim to regulate different aspects of digital.
| AI Act | GDPR | |
|---|---|---|
| Effective date | August 1, 2024 | May 25, 2018 |
| Scope of application | AI development, commercialization and deployment | Any processing of personal data (including for the purpose of developing or training an AI) |
| Approach | Risk-based approach: unacceptable, high, limited, and minimal. | Approach based on the application of key principles, risk assessment and Accountability |
| Target players | Mainly suppliers and deployers of AI systems | Processing managers and processors |
| Conformity assessment | Internal or third-party assessment through a risk management system | Accountability principle (legal documentation in-house) and compliance tools |
| Control authority | Not yet defined | CNIL (National Commission for Information Technology and Civil Liberties) ) (Commission nationale de l'informatique et des libertés) |
| Sanctions | Market withdrawal or product recall + administrative fines of up to 35 million euros or 7% of worldwide annual sales . | Formal notice + administrative fines of up to 20 million euros or 4% of worldwide annual sales . |
How does RIA fit in with GDPR ?
🔍 Four scenarios are possible:
1️⃣ RIA only: RIA only applies when, for example, a high-risk AI system is used without involving the processing of personal data.
2️⃣ GDPR only: The GDPR applies in situations where personal data is processed, but the AI system used is not subject to the requirements of the RIA.
3️⃣ RIA and GDPR combined: The two regulations apply in conjunction when, for example, a high-risk AI system requires personal data for its development or implementation.
4️⃣ Neither applies: Neither the RIA nor the GDPR are required in cases where an AI system presents minimal risk and does not process personal data.
Here are some concrete examples:

What does this mean for your company?
The coming into force of the RIA requires companies to review their AI processes and systems to ensure compliance with the new regulations.
For you as a company, it's an opportunity to strengthen your credibility and competitiveness in the marketplace by adopting practices that comply with European expectations.
Here are four best practices to implement right away:
- Audit and evaluate your AI systems: Companies need to identify the AI systems in use, assess their level of risk, and implement the necessary corrective measures.
- Update your internal policies: It's essential to develop clear internal policies regarding the use of AI, ensuring that all practices comply with RIA requirements.
- Train your teams: Employees must be trained in the new AI obligations and best practices to avoid any unintentional breaches of the regulations.
- Collaborate with experts: Companies can benefit from the expertise of outsourced DPOs or specialist firms to navigate the complexities of RIA and ensure ongoing compliance.
Dipeeo support
At Dippeo, as an external DPO for nearly 530 clients, we support you in securing your AI projects:
✔️ Documentation & traceability: We make AI explainable: your decision criteria are clarified, traceable, and compliant with AI Regulation requirements.
✔️ User control verification: We ensure that your processes include the right to challenge and effective human intervention.
✔️ Audit and continuous updating of your AI practices: We identify biases, ensure regulatory adjustments, and continuously manage your compliance.
✔️ AI-specific impact assessment: We conduct a thorough assessment of the risks associated with your AI systems, tailored to the requirements of the AI Act.
✔️ Transparent user information: We help you clearly inform those affected about the use of an AI system.
✔️ Guidance & awareness-raising through an AI charter: We develop an AI charter tailored to your company to formalize usage and raise awareness among your teams.
Dipeeo guides you toward responsible, controlled, and compliant AI.
Conclusion
The Artificial Intelligence Regulation (AIR) represents a turning point for the AI industry in Europe, imposing high standards in terms of safety and respect for human rights. Its entry into force on August 1, 2024 is a key step towards a more responsible and ethical use of artificial intelligence.
For companies, this means not only a need to adapt quickly, but also an opportunity to strengthen their credibility and competitiveness in the marketplace by adopting practices that comply with European expectations. The future of AI is promising, but it's up to economic players to embrace it safely and responsibly.
To go further:
Discover our webinar on the RIA/AI Act: YouTube link.