AI Assessment Rules & Trustworthy Artificial Intelligence
Oleksandra Mali?henko, Ph.D in Law, LL.M.
?? Global Tech IT Commercial Product Lawyer, AI Lawyer, General Counsel, Head of Legal (Multi-regions, worldwide) ??Black Belt in International IT SaaS Contracting, Compliance, AI Regulation & Data Protection Expert
I'm Oleksandra, an experienced IT/Tech Global Commercial Lawyer with over 18 years of international expertise,? and I’m sharing my legal practice knowledge and insights here.
??Stay Up-to-Date with the Latest IT and Tech Legal Trends for Your business! ??
AI principles was translated into an accessible and dynamic checklist that guides developers and deployers of AI
The EU AI Act, which is part of a broader regulatory framework governing artificial intelligence, aims to ensure that AI technologies are used safely and ethically within the EU.
For example, The CNIL - Commission Nationale de l'Informatique et des Libertés offers organisations a very useful ?? analysis grid through which to assess by themselves the maturity of their artificial intelligence systems https://www.cnil.fr/en/self-assessment
Therefore, the main AI assessment rules and guidelines can be summarized as follows:
?? Risk-Based Classification. The AI Act categorizes AI systems into four risk levels:?
- Unacceptable Risk: Prohibited AI practices (e.g., social scoring by governments).
??- High Risk: AI systems that have significant implications for safety or fundamental rights (e.g., biometric identification, critical infrastructure).
? - Limited Risk: AI systems with minimal risk that require transparency obligations (e.g., chatbots).
?? - Minimal Risk: AI systems that pose little to no risk, with few regulatory obligations.
?? Compliance Requirements for High-Risk AI Systems. High-risk AI systems must comply with specific requirements, including
?? ? - Risk Management System: Implement measures to identify, assess, and mitigate risks.
?? ? - Data Governance: Ensure high-quality datasets to minimize bias and discrimination.
?? ? - Documentation and Reporting: Maintain detailed technical documentation and logs for transparency and accountability.
?? ? - Human Oversight: Incorporate mechanisms for human oversight to mitigate risks effectively.
?? ? - Testing and Validation: Regularly test and validate the AI system to ensure it operates as intended.
?? Transparency and Information Requirements
Developers and operators of AI systems must provide clear information to users about the capabilities and limitations of the system, particularly for high-risk applications.
?? Post-Market Monitoring:
Continuous monitoring of AI systems after deployment to ensure compliance with regulations and to address any emerging risks or issues.
?? Accountability
Clear assignment of responsibility for AI systems, including obligations for providers, users, and any third parties involved in the deployment or use of the AI system.
Assessing and Auditing Process of AI Systems:
To assess and audit AI systems effectively, organizations should follow a structured approach:
1. Risk Assessment: Conduct a comprehensive risk assessment to determine the risk category of the AI system and its potential impact on individuals and society.
2. Documentation: Maintain detailed documentation that includes technical specifications, design choices, data handling procedures, and compliance measures.
3. Data Quality and Governance: Assess the quality and representativeness of the datasets used for training and validating the AI system. Implement data governance protocols to ensure compliance with data protection regulations.
4. Performance Evaluation: Regularly test and evaluate the AI system's performance using benchmarks and real-world data to ensure it functions as intended and meets safety standards.
5. Human Oversight Mechanisms: Establish procedures for human oversight, including decision-making processes and the ability to intervene when necessary.
6. Internal Audits: Conduct periodic internal audits to review compliance with the AI Act requirements, evaluate the effectiveness of risk management strategies, and identify areas for improvement.
7. External Audits and Certification: For high-risk AI systems, consider engaging external auditors or certifying bodies to validate compliance with the EU AI Act requirements.
??????
Organizations should stay abreast of developments in the legislation and adapt their practices accordingly to ensure compliance and enhance the ethical use of AI technologies.
??????
?? If you’re interested to know more, don’t hesitate to get in touch!
????? #AI #AIaw #ITLawyer #LegalIndustry #AITrends
Delivery Head | Project Management Specialist | Agile
1 个月Diving into AI ethics is a must, right? It’s all about keeping things transparent and reliable. What's your take on it?
Head of Custom Brokerage, Import & Export Expert at the cmp.kiev.ua
1 个月Well, seems navigating rules and guidelines for AI is a great challenge ??
Insurance Coverage & Commercial Trial Lawyer; Co-Chair, Data Security & Technology Practice Area; Host, "Cyber Sip"
1 个月An excellent overview, Oleksandra Malichenko, Ph.D in Law, LL.M.!
The quest for trustworthy AI—because who wouldn’t want a transparent robot making decisions for them? ?? At Monyble, we make it easy to build AI solutions that not only follow the rules but also bring your wildest tech dreams to life without requiring a PhD in coding. Let’s just say, we’re here to help you navigate this journey with a bit more flair and a lot less headache!?
Meet the HUMANS behind AI
1 个月It’s not just about using AI, but using it correctly Oleksandra.