Following a piloting process where over 350 stakeholders participated, an earlier prototype of the list was revised and translated into a tool to support AI developers and deployers in developing Trustworthy AI.
The tool supports the actionability the key requirements outlined by the Ethics Guidelines for Trustworthy Artificial Intelligence (AI), presented by the High-Level Expert Group on AI (AI HLEG) presented to the European Commission, in April 2019. The Ethics Guidelines introduced the concept of Trustworthy AI, based on seven key requirements:
- human agency and oversight
- technical robustness and safety
- privacy and data governance
- diversity, non-discrimination and fairness
- environmental and societal well-being and
Through the Assessment List for Trustworthy AI (ALTAI), AI principles are translated into an accessible and dynamic checklist that guides developers and deployers of AI in implementing such principles in practice. ALTAI will help to ensure that users benefit from AI without being exposed to unnecessary risks by indicating a set of concrete steps for self-assessment.
The ALTAI is also available in a web-based tool version.