Examples:
- Generating art
- Music composition
- Style transfer
This tool responds to the GC AI Strategy for the Federal Public Service 2025-2027, the GC Data Strategy for the Federal Public Service (2023-2026) Mission 3.3 for responsible, transparent and ethical data stewardship to maintain trust, the GC Directive on automated decision-making , the GC Guide on the use of generative artificial intelligence , the GC Algorithmic impact assessment tool , NIST AI Risk Management Framework , the International Scientific Report on the Safety of Advanced AI , and the EU Artificial Intelligence Act.
This tool is suitable for a first screening level estimation of the AI safety of a particular AI application.
You may also find the following companion tools useful:
Do you work with Artificial Intelligence (AI)? Are you looking to make it future proof? The FAIRER data principles will help you!
This tool helps you to assess the AI safety level of an AI application and get tips on how you could increase the AI safety.
The tool is discipline-agnostic, making it relevant to any field.
The checklist will take 15-30 minutes to complete, after which you will receive a quantitative summary of the AI safety level your application, and tips on how you can improve the level of AI safety. No information is saved on our servers, but you will be able to save the results of the assessment, including tips for improvement, to your local computer and add notes for future reference.
Version 1.0
ARTIFICIAL INTELLIGENCE (AI) BACKGROUND
Artificial Intelligence (AI) simulates human-like intelligence in machines, enabling them to learn, reason, solve problems, perceive environments, and make autonomous or semi-autonomous decisions. AI is a machine-based system that, for explicit or implicit objectives, infers from the input it receives, how to generate outputs such as predictions, content, recommendations, or decisions that can influence physical or virtual environments. AI systems differ in their autonomy (ability to operate without human intervention) and adaptiveness (capacity to improve or modify behavior after deployment). (OECD 2024)
AI includes knowledge-based systems (which use human-curated domain knowledge, rules, facts, and relationships to simulate expert decision-making) and machine learning systems (which learn patterns from data and generalize to new tasks without being explicitly programmed). The latter include neural networks (algorithms inspired by the structure and function of the human brain and used for pattern recognition and classification tasks) and deep learning (a subfield of neural networks that uses many layers of representation to learn complex features from large datasets).
AI application areas include natural language processing (NLP), computer vision, speech recognition, intelligent decision support systems, intelligent robotic systems, predictive analytics, and recommendation systems.
Generative AI is a subfield of artificial intelligence focused on creating new content (e.g., text, images, audio, video, music, speech, computer code, or synthetic data) based on patterns learned from input data.
Automation (not AI) uses technology to execute predefined, often repetitive tasks with minimal human intervention. Automation systems do not learn, adapt, or infer; they follow static rules or programmed sequences. Examples include robotic process automation (RPA), scripting and macros, workflow engines and batch processing, industrial robotics and control systems and algorithmic tasks that do not involve learning or adaptation.
NOTE: This checklist is not designed for assessment of automation applications that do not use AI.
1. I am: | |
---|---|
A data steward | |
A policy advisor | |
A researcher | |
A scientist | |
An AI developer | |
An AI end-user | |
Other | |
2. I use this AI application during one or more stages of the AI lifecycle (NIST): | |
![]() Lifecycle and Key Dimensions of an AI System. The two inner circles show AI systems’ key dimensions and the outer circle shows AI lifecycle stages: Plan & Design, Collect & Process Data, Build & Use Model, Verify & Validate, Deploy & Use, Operate & Monitor. Ideally, risk management efforts start with the Plan & Design function in the application context and are performed throughout the AI system lifecycle. SOURCE: NIST, modified from OECD (2022) Framework for the Classification of AI systems. |
|
Plan and Design | |
Collect and Process Data | |
Build and Use Model | |
Verify and Validate | |
Deploy and Use | |
Operate and Monitor | |
Use or Impacted by | |
3. This AI application is used in: | |
Administrative decision-making | |
Policy formulation | |
Project prioritization | |
Scenario analysis | |
Science based operations | |
Research | |
A system operating in a test environment | |
Other | |
4. I have completed an Algorithmic Impact Assessment and the score was: | |
0-25% (level I, Little to no impact). | |
26-50% (Level II, moderate impact). | |
51-75% (Level III, high impact). | |
76%-100% (Level IV, very high impact). | |
Not completed | |
5. The practical use of this AI application is for: | |
Art and creativity | |
Autonomous vehicles | |
Finance | |
Healthcare | |
Image and video recognition | |
Natural Language Processing (NLP) | |
Generative AI | |
Other | |
6. This AI application embodies principles characteristic of trustworthy AI: | |
![]() Trustworthiness: nine
principles characteristic of trustworthy AI systems. Valid &
Reliable is a necessary condition of
trustworthiness and is shown as the base for the principles: safe, secure &
resilient, explainable & interpretable,
privacy-enhanced, and fair & managed bias. Accountable & Transparent is
shown as a vertical box relating to those six
principles. Altogether, 150 properties of trustworthiness have been
identified across all seven principles
(Newman, 2023) . The 8th and 9th principles
(Human-centred values and Inclusive growth, sustainable development, and
well-being)
are shown as underlying all the other principles. Figure adapted from
NIST,
European Commission, and
OECD.
|
|
Safe | |
Secure & resilient | |
Explainable & interpretable | |
Privacy-enhanced | |
Fairness and management of bias | |
Valid & Reliable | |
Accountable & Transparent | |
Human-centred values | |
Inclusive growth, sustainable development, and well-being | |
7. This AI application was designed taking into account: | |
The positive and negative impacts on end users | |
All foreseeable use cases of the AI application | |
8. Designated humans have the ultimate responsibility for all decisions and outcomes from this AI application: | |
Responsibilities are explicitly defined between the AI system and human(s), and how they are shared. | |
Human responsibility will be preserved for final decisions that affect a person’s life, quality of life, health, or reputation. | |
Humans are always able to monitor, control, and deactivate systems. | |
Significant decisions made by the AI system are explained. | |
Significant decisions made by the AI system are able to be overridden. | |
Significant decisions made by the AI system are appealable. | |
Significant decisions made by the AI system are reversible. | |
9. The design and use of this AI application embodies transparency and engenders trust: | |
The purpose, limitations, and biases of the AI system are explained in plain language. | |
Data sources have unambiguous respected sources, and biases are known and explicitly stated. | |
Data used for training are updated to suit the appropriate use cases | |
Algorithms and models are appropriate and verifiable. | |
Confidence level and context are presented for humans to base decisions on. | |
Transparent justification for recommendations and outcomes are provided. | |
Straightforward and interpretable monitoring systems are provided. | |
Humans are aware when they are being monitored or surveilled for the purpose of data collection or performance | |
Humans can easily discern when they are interacting with the AI system vs. a human. | |
Humans can easily discern when and why the AI system is taking action and/or making decisions. | |
Improvements are made regularly to meet human needs and technical standards. | |
10. The most relevant method used for this AI application is (ordered from safest to least safe): | |
Rule-based expert system | |
Machine Learning | |
Neural Network | |
Deep Learning | |
Computer Vision | |
Natural Language Processing (NLP) or Large Language Model (LLM) | |
Don't know | |
11. This AI application uses tools and methods that are inherently designed for AI tasks to build, train, and deploy AI models: | |
AI Development and Deployment Platforms | |
AI Frameworks for Edge and Mobile | |
AI Research and Experimentation Platforms | |
AI-Specific Hardware | |
Automated Machine Learning (AutoML) | |
Computer Vision Libraries | |
Deep Learning Libraries | |
Machine Learning Frameworks | |
Natural Language Processing (NLP) Tools | |
Reinforcement Learning Libraries | |
Simple AI tasks | |
Recommendation System | |
Robotics | |
Speech recognition | |
Other | |
12. The capability vs. inherent limitation of this AI application is properly managed to mitigate negative impacts: | |
Data-Driven Decision Making vs. Human Judgment | |
Efficiency vs. Ethical Considerations | |
Personalization vs. Privacy | |
Automation vs. Human Interaction | |
Common Sense Reasoning vs. Causality | |
Data Processing vs. Data Quality | |
Creativity vs. Hallucination | |
Knowledge vs. Inconsistency | |
Optimization vs. Context Awareness | |
Predictability vs. Innovation | |
Other (Choices 11-26) | |
13. The data used in this AI application are available: | |
Raw input dataset(s) are openly available. | |
Input data provenance is identified | |
AI generated dataset(s) are openly available | |
14. The data used in this AI application comply with FAIRER principles as described by the FAIRER-Aware Data Assessment Tool | |
FAIRER = Findable, Accessible, Interoperable, Ethical, and Reproducible | |
15. I have completed FAIRER-Aware Reproducibility Checklist and the score was: | |
0-25% (Poor) | |
26-50% (Low) | |
51-75% (Good) | |
76%-100% (High) | |
Not completed | |
16. This AI application has been tested for model performance using: | |
Benchmark measurements | |
Adversarial attack | |
Auditing | |
Field testing | |
Human evaluation | |
Other | |
Not tested | |
17. I have read AI legislation, international and government references, and other resources: | |
International | |
Government of Canada | |
Europe | |
U.S. Government | |
National Institute of Standards and Technology (NIST) | |
National Academies of Sciences, Engineering, and Medicine (NASEM) | |
Other |
Scoring Rules
Your Notes