This tool responds to the GC AI Strategy for the Federal Public Service 2025-2027, the GC Data Strategy for the Federal Public Service (2023-2026) Mission 3.3 for responsible, transparent and ethical data stewardship to maintain trust, the GC Directive on automated decision-making , the GC Guide on the use of generative artificial intelligence , the GC Algorithmic impact assessment tool , NIST AI Risk Management Framework , the International Scientific Report on the Safety of Advanced AI , and the EU Artificial Intelligence Act.

This tool is suitable for a first screening level estimation of the AI safety of a particular AI application.

You may also find the following companion tools useful:



Do you work with Artificial Intelligence (AI)? Are you looking to make it future proof? The FAIRER data principles will help you!

This tool helps you to assess the AI safety level of an AI application and get tips on how you could increase the AI safety.

The tool is discipline-agnostic, making it relevant to any field.

The checklist will take 15-30 minutes to complete, after which you will receive a quantitative summary of the AI safety level your application, and tips on how you can improve the level of AI safety. No information is saved on our servers, but you will be able to save the results of the assessment, including tips for improvement, to your local computer and add notes for future reference.

Version 1.0

CRediT Author statement



ARTIFICIAL INTELLIGENCE (AI) BACKGROUND

Artificial Intelligence (AI) simulates human-like intelligence in machines, enabling them to learn, reason, solve problems, perceive environments, and make autonomous or semi-autonomous decisions. AI is a machine-based system that, for explicit or implicit objectives, infers from the input it receives, how to generate outputs such as predictions, content, recommendations, or decisions that can influence physical or virtual environments. AI systems differ in their autonomy (ability to operate without human intervention) and adaptiveness (capacity to improve or modify behavior after deployment). (OECD 2024)

AI includes knowledge-based systems (which use human-curated domain knowledge, rules, facts, and relationships to simulate expert decision-making) and machine learning systems (which learn patterns from data and generalize to new tasks without being explicitly programmed). The latter include neural networks (algorithms inspired by the structure and function of the human brain and used for pattern recognition and classification tasks) and deep learning (a subfield of neural networks that uses many layers of representation to learn complex features from large datasets).

AI application areas include natural language processing (NLP), computer vision, speech recognition, intelligent decision support systems, intelligent robotic systems, predictive analytics, and recommendation systems.

Generative AI is a subfield of artificial intelligence focused on creating new content (e.g., text, images, audio, video, music, speech, computer code, or synthetic data) based on patterns learned from input data.

Automation (not AI) uses technology to execute predefined, often repetitive tasks with minimal human intervention. Automation systems do not learn, adapt, or infer; they follow static rules or programmed sequences. Examples include robotic process automation (RPA), scripting and macros, workflow engines and batch processing, industrial robotics and control systems and algorithmic tasks that do not involve learning or adaptation.

NOTE: This checklist is not designed for assessment of automation applications that do not use AI.


Checklist Questions

1. I am:
A data steward
A policy advisor
A researcher
A scientist
An AI developer
An AI end-user
Other
2. I use this AI application during one or more stages of the AI lifecycle (NIST):

Lifecycle and Key Dimensions of an AI System. The two inner circles show AI systems’ key dimensions and the outer circle shows AI lifecycle stages: Plan & Design, Collect & Process Data, Build & Use Model, Verify & Validate, Deploy & Use, Operate & Monitor. Ideally, risk management efforts start with the Plan & Design function in the application context and are performed throughout the AI system lifecycle.

SOURCE: NIST, modified from OECD (2022) Framework for the Classification of AI systems.
  Plan and Design
  Collect and Process Data
  Build and Use Model
  Verify and Validate
  Deploy and Use
  Operate and Monitor
  Use or Impacted by
3. This AI application is used in:
  Administrative decision-making
Policy formulation
Project prioritization
Scenario analysis
Science based operations
Research
A system operating in a test environment
Other
4. I have completed an Algorithmic Impact Assessment and the score was:
0-25% (level I, Little to no impact).
26-50% (Level II, moderate impact).
51-75% (Level III, high impact).
76%-100% (Level IV, very high impact).
Not completed
5. The practical use of this AI application is for:
  Art and creativity
  Autonomous vehicles
  Finance
  Healthcare
  Image and video recognition
  Natural Language Processing (NLP)
  Generative AI
Other
6. This AI application embodies principles characteristic of trustworthy AI:

Trustworthiness: nine principles characteristic of trustworthy AI systems. Valid & Reliable is a necessary condition of trustworthiness and is shown as the base for the principles: safe, secure & resilient, explainable & interpretable, privacy-enhanced, and fair & managed bias. Accountable & Transparent is shown as a vertical box relating to those six principles. Altogether, 150 properties of trustworthiness have been identified across all seven principles (Newman, 2023) . The 8th and 9th principles (Human-centred values and Inclusive growth, sustainable development, and well-being) are shown as underlying all the other principles. Figure adapted from NIST, European Commission, and OECD.
  Safe
  Secure & resilient
  Explainable & interpretable
  Privacy-enhanced
  Fairness and management of bias
  Valid & Reliable
  Accountable & Transparent
  Human-centred values
  Inclusive growth, sustainable development, and well-being
7. This AI application was designed taking into account:
The positive and negative impacts on end users
All foreseeable use cases of the AI application
8. Designated humans have the ultimate responsibility for all decisions and outcomes from this AI application:
Responsibilities are explicitly defined between the AI system and human(s), and how they are shared.
Human responsibility will be preserved for final decisions that affect a person’s life, quality of life, health, or reputation.
Humans are always able to monitor, control, and deactivate systems.
Significant decisions made by the AI system are explained.
Significant decisions made by the AI system are able to be overridden.
Significant decisions made by the AI system are appealable.
Significant decisions made by the AI system are reversible.
9. The design and use of this AI application embodies transparency and engenders trust:
The purpose, limitations, and biases of the AI system are explained in plain language.
Data sources have unambiguous respected sources, and biases are known and explicitly stated.
Data used for training are updated to suit the appropriate use cases
Algorithms and models are appropriate and verifiable.
Confidence level and context are presented for humans to base decisions on.
Transparent justification for recommendations and outcomes are provided.
Straightforward and interpretable monitoring systems are provided.
Humans are aware when they are being monitored or surveilled for the purpose of data collection or performance
Humans can easily discern when they are interacting with the AI system vs. a human.
Humans can easily discern when and why the AI system is taking action and/or making decisions.
Improvements are made regularly to meet human needs and technical standards.
10. The most relevant method used for this AI application is (ordered from safest to least safe):
  Rule-based expert system
  Machine Learning
  Neural Network
  Deep Learning
  Computer Vision
  Natural Language Processing (NLP) or Large Language Model (LLM)
  Don't know
11. This AI application uses tools and methods that are inherently designed for AI tasks to build, train, and deploy AI models:
  AI Development and Deployment Platforms
  AI Frameworks for Edge and Mobile
  AI Research and Experimentation Platforms
  AI-Specific Hardware
  Automated Machine Learning (AutoML)
  Computer Vision Libraries
  Deep Learning Libraries
  Machine Learning Frameworks
  Natural Language Processing (NLP) Tools
  Reinforcement Learning Libraries
  Simple AI tasks
  Recommendation System
  Robotics
  Speech recognition
Other
  12. The capability vs. inherent limitation of this AI application is properly managed to mitigate negative impacts:
  Data-Driven Decision Making vs. Human Judgment
  Efficiency vs. Ethical Considerations
  Personalization vs. Privacy
  Automation vs. Human Interaction
  Common Sense Reasoning vs. Causality
  Data Processing vs. Data Quality
  Creativity vs. Hallucination
  Knowledge vs. Inconsistency
  Optimization vs. Context Awareness
  Predictability vs. Innovation
  Other (Choices 11-26)
13. The data used in this AI application are available:
Raw input dataset(s) are openly available.
Input data provenance is identified
AI generated dataset(s) are openly available
14. The data used in this AI application comply with FAIRER principles as described by the FAIRER-Aware Data Assessment Tool
FAIRER = Findable, Accessible, Interoperable, Ethical, and Reproducible
15. I have completed FAIRER-Aware Reproducibility Checklist and the score was:
0-25% (Poor)
26-50% (Low)
51-75% (Good)
76%-100% (High)
Not completed
16. This AI application has been tested for model performance using:
  Benchmark measurements
  Adversarial attack
  Auditing
  Field testing
  Human evaluation
  Other
Not tested
17. I have read AI legislation, international and government references, and other resources:
  International
  Government of Canada
  Europe
  U.S. Government
  National Institute of Standards and Technology (NIST)
  National Academies of Sciences, Engineering, and Medicine (NASEM)
  Other

Your Notes

  • Add any notes you may have here. These notes will be included when you print and save your results to your local computer. No information will be saved to our server. Feel free to capture any thoughts or insights that you'd like to remember or revisit later.
Your use of this AI application is 0% safe



Scoring Rules