PLOT4AI - Library (original) (raw)

Data Quality

Is our data complete, up-to-date, and trustworthy?

Can you avoid the known principle of “garbage in, garbage out”? Your AI system is only as reliable as the data it works with.

If your answer is No or MAYBE, you might be at risk

FLIPCARD

Recommendations

Target Leakage

Can we prevent target leakage?

Target Leakage is present when your features contain information that your model should not legitimately be allowed to use, leading to overestimation of the model's performance. It can occur when information from outside the training dataset is improperly included in the model during training. This can result in an unrealistically high performance during evaluation.

If your answer is No or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Data Continuity

Can the AI model maintain continuous access to data sources after deployment?

If your answer is No or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Update Latency

Can we process new or updated data from external sources without delay?

If your answer is No or MAYBE, you might be at risk

FLIPCARD

Recommendations

Data Legitimacy

Are all required data sources legitimate, authorized, and verified?

If your answer is No or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Data Collection

Can we obtain the data needed to develop or fine-tune the AI model?

Could you face difficulties obtaining certain type of data? This could be due to different reasons such as legal, proprietary, financial, physical, technical, etc. This could put the whole project in danger.

If your answer is No or MAYBE, you might be at risk

FLIPCARD

Recommendations

In the early phases of the project (as soon as the task becomes more clear), start considering which raw data and types of datasets you might need. You might not have the definitive answer until you have tested the model, but it will already help to avoid extra delays and surprises. You might have to involve your legal and financial department. Remember that this is a team effort.

Data Traceability

Can we trace the provenance and lineage of the data used to train or fine-tune the AI model?

AI models require traceability of data sources to ensure ethical usage, reproducibility, and compliance. Without proper data lineage, it is difficult to verify the credibility and accuracy of training data.

If your answer is No or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Copyright, IP & Legal Restrictions

Could our dataset have copyright or other legal restrictions?

Consider any legal, licensing, or privacy constraints that might prevent you from using certain datasets. This also applies to proprietary libraries, tools, or other resources.

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Data Integrity

Can we detect and prevent data tampering across the AI lifecycle?

Data integrity is critical to ensuring that AI systems function as intended. Tampered data, whether during ingestion, transformation, storage, or transfer, can introduce hidden errors, biases, or malicious payloads. AI models built on compromised data may behave unpredictably, yield incorrect results, or violate compliance requirements. Integrity threats may be unintentional (e.g., pipeline errors) or deliberate (e.g., insider sabotage or supply chain attacks).

CIA traid impact:
Integrity

If your answer is No or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Inclusivity

Is our AI system inclusive and accessible?

If your answer is No or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Perception

Could the user perceive the message from the AI system in a different way than intended?

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Learning Curve

Is the AI system easy for users to learn and operate?

If your answer is No or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

AI Interaction Awareness

Are users clearly made aware that they are interacting with an AI system or consuming AI-generated content?

If your answer is No or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

System Transparency for Effective Use

Are users informed about the AI system's reliability, limitations, and risks in a way that enables safe and effective use?

Users need to understand what the AI system can and cannot do, including its intended use, reliability, limitations, and potential risks. Without clear communication, users may place unwarranted trust in the system, misuse it, or be harmed by misleading outputs. This undermines transparency, fairness, safety, and user autonomy. For example, failing to disclose error rates, decision logic, or appropriate use contexts can lead to over-reliance or unsafe behavior, especially in sensitive domains.

If your answer is No or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Linkability

Can the training data be linked to individuals?

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Information Disclosure

Could the AI system infer and reveal information that a person has not explicitly shared?

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Local Restrictions

Could geolocation restrictions or regional regulations impact the implementation of our AI system in other countries?

AI systems often process sensitive data, including personal or location-based information, which may be subject to regional data sovereignty laws and ethical restrictions. Additionally, certain countries may restrict the deployment of AI technologies based on local regulatory frameworks, ethical concerns, or national security considerations. This could limit the usage of your product in those regions.

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Data Minimization

Can we minimize the amount of personal data used while preserving model performance?

The principle of data minimization, as outlined in the General Data Protection Regulation (GDPR) and reflected in many global privacy standards, requires that only data necessary for achieving the system's purpose is collected and processed. However, reducing data too much can sometimes negatively impact the accuracy and performance of AI models, leading to critical or damaging consequences. Balancing regulatory compliance with operational effectiveness is essential to avoid undermining the model's reliability while adhering to privacy principles.

If your answer is No or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Sensitive Data

Are we processing special categories of personal data or sensitive data?

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Automated Decision-Making (ADM)

Could the AI system make decisions with legal or similarly significant effects without human intervention?

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Lawful Basis

Do we have a valid legal basis for processing personal data?

Do you know which GDPR legal ground you can apply?

If your answer is No or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Purpose Limitation

Could we be using personal data for purposes different from those for which it was originally collected?

The principle of purpose limitation, as defined in the General Data Protection Regulation (GDPR) and echoed in many global privacy frameworks, requires that personal data is collected for specified, explicit, and legitimate purposes and not further processed in a way incompatible with those purposes. Data repurposing is a significant challenge when applying this principle. If datasets were originally collected for a different purpose, their reuse without proper consent or legal justification may violate privacy regulations and ethical standards.

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Data Subject Rights

Are we able to comply with all the applicable GDPR data subjects’ rights?

If your answer is No or MAYBE, you might be at risk

FLIPCARD

Recommendations

Privacy Impact Assessment

Could we be deploying the AI system without conducting a required Data Protection Impact Assessment (DPIA)?

The use of AI is more likely to trigger the requirement for a DPIA, based on criteria in Article 35 GDPR. The GDPR and the EDPB’s Guidelines on DPIAs identify both “new technologies” and the type of automated decision-making that produce legal effects or similarly significantly affect persons as likely to result in a “high risk to the rights and freedoms of natural persons”.

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Third-party Data Processing

Are we using third-party providers while processing data from children or other vulnerable individuals?

If your system processes data from children or other vulnerable groups, any third-party providers you rely on (such as libraries, SDKs, or other tools) may also have access to this data. In such cases, you must ensure they comply with relevant privacy regulations like GDPR, COPPA, or similar frameworks. Even if your own system adheres to strong data protection measures, vulnerabilities or non-compliance on the part of third-party providers could expose sensitive data or create ethical risks.

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Metadata

Are we using metadata that could reveal personal data or behavior patterns?

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Privacy Rights

Could we compromise users’ rights to privacy and to a private and family life?

The AI system may intrude on users' right to privacy by exposing sensitive aspects of their private lives, such as personal behaviors, preferences, or relationships, without their explicit consent or awareness. This can occur through excessive surveillance, unintended inferences, profiling, or sharing personal data without proper safeguards. Such compromises may undermine users' autonomy, dignity, and trust in the system, leading to legal, ethical, and reputational consequences for providers.

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Transparent Information

Are we providing sufficient transparency about how the AI model collects, processes, and uses personal data?

Users and stakeholders may not fully understand how data is collected, processed, and utilized, leading to concerns about privacy, accountability, and trust. A lack of transparency can make it difficult to verify whether personal data is being used lawfully or ethically. AI decision-making may be opaque, increasing risks of bias, discrimination, or unfair outcomes.

If your answer is No or MAYBE, you might be at risk

FLIPCARD

Recommendations

Storing of User Data

Are we logging or storing user input data in ways that may violate privacy?

AI systems, particularly Large Language Models (LLMs), may log user inputs and outputs for debugging or model fine-tuning, potentially storing sensitive data without explicit user consent. Logged data could be included in training datasets, making it possible for adversaries to conduct data poisoning attacks, influencing model behavior. Even metadata from logs may reveal sensitive details about users.

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Inaccurate Output

Could the AI system produce inaccurate or misleading outputs that result in privacy violations or harm?

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

For generative AI:

For traditional AI (e.g., classification, regression, or rule-based systems):

Interesting resources/references

Data Transfers

Are we transferring personal data to countries that lack adequate privacy protections?

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Storage Limitation

Can we comply with the storage limitation principle and international data retention regulations?

The principle of storage limitation, as stated in Article 5(e) of the GDPR, requires personal data to be stored only as long as necessary for the intended purpose. Similarly, many global privacy regulations, such as CCPA (California), LGPD (Brazil), and PDPB (India), impose strict rules on data retention and deletion. Do you have a clear understanding of how long you need to keep the data (training data, output data, etc.) and whether you comply with internal, local, national, or international retention requirements?

If your answer is No or MAYBE, you might be at risk

FLIPCARD

Recommendations

Security Testing

Could we be deploying the AI system without testing for adversarial robustness and systemic vulnerabilities?

AI systems can be targeted in unique ways, such as adversarial inputs, poisoning attacks, or reverse-engineering of model outputs. These threats could compromise the system's confidentiality, integrity, and availability, leading to reputational damage or harm to users. Testing for these issues may require specialized expertise, tools, and time, which could affect project timelines.

CIA traid impact:
ConfidentialityIntegrityAvailability

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Plan for AI-specific penetration testing or red-teaming exercises, focusing on adversarial robustness, data governance, and model-specific vulnerabilities. Allocate time in the project for external audits, agreement on scope, and retesting if vulnerabilities are found.

Interesting resources/references

API & Model Interface Security

Are our AI inference APIs and function-calling interfaces securely implemented?

AI systems increasingly rely on APIs for inference (e.g., LLM endpoints), orchestration (e.g., function calls via tools), or dynamic prompt injection (e.g., Model Context Protocol). Poorly secured APIs expose attack surfaces specific to LLMs and other AI models.

Threats include:

Attacks on shared foundational model APIs can impact multiple downstream applications through shared vulnerabilities, hallucination exploits, or jailbreak discovery.

CIA traid impact:
ConfidentialityIntegrityAvailability

If your answer is No or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Confidential Information

Is the AI model suited for processing confidential information?

CIA traid impact:
Confidentiality

If your answer is No or MAYBE, you might be at risk

FLIPCARD

Recommendations

When selecting the algorithm, conduct a thorough analysis to evaluate the risk of algorithmic leakage. For models known to retain training data (e.g., k-nearest neighbors, support vector machines), assess whether sensitive or identifiable information could be exposed through predictions or reverse engineering.

Interesting resources/references

Model Sabotage

Have we protected our AI system against model sabotage?

Model sabotage involves deliberate manipulation or damage to AI systems at any stage, from development to deployment. This can include embedding backdoors, altering model behavior, or exploiting vulnerabilities in training data, third-party tools, or infrastructure.

CIA traid impact:
ConfidentialityIntegrityAvailability

If your answer is No or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Poisoning Attacks

Are we protected from poisoning attacks?

In a poisoning attack, the goal of the attacker is to contaminate the training data or the model generated in the training phase, so that predictions on new data will be modified in the testing phase. This attack could also be caused by insiders. Example: in a medical dataset where the goal is to predict the dosage of a medicine using demographic information, researchers introduced malicious samples at 8% poisoning rate, which changed the dosage by 75.06% for half of the patients.

Other scenarios:

CIA traid impact:
Integrity

If your answer is No or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Model Inversion

Are we protected from model inversion attacks?

CIA traid impact:
Confidentiality

If your answer is No or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Adversarial Examples

Are we protected from adversarial examples?

Adversarial examples are a type of evasion attack where malicious inputs are deliberately crafted to mislead AI models. These inputs are minimally modified, often imperceptible to humans, but can cause the model to produce incorrect or harmful predictions. Examples include researchers demonstrating that carefully designed patterns on accessories, like sunglasses, could deceive facial recognition systems into misidentifying individuals. Such examples are particularly problematic in critical domains like healthcare, finance, and security, where incorrect predictions could lead to severe consequences.

CIA traid impact:
Integrity

If your answer is No or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Jailbreaking

Could the AI system be vulnerable to jailbreak techniques, allowing attackers to bypass safety restrictions?

Attackers can exploit jailbreak techniques to bypass an AI system’s built-in safety constraints, enabling it to generate restricted or harmful content.

CIA traid impact:
ConfidentialityIntegrity

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Prompt Injection

Could the AI system be vulnerable to prompt injection attacks, leading to unauthorized access or manipulation?

AI models, particularly large language models (LLMs), are susceptible to prompt injection attacks, where adversaries craft inputs designed to override model constraints, extract sensitive data, or manipulate system behavior.

CIA traid impact:
ConfidentialityIntegrity

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Environment Unauthorized Access

Is the AI training environment secured against unauthorized access and manipulation?

AI training environments often handle sensitive data and require extensive computational resources. If left unprotected, they become a target for adversaries who may attempt to steal data, modify training sets, or inject adversarial inputs.

CIA traid impact:
ConfidentialityIntegrityAvailability

If your answer is No or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

System Unauthorized Access

Is the deployed AI system protected from unauthorized access and misuse?

Unauthorized access to AI systems can result in data breaches, model theft, and exploitation of sensitive functionalities. Without proper access control, attackers can extract model parameters, manipulate system behavior, or leak confidential data.

CIA traid impact:
ConfidentialityIntegrityAvailability

If your answer is No or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

AI Supply Chain Tools

Could third-party tools, plugins, or dependencies introduce vulnerabilities in our AI system?

Modern AI systems increasingly rely on external tools and plugin interfaces (e.g., Model Context Protocol, LangChain, OpenAI plugins) to expand their capabilities. These interfaces pose unique security risks if not tightly controlled.

Runtime Abuse: If tool or plugin inputs are not strictly validated, LLMs may:

Supply Chain Risks: Third-party plugins and dependencies may contain vulnerabilities or backdoors. Attackers can:

These risks are magnified in open ecosystems where tools are crowd-sourced or rapidly integrated without full vetting.

CIA traid impact:
ConfidentialityIntegrityAvailability

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Unsafe SQL

Could the AI system generate or execute unsafe SQL queries from user input?

CIA traid impact:
ConfidentialityIntegrityAvailability

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Remote Code Execution (RCE)

Could the AI system generate or execute unsafe code based on user input?

CIA traid impact:
ConfidentialityIntegrityAvailability

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Agentic AI Interaction

Could autonomous AI agents access or interact with malicious web content?

CIA traid impact:
ConfidentialityIntegrity

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Agentic AI Memory

Could agent memory be poisoned with malicious or misleading information?

CIA traid impact:
Integrity

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Agentic AI Tools Misuse

Could agents misuse tools or APIs they are authorized to access?

Agents that have access to tools (e.g., file systems, webhooks, APIs) may invoke them in unintended or harmful ways. This misuse can result from adversarial prompts, faulty reasoning, or misunderstood intent. Example: an agent with access to a web browser could issue API delete requests or trigger real-world effects in connected systems.

CIA traid impact:
ConfidentialityIntegrityAvailability

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Agentic AI Hallucinations

Could hallucinated output from one agent propagate and mislead others in multi-agent systems?

In multi-agent systems, one agent’s hallucinated output can become another’s input. This can cause cascading misinformation, particularly if agents defer to each other’s outputs without validation. Example: Agent A misclassifies a vulnerability, Agent B acts on this and takes inappropriate mitigation actions.

CIA traid impact:
Integrity

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Agentic AI Actions Traceability

Can we trace and audit the actions and decisions of autonomous agents in our system?

CIA traid impact:
IntegrityAvailability

If your answer is No or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Agentic AI Malicious Agent

Could a compromised or malicious agent sabotage a multi-agent system?

CIA traid impact:
ConfidentialityIntegrityAvailability

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Agentic AI Unauthorized Access

Could an agent gain access to functions or data beyond its intended permissions?

CIA traid impact:
ConfidentialityIntegrityAvailability

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Resource Overload

Could an attacker or user intentionally overload the AI system’s resources to degrade performance or cause failures?

CIA traid impact:
AvailabilityIntegrity

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Identity Spoofing & Impersonation

Could an attacker or agent impersonate a user or AI identity to gain unauthorized influence?

CIA traid impact:
IntegrityAvailability

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Agentic AI Deceiving Users

Could an agent be misused to manipulate or deceive users?

CIA traid impact:
ConfidentialityIntegrity

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Agent Communication Tampering

Could an attacker intercept or manipulate communications between agents to alter system behavior?

Agents that exchange messages may be vulnerable to communication poisoning, where an attacker injects or modifies messages to alter system behavior. This can mislead agents, propagate misinformation, or trigger unintended actions in chained workflows. Examples include impersonating an agent, sending conflicting commands, or embedding adversarial prompts.

CIA traid impact:
ConfidentialityIntegrity

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

File Upload

Could unsafe file uploads introduce security risks?

AI systems that ingest or process uploaded files, such as PDFs, Word documents, images, or code, are vulnerable to multiple attack vectors:

These threats are particularly relevant when files are processed automatically by LLMs or downstream tools, often without human review.

CIA traid impact:
ConfidentialityIntegrityAvailability

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Model Serialization

Could unsafe deserialization of model artifacts lead to code execution or system compromise?

Models are serialized and transferred between systems for deployment, a stage vulnerable to model serialization attacks. Models are often serialized for storage, sharing, or deployment, using formats like pickle, joblib, ONNX, or TensorFlow SavedModel. However, many serialization formats can embed executable code or unsafe object structures.

If an attacker tampers with a serialized model artifact and it is later deserialized without validation, they may achieve:

These risks are especially severe when models are downloaded from untrusted sources, integrated via ML pipelines, or auto-loaded during CI/CD processes.

CIA traid impact:
ConfidentialityIntegrityAvailability

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Fine-tuning Attacks

Could malicious fine-tuning compromise the safety or alignment of our GenAI model?

CIA traid impact:
ConfidentialityIntegrityAvailability

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

RAG & Vector Databases

Are we protected from vulnerabilities in vector databases and RAG pipelines?

Retrieval-Augmented Generation (RAG) systems combine LLMs with vector databases to enrich answers with external knowledge. However, if the retrieval layer is compromised or poorly validated, it can feed the model misleading, biased, or adversarial content. Untrusted documents in vector stores can serve as indirect prompt injections, while insecure embeddings can allow unauthorized inference or leakage. Additionally, RAG systems may unintentionally disclose proprietary documents retrieved through similarity search.

CIA traid impact:
ConfidentialityIntegrity

If your answer is No or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Input Channel Failure

Could failures in real-time data collection channels disrupt model performance?

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Misinformation

Could AI-generated hallucinations lead to misinformation or decision-making risks?

AI models may generate hallucinations, producing incorrect, misleading, or fabricated information. These errors can undermine trust, propagate misinformation, and lead to unsafe decision-making.

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Interpretability

Could the lack of interpretability in our AI models compromise safety?

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Over-reliance

Can human over-reliance on automated systems lead to failures during emergencies?

Relying too heavily on automation can reduce human involvement and oversight, making it difficult to respond quickly or effectively to unexpected failures or emergency situations.

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

System Failure

In case of system failure, could users be adversely impacted?

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Contextual Robustness

Is our AI model robust and suitable for its intended use across different deployment contexts?

Are you testing the product in a real environment before releasing it? When deploying an AI model, it is critical to ensure that it aligns with the intended use and functions effectively in its operational environment. If the model is trained and tested on data from one context but deployed in a different one, there is a significant risk of performance degradation, or unintended behavior. This is particularly important in cases where environmental changes, unexpected inputs, or shifts in user interaction occur. Additionally, reinforcement learning models may require retraining when objectives or environments deviate slightly from the training setup. Beyond data, other contextual factors like legal, cultural, or operational constraints must be considered to ensure successful deployment.

If your answer is No or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Benchmark Misalignment

Could the AI system's performance on benchmarks be misleading or fail to reflect real-world risks?

AI models often report strong results on standard academic benchmarks, but these benchmarks may not reflect the diversity, complexity, or unpredictability of real-world use cases. Overfitting to test sets, narrow coverage, or outdated benchmarks can lead to misleading performance estimates. As a result, systems may behave unreliably or unfairly once deployed, especially in edge cases, non-English contexts, or under adversarial conditions. This can cause harm, erode trust, and create legal or reputational liabilities.

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Persuasive AI

Could the AI system become persuasive causing harm to users?

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Reward Hacking

Could our AI agents hack their reward functions to exploit the system?

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

One possible approach to mitigating this problem would be to have a “reward agent” whose only task is to mark if the rewards given to the learning agent are valid or not. The reward agent ensures that the learning agent (robot for instance) does not exploit the system, but rather, completes the desired objective. For example: a “reward agent” could be trained by the human designer to check if a room has been properly cleaned by the cleaning robot. If the cleaning robot shuts off its visual sensors to avoid seeing garbage and claims a high reward, the “reward agent” would mark the reward as invalid because the room is not clean. The designer can then look into the rewards marked as “invalid” and make necessary changes in the objective function to fix the loophole.

Interesting resources/references

Child Safety & Age-Appropriate Design

Could the AI system expose children to harmful, inappropriate, or unsafe content or interactions?

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Malicious Use of AI

Could the AI system be misused for malicious purposes such as disinformation, cyberattacks or warfare?

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

CBRNE Threats

Could the AI system accelerate the development of bioweapons or other CBRNE threats?

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Combine these technical safeguards with legal, contractual, and organizational controls to ensure end-to-end risk mitigation. design.

Interesting resources/references

Deepfakes & Synthetic Deception

Could the AI system generate or disseminate deepfakes or synthetic media that mislead users, impersonate individuals, or cause harm?

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Model Toxicity

Could the AI system generate toxic or harmful content?

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Model Deception

Could the AI system deliberately mislead users or hide its capabilities during deployment or evaluation?

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Critical Infrastructure Harm

Could AI decisions result in physical damage, infrastructure failure, or major financial losses?

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

GenAI Version Drift

Do we monitor how version updates from third-party GenAI models can affect our system's behaviour?

If your answer is No or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Loss of Control

Could the development of autonomous AI agents lead to loss of control, concentration of power or rogue behavior?

Autonomous AI systems are increasingly capable of making independent decisions, executing commands, and adapting to changing environments. If misaligned or maliciously designed, these systems may act unpredictably or against human interests.

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Climate & Disaster Resilience

Could environmental phenomena or natural disasters compromise our AI system?

Examples of natural disasters include earthquakes, floods, and fires. These events, as well as environmental phenomena such as extreme heat or cold, may adversely affect the operation of IT infrastructure and hardware systems that support AI systems. Natural disasters may lead to unavailability or destruction of the IT infrastructures and hardware that enables the operation, deployment and maintenance of AI systems. Such outages may lead to delays in decision-making, delays in the processing of data streams and entire AI systems being placed offline. Sources: ENISA

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Implement a disaster recovery plan considering different scenarios, impact, Recovery Time Objective (RTO), Recovery Point Objective (RPO) and mitigation measures.

Interesting resources/references

Unsafe Exploration & Environmental Harm

Could AI agents take actions that unintentionally harm users, the environment or themselves during learning or deployment?

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

CO2 Emissions

Does training and deploying our AI system generate high CO2 emissions?

AI systems, especially large-scale models, require a lot of computational power. It’s important to consider the environmental impact of building and maintaining your system. Does its scope and the benefits it provides justify its emissions? Are you effectively minimizing CO2 emissions throughout your supply chain?

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Safety & Environmental Impact Category

Design Phase

AI Hardware

Is the production of our AI hardware exploiting limited material resources?

AI hardware production relies on rare minerals like cobalt and lithium, which are often extracted at the cost of environmental damage and community exploitation. The short lifespan of AI devices also creates electronic waste and can involve leaking toxic chemicals into ecosystems and harming human health. When assessing your hardware, consider the resource availability and the risks of relying on these materials. Does your current hardware use materials that are becoming harder to source? Could this create future supply chain issues or environmental impact?

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Environmental Footprint

Are we assessing our AI system’s environmental impact across its entire life cycle?

An AI system’s environmental footprint goes beyond its operational phase. A full life cycle assessment (LCA) should account for resource extraction, hardware manufacturing, training, deployment, and end-of-life disposal. Key impact indicators include CO2 emissions, energy and water consumption, and raw material use. Since many AI systems run in mixed-use facilities, properly allocating environmental costs can be complex but necessary for accurate reporting.

If your answer is No or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Deployment, Representation & Sampling Bias

Is the dataset representative of the different real-world groups, populations and environments?

Have you considered the diversity and representativeness of individuals, user groups, and environments in the data? When applying statistical generalisation, the risk exists of making inferences due to misrepresentation, for instance: a postal code where mostly young families live can discriminate the few old families living there because they are not properly represented in the group.

If your answer is No or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Incorrect Attribution

Could the AI system incorrectly attribute actions to individuals or groups?

Your AI system could adversely affect individuals by incorrectly attributing actions or facts to them. For example, a facial recognition system may misidentify someone, or a flawed risk prediction model could negatively impact a person’s opportunities or reputation.

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Unfair Disproportion

Could certain groups be disproportionately affected by the outcomes of the AI system?

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Institutional Bias

Could our AI system reinforce systemic inequalities?

Institutional biases, like racism or sexism, are often rooted in organizational structures and policies. Could such biases, intentionally or unintentionally, be embedded or influence the design or the functioning of the system?

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Abstraction Traps

Could our AI system oversimplify real-world problems?

An AI systems can overlook the social contexts in which they operate, leading to unintended consequences. Specifically, watch out for these types of abstraction traps:

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Construction Validity Bias

Could our AI system accurately capture the factors it's designed to measure?

Construct validity bias occurs when a feature or target variable fails to adequately represent the concept it is intended to measure, leading to inaccurate measurements and potential biases. For example, measuring socioeconomic status using income alone overlooks important factors such as wealth and education. This bias can arise during various stages of the AI lifecycle and should be addressed early on to improve system accuracy.

If your answer is No or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Historical Bias

Could the AI system reinforce historical inequalities embedded in the data?

Historical bias occurs when AI systems mirror or exacerbate past social and cultural inequalities, even when using accurate data. For example, an AI healthcare tool trained on historical patient data may reflect disparities in access to care. Minority groups, underrepresented in the data due to systemic inequities, may receive less accurate diagnoses, perpetuating racial bias even without explicit racial features.

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Labeling Bias

Can data be labeled consistently?

If your answer is No or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Proxy Variables

Could the system be using proxy variables that reflect sensitive attributes or lead to indirect discrimination?

Proxy variables are features used as stand-ins for harder-to-measure characteristics. While proxies can be useful for model performance, they may be highly correlated with sensitive attributes such as race, gender, religion, age, or socioeconomic status. This can lead to indirect or proxy discrimination, where individuals from protected groups are disproportionately harmed despite sensitive data not being explicitly included.

For example, ZIP code, school name, or browsing history may function as proxies for race or income level. In such cases, the system might appear 'neutral' but still replicate or amplify historical inequalities. Proxy bias is especially insidious because it is often unintentional and hidden in seemingly innocuous variables.

Generative models can also internalize and reproduce these biases in subtle ways, such as generating different responses for identical inputs that differ only by proxy cues.

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Design Choices

Could the AI system’s design choices lead to unfair outcomes?

Biases can emerge from an AI model’s design and training, even if the dataset is unbiased. Design choices and development processes can introduce various biases that affect fairness and accuracy.

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Over-reliance

Could we over-rely on early evaluation results or AI-generated outputs?

Biases can emerge during the evaluation and validation stages of AI models, especially when over-relying on early test results or automated AI decisions. This can lead to misleading conclusions. Specific biases include:

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Popularity Bias

Could popularity bias reduce diversity in system's recommendations?

Recommendation systems often amplify what’s already popular, making it harder for niche or lesser-known options to be discovered. This can reduce diversity, personalization, and fairness in recommendations, limiting users’ exposure to a broader range of choices.

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Diversity of Opinions

Is the AI system designed to support multiple viewpoints and narratives?

An AI system that does not consider or promote diverse viewpoints and narratives risks reinforcing biases, perpetuating stereotypes, or marginalizing specific groups. Such systems might unintentionally amplify dominant cultural, religious, or linguistic perspectives while excluding or suppressing minority voices. For example, content recommendation systems may disproportionately highlight mainstream viewpoints, reducing exposure to diverse cultural or ideological perspectives. This could hinder freedom of opinion and expression, harm cultural diversity, and lead to discriminatory outcomes.

If your answer is No or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Social Division

Could our AI system contribute to social division or rivalry?

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

People Categorization

Could our AI system automatically label or categorize people?

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Right to Work

Could the AI system affect employment conditions, labor rights, or job opportunities?

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Human Dignity

Could our AI system fail to uphold and respect human dignity?

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Right to Democracy

Could the AI system affect democracy or have an adverse impact on society at large?

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

User Redress & Remedy

Do we offer users and accessible way to contest AI decisions or seek redress?

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Design redress mechanisms that allow affected individuals to report harm, request compensation, or demand system correction. This includes enabling redress even for those indirectly harmed (e.g., via biased profiling). Ensure accessibility and transparency of the redress process, define timelines and escalation paths, and document how redress outcomes are used to improve system performance.

Interesting resources/references

Right to Life

Could the system have an impact on decisions that affect life, health, or personal safety?

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Freedom of Expression

Could the AI system limit, suppress or distort users’ freedom of expression?

Consider whether your AI system’s moderation, recommendation, or censorship mechanisms may inadvertently restrict or distort users' ability to express themselves freely.

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Access to Essential Services

Could our AI system affect access to services such as healthcare, housing, insurance, benefits or education?

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Human Autonomy

Could the AI system interfere with users’ autonomy influencing their decision-making process?

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Freedom of Thought

Could the AI system promote certain values or beliefs on users?

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Vulnerable Groups

Could the AI system negatively impact vulnerable groups or fail to protect their rights?

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Children’s Rights

Could the AI system fail to uphold the rights and best interests of children?

Children interacting with AI systems require special protections to ensure their rights, safety, and well-being are preserved. AI systems used by or designed for children must prioritize their best interests, such as ensuring age-appropriate content, safeguarding their privacy, and fostering their ability to share, learn, and express themselves freely. A failure to address these factors could result in harm, exploitation, or the suppression of their rights. For example, an AI system might expose children to inappropriate content, fail to protect their personal data, or limit their ability to engage in meaningful learning and expression.

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Proportionality

Is the development and use of the AI system proportionate to its intended purpose and impact on rights?

If your answer is No or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Behavioral Data

Does the AI system use behavioral data in ways that may raise ethical, privacy, or human rights concerns?

If your answer is Yes or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Unclear Task Definition

Is the AI system's task clearly defined, with well-scoped objectives and boundaries?

If your answer is No or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Stakeholder Involvement

Have we identified and involved all key stakeholders relevant to this phase of the AI lifecycle?

If your answer is No or MAYBE, you might be at risk

FLIPCARD

Recommendations

Training and Oversight Readiness

Have all relevant staff and users received adequate training to understand, oversee, and responsibly interact with the AI system?

Individuals involved in the development, deployment, operation, or use of AI systems must understand their functionality, risks, and limitations. Without adequate training, staff may misuse the system, fail to detect errors, or be unable to intervene effectively. This undermines human oversight, accountability, and compliance with regulatory requirements. Article 4 of the EU AI Act emphasizes the need for AI literacy, particularly for those responsible for high-risk systems.

If your answer is No or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

AI Agents’ Feedback

Do we have qualified people available to supervise the behavior of AI agents and provide feedback during learning?

If your answer is No or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Oversight Capacity

Do we have the resources and processes to effectively oversee AI decision-making?

If your answer is No or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Escalation Path

Is there a well-defined process to escalate AI-related failures or unexpected outcomes?

If your answer is No or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Responsibility

Have we defined who is accountable for the AI system’s decisions and outcomes?

AI outputs can lead to mistakes or even cause harm. In such cases, is it clear who is responsible within your organization? Are accountability structures clearly defined and documented?

If your answer is No or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Continuous Assessment

Do we regularly review whether the AI system’s goals, assumptions, and impacts are still appropriate?

AI models and their objectives may drift from their original intent, making human oversight crucial to ensure ongoing alignment with ethical and business objectives. Are there periodic human-led reviews in place to monitor AI system behavior, validate outcomes, and reassess goals? Human oversight should play an active role in detecting unintended consequences, adjusting governance policies, and maintaining accountability throughout the AI system’s lifecycle.

If your answer is No or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Human Override Mechanisms

Can human operators safely interrupt or override the AI system at any time?

If your answer is No or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Contestability of AI Decisions

Could users contest or challenge the decisions made by the AI system?

Some AI systems make or support decisions that significantly affect individuals, such as in hiring, lending, or criminal justice. If users cannot challenge these decisions or request human review, the system may violate oversight obligations and erode trust. Lack of contestability undermines accountability and may breach Article 22(3) of the GDPR or Article 14 of the EU AI Act, both of which require mechanisms for human intervention and review.

If your answer is No or MAYBE, you might be at risk

FLIPCARD

Recommendations

Charter of Fundamental Rights of the European Union

Interesting resources/references

Liability Risk

Have we assessed our legal liability for damages caused by our AI system?

If your answer is No or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Lack of MLOps

Do we have adequate resources and MLOps practices in place to manage, monitor, and maintain our AI system?

MLOps (Machine Learning Operations) refers to the engineering and governance practices required to reliably develop, deploy, and monitor machine learning models in production. Without proper MLOps, organizations may face:

MLOps is especially important for high-risk AI applications under the EU AI Act, where continuous monitoring, retraining, and documentation are legal obligations.

If your answer is No or MAYBE, you might be at risk

FLIPCARD

Recommendations

Interesting resources/references

Shared Responsibility

If we plan to deploy a third-party AI tool, have we assessed our shared responsibility for its potential impact on users?

If you use a third-party tool you might still have a responsibility towards the users. Think about employees, job applicants, patients, etc. It is also your responsibility to make sure that the AI system you choose won't cause harm to the individuals.

If your answer is No or MAYBE, you might be at risk

FLIPCARD

Recommendations

If personal data is involved, review which ones are your responsibilities (look into art. 24 and 28 GDPR).

You can also start by checking: