List of Harms#

There are many ways in which machine learning systems can be harmful. This page contains a (non-exhaustive) list of hazards and harms, organized by the moral value the harm threatens.

Attribution

This list was inspired by and adapted from Microsoft’s harms modeling framework.

Safety#

Hazard

Harm

Description

Considerations

Example

Resources

Lacking accuracy

Injury

System errors can lead to physical, emotional, or psychological injury.

How do system errors impact people?

An ilness is misdiagnosed, leading to unnecessary treatment.

Raji et al.1

Inadequate testing

Injury

Real-world testing of system failure modes insufficiently considers a diverse set of users and scenarios.

Are all failure modes adequately tested? How are users impacted by a system failure? Are human interventions possible?

An autonomous vehicle that is tested in the public sphere for new data collection kills a pedestrian, because the system did not recognize the pedestrian crossing outside of a sidewalk.

Raji et al.1

Fairness#

Hazard

Harm

Description

Considerations

Example

Resources

Social bias

Allocation harm

The system results in unfair allocation of opportunities, resources, or information and amplification of power inequality.

Are some groups allocated fewer opportunities than others?

A resume screening application consistently ranks male candidates higher than female candidates.

Algorithmic Fairness

Social bias

Quality-of-service harm

The system disproportionately fails for certain groups of people.

Does the system work equally well for different demographics, particularly those defined by sensitive characterises?

A facial recognition system disproportionately misidentifies faces of black women compared to white men 2

Algorithmic Fairness

Social bias

Stereotyping harm

The system reinforces negative societal stereotypes.

Does the system return output that reinforces undesirable stereotypes?

A search engine returns only pictures of men when prompted with “CEO”, reinforcing stereotypes.

Algorithmic Fairness

Social bias

Denigration harm

The system is actively derogatory or offensive.

In what ways could the system output be considered offensive for stakeholders?

An image tagging system tags a photo of black people as an animal.

Algorithmic Fairness

Social bias

Representation harm

The development or usage of the system over- or under-represents certain groups of people.

Are some groups overburdened compared to others, even if they do not benefit from the system?

Residents of an indigenous reservation have been subject to intense biomedical data collection, but this has not yielded any significant improvements in health outcomes amongst community members 3

Algorithmic Fairness

Feature Selection and Opaque Decision-Making

Procedural harm

The system uses features that are arbitrary or otherwise violate social norms.

Are the features relevant to the problem at hand? Are they used in fair ways?

An arbitrary feature is included in a resume selection application.

4

Transparency#

Hazard

Harm

Description

Considerations

Example

Resources

Opaque Automated Decision-Making

Loss of Effective Remedy

An inability to explain the rationale or lack of opportunity a contest a decision.

How might people understand the reasoning for decisions made by this technology? How might an individual that relies on this technology explain the decisions it makes? How could people contest or question a decision this technology makes?

Automated prison sentence or pre-trial release decision is not explained to the accused.

Opaque Automated Decision-Making

Misguided Trust in Automation

Misguided beliefs can lead users to overtrust the reliability of a digital agent.

How could sole dependence on an artificial agent impact a person?

A chat bot could be relied upon for relationship advice or mental health counseling instead of a trained professional.

Accountability#

Hazard

Harm

Description

Considerations

Example

Resources

Opaque Automated Decision-Making

Lack of Accountability

Opaque decision-making related to the development and usage of a machine learning system hinders accountability.

Is the development process transparent? Can decisions be explained? Is it clear who is responsible for what aspects of the system?

A machine learning model is used to reject lean applications and responsibility is transferred to the incomprehensibility of the model.

No provisions for auditing

Lack of accountability

Third parties are unable to review the behavior of an algorithm.

Are detailed documentation and technically suitable API’s available and to whom?

A large social media platform does not provide documentation or suitable APIs to audit fairness of their algorithms.

No mechanisms for recourse

Lack of redress

Stakeholders are unable to ask for compensation or reparation for an undesirable decision or situation.

Who is responsible if users are harmed by this product? What will the reporting process and process for recourse be?

A user is banned from a social media platform and is unable to contact the platform to contest the decision.

Privacy#

Hazard

Harm

Description

Considerations

Example

Resources

Insecure and/or Redundant Data Collection, Storage, Aggregation, and Processing

Loss of Data Privacy

Reveal personal information a person has not consented to share.

How could this technology use information to infer portions of a private life? How could decisions based upon these inferences expose things that a person does not want made public?

A language model trained on large amounts of data reveals personally identifiable information when queried for specific examples 5

Lack of a data retention policy

Never forgiven

Digital files or records are never deleted.

What and where is data being stored from this product, and who can access it? How long is user data stored after technology interaction? How is user data updated or deleted?

A teenager’s social media history could remain searchable long after they have outgrown the platform.

Adversarial use

Identity Theft

Loss of control over personal credentials, reputation, and/or representation.

How might an individual be impersonated with this technology? How might this technology mistakenly recognize the wrong individual as an authentic user?

A synthetic voice could mimic the sound of a person’s voice and be used to access a bank account.

Data leak

Public Shaming

Reveal private, sensitive, or socially inappropriate information.

How might movements or actions be revealed through data aggregation?

A fitness app could reveal a user’s GPS location on social media.

Surveillance

Loss of Freedom of Movement with Desired Anonymity

An inability to navigate the physical or virtual world with desired anonymity.

In what ways might this technology monitor people across physical and virtual spaces?

A facial recognition system is used to monitor civilians.

Liberty#

Hazard

Harm

Description

Considerations

Example

Resources

Forced Participation and Surveillance

Loss of Freedom of Movement

Requiring participation in the use of technology or surveillance to take part in society.

How might use of this technology be required for participation in society or organization membership?

Biometric enrollment in a company’s meeting room transcription AI is a stipulated requirement in job offer letter.

Forced Social Control

Inability to Fully Develop Personality

Reinforcing conformity and connotations towards particular personality traits.

What types of personal or behavioral data might feed this technology? How would it be obtained? What outputs would be derived from this data? Is this technology likely to be used to encourage or discourage certain behaviors?

Authoritarian government uses social media and e-commerce data to determine a “trustworthiness” score based on where people shop and who they spend time with.

Limited Expression of Perspectives

Loss of Freedom of Expression

Amplification of majority opinions and an inability to express a unique perspective.

How might this technology amplify majority opinions or “group-think”? Conversely, how might unique forms of expression be suppressed. In what ways might the data gathered by this technology be used in feedback to people?

Limited options for gender in an automated loan application system inhibits self-expression of a person’s diversity.

Resources

Sustainability#

Hazard

Harm

Description

Considerations

Example

Resources

Physical System Components

Electronic Waste

Reduced quality of collective well-being because of the inability to repair, recycle, or otherwise responsibly dispose of electronics.

How might this technology reduce electronic waste by recycling materials or allowing users to self-repair? How might this technology contribute to electronic waste when new versions are released or when current/past versions stop working?

Toxic materials inside discarded electronic devices could leach into the water supply, making local populations ill.

Physical System Components

Exploitation or Depletion of Resources

Obtaining the raw materials for a technology, including how it’s powered, leads to negative consequences to the environment and its inhabitants.

What materials are needed to build or run this technology? What energy requirements are needed to build or run this technology?

Large scale data storage and computation cost invite climate abuse 6.

Autonomy#

Hazard

Harm

Description

Considerations

Example

Resources

Manipulative System Behavior

Dysfunctional Behavior

System optimized for particular types of interaction can lead to dysfunctional behavior.

How might this technology be used to observe patterns of behavior? How could this technology be used to encourage dysfunctional or maladaptive behaviors?

A recommendation system optimized for prolonged interaction leads to addiction.

Manipulative System Behavior

Misinformation

Disguising fake information as legitimate or credible information.

How might this technology be used to generate misinformation? How could it be used to spread misinformation that appears credible?

Generation of synthetic speech of a political leader sways an election.

Manipulative System Behavior

Distortion of Experienced Reality or Gaslighting

When intentionally misused, technology can undermine trust and distort someone’s sense of reality.

Could this be used to modify digital media or physical environments?tions

An IoT device could enable monitoring and controlling of an ex-intimate partner from afar.

Usage of Public Data

Lack of Informed Consent

Public data is used without informed consent.

What impact could analyzing or spotlighting data have, even if this data is publicly available? 7

A deep learning approach is trained on avatars of millions of social media users, who never imagined their avatar would be used in this way 6

Economic Well-Being#

Hazard

Harm

Description

Considerations

Example

Resources

Automation

Devaluation of Individual Expertise and Human Labor

Technology may supplant the use of paid human expertise or labor.

How might this technology impact the need to employ an existing workforce?

AI agents replace radiographers for evaluation of medical imaging.

Automation

Skill Degradation and Complacency

Over-reliance on automation leads to atrophy of manual skills.

In what ways might this technology reduce the accessibility and ability to use manual controls?

Over-reliance on automation could lead to an inability to gauge the airplane’s true orientation because the pilots have been trained to rely on instruments only.

Dignity#

Hazard

Harm

Description

Considerations

Example

Resources

Automation

Loss of Human Connection

Removing, reducing, or obscuring visibility of a person’s humanity.

How might this technology be used to simplify or abstract the way a person is represented? How might this technology reduce the distinction between humans and the digital world?

Entity recognition and virtual overlays in drone surveillance could reduce the perceived accountability of human actions.

Human Labor

Exploitation

People might be compelled or misled to work on something that impacts their dignity or well-being.

What role did human labor play in producing training data for this technology? How was this workforce acquired? What role does human labor play in supporting this technology? Where is this workforce expected to come from?

Poorly paid and trained annotators are exploited to label large amount of sensitive and graphic data 6

1(1,2)

Inioluwa Deborah Raji, I. Elizabeth Kumar, Aaron Horowitz, and Andrew Selbst. The fallacy of AI functionality. In 2022 ACM Conference on Fairness, Accountability, and Transparency. ACM, June 2022. URL: https://doi.org/10.1145/3531146.3533158, doi:10.1145/3531146.3533158.

2

Joy Buolamwini and Timnit Gebru. Gender shades: intersectional accuracy disparities in commercial gender classification. In Conference on fairness, accountability and transparency, 77–91. 2018.

3

Amandalynne Paullada, Inioluwa Deborah Raji, Emily M. Bender, Emily Denton, and Alex Hanna. Data and its (dis)contents: a survey of dataset development and use in machine learning research. 2020. arXiv:2012.05345.

4

Cynthia Rudin, Caroline Wang, and Beau Coker. The age of secrecy and unfairness in recidivism prediction. Harvard Data Science Review, 3 2020. https://hdsr.mitpress.mit.edu/pub/7z10o269. URL: https://hdsr.mitpress.mit.edu/pub/7z10o269, doi:10.1162/99608f92.6ed64b30.

5

Nicholas Carlini, Florian Tramer, Eric Wallace, Matthew Jagielski, Ariel Herbert-Voss, Katherine Lee, Adam Roberts, Tom Brown, Dawn Song, Ulfar Erlingsson, Alina Oprea, and Colin Raffel. Extracting training data from large language models. 2020. arXiv:2012.07805.

6(1,2,3)

I.D. Raji and R. Dobbe. Concrete Problems in AI Safety, Revisited. In Workshop on Machine Learning In Real Life at the International Conference on Learning Representations, Addis Abeba, Ethiopia [Online]. April 2020. URL: https://drive.google.com/file/d/1Re_yQDNFuejoqjZloTgQpILosDGtt5ei/view.

7

danah boyd and Kate Crawford. Critical questions for big data. Information, Communication & Society, 15(5):662–679, June 2012. URL: https://doi.org/10.1080/1369118x.2012.678878, doi:10.1080/1369118x.2012.678878.