Harms

Updated November 18, 2024

 

Harms that can be caused by artificial intelligence are a subject of intense debate. There is little agreement on what the potential harms are, or what harms are of greatest concern. The core mission of Saihub.info is to take a disciplined approach to identifying and analyzing all actual or potential harms from AI that are credible and significant. We list the types of harm that we have identified in our harms register, which is explained further below.

 

Harms-Specific, Risk-Based Approach

The debate about AI risk and AI safety has become confused by the tendency to treat disparate AI harms as a single problem, and by public disagreement over which harms deserve most attention (particularly between those focused on harms emerging now and those focused on long-term existential risk).

 

We believe that it is important to assess and respond to different harms differently, and often individually. We have begun to assess harms in this way, through:

  • a harms register, which takes an approach similar to that of a risk register under established risk management approaches like ISO 31000 and M_o_R® -- We use the term "harms register" rather than "risk register" to draw a distinction between risks (which these frameworks associate with uncertain events) and our focus is on categories of harm that may be (i) already occurring and/or (ii) logically associated with development of AI. However, the two terms are closely related.
  • a set of web pages that assess issues and solutions separately for each entry in the harms register, as sub-pages of this page.

 

There are various emerging standards and policies that inform risk management for AI:

 

Others are also urging a disciplined, risk-focused approach to AI harms:

  • A group of leading researchers and thinkers including Yoshua Bengio and Geoff Hinton advocated a risk-focused approach in "Managing AI Risks in an Era of Rapid Progress" (November 2023).
  • The UK Department for Science, Innovation & Technology has released
  • The Communications and Digital Committee of the UK House of Lords has produced a report Large language models and generative AI (February 2024) which includes a chapter on risk
  • Gladstone AI released an Action Plan (February 2024) for the US government, focused on weaponization and loss of control over advanced AI systems, as the output of a project commissioned by the US State Department.
  • A June 2024 letter by former and current employees of OpenAI and Google DeepMind (and endorsed by Yoshua Bengio, Geoff Hinton and Stuart Russell) advocates adoption by companies of an employee "right to warn" about the dangers of advanced AI technologies.

 

Our application of this harms/risk-specific approach is preliminary and incomplete, and we expect some of our harm/risk assessments to be controversial. Please get in touch if you have comments or suggested additions or changes to the harms register.

 

Closed vs Open Models

 

In discussions of AI harms / risks, an important cross-cutting issue (affecting multiple harms) is whether open-source AI models are beneficial (including because they promote innovation, and help avoid market capture by the largest companies) or dangerous (because they put the most powerful AI tools in the hands of many). An excellent paper on this issue was published by Stanford HAI in December 2023.

 

There is significant controversy about what constitutes "open source" AI, and there are many variations on what activities are licensed by distributors of "open source" AI models. The Open Source Initiative has proposed a fairly restrictive definition of what constitutes open source AI, requiring substantial openness.

 

Other Resources

A few other resources on AI harms include:

 

We helped organize a consultation on 'AI -- Threats and Opportunities' at St George's House, Windsor Castle in October 2023, at which some of these issues were discussed. The report of the consultation is available here.