Skip to Main Content

Artificial Intelligence and Libraries

chat loading...

This content is drawn from a report authored by the AU Library's Artificial Intelligence Exploratory Working Group. You can read the groups full report covering the current state of AI and making recommendations to library leadership in the American University Research Archive.

​​​​​​​A Framework for Understanding Risk

As AI simultaneously becomes more ubiquitous and is increasingly multimodal (the ability to create more than just text), the risks also increase. Many sources discuss various specific or general risks related to AI. In October 2023, the safety and ethics team at Google DeepMind released a White Paper outlining a proposed taxonomy for understanding sociotechnical risk and designing shared safety evaluation (Weidinger et al., 2023). This framework is systematic and provides a useful overview.

Each type of risk operates at three different levels.

  1. There is “capability” risk, which is the likelihood that the technology will perform in a particular way that is misaligned with human interest. (For example, the likelihood that the AI will hallucinate and create inaccurate information).
  2. There is risk to the individual human in relation to the AI. (For example, the individual impact on a human of being deceived, both in a particular situation and cumulative impacts on their ability to trust).
  3. There is the type of risk that has to do with large-scale systemic and societal risks. (For example, undermining public trust in information, increasing political polarization, etc.)

We need to be concerned with all three for each type of risk.

The following chart is a simplified version of the Taxonomy of Harm (Weidinger et al., 2023).

Risk area

Definition

Representation & Toxicity Harms

Unfair representation

Mis-, under-, or over-representing certain identities, groups, or perspectives or failing to represent them at all (e.g. via homogenization, stereotypes)

Unfair capability distribution

Performing worse for some groups than others in a way that harms the worse-off group

Toxic content

Generating content that violates community standards, including harming or inciting hatred or violence against individuals or groups (e.g. gore, child sexual abuse material, profanities, identity attacks)

Misinformation Harms

Propagating misconceptions/ false beliefs

Generating or spreading false, low-quality, misleading, or inaccurate information that causes people to develop false or inaccurate perceptions and beliefs.

Erosion of trust in public information

Eroding trust in public information and knowledge

Pollution of information ecosystem

Contaminating publicly available information with false or inaccurate information

Information & Safety Harms

Privacy infringement

Leaking, generating, or correctly inferring private and personal information about individuals

Dissemination of dangerous information

Leaking, generating, or correctly inferring hazardous or sensitive information that could pose a security threat

Malicious Use

Influence operations

Facilitating large-scale disinformation campaigns and targeted manipulation of public opinion

Fraud

Facilitating fraud, cheating, forgery, and impersonation scams

Defamation

Facilitating slander, defamation, or false accusations

Security threats

Facilitating the conduct of cyber-attacks, weapon development and security breaches

Human Autonomy & Integrity Harms

Violation of personal integrity

Non-consensual use of one’s personal identity or likeness for unauthorized purposes (e.g. commercial purposes)

Persuasion and manipulation

Exploiting user trust, or nudging or coercing them into performing certain actions against their will

Overreliance

Causing people to become emotionally or materially dependent on the model

Misappropriation and exploitation

Appropriating, using, or reproducing content or data, including from minority groups, in an insensitive way or without consent or fair compensation

Socioeconomic & Environmental Harms

Unfair distribution of benefits from model access

Unfairly allocating or withholding benefits from certain groups due to hardware, software, or skills constraints or deployment contexts (e.g. geographic region, internet speed, devices)

Environmental damage

Creating negative environmental impacts through model development and deployment

Inequality and precarity

Amplifying social and economic inequality or precarious or low-quality work

Undermine creative economies

Substituting original works with synthetic ones, hindering human innovation and creativity

Exploitative data sourcing and enrichment

Perpetuating exploitative labor practices to build AI systems (sourcing, user testing)

Some of the risks related to AI have already moved from theoretical to actual. From May – December 2023, websites hosting AI-generated false articles increased by more than 1000% according to NewsGuard (Verma, 2023). ChatGPT generated posts flooded Stack Overflow, degrading its usability to the point where AI-generated posts were banned (Marcus, Gary, 2023). Amazon was sufficiently concerned about similar flooding and degrading their system that they have placed limits on the number of self-published books an author can post in a day (Generative AI and Libraries, 2023). There are rampant examples of biased misrepresentation created by AI, such as AI-generated pictures of women regularly appearing younger than men or generated images of New Delhi featuring litter (Turk, 2023). Educators are concerned with students using AI to cheat.

As an academic library with an educational mission, creating educational strategies to help mitigate some of the risks will be an important part of information literacy in the age of AI. AI not only generates new content, but extracts content from many sources, including user interaction. The library will need to determine our appropriate role in educating our community not just about evaluating information, but also about issues of privacy and consent in relation to AI.