What do we do?
As algorithms and generative AI become more integrated into our social systems and processes, a failure to recognize and combat bias will have damaging consequences. Oketunji, Anas, and Saina write:
Bias in LLMs is not merely a technical anomaly but a reflection of deeper societal and cultural imbalances. Organisations train these models on vast datasets derived from human language, which inherently contain societal biases (Bender et al., 2021). As a result, LLMs can perpetuate and even amplify biases, leading to skewed responses with significant implications, especially when these models are deployed in decision-making processes or as interfaces in various sectors (Caliskan et al., 2017).
Kate Crawford (author of the excellent book Atlas of AI) and Trevor Paglen write: “There is much at stake in the architecture and contents of the training sets used in AI. They can promote or discriminate, approve or reject, render visible or invisible, judge or enforce. And so we need to examine them—because they are already used to examine us—and to have a wider public discussion about their consequences, rather than keeping it within academic corridors. As training sets are increasingly part of our urban, legal, logistical, and commercial infrastructures, they have an important but underexamined role: the power to shape the world in their own images.”
Awareness of these issues is a key first step in addressing the potential impacts of bias in LLMs/generative AI. Data literacy has been called a “survival skill” for the age of AI because it involves, “understanding the ethical implications of data usage, recognizing biases in data collection and analysis, and being able to question and validate the sources and methodology of data generation” (Acalde, 2023) In addition to honing our awareness of biases in place in current LLMs, we should be vigilant about what this might look like in the future. For example, we know that companies and governments around the world are creating their own LLMs and GPTs/chatbots based on more specialized or curated data sets. This might look like a country creating its own LLM to reduce political influences of another country, or a company creating an AI training chatbot that draws from internal documentation sources to orient staff members. As approaches to data selection, collection, and training for LLMs change, our data literacies will need to evolve to be able to recognize and counteract inherent biases.
Researchers and tech companies are exploring possible responses to deal with the issues of bias and overall poor data quality in LLMs and generative AI. Some researchers suggest that there are things we as users can do to reduce bias in LLM responses. For example, a study of bias in Anthropic’s Claude 2 suggested several strategies users can try to reduce bias in LLM responses, including telling the LLM that it’s important not to discriminate in its answers, asking the LLM to verbalize rationales for its responses specifically to avoid discrimination, and avoiding emotional language in prompts. Other researchers have explored responses and accountability at the technological and corporate levels by, for example, creating frameworks for quantifying and counteracting biases in LLMs, such as the Large Language Model Bias Index (LLMBI). Agreements about what bias is and how to measure it will be complex, will evolve over time, and will vary between cultures. Undoubtedly, a large amount of bias analysis will be performed programmatically simply due to the scale and complexity of the data involved and the variety of expressions of bias. When AI is evaluating AI for bias, one cannot help but wonder who will watch the watchers?