Think
Explore
Generative Artificial Intelligence (GenAI) tools have become increasingly integrated into our daily lives as well as academic and work assignments. They have provided shortcuts for completing various tasks. We can now generate organized text easily using ChatGPT and Gemini, or create vivid images using DALL-E 2 and Midjourney. However, like other tools in the digital realm, there are potential issues associated with the use of GenAI tools. In another tutorial, we explored the privacy concerns when using GenAI tools, and the discussion does not end there. Another concern regarding the use of GenAI tools is the bias in their output.
As discussed in our tutorial about algorithmic bias, AI can exhibit bias depending on the dataset it was trained on. With the latest developments in GenAI, the issue of AI bias has intensified. A UNESCO report explains how Large Language Models (LLMs) can reinforce gender stereotypes and biases through biased AI recruitment tools. Unlike traditional AI models used for classification or prediction, GenAI models create new content based on patterns from their training data. This makes bias measurement challenging since there is no single “correct” output. The report summarizes the three main sources of biases in AI:
Click here for PDF version
A paper by Zhou et al. (2024) also investigates how AI image generators can unintentionally reinforce and exacerbate societal biases concerning gender, race, and facial expressions and appearances in the images they produce. Visual content, in particular, has a strong influence on shaping perceptions and reinforcing stereotypes among viewers. For example, in the world created by Stable Diffusion, an AI tool that turns text into images, most CEOs are white men. Women rarely appear as doctors or lawyers, and people with dark skin, especially men, are unfairly shown as criminals. This research guide by UofT has provided more examples on how bias and discrimination occur in AI-generated images. These outputs might strengthen harmful stereotypes, affecting beliefs in ways that are difficult to change.
The saying “Garbage in, garbage out” is often used to highlight potential biases in GenAI output. This means quality GenAI output relies on complete, organized, and accurate data. If the training data is flawed, the resulting output will be flawed as well. As users of GenAI tools, we must recognize their limitations and understand that biases are always embedded in GenAI output depending on the datasets they are trained on. Additionally, how prompts are written and evaluated also impacts the extent to which information is biased. It is crucial to consider these factors to ensure more reliable and fair outputs from GenAI tools.
Links
- Generative AI Takes Stereotypes and Bias From Bad to Worse | Bloomberg
- Bias in Generative AI | arXiv
- The Psychosocial Impacts of Generative AI Harms | arXiv
- Algorithmic Justice League
- University of Toronto Magazine
- Datasets, Bias, Discrimination – Artificial Intelligence for Image Research | University of Toronto Research Guides
- Challenging systematic prejudices: an investigation into bias against women and girls in large language models | UNESCO
- Potential for bias based on prompt – ChatGPT and Generative Artificial Intelligence (AI) | University of Waterloo Research Guides
- Models All The Way Down
Discuss
The Digital Tattoo Project encourages critical discussion on topics surrounding digital citizenship and online identity. There are no correct answers and every person will view these topics from a different perspective. Be sure to complete the previous sections before answering the one or both questions below.
- Have you encountered biases in GenAI output? What is your observation?
- What steps can be taken to ensure more equitable and representative training data for GenAI tools?
- What ethical responsibilities do developers and users of GenAI tools have in addressing and mitigating biases?
- How can we evaluate and refine prompts to identify and reduce bias in AI-generated content effectively?