Table of Contents

Do not index
Chapter 9: Algorithmic bias
Biases in AI systems
Provocation:
- “Algorithmic Bias | Ethics Defined” Ethics Unwrapped @ McCombs School of Business, University of Texas at Austin
Required reading:
- RAI ch9
- Aylin Caliskan, Joanna J. Bryson, and Arvind Narayanan. “Semantics derived automatically from language corpora contain human-like biases”. Science 356, 183-186(2017). DOI: 10.1126/science.aal4230
Suggested materials:
- “A Story of Discrimination and Unfairness (33c3)” Aylin Caliskan @ media.ccc.de, Dec 2016
- T. J. Sejnowski, ‘Large Language Models and the Reverse Turing Test’, Neural Computation, vol. 35, no. 3, pp. 309–342, Feb. 2023, doi: 10.1162/neco_a_01563
- D. E. O’Leary, ‘Confirmation and Specificity Biases in Large Language Models: An Explorative Study’, IEEE Intell. Syst., vol. 40, no. 1, pp. 63–68, Jan. 2025, doi: 10.1109/MIS.2024.3513992
- A. Acerbi and J. M. Stubbersfield, ‘Large language models show human-like content biases in transmission chain experiments’, Proc. Natl. Acad. Sci. U.S.A., vol. 120, no. 44, p. e2313790120, Oct. 2023, doi: 10.1073/pnas.2313790120.
- A. K. Singh, B. Lamichhane, S. Devkota, U. Dhakal, and C. Dhakal, ‘Do Large Language Models Show Human-like Biases? Exploring Confidence—Competence Gap in AI’, Information, vol. 15, no. 2, p. 92, Feb. 2024, doi: 10.3390/info15020092
- Luyang Lin, Lingzhi Wang, Jinsong Guo, and Kam-Fai Wong. “Investigating Bias in LLM-Based Bias Detection: Disparities between LLMs and Human Perception”. Version 2, https://arxiv.org/abs/2403.14896, retrieved 10 Dec 2024
Exercises:
- List three examples in the past 24 hours where the algorithms behind your social media feeds, newsfeeds, chatbot interactions, or search or recommendation engines have learned your biases.

