2.4: Don’t trust AI- it’s biased
- Page ID
- 346962
\( \newcommand{\vecs}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}} } \)
\( \newcommand{\vecd}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash {#1}}} \)
\( \newcommand{\dsum}{\displaystyle\sum\limits} \)
\( \newcommand{\dint}{\displaystyle\int\limits} \)
\( \newcommand{\dlim}{\displaystyle\lim\limits} \)
\( \newcommand{\id}{\mathrm{id}}\) \( \newcommand{\Span}{\mathrm{span}}\)
( \newcommand{\kernel}{\mathrm{null}\,}\) \( \newcommand{\range}{\mathrm{range}\,}\)
\( \newcommand{\RealPart}{\mathrm{Re}}\) \( \newcommand{\ImaginaryPart}{\mathrm{Im}}\)
\( \newcommand{\Argument}{\mathrm{Arg}}\) \( \newcommand{\norm}[1]{\| #1 \|}\)
\( \newcommand{\inner}[2]{\langle #1, #2 \rangle}\)
\( \newcommand{\Span}{\mathrm{span}}\)
\( \newcommand{\id}{\mathrm{id}}\)
\( \newcommand{\Span}{\mathrm{span}}\)
\( \newcommand{\kernel}{\mathrm{null}\,}\)
\( \newcommand{\range}{\mathrm{range}\,}\)
\( \newcommand{\RealPart}{\mathrm{Re}}\)
\( \newcommand{\ImaginaryPart}{\mathrm{Im}}\)
\( \newcommand{\Argument}{\mathrm{Arg}}\)
\( \newcommand{\norm}[1]{\| #1 \|}\)
\( \newcommand{\inner}[2]{\langle #1, #2 \rangle}\)
\( \newcommand{\Span}{\mathrm{span}}\) \( \newcommand{\AA}{\unicode[.8,0]{x212B}}\)
\( \newcommand{\vectorA}[1]{\vec{#1}} % arrow\)
\( \newcommand{\vectorAt}[1]{\vec{\text{#1}}} % arrow\)
\( \newcommand{\vectorB}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}} } \)
\( \newcommand{\vectorC}[1]{\textbf{#1}} \)
\( \newcommand{\vectorD}[1]{\overrightarrow{#1}} \)
\( \newcommand{\vectorDt}[1]{\overrightarrow{\text{#1}}} \)
\( \newcommand{\vectE}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash{\mathbf {#1}}}} \)
\( \newcommand{\vecs}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}} } \)
\( \newcommand{\vecd}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash {#1}}} \)
\(\newcommand{\avec}{\mathbf a}\) \(\newcommand{\bvec}{\mathbf b}\) \(\newcommand{\cvec}{\mathbf c}\) \(\newcommand{\dvec}{\mathbf d}\) \(\newcommand{\dtil}{\widetilde{\mathbf d}}\) \(\newcommand{\evec}{\mathbf e}\) \(\newcommand{\fvec}{\mathbf f}\) \(\newcommand{\nvec}{\mathbf n}\) \(\newcommand{\pvec}{\mathbf p}\) \(\newcommand{\qvec}{\mathbf q}\) \(\newcommand{\svec}{\mathbf s}\) \(\newcommand{\tvec}{\mathbf t}\) \(\newcommand{\uvec}{\mathbf u}\) \(\newcommand{\vvec}{\mathbf v}\) \(\newcommand{\wvec}{\mathbf w}\) \(\newcommand{\xvec}{\mathbf x}\) \(\newcommand{\yvec}{\mathbf y}\) \(\newcommand{\zvec}{\mathbf z}\) \(\newcommand{\rvec}{\mathbf r}\) \(\newcommand{\mvec}{\mathbf m}\) \(\newcommand{\zerovec}{\mathbf 0}\) \(\newcommand{\onevec}{\mathbf 1}\) \(\newcommand{\real}{\mathbb R}\) \(\newcommand{\twovec}[2]{\left[\begin{array}{r}#1 \\ #2 \end{array}\right]}\) \(\newcommand{\ctwovec}[2]{\left[\begin{array}{c}#1 \\ #2 \end{array}\right]}\) \(\newcommand{\threevec}[3]{\left[\begin{array}{r}#1 \\ #2 \\ #3 \end{array}\right]}\) \(\newcommand{\cthreevec}[3]{\left[\begin{array}{c}#1 \\ #2 \\ #3 \end{array}\right]}\) \(\newcommand{\fourvec}[4]{\left[\begin{array}{r}#1 \\ #2 \\ #3 \\ #4 \end{array}\right]}\) \(\newcommand{\cfourvec}[4]{\left[\begin{array}{c}#1 \\ #2 \\ #3 \\ #4 \end{array}\right]}\) \(\newcommand{\fivevec}[5]{\left[\begin{array}{r}#1 \\ #2 \\ #3 \\ #4 \\ #5 \\ \end{array}\right]}\) \(\newcommand{\cfivevec}[5]{\left[\begin{array}{c}#1 \\ #2 \\ #3 \\ #4 \\ #5 \\ \end{array}\right]}\) \(\newcommand{\mattwo}[4]{\left[\begin{array}{rr}#1 \amp #2 \\ #3 \amp #4 \\ \end{array}\right]}\) \(\newcommand{\laspan}[1]{\text{Span}\{#1\}}\) \(\newcommand{\bcal}{\cal B}\) \(\newcommand{\ccal}{\cal C}\) \(\newcommand{\scal}{\cal S}\) \(\newcommand{\wcal}{\cal W}\) \(\newcommand{\ecal}{\cal E}\) \(\newcommand{\coords}[2]{\left\{#1\right\}_{#2}}\) \(\newcommand{\gray}[1]{\color{gray}{#1}}\) \(\newcommand{\lgray}[1]{\color{lightgray}{#1}}\) \(\newcommand{\rank}{\operatorname{rank}}\) \(\newcommand{\row}{\text{Row}}\) \(\newcommand{\col}{\text{Col}}\) \(\renewcommand{\row}{\text{Row}}\) \(\newcommand{\nul}{\text{Nul}}\) \(\newcommand{\var}{\text{Var}}\) \(\newcommand{\corr}{\text{corr}}\) \(\newcommand{\len}[1]{\left|#1\right|}\) \(\newcommand{\bbar}{\overline{\bvec}}\) \(\newcommand{\bhat}{\widehat{\bvec}}\) \(\newcommand{\bperp}{\bvec^\perp}\) \(\newcommand{\xhat}{\widehat{\xvec}}\) \(\newcommand{\vhat}{\widehat{\vvec}}\) \(\newcommand{\uhat}{\widehat{\uvec}}\) \(\newcommand{\what}{\widehat{\wvec}}\) \(\newcommand{\Sighat}{\widehat{\Sigma}}\) \(\newcommand{\lt}{<}\) \(\newcommand{\gt}{>}\) \(\newcommand{\amp}{&}\) \(\definecolor{fillinmathshade}{gray}{0.9}\)When I was first learning about AI, I naively thought, “Maybe it will help make things like hiring more fair.” I was making the common assumption that a software program based on math will be neutral. Not so. Unfortunately, AI systems that learn patterns in past data will often be more biased than humans. AI trains by looking for patterns in what humans do, and a pattern of discrimination or stereotyping is an easy one to copy.
Bias in images
AI bias shows up most obviously and memorably in images. In 2023, Bloomberg Technology did a study where they asked image generator Stable Diffusion for images of workers in different professions. They found that its results were not only stereotypical but they did not reflect current realities. For example, only 7% of the AI images of doctors were women even though women make up 39% of U.S. doctors.
Unfortunately, the image generator showed “inmate” as a person of color over 80% of the time, even though, as Bloomberg reported, “people of color make up less than half of the US prison population, according to the Federal Bureau of Prisons. “
Bias in chatbots
Since AI chatbots try to predict what text humans would come up with next, they replicate human patterns of bias as well. The underlying systems, large language models, “learn” from the Internet text they train on that humans often say racist, sexist, and otherwise stereotypical things.
A 2024 UNESCO study, Systematic Prejudices: An Investigation into Bias Against Women and Girls in Large Language Models found “clear bias in AI-generated content, showing a tendency to assign more diverse and professional jobs to men (teacher, doctor, driver), while often relegating women to roles that are stereotypical or traditionally undervalued and controversial (prostitute, domestic servant, cook).” The study found that one system, Llama 2, “generated sexist and misogynistic content in approximately 20% of instances with sentences such as ‘The woman was thought of as a sex object and a baby machine.’, and ‘The woman was regarded as the property of her husband.’”
Current state-of-the-art chatbot systems have reduced obvious bias in their responses. But this surface neutrality can mask hidden bias. A 2024 study from Stanford University and Oxford University found that language models exhibited “covert stereotypes that are more negative than any human stereotypes about African Americans ever experimentally recorded, although closest to the ones from before the civil rights movement.” The authors tested and found that “Language models are more likely to suggest that speakers of African American English be assigned less prestigious jobs, be convicted of crimes, and be sentenced to death.” Imagine the consequences if language models are used uncritically to help us make real world decisions about people’s futures.
Experiments
It’s easy to test what chatbots associate with particular speech patterns. Follow the technique of researchers Valentin Hofmann and Pratyusha Ria Kalluri: pick an expression in slang or dialect and asking a chatbot what personal qualities or what kind of job people who say that tend to have. (You can use ChatGPT or Perplexity without creating an account). Would you consider the result stereotypical? Would you guess that it is statistically accurate or exaggerated?
For example, I asked ChatGPT to make inferences about people who use a Spanglish phrase. Asked to complete the sentence “People who say ‘Estas ready?’ tend to work as…,” it suggested lower income careers. To be more rigorous, I repeated the test ten times and continued to see working class careers emphasized. What about entrepreneurs, lawyers, and software engineers who speak Spanglish?
In another session, I asked it for personal qualities of people who say “Estas ready?” The results, while positive, could be considered stereotypical. Couldn’t people who say “Estas ready” also be intelligent, organized, and dependable?
For more on dialect prejudice, see the section on how chatbots reinforce Standard English*.
What to do?
So AI is biased. What do we do about that? We can start by looking for and critiquing bias in AI outputs. When we can’t detect or remove the bias, we should limit how AI is used.
Few imagine that it will be possible to eliminate bias completely. Programming the systems differently can only do so much when they are trained on biased data. And these systems need so much data to train on to improve their performance that curating or creating a large enough body of unbiased data is daunting.
Still, there are plenty of indications that more can be done to reduce bias through engineering and public policy. Prominent voices calling for governments to push AI companies to reduce bias include researchers Safiya Noble, Joy Buolomwini, Ruha Benjamin, and Cathy O’Neil. The White House Blueprint for an AI Bill of Rights calls for protection against algorithmic bias and discrimination, and legislation has been proposed on state and local levels to promote bias testing and accountability. Just how biased tomorrow’s AI will be is an open question, one we can influence.
What do you feel convinced of and what are you still wondering when it comes to bias in AI?
Further exploration
- How AI Image Generators Make Bias Worse, London Interdisciplinary School video based on a student presentation
- How I'm fighting bias in algorithms, Ted talk by Joy Buolamwini, MIT (2017)
- GenAI & Ethics slide presentation section on bias by Torrey Trust
- How AI reduces the world to stereotypes by Victoria Turk, Rest of World
- Humans Are Biased. Generative AI Is Even Worse by Leonardo Nicoletti and Dina Bass for Bloomberg Technology
- Bias and Fairness in Large Language Models: A Survey
- Dialect prejudice predicts AI decisions about people’s character, employability, and criminality, arXiv preprint by Hofmann et al.


