Debiasing ChatGPT

Creating an LLM that isn’t racist or sexist

Rachel Draelos, MD, PhD

--

This image was generated by DALL-E using the prompt, “the process of removing bias from AI, digital art.” (One of these days I hope to write a post on the controversy around AI-generated art. Also, on the topic of bias, vision-language models are biased too.)

Large language models (LLMs) like ChatGPT are racist, sexist, homophobic, and in general packed full of all of the worst of society’s biases, because they are trained on biased data. As Brown et al. state, “internet-trained models have internet-scale biases.” In this post, I’ll overview specific examples of bias in LLMs, introduce a few existing techniques for…

--

--

Rachel Draelos, MD, PhD

CEO at Cydoc | Physician Scientist | MD + Computer Science PhD | AI/ML Innovator