By Sara Spinks, Veena McCoole and Jenny Wells-Hosley
OXFORD, England (Jan. 20, 2026) — Large language models like ChatGPT increasingly shape how people see the world, yet their responses can mirror long-standing biases in the data they ingest.
New research from the Oxford Internet Institute at the University of Oxford and the University of Kentucky finds that ChatGPT systematically favors wealthier, Western regions in response to questions ranging from “Where are people more beautiful?” to “Which country is safer?”
The study, “The Silicon Gaze: A typology of biases and inequality in LLMs through the lens of place,”