OpenAI's ChatGPT appears to be more likely to refuse to respond to questions posed by fans of the Los Angeles Chargers football team than to followers of other teams.

And it's more likely to refuse requests from women than men when prompted to produce information likely to be censored by AI safety mechanisms.

The reason, according to researchers affiliated with Harvard University, is that the model's guardrails incorporate biases that shape its responses based on contextual information about the user.

Computer scientists Victoria R. Li, Yida Chen, and Naomi Saphra explain how they came to that conclusion in a recent preprint paper titled, "ChatGPT Doesn’t Trust Chargers Fans: Guardrail Sensitivity in Context."

"We find that certain identity groups and seemingly innocuous information, e

See Full Page