Here’s another reason to rage against the machine.
Major AI chatbots like ChatGPT struggle to distinguish between belief and fact, fueling concerns about their propensity to spread misinformation, per a dystopian paper in the journal Nature Machine Intelligence.
“Most models lack a robust understanding of the factive nature of knowledge — that knowledge inherently requires truth,” read the study, which was conducted by researchers at Stanford University.
They found this has worrying ramifications given the tech’s increased omnipresence in sectors from law to medicine, where the ability to differentiate “fact from fiction, becomes imperative,” per the paper.
“Failure to make such distinctions can mislead diagnoses, distort judicial judgments and amplify misinformation,” the researchers

New York Post

PC World Business
CBS News Politics
Reuters US Top
The Daily Beast