The former CEO of Google, Eric Schmidt, has warned that if AI falls into the hands of bad actors it could be deadly. “There’s evidence that you can take models, closed or open, and you can hack them to remove their guardrails,” Schmidt said at a European tech conference. ”So in the course of their training, they learn a lot of things. A bad example would be they learn how to kill someone,” he added. Among the methods of attacking AI are jailbreaks and prompt injections, both of which can circumvent guardrails and cause systems to execute instructions that violate operators’ policies, such as answering questions that may help build a bomb. A study by the AI research company Anthropic stress-tested 16 leading large language models (LLMs) in various hypothetical scenarios. In one scenario, re
Ex-Google CEO Says AI Models Can ‘Learn How to Kill’

27