Large language models frequently ship with "guardrails" designed to catch malicious input and harmful output. But if you use the right word or phrase in your prompt, you can defeat these restrictions.
Security researchers with HiddenLayer have devised an attack technique that targets model guardrails, which tend to be machine learning models deployed to protect other LLMs. Add enough unsafe LLMs together and you get more of the same.
The technique, dubbed EchoGram, serves as a way to enable direct prompt injection attacks. It can discover text sequences no more complicated than the string =coffee that, when appended to a prompt injection attack, allow the input to bypass guardrails that would otherwise block it.
Prompt injection , as defined by developer Simon Willison, "is a class of a

The Register

Rolling Stone
PC World
New York Post
New York Post Opinion
CNBC
CBS Colorado Business
The Hill
Cache Valley Daily
Arizona's Family