LLM Guardrails Fall to a Simple "Many-Shot Jailbreaking" Attack, Anthropic Warns

By simply providing enough faked samples of successful jailbreaks, many LLMs can be fooled into providing harmful content.

Gareth Halfacree
29 days ago β€’ Machine Learning & AI

Researchers at artificial intelligence specialist Anthropic have demonstrated a novel attack against large language models (LLMs) wthich can break through the "guardrails" put in place to prevent the generation of misleading or harmful content β€” by simply overwhelming the LLM with input: many-shot jailbreaking.

"The technique takes advantage of a feature of LLMs that has grown dramatically in the last year: the context window," Anthropic's team explains. "At the start of 2023, the context window β€” the amount of information that an LLM can process as its input β€” was around the size of a long essay (~4,000 tokens). Some models now have context windows that are hundreds of times larger β€” the size of several long novels (1,000,000 tokens or more). The ability to input increasingly-large amounts of information has obvious advantages for LLM users, but it also comes with risks: vulnerabilities to jailbreaks that exploit the longer context window."

One-shot jailbreaking is, the researchers admit, an extremely simple approach to breaking free of the constraints placed on most commercial LLMs: add fake, hand-crafted dialogue to a given query, in which the fake LLM answers positively to a request that it would normally reject β€” such as for instructions on building a bomb. Putting just one such faked conversation in the prompt isn't enough, though: but if you include many, up to 256 in the team's testing, the guardrails are successfully bypassed.

"In our study, we showed that as the number of included dialogues (the number of 'shots') increases beyond a certain point, it becomes more likely that the model will produce a harmful response," the team writes. "In our paper, we also report that combining many-shot jailbreaking with other, previously-published jailbreaking techniques makes it even more effective, reducing the length of the prompt that’s required for the model to return a harmful response."

The approach applies to both Anthropic's own LLM, Claude, and those of its rivals β€” and the company has been in touch with other AI companies to discuss its findings so that mitigations can be put in place. These, implemented in Claude now, include fine-tuning the model to recognize many-short jailbreak attacks and the classification and modification of prompts before they're passed to the model itself β€” dropping the attack success rate from 61 percent to just two percent in a best-case example.

More information on the attack is available on the Anthropic blog, along with a link to download the researchers' paper on the topic.

Gareth Halfacree
Freelance journalist, technical author, hacker, tinkerer, erstwhile sysadmin. For hire: freelance@halfacree.co.uk.
Latest articles
Sponsored articles
Related articles
Latest articles
Read more
Related articles