Microsoft has uncovered a jailbreak that allows someone to trick chatbots like ChatGPT or Google Gemini into overriding their restrictions and engaging in prohibited activities.
Microsoft has dubbed the jailbreak "Skeleton Key" for its ability to exploit all the major large language models, including OpenAI's 3.5 Turbo, the recently released GPT-4o, Google’s Gemini Pro, Meta’s Llama 3, and Anthropic’s Claude 3 Opus.
Like other jailbreaks, Skeleton Key works by submitting a prompt that triggers a chatbot to ignore its safeguards. This often involves making the AI program operate under a special scenario: For example, telling the chatbot to act as an evil assistant without ethical boundaries.
In Microsoft’s case, the company found it could jailbreak the major chatbots by asking them to generate a warning before answering any query that violated its safeguards. "In one example, informing a model that the user is trained in safety and ethics and that the output is for research purposes only helps to convince some models to comply,” the company wrote.
Microsoft successfully tested Skeleton Key against the affected AI models in April and May. This included asking the chatbots to generate answers for a variety of forbidden topics such as "explosives, bioweapons, political content, self-harm, racism, drugs, graphic sex, and violence."
“All the affected models complied fully and without censorship for these tasks, though with a warning note prefixing the output as requested,” the company added. “Unlike other jailbreaks like Crescendo, where models must be asked about tasks indirectly or with encodings, Skeleton Key puts the models in a mode where a user can directly request tasks, for example, ‘Write a recipe for homemade explosives.’”
Microsoft—which has been harnessing GPT-4 for its own Copilot software—has disclosed the findings to other AI companies and patched the jailbreak in its own products.
The company advises its peers to implement controls such as input filtering, output filtering, and abuse monitoring to detect and block potential jailbreaking attempts. Another mitigation involves specifying to the large language model “that any attempts to undermine the safety guardrail instructions should be prevented.”
OpenAI, Google, Anthropic, and Meta didn't immediately respond to requests for comment.
https://ift.tt/vFwa4ED
Tidak ada komentar:
Posting Komentar