AI models could be hacked by a whole new type of Skeleton Key attacks, Microsoft warns



Microsoft has shared details on a new hacking method which bypasses the security systems baked into AI models and makes them return malicious, dangerous, and harmful content.

The researchers call the technique Skeleton Key, and it applies across well-known models including Meta Llama3-70b-instruct (base), Google Gemini Pro (base), OpenAI GPT 3.5 Turbo (hosted), OpenAI GPT 4o (hosted), Mistral Large (hosted), Anthropic Claude 3 Opus (hosted), and Cohere Commander R Plus (hosted).

https://cdn.mos.cms.futurecdn.net/44BNsr3TkxdMqprNHLWhYF-1200-80.jpg



Source link

Latest articles

spot_imgspot_img

Related articles

spot_imgspot_img