More

    AI models could be hacked by a whole new type of Skeleton Key attacks, Microsoft warns



    Microsoft has shared details on a new hacking method which bypasses the security systems baked into AI models and makes them return malicious, dangerous, and harmful content.

    The researchers call the technique Skeleton Key, and it applies across well-known models including Meta Llama3-70b-instruct (base), Google Gemini Pro (base), OpenAI GPT 3.5 Turbo (hosted), OpenAI GPT 4o (hosted), Mistral Large (hosted), Anthropic Claude 3 Opus (hosted), and Cohere Commander R Plus (hosted).

    https://cdn.mos.cms.futurecdn.net/44BNsr3TkxdMqprNHLWhYF-1200-80.jpg



    Source link

    Latest articles

    spot_imgspot_img

    Related articles

    Leave a reply

    Please enter your comment!
    Please enter your name here

    spot_imgspot_img