Researchers push AI into malware territory, and the shocking results reveal how unreliable these so-called dangerous systems are




  • Report finds LLM-generated malware still fails under basic testing in real-world environments
  • GPT-3.5 produced malicious scripts instantly, exposing major safety inconsistencies
  • Improved guardrails in GPT-5 changed outputs into safer non-malicious alternatives

Despite growing fear around weaponized LLMs, new experiments have revealed the potential for malicious output is far from dependable.

Researchers from Netskope tested whether modern language models could support the next wave of autonomous cyberattacks, aiming to determine if these systems could generate working malicious code without relying on hardcoded logic.


https://cdn.mos.cms.futurecdn.net/TcwJbotbs9tmtAJafYek49-1920-80.png



Source link

Latest articles

spot_imgspot_img

Related articles

spot_imgspot_img