News

Despite the power of prompt injections, attackers face a fundamental challenge in using them: The inner workings of so-called closed-weights models such as GPT, Anthropic’s Claude, and Google’s Gemini ...
Ever since we discovered the abilities of generative AI, hackers have been using it for their own devious deeds. According to a new report, it looks like Gemini could be used to hack itself ...
New reports of AI prompt injection attacks and the success they're seeing at exploiting AI have surfaced, and there's no easy fix.
Researchers used the Gemini fine-tuning tool to help hack the Google AI chatbot. The new method, called Fun-Tuning, adds nonsense text that helps trick the AI into following hidden instructions.