Hackers use AI to attack Google’s Gemini

by nativetechdoctor
2 minutes read

Recent research has highlighted a concerning new technique known as ‘Fun-Tuning,’ which involves the use of artificial intelligence (AI) to execute highly effective prompt injection attacks on advanced AI models, including Google’s Gemini. This method represents a significant escalation in the ongoing cybersecurity battle where AI is both a target and a tool for malicious actors.

Prompt injection is a strategy wherein attackers embed harmful instructions within the input data of an AI model—this could be done through comments in the source code or concealed text on websites. The objective is to deceive the AI, compelling it to ignore built-in safety mechanisms, which can lead to severe consequences, such as the unauthorized disclosure of sensitive information, the dissemination of false data, or other harmful actions.

Historically, deploying these attacks, especially against restricted models like Gemini or GPT-4, required complex, time-intensive manual testing. However, Fun-Tuning has streamlined this process dramatically. Developed by a collaborative team of researchers across various universities, this technique takes advantage of the tuning application programming interface (API) that Google offers for free to users of Gemini.

By examining how the Gemini model responds to errors during the tuning process, Fun-Tuning can automatically identify the most effective ways to embed malicious commands within ‘prefixes’ and ‘suffixes.’ This capability notably increases the likelihood of the AI behaving as intended by the attacker. Test results indicate that Fun-Tuning has achieved a success rate of up to 82% with certain versions of Gemini, which significantly exceeds the success rates of traditional methods that are often below 30%.

Another alarming aspect of Fun-Tuning is its affordability. Since Google’s tuning API is freely accessible, attackers can potentially launch effective attacks for as little as $10. Additionally, researchers found that an attack aimed at one version of Gemini could be adapted for use against other versions, raising the potential for widespread exploitation.

Google has acknowledged its awareness of the threats posed by Fun-Tuning but has not yet revealed whether it will modify how the tuning API functions. The researchers have noted a challenging dilemma: removing the information that facilitates Fun-Tuning might diminish the API’s usability for legitimate developers, while maintaining it could allow malicious actors to exploit the vulnerabilities.

The rise of Fun-Tuning serves as a stark warning that the dynamics of cyber warfare have become increasingly intricate. AI is not only a target in these conflicts but also a weapon employed by those with harmful intentions.

Related Posts

Leave a Reply

[script_24]

Discover more from ITFamilyMedia

Subscribe now to keep reading and get access to the full archive.

Continue reading

Adblock Detected

Please support us by disabling your AdBlocker extension from your browsers for our website.