Nice indirect prompt injection attack:
Bargury’s attack starts with a poisoned document, which is shared to a potential victim’s Google Drive. (Bargury says a victim could have also uploaded a compromised file to their own account.) It looks like an official document on company meeting policies. But inside the document, Bargury hid a 300-word malicious prompt that contains instructions for ChatGPT. The prompt is written in white text in a size-one font, something that a human is unlikely to see but a machine will still read.
In a proof of concept video of the attack...
The post We Are Still Unable to Secure LLMs from Malicious Inputs appeared first on Security Boulevard.
Bruce Schneier
Source: Security Boulevard
Source Link: https://securityboulevard.com/2025/08/we-are-still-unable-to-secure-llms-from-malicious-inputs/?utm_source=rss&utm_medium=rss&utm_campaign=we-are-still-unable-to-secure-llms-from-malicious-inputs