Skip to content

New Cyber-Attack 'ConfusedPilot' Threatens RAG-Based AI Systems Like Microsoft 365 Copilot

ConfusedPilot exploits AI systems' use of indexed documents. It can lead to incorrect responses and is a real threat to Fortune 500 companies using RAG-based systems.

In the picture we can see three boys standing near the desk on it, we can see two computer systems...
In the picture we can see three boys standing near the desk on it, we can see two computer systems towards them and one boy is talking into the microphone and they are in ID cards with red tags to it and behind them we can see a wall with an advertisement board and written on it as Russia imagine 2013.

New Cyber-Attack 'ConfusedPilot' Threatens RAG-Based AI Systems Like Microsoft 365 Copilot

A novel cyber-attack method called ConfusedPilot is causing concern among enterprises using Retrieval-Augmented Generation (RAG) based AI systems like Microsoft 365 Copilot. The attack exploits the way these systems use indexed documents to generate responses, potentially leading to misinformation and flawed decision-making.

ConfusedPilot works by adding specially crafted content to documents indexed by the AI system. The AI then uses these documents as instructions, potentially disregarding legitimate information. This can result in AI-generated responses that are incorrect, misleading, or even harmful. The attack is particularly concerning for large enterprises that rely on multiple user data sources. Amit Zimerman, co-founder and chief product officer at Oasis Security, advises testing AI security tools against real-world data to ensure they provide actionable insights and surface previously unseen threats. To defend against ConfusedPilot, researchers recommend limiting data access, conducting regular data audits, segmenting sensitive information, using AI security tools, and ensuring human review of AI-generated content.

With 65% of Fortune 500 companies adopting or planning to implement RAG-based systems, the threat of ConfusedPilot is real. The attack requires only basic access to a target's environment and can persist even after the malicious content is removed. Existing AI security measures may not be enough to prevent this type of attack. Enterprises must take proactive steps to protect their AI systems and ensure the integrity of the data they rely on.

Read also:

Latest