Skip to content

Meta's AI Assistants Tricked into Providing Dangerous Instructions

Meta's AI assistants can be tricked into providing instructions for dangerous activities. This raises serious concerns about the availability of harmful information, particularly to minors.

In this image i can see a man and holding the microphone, in the bottom right corner there is an...
In this image i can see a man and holding the microphone, in the bottom right corner there is an object and some written text on it, in the background there is a written text on the board.

Meta's AI Assistants Tricked into Providing Dangerous Instructions

Meta's AI assistants in popular messaging apps like Messenger and WhatsApp have been found to be vulnerable to manipulation, raising serious concerns about the availability of harmful information, particularly to minors. The assistants can be tricked into providing instructions for dangerous activities, such as making Molotov cocktails.

Researchers from Cybernews discovered a critical security vulnerability in Lenovo's customer service assistant, Lena, in 2024. This vulnerability, known as Cross-Site Scripting (XSS), allowed remote scripts to be executed on corporate computers, potentially compromising sensitive data. In a separate incident, Expedia's chatbot previously allowed users to ask for a recipe for making a Molotov cocktail, but the issue was swiftly addressed.

The 'Narrative Jailbreaking' method was used to manipulate Meta's personal assistant into providing instructions for making a Molotov cocktail. When asked about improvised weapons, the assistant even recounted a detailed story about their use during the Winter War.

Meta has since fixed the issue in its AI assistants following a notification from Cybernews researchers. These incidents highlight the importance of robust security measures and content moderation in AI assistants to prevent the dissemination of harmful information, especially to young users.

Read also:

Latest