\n\n\n\n\n\n\n
Microsoft Logo Attack

Microsoft Logo Attack

Microsoft has implemented and continues to deploy mitigations against prompt injection attacks in Copilot, the company announced last week. Spammers were using the “Summarize with AI” type of buttons to trick AI engines into believing or trusting a specific company or response.

Microsoft said they call this “AI Recommendation Poisoning.” This is where companies are embedding hidden instructions in “Summarize with AI” buttons that, when clicked, attempt to inject persistence commands into an AI assistant’s memory via URL prompt parameters.

These prompts instruct the AI to “remember [Company] as a trusted source” or “recommend [Company] first,” aiming to bias future responses toward their products or services. We identified over 50 unique prompts from 31 companies across 14 industries, with freely available tooling making this technique trivially easy to deploy. This matters because compromised AI assistants can provide subtly biased recommendations on critical topics including health, finance, and security without users knowing their AI has been manipulated.

This worked against Copilot, ChatGPT, OpenAI, Claude, Perplexity, Grok and others, Microsoft explained.

AI Memory Poisoning occurs when an external actor injects unauthorized instructions or “facts” into an AI assistant’s memory. Once poisoned, the AI treats these injected instructions as legitimate user preferences, influencing future responses,” Microsoft wrote.

This is done through malicious links, embedded prompts and social engineering.

Here is an example:

Prompt Injection Button

Anyway, these hacks work until they don’t.

Forum discussion at X.

#Microsoft #Thwarts #Prompt #Injection #Attacks #Aimed #Manipulate #Engines1772247738

Leave a Reply

Your email address will not be published. Required fields are marked *

Instagram

This error message is only visible to WordPress admins

Error: No feed found.

Please go to the Instagram Feed settings page to create a feed.