Discussion about this post

User's avatar
Peter H. Schmidt's avatar

Allow me to politely suggest that people who offer robots controlled by LLMs should be held strictly and personally liable for all harms caused by such robots as if the actions were performed with intent by the offeror.

Expand full comment
sj_ben's avatar

There's actually quite a few scary things in that first paper. The visual version of the attack has the benefit that users aren't even alerted to the fact that they're entering unknown gibberish into the prompt.

The text versions from the paper appear to be undecipherable and a cautious (informed) user might refrain from entering it, just as a cautious user might refrain from clicking on a phishing link. But, presumably the attack could be optimized to make it look less threatening (but still obscuring it) as part of a larger "helpful" pre-made prompt. It could even be as simple as making a request to an attacker's site for a larger and more nefarious prompt injection attack. Maybe LLMs need a version of XSS security too.

Expand full comment
52 more comments...

No posts