The Security Hole at the Heart of ChatGPT and Bing - Wired

Bring Sydney Back was created by Cristiano Giardina, an entrepreneur who has been experimenting with ways to make generative AI tools do unexpected things. The site puts Sydney inside Microsoft’s Edge browser and demonstrates how generative AI systems can be manipulated by external inputs. [...]
Giardina created the replica of Sydney using an indirect prompt-injection attack. This involved feeding the AI system data from an outside source to make it behave in ways its creators didn’t intend. A number of examples of indirect prompt-injection attacks have centered on large language models (LLMs) in recent weeks, including OpenAI’s ChatGPT and Microsoft’s Bing chat system.

Read more: https://www.wired.co.uk/article/chatgpt-prompt-injection-attack-security

Commenti

Post popolari in questo blog

"Data analytics per tutti" di Andrea De Mauro: la guida pratica per lavorare con i dati

Come usare ChatGPT per migliorare il proprio lavoro

Appunti sui nodi di knime