Kevin Lu, 17, is working on ways to protect AI from sneak attacks meant to steal sensitive data or do other harm.
Today, people are using AI agents to perform a growing mix of tasks — from drafting emails to handling files or searching the web. But these agents can be vulnerable to something known as prompt injection attacks. That’s when a hacker hides instructions inside a seemingly innocent input, such as a piece of text. When an AI model encounters that input, it can be coaxed to spill private data, spread fake news and more.
Read Entire Article


English (US)