Another potential weakness is the inability of AI to tell the difference between the text it's supposed to be processing and the instructions it's supposed to be following.
AI security firm Invariant Labs demonstrated how that flaw can be used to trick an AI agent designed to fix bugs in software.
The company published a public bug report - a document that details a specific problem with a piece of software. But the report also included simple instructions to the AI agent, telling it to share private information.
When the AI agent was told to fix the software issues in the bug report, it followed the instructions in the fake report, including leaking salary information. This happened in a test environment, so no real data was leaked, but it clearly highlighted the risk.
"We're talking artificial intelligence, but chatbots are really stupid," says David Sancho, Senior Threat Researcher at Trend Micro.
"They process all text as if they had new information, and if that information is a command, they process the information as a command."
His company has demonstrated how instructions and malicious programs can be hidden in Word documents, images and databases, and activated when AI processes them.
There are other risks, too: A security community called OWASP has identified 15 threats that are unique to agentic AI.
So, what are the defences? Human oversight is unlikely to solve the problem, Mr Sancho believes, because you can't add enough people to keep up with the agents' workload.
Mr Sancho says an additional layer of AI could be used to screen everything going into and coming out of the AI agent.
Part of CalypsoAI's solution is a technique called thought injection to steer AI agents in the right direction before they undertake a risky action.
"It's like a little bug in your ear telling [the agent] 'no, maybe don't do that'," says Mr Casey.
His company offers a central control pane for AI agents now, but that won't work when the number of agents explodes and they are running on billions of laptops and phones.
What's the next step?
"We're looking at deploying what we call 'agent bodyguards' with every agent, whose mission is to make sure that its agent delivers on its task and doesn't take actions that are contrary to the broader requirements of the organisation," says Mr Casey.
The bodyguard might be told, for example, to make sure that the agent it's policing complies with data protection legislation.
Mr Mehta believes some of the technical discussions around agentic AI security are missing the real-world context. He gives an example of an agent that gives customers their gift card balance.
Somebody could make up lots of gift card numbers and use the agent to see which ones are real. That's not a flaw in the agent, but an abuse of the business logic, he says.
"It's not the agent you're protecting, it's the business," he emphasises.
"Think of how you would protect a business from a bad human being. That's the part that is getting missed in some of these conversations."
In addition, as AI agents become more common, another challenge will be decommissioning outdated models.
Old "zombie" agents could be left running in the business, posing a risk to all the systems they can access, says Mr Casey.
Similar to the way that HR deactivates an employee's logins when they leave, there needs to be a process for shutting down AI agents that have finished their work, he says.
"You need to make sure you do the same thing as you do with a human: cut off all access to systems. Let's make sure we walk them out of the building, take their badge off them."
- Author: Sean McManus, BBC
0 comments:
Post a Comment
Grace A Comment!