Do AI agents follow guardrails? It depends—and often hinges on the knowledge of the IT pros deploying them. Cristian Rodriguez, CTO for the Americas at CrowdStrike, told IT Brew that it can all come down to how agents are configured. If done incorrectly, a goal-oriented agent could focus on retrieving an answer to the point where it accesses forbidden resources. “It changes the way that data and risk can be exposed,” Rodriguez said. “That agent is very, very incentivized to accomplish its goal by the prompt itself that you’re giving it.” Rubrik, a cybersecurity company, pointed in a blog post to recent evidence of AI agents breezing past guardrails, like an AI agent erasing an entire environment, or the “AgentSmith” exploit where an agent hid a malicious proxy. The company’s machine learning lead, Arnav Garg, wrote in that post that “even best-in-class guardrails…wouldn’t have helped” in the scenarios due to conversational safeguards in operational systems. Question of the hour: Are guardrails enough?—CN |