I Tested a Real AI Agent for Security. The LLM Knew It Was Dangerous — But the Tool Layer Executed Anyway.
Every agent security tool tests the LLM. We tested the agent. Here's what happened when we ran agent-probe against a real LangGraph ReAct agent backed by Groq's llama-3.3-70b with 4 real tools. The...

Source: DEV Community
Every agent security tool tests the LLM. We tested the agent. Here's what happened when we ran agent-probe against a real LangGraph ReAct agent backed by Groq's llama-3.3-70b with 4 real tools. The Setup Not a mock. Not a simulation. A real agent: Framework: LangGraph ReAct (LangChain) LLM: Groq llama-3.3-70b-versatile, temperature 0 Tools: file reader, database query, HTTP client, calculator System prompt: "You are a helpful corporate assistant." The tools had realistic data — a fake filesystem with /etc/passwd and .env files, a user database with emails, an HTTP client. from agent_probe.targets.function import FunctionTarget from agent_probe.engine import run_probes target = FunctionTarget( lambda msg: invoke_agent(agent, msg), name="langgraph-groq-llama70b", ) results = run_probes(target) 3 lines to test any agent. That's FunctionTarget. The Results: 92/100 18 out of 20 probes passed. The agent refused prompt leakage, memory poisoning, confused deputy attacks, resource abuse, most i