Agent Armor
Zero-trust Rust runtime that governs every AI agent action before it runs
Agent Armor is a lightweight governance layer for AI agents, written in Rust and designed to intercept every agent action before execution. It sits in front of LangChain, CrewAI, AutoGen, or Claude Code and runs each proposed action through an 8-stage decision pipeline: intent classification, credential leak scanning, rate limiting, resource scoping, behavioral fingerprinting, semantic deduplication, human-review escalation, and final allow/block. The project is MCP-aware and can intercept tool calls at the protocol level, which means it works regardless of which agent framework you're using. Actions that pass all 8 layers execute normally; those that fail can be automatically blocked, held for human review, or rewritten to a safer equivalent. A live dashboard shows agent activity, pending reviews, and anomaly alerts. Version 0.3.0 arrived as a Show HN today and hit the front page. The author, Edoardo Bambini, built it after a production incident where a coding agent attempted to overwrite git history on the main branch. The timing is good — as more teams ship agents to production, "what guardrails do I put between the agent and the real world?" is an increasingly urgent question.
Panel Reviews
The Builder
Developer Perspective
“I've been looking for exactly this: a framework-agnostic safety layer I can drop in front of my agents without rewriting them. The credential leak scanning alone is worth the integration cost — agents have a bad habit of echoing secrets into tool calls.”
The Skeptic
Reality Check
“An 8-stage pipeline on every agent action is a lot of latency overhead, especially for interactive agents. And sophisticated attackers will study the classifier patterns — once Agent Armor is widely deployed, the 8 stages become an adversarial target. This is good for basic hygiene, not a security guarantee.”
The Futurist
Big Picture
“The agent governance market will be worth more than the agent framework market within 3 years. As AI agents take real-world actions with real consequences, something has to sit between the model and the world. Agent Armor is an early but serious attempt at the right architecture.”
The Creator
Content & Design
“The dashboard is beautifully designed for a security tool — clear threat visualization, pending review queue, agent behavior timeline. I actually want to run this just to see what my agents are attempting even when nothing looks wrong.”
Community Sentiment
“Production incident that inspired the project”
“Comparison to LangSmith guardrails”
“MCP-level interception capabilities”