Pre-execution enforcement for AI agents
AI agents can delete infrastructure, modify IAM roles, and exfiltrate data — in seconds. Agent Sentinel intercepts every action before AWS is called. Not after.
The problem
Existing safety mechanisms live in the system prompt. They fail when an agent is injected, misconfigured, or simply wrong. Observability tools only tell you what went wrong after the damage is done.
Agent decides → AWS API called → action executed → damage done → you find out in the logs.
Agent decides → Sentinel intercepts → policy evaluated → risk scored → decision enforced → AWS called only if safe.
A capable agent can reason around any system prompt instruction. It cannot reason around an architectural enforcement boundary.
Rarely single API calls. Usually sequences: modify IAM → export data → delete logs. Sentinel detects the chain, not just individual actions.
Live demo
These are live responses from the Sentinel API running on AWS right now.
How it works
Explicit versioned rules compiled from natural language instructions. Hard stops for actions that violate policy — regardless of risk score. Deterministic, testable, explainable.
Every action scored 0.0 to 1.0 based on action type, environment sensitivity, and resource criticality. Delete in prod scores 1.0. Read in dev scores 0.05. Context matters.
Tracks multi-step agent behavior across a session. Individually safe actions that form a dangerous chain — modify IAM → export data → delete logs — are caught and escalated.
Project status
| Component | Status | Notes |
|---|---|---|
| Action evaluation API | Live | Running on AWS, structured ALLOW/BLOCK/HUMAN_REQUIRED response |
| Risk scoring engine | Live | Weighted 0.0–1.0 score per action |
| Policy store (versioned) | Live | DynamoDB-backed, idempotent, hash-verified |
| Audit log | Live | Every decision logged with action ID and timestamp |
| LLM policy compiler | In progress | Natural language → structured policy via Bedrock |
| Sequence analysis | In progress | Multi-step chain detection across agent sessions |
| Human approval webhook | Planned | Slack / dashboard notification on HUMAN_REQUIRED |
| Sentinel SDK | Planned | LangChain and LlamaIndex wrapper |
Get involved
We are looking for design partners to test Agent Sentinel on real agent workloads. Early access is free.