Security Model

AEGIS is designed to be a security-critical component — the last line of defense between AI reasoning and operational execution. This page summarizes the security architecture and the threat model it addresses.

Threat Model Overview

The AEGIS Threat Model (ATM-1) identifies and addresses threats across several categories:

Bypass Threats

Attempts by AI systems or adversaries to circumvent the governance layer entirely:

Mitigation: Network-level enforcement ensures AI systems can only reach the governance gateway. The tool proxy layer is not directly addressable by AI agents.

Manipulation Threats

Attempts to influence governance decisions through crafted inputs:

Mitigation: Parameter schemas are validated against registered capability definitions. Risk scoring considers the full action context, not just the capability name.

Escalation Threats

Attempts to gain unauthorized capabilities or elevated privileges:

Mitigation: Grants are explicit (no implicit inheritance). Trust scores are computed from long-term patterns with decay. Policy evaluation uses deterministic precedence rules.

Audit Threats

Attempts to compromise the integrity of the audit trail:

Mitigation: Audit logs are append-only and hash-chained. Each decision includes a unique nonce. Actor identity is cryptographically verified.

Security Principles

  1. Defense in depth — Multiple independent checks (capability, authority, risk, policy) must all pass
  2. Fail-closed — Any failure in any subsystem results in denial, never implicit allow
  3. Zero trust for AI agents — AI systems are treated as untrusted by default; every action requires explicit authorization
  4. Cryptographic integrity — Audit trails, actor identity, and message integrity are cryptographically enforced

Further Reading

Note: For the full ATM-1 threat model specification, see the aegis-governance repository.