If an agent only works on curated prompts, you do not have a system. You have a demo.

Question

What must be true before an AI agent is safe to ship to production?

Quick answer

Before release, verify five controls:

  1. input validation,
  2. bounded tool permissions,
  3. deterministic fallback behavior,
  4. human escalation path,
  5. traceable action logs.

Reliability gate

Ask this question:

Can the agent fail safely when context is incomplete, contradictory, or adversarial?

If answer is no, block deployment.

Minimal launch standard

  1. Define non-negotiable no-action conditions.
  2. Add retry limits and timeout ceilings.
  3. Require confidence thresholds for irreversible actions.
  4. Capture full reasoning artifacts for audit.

Without these, your incident timeline becomes guesswork.

5-minute launch rubric

GatePass signalBlock signal
Input controlSchema validation + reject list is activeFree-form input goes straight to tool calls
Action controlTool scopes are least-privilege and explicitAgent has broad unbounded permissions
Fallback behaviorKnown-safe fallback path is documented and testedFailure path is undefined or human-dependent
AuditabilityRequest, context, decision, and action logs correlate by IDLogs are partial and cannot reconstruct incidents

If any row is "Block signal", delay launch and fix that row first.

10-minute action step

  1. Choose one real workflow where this decision applies today.
  2. Define one pass/fail metric before you test (cost, latency, reliability, or risk).
  3. Run 10 realistic examples and log misses with root cause tags.
  4. Ship only the smallest fix that moves your chosen metric.

Success signal

You can show a before/after metric change with a written decision rule the team can reuse.