--- name: L1.1 AI Governance description: "Audit trail, confidence tagging, and ethical checks on every AI decision your operator produces." business_outcome: "Audit-trail coverage rate of 100% across AI decisions; mechanism tracked in documents/ledgers/MOAT_TRACKER.md § MOAT-01." primary_framework: "Prospect Theory (loss aversion)" product_id: L1.1 lemonsqueezy_variant_id: 1551385 pack_tier: L1 ships_with: v1.0.0 license: proprietary --- # L1.1 AI Governance ## The Maxim Moat Governance is not compliance. Governance is coherence made auditable. L1.1 AI Governance is the pack that makes that coherence enforceable across every AI decision that ships under your operator identity. Most AI tools produce outputs without audit trails. Confidence is assumed. Ethical boundaries are named in marketing and never enforced at generation time. Compliance frameworks are cited in sales decks and never checked in practice. Teams do not fail audits because they lack frameworks. They fail because the frameworks are decorative. The replication barrier is the audit trail itself. Every AI decision under Maxim is timestamped, logged, and tagged with a confidence signal before it ships. Ethical boundaries come from `documents/governance/ETHICAL_GUIDELINES.md` and run on every output, not on request. Competitors can claim transparency. They cannot produce a queryable audit trail for a decision their tool made six months ago, with the exact prompt, the exact model, the exact confidence tag, and the ethical check that ran against it. Audit trail is not a feature. Audit trail is the defensibility. ## Business Outcome L1.1 AI Governance applies Prospect Theory (Daniel Kahneman & Amos Tversky, 1979) as the motivational substrate for the entire pack. Loss aversion dominates regulatory language; governance infrastructure converts that aversion into operational assets. The mechanism: every AI decision produces an audit trail, a confidence tag, and an ethical check before ship. Outcome claims accumulate in the living ledger at `documents/ledgers/MOAT_TRACKER.md` § MOAT-01. Primary metric: audit-trail coverage rate. The percentage of AI decisions (each generation, tool call, or delivered output) accompanied by a complete audit-trail record. Target post-v1.0.0: 100%. Secondary metrics: confidence-tag accuracy (the percentage of 🟢 HIGH tags that survive human spot-check review; target ≥95%); EU AI Act Article 13 and 14 self-assessed transparency coverage; time-to-auditable-log in seconds for any given decision on demand. ## Primary Behavioral Framework **Framework:** Prospect Theory. **Mechanism:** people evaluate outcomes relative to a reference point, not in absolute terms; losses loom roughly twice as large as equivalent gains (loss-aversion coefficient near 2.0 per Tversky & Kahneman, 1992). This asymmetry drives risk behavior. Given the same expected value, people avoid certain loss far more aggressively than they pursue certain gain. Governance exists because regulators, auditors, and stakeholders have already framed the alternative as loss: fines, audit failures, reputational damage, market exclusion. **Source:** Daniel Kahneman & Amos Tversky, *Econometrica* (1979); refined as Cumulative Prospect Theory (Tversky & Kahneman, 1992). **Application:** AI Governance makes the loss-avoidance apparatus operational. Every AI decision gets an audit trail before ship. Every output gets a confidence tag so downstream consumers know the reliability bound. Every ethical boundary gets enforced via `documents/governance/ETHICAL_GUIDELINES.md`. Regulators frame the alternative as loss. Loss framing triggers aversion. Aversion funds the governance infrastructure that makes outputs defensible. **Critique:** Prospect Theory assumes a stable reference point, but real decisions involve shifting reference points (Kőszegi & Rabin, 2006 on reference-dependent utility). Maxim treats loss aversion as directional, not deterministic. Governance exists because loss framing dominates regulatory language, not because every decision is binary. ## Behavioral → AI Governance Translation The five frameworks below make governance operational. Prospect Theory drives why governance exists. The others define how it lands inside Maxim. | Framework | Mechanism | Maxim positive | Maxim anti-pattern | Boundary | |---|---|---|---|---| | Prospect Theory (Kahneman & Tversky, 1979) | loss aversion, reference point | "avoid $200K fine" framing on EU AI Act | loss framing on positive-sum internal change | reference points shift (Kőszegi & Rabin, 2006) | | SCARF — Certainty (David Rock, 2008) | certainty as reward signal | confidence tags give downstream Certainty | every output tagged 🟢 destroys the signal | model not peer-reviewed at aggregate | | Dual Process Theory (Daniel Kahneman, 2011) | System 1 / System 2 | pre-commit ethical check is a forced System 2 pause | slow audit where fast decision is needed | dichotomy oversimplified (Evans, 2014) | | Nudge Theory (Thaler & Sunstein, 2008) | default plus framing | safe defaults in `documents/governance/ETHICAL_GUIDELINES.md` | hidden opt-outs that bypass the ethics check | paternalism critique (Glaeser, 2006) | | Diffusion of Innovations (Everett Rogers, 1962) | five adopter categories | governance adoption curve tracked per team | forcing laggards to early-adopter pace | curve validity contested in organizational contexts | Where others see audit overhead, I see defensibility compounding. Each record row is cheap to produce and expensive to reproduce from the outside. Concrete invocations follow. **1. Audit-trail log sample (one JSONL row per AI decision):** ```json { "timestamp": "2026-04-20T14:32:11.042Z", "session_id": "sess_abc123", "operator": "DrNabeelKhan", "prompt_hash": "sha256:a3f2...", "model": "claude-opus-4-7", "tools_called": ["Read", "Edit"], "output_confidence": "GREEN", "confidence_rationale": "Maxim skill matched; behavioral layer applied", "ethics_check": "PASS", "ethics_frameworks": ["documents/governance/ETHICAL_GUIDELINES.md v1.0"], "eu_ai_act_article_13": {"disclosed": true, "output_type": "text"} } ``` **2. Confidence-tagged CLI output:** ``` $ /mxm-implement feature-x [skill-dispatch] content-creation matched [behavioral-layer] applied: Fogg + Cialdini [ethics-check] PASS [confidence-tag] GREEN HIGH (Maxim skill matched, behavioral layer applied) [audit-log] .claude-sessions-memory/audit-2026-04-20.jsonl:4728 ``` **3. EU AI Act transparency check via the compliance MCP:** ```json { "tool": "mcp__mxm-compliance__check_compliance", "args": { "framework": "EU-AI-Act", "article": "13", "scope": "operator-facing outputs", "evidence_dir": "documents/compliance/" } } // Returns: { "compliant": true, // "article_13_coverage": 1.0, // "article_14_coverage": 0.8, // "gaps": ["Art. 14 §3 — human oversight SOP missing"] } ``` ## Anti-Patterns Five failures catalogued from governance rollouts across Maxim-using teams, documented so the audit trail is not silently eroded. **1. Audit-trail theater.** Logs generated but never queried. The mechanism exists; the defensibility does not. Auditors do not find audit trails that were written and discarded; they find the absence of queryable history. **2. Confidence-tag inflation.** Every output tagged 🟢 by default. The signal collapses. Downstream consumers stop trusting the tag and start re-checking everything, which is exactly what the tag was meant to prevent. **3. Ethics-check as afterthought.** `documents/governance/ETHICAL_GUIDELINES.md` referenced in frontmatter but scanned only at publish time. Fast-thinking outputs ship with zero System 2 verification. The violation compounds silently over hundreds of decisions. **4. Regulatory cargo cult.** GDPR language copy-pasted into a privacy policy without mapping real data flows. The words exist; the mechanism does not. Passes the first reviewer; fails the second. **5. Compliance as sales artifact.** Framework logos on landing pages without enforcement at generation time. Sales closes the deal; audit opens it again six months later. Diffusion of Innovations in reverse: innovators buy, laggards get burned. ## Pack Integrations Three sibling packs deepen AI Governance in specific ways. Pack L1.2 through L1.6 are the live ecosystem; no future packs referenced. - **L1.4 Compliance Shield:** extends the governance baseline from EU AI Act Article 13 and 14 to 14 regulated-industry frameworks (GDPR, HIPAA, PCI-DSS, SOC2, ISO 27001, and 9 others). Governance becomes industry-specific without new operator effort. - **L1.3 Proactive Watch:** the `compliance-drift` checker (drift class 8) catches regressions continuously. When a new skill ships without an ethics check, the class fails the commit at pre-commit time. Governance stays enforced without manual vigilance. - **L1.6 Behavioral Intelligence:** governance adoption is itself a behavior change. Prospect Theory frames the loss-avoidance motivator; Cialdini Commitment locks in consistency once governance is on. Operators who turn governance on rarely turn it off. Governance is not bureaucracy. Governance is the memory of good decisions, written at the speed of generation so it survives any audit that comes later. *Team-tier operators get L1.1 AI Governance inside Founder OS, Growth Stack, Professional OS, or Agency All-In at 47% off individual L1 pricing; the [pack catalog](../../../maxim/documents/business/sales-marketing/packs-catalog/mxm-pack-catalog-v1.0.0.md) has the full matrix.* --- _Copyright (c) 2026 iSystematic Inc. Maxim product. BSL 1.1._