The Open-Box Fallacy: Why AI Deployment Needs a Calibrated Verification Regime

Under review at NeurIPS 2026, 2026

Abstract

AI deployment in sensitive domains is often treated as unsafe to authorize until model internals can be explained. We argue that the gate should instead be calibrated verification: domain-scoped, independently checkable, monitored, accountable, contestable, and revocable. We propose Verification Coverage, a six-component reportable standard with a minimum-composition rule, intended to sit beside capability scores in model cards, leaderboards, and regulatory disclosures.

Download Paper