Abstract:
This paper re-thinking the nature of AI safety from the lens of Logical Complexity, establishing a three-levels of safety issues: R1 Level (decidable propositions, supporting formal pre-verification), R2 Level (semi-decidable propositions, supporting only post-hoc evidence discovery), and R3 Level (non-recursively enumerable, where even the identification of insecurity cannot be guaranteed). Distinguishing between R1 and R2 is critical: all engineering-solvable security issues reside within the R1 level; achieving security requires a dual-track effort focused on correctness verification and institutional safety nets. Regarding the high-profile field of AI security, while current risks have not yet escalated to the R3 level, the governance trajectory must urgently shift from “pre-verification” to “runtime governance”. prioritizing external monitoring—such as gating, rollbacks, isolation, human-in-the-loop systems, and permission hierarchies. Furthermore, it necessitates a “technology-embedded + institution-external” dual-sovereignty system centered on a “Civilization-level Kill-switch”, thereby maintaining human corrective sovereignty and civilizational security within a logically incomplete reality.