Explorerβ€ΊArtificial Intelligenceβ€ΊAI
Research PaperResearchia:202604.28065

Governing What You Cannot Observe: Adaptive Runtime Governance for Autonomous AI Agents

German Marin

Abstract

Autonomous AI agents can remain fully authorized and still become unsafe as behavior drifts, adversaries adapt, and decision patterns shift without any code change. We propose the \textbf{Informational Viability Principle}: governing an agent reduces to estimating a bound on unobserved risk $\hat{B}(x) = U(x) + SB(x) + RG(x)$ and allowing an action only when its capacity $S(x)$ exceeds $\hat{B}(x)$ by a safety margin. The \textbf{Agent Viability Framework}, grounded in Aubin's viability theory, ...

Submitted: April 28, 2026Subjects: AI; Artificial Intelligence

Description / Details

Autonomous AI agents can remain fully authorized and still become unsafe as behavior drifts, adversaries adapt, and decision patterns shift without any code change. We propose the \textbf{Informational Viability Principle}: governing an agent reduces to estimating a bound on unobserved risk B^(x)=U(x)+SB(x)+RG(x)\hat{B}(x) = U(x) + SB(x) + RG(x) and allowing an action only when its capacity S(x)S(x) exceeds B^(x)\hat{B}(x) by a safety margin. The \textbf{Agent Viability Framework}, grounded in Aubin's viability theory, establishes three properties -- monitoring (P1), anticipation (P2), and monotonic restriction (P3) -- as individually necessary and collectively sufficient for documented failure modes. \textbf{RiskGate} instantiates the framework with dedicated statistical estimators (KL divergence, segment-vs-rest zz-tests, sequential pattern matching), a fail-secure monotonic pipeline, and a closed-loop Autopilot formalised as an instance of Aubin's regulation map with kill-switch-as-last-resort; a scalar Viability Index VI(t)∈[βˆ’1,+1]VI(t) \in [-1,+1] with first-order tβˆ—t^* prediction transforms governance from reactive to predictive. Contributions are the theoretical framework, the reference implementation, and analytical coverage against published agent-failure taxonomies; quantitative empirical evaluation is scoped as follow-up work.


Source: arXiv:2604.24686v1 - http://arxiv.org/abs/2604.24686v1 PDF: https://arxiv.org/pdf/2604.24686v1 Original Link: http://arxiv.org/abs/2604.24686v1

Please sign in to join the discussion.

No comments yet. Be the first to share your thoughts!

Access Paper
View Source PDF
Submission Info
Date:
Apr 28, 2026
Topic:
Artificial Intelligence
Area:
AI
Comments:
0
Bookmark
Governing What You Cannot Observe: Adaptive Runtime Governance for Autonomous AI Agents | Researchia