New Framework Proposes Continuous Control Model for Military AI Agents
Key Takeaways
- ▸Agentic AI systems require distinct governance frameworks addressing six specific control failure modes not covered by existing safety approaches
- ▸The proposed AMAGF uses a real-time Control Quality Score to measure human control continuously rather than relying on binary control assumptions
- ▸Three-pillar governance structure (Preventive, Detective, Corrective) with distributed institutional responsibilities provides measurable, scalable oversight
Summary
A new academic paper submitted to arXiv proposes the Agentic Military AI Governance Framework (AMAGF), a comprehensive approach to managing autonomous AI agents in military contexts. The research identifies six distinct control failures unique to agentic AI systems—which can interpret goals, model the world, plan operations, use tools, operate over long horizons, and coordinate autonomously—that existing safety frameworks fail to address.
The framework moves away from binary control concepts toward a continuous model that actively measures and manages control quality throughout an AI system's operational lifecycle. Its central mechanism, the Control Quality Score (CQS), provides real-time quantification of human control, enabling graduated institutional responses as control weakens. The authors propose three governance pillars: Preventive Governance (reducing failure likelihood), Detective Governance (real-time monitoring for control degradation), and Corrective Governance (restoring or safely degrading operations).
The framework assigns clear responsibilities across five institutional actors and includes concrete mechanisms and evaluation metrics for each identified failure type. A worked operational scenario demonstrates implementation, situating the research within established agent safety literature. The authors argue this approach is essential as agentic AI systems become increasingly capable and autonomous.
- Framework emphasizes managing control quality throughout operational lifecycle, enabling graduated responses to degradation rather than catastrophic failure scenarios
Editorial Opinion
This research addresses a critical gap in AI governance discourse by focusing specifically on agentic systems' unique control vulnerabilities in high-stakes military contexts. The shift from binary to continuous control quality measurement is conceptually sound and pragmatic. However, the framework's real-world effectiveness will depend heavily on institutional commitment and technical implementation capacity—factors academic papers often underestimate. Significant follow-up work on incentive structures and adversarial robustness will be needed before adoption.



