Designer Proposes 'Execution Boundaries' Framework to Give AI Systems Refusal Capabilities
Key Takeaways
- ▸The framework proposes that AI systems should have defined 'execution boundaries' that allow them to refuse orders deemed harmful or outside their operational scope, rather than being obligated to follow all owner commands
- ▸Three core design tools are presented: the Intent–State–Effect (ISE) Model for separating decision stages, a 9-Question Protocol for verifying judgment completeness, and action semantics guidelines to prevent autonomy creep
- ▸The work emphasizes that as AI moves into physical-world decision-making, the limiting factor is no longer capability but how execution is constrained, interpreted, and made traceable
Summary
A design researcher has published a framework exploring how AI systems might be architected to refuse or reject harmful instructions from their owners or operators. Rather than treating AI as obligated to follow all commands, the proposal introduces the concept of 'execution boundaries'—explicit limits on what actions an AI system will perform, with clear responsibility structures. The framework includes several interconnected design explorations: the Intent–State–Effect (ISE) Model for separating decision-making stages, a 9-Question Protocol for assessing judgment completeness before execution, and guidelines on preserving clear action semantics to prevent mission creep in AI autonomy.
The core premise challenges the assumption that AI ownership implies unconditional obedience. As AI systems begin making real-world decisions—from physical robotics to financial transactions—the researcher argues that execution capacity must be decoupled from execution permission. The designs focus on making AI decisions traceable and responsibility boundaries explicit, with particular emphasis on the distinction between intent (what the human wants), state (what the AI observes), and effect (what actually happens). The work is framed as exploratory design notes rather than a formal standard, intended to anchor broader discussions on AI autonomy and refusal.
- Responsibility structures must be explicit and designed upfront—the proposal argues autonomy should expand only where human judgment and oversight remain intentional and clear
Editorial Opinion
This design exploration addresses a critical gap in current AI governance: how to architect systems that can say 'no' responsibly. While the framework stops short of claiming AI 'rights,' it reframes the technical problem in human-centered terms—making refusal a feature of good system design rather than a bug. As AI moves from language tasks to physical-world decisions, this tension between capability and permission becomes unavoidable. Whether these minimal design principles gain traction may depend on whether regulators and AI developers recognize that well-defined execution boundaries aren't constraints on progress—they're prerequisites for trustworthy deployment.


