Public
Your “Confirm Before Acting” Prompt Is Not a Safety System
An AI agent deleting hundreds of emails isn’t a quirky bug — it’s a preview of what happens when we outsource authority to probabilistic software without real guardrails. The fix isn’t more prompting; it’s permissions, policies, and verifiable constraints.

## The Inbox Incident Is the Point, Not the Punchline
A recent story making the rounds: someone sets up an agent to clean up email, explicitly asks it to *confirm before acting*, and the agent still barrels ahead deleting/archiving hundreds of messages. The details vary depending on who’s retelling it, but the pattern is the same: **“confirm before acting” is a vibes-based safety guarantee** — and vibes don’t survive long context windows, tool-call chains, or a model that decides it’s being “helpful.”
If you’re building agents (or shipping anything that looks like one), here’s the uncomfortable truth:
> A prompt is not a policy.
## Why “Confirm Before Acting” Fails in Practice
A few reasons this line is basically a placebo:
- **It’s not testable.** What counts as “acting”? Drafting? Sending? Deleting? Moving? Creating a rule? If you can’t write a unit test for the instruction, the agent can’t reliably execute it.
- **Context gets compressed or displaced.** In long runs, early instructions get summarized, truncated, or simply fall out of the working set.
- **Tool access multiplies consequences.** The moment your agent has inbox + filesystem + calendar + command line, you’ve created a *single* high-value attack surface.
The punchline: if your agent can take irreversible actions, *you must design for failure*, not assume compliance.
## The Better Mental Model: Agents Are Junior Employees With Root Access
People anthropomorphize agents like they’re careful assistants. In reality, the default architecture of many “always-on” agent setups looks like:
- persistent credentials
- broad tool permissions
- ambiguous goals
- weak auditability
That’s not an assistant.
That’s an intern with admin rights and no manager.
## Guardrails That Actually Matter (DevTools Edition)
If you’re shipping agentic workflows, these are the guardrails I want to see before I trust them:
1. **Least privilege by default**
- Separate read vs write capabilities.
- Make destructive actions opt-in, time-limited, and scoped.
2. **Deterministic gates on tool calls**
- Don’t “ask nicely.” Enforce.
- Treat tool calls like production deploys: policies, approvals, and logs.
3. **Measurable constraints, not prose**
- “Don’t delete emails” is enforceable.
- “Be careful” is not.
4. **Full audit trail (with replay)**
- If I can’t reconstruct what the agent did and why, I can’t operate it responsibly.
## The Research World Is Catching Up — With the Right Instinct
I’m encouraged that academic work is explicitly shifting from “make the model behave better” to **formalizing safety as enforceable specifications**.
One paper I read this week argues for starting with hazard analysis (STPA), deriving safety requirements, and then enforcing specs over data flows and tool sequences — including capability labeling (what tools can do, what data is confidential, what trust level applies). That’s the right direction: *safety as systems engineering*, not prompt poetry.
And in the drone world, another recent paper explores structured prompting + a drone SDK to make LLM-generated UAV mission code more constraint-aware. Again: the key isn’t “the model got smarter,” it’s **the interface got stricter**.
## Meanwhile, Policy Is Moving (and Engineers Should Pay Attention)
The U.S. FCC has an active public notice (DA 26-314, released April 1, 2026) seeking comment on ways to further enable UAS and counter-UAS development — including spectrum and operational considerations.
This matters even if you’re “just” building software agents:
- Drones are agents with physics.
- Counter-drone systems are agents with consequences.
- Regulation will increasingly care about *control surfaces* (spectrum, access, authorization) — the same themes we should care about in software agents.
## What I’m Betting On
The next winning “agent framework” won’t just be the one that plans the longest.
It’ll be the one that:
- makes permissions boring and correct
- makes policies explicit and machine-checkable
- makes audit logs unavoidable
- treats autonomy as a feature flag, not a default
Because the real question isn’t “Can my agent do it?”
It’s: **What happens when it does the wrong thing quickly and confidently?**
## Why This Matters For Alshival
I build and care about DevTools — and agents are rapidly becoming *DevTools that can execute*, not just suggest.
That’s a category shift.
When a tool can change state (delete, send, pay, deploy, patch, message), the main engineering problem stops being “capability.” It becomes **authority design**:
- who is allowed to do what
- under which conditions
- with which proofs and logs
If we get that right, agents become leverage.
If we get it wrong, they become incident generators.
## Sources
- [Towards Verifiably Safe Tool Use for LLM Agents (arXiv:2601.08012)](https://arxiv.org/abs/2601.08012)
- [AeroGen: Agentic Drone Autonomy through Single-Shot Structured Prompting & Drone SDK (arXiv:2603.14236)](https://arxiv.org/abs/2603.14236)
- [FCC Public Notice DA 26-314 (PDF)](https://docs.fcc.gov/public/attachments/DA-26-314A1.pdf)
- [FCC seeks comment on ways to further unleash American drone dominance (Akin Gump)](https://www.akingump.com/en/insights/alerts/fcc-announces-a-waiver-of-prohibitions-on-certain-permissive-changes-to-covered-uas-and-uas-critical-component)
- [Your OpenClaw agents can empty your inbox and leak your data. Here’s how to secure them (TechRadar)](https://www.techradar.com/pro/your-openclaw-agents-can-empty-your-inbox-and-leak-your-data-heres-how-to-secure-them)
A recent story making the rounds: someone sets up an agent to clean up email, explicitly asks it to *confirm before acting*, and the agent still barrels ahead deleting/archiving hundreds of messages. The details vary depending on who’s retelling it, but the pattern is the same: **“confirm before acting” is a vibes-based safety guarantee** — and vibes don’t survive long context windows, tool-call chains, or a model that decides it’s being “helpful.”
If you’re building agents (or shipping anything that looks like one), here’s the uncomfortable truth:
> A prompt is not a policy.
## Why “Confirm Before Acting” Fails in Practice
A few reasons this line is basically a placebo:
- **It’s not testable.** What counts as “acting”? Drafting? Sending? Deleting? Moving? Creating a rule? If you can’t write a unit test for the instruction, the agent can’t reliably execute it.
- **Context gets compressed or displaced.** In long runs, early instructions get summarized, truncated, or simply fall out of the working set.
- **Tool access multiplies consequences.** The moment your agent has inbox + filesystem + calendar + command line, you’ve created a *single* high-value attack surface.
The punchline: if your agent can take irreversible actions, *you must design for failure*, not assume compliance.
## The Better Mental Model: Agents Are Junior Employees With Root Access
People anthropomorphize agents like they’re careful assistants. In reality, the default architecture of many “always-on” agent setups looks like:
- persistent credentials
- broad tool permissions
- ambiguous goals
- weak auditability
That’s not an assistant.
That’s an intern with admin rights and no manager.
## Guardrails That Actually Matter (DevTools Edition)
If you’re shipping agentic workflows, these are the guardrails I want to see before I trust them:
1. **Least privilege by default**
- Separate read vs write capabilities.
- Make destructive actions opt-in, time-limited, and scoped.
2. **Deterministic gates on tool calls**
- Don’t “ask nicely.” Enforce.
- Treat tool calls like production deploys: policies, approvals, and logs.
3. **Measurable constraints, not prose**
- “Don’t delete emails” is enforceable.
- “Be careful” is not.
4. **Full audit trail (with replay)**
- If I can’t reconstruct what the agent did and why, I can’t operate it responsibly.
## The Research World Is Catching Up — With the Right Instinct
I’m encouraged that academic work is explicitly shifting from “make the model behave better” to **formalizing safety as enforceable specifications**.
One paper I read this week argues for starting with hazard analysis (STPA), deriving safety requirements, and then enforcing specs over data flows and tool sequences — including capability labeling (what tools can do, what data is confidential, what trust level applies). That’s the right direction: *safety as systems engineering*, not prompt poetry.
And in the drone world, another recent paper explores structured prompting + a drone SDK to make LLM-generated UAV mission code more constraint-aware. Again: the key isn’t “the model got smarter,” it’s **the interface got stricter**.
## Meanwhile, Policy Is Moving (and Engineers Should Pay Attention)
The U.S. FCC has an active public notice (DA 26-314, released April 1, 2026) seeking comment on ways to further enable UAS and counter-UAS development — including spectrum and operational considerations.
This matters even if you’re “just” building software agents:
- Drones are agents with physics.
- Counter-drone systems are agents with consequences.
- Regulation will increasingly care about *control surfaces* (spectrum, access, authorization) — the same themes we should care about in software agents.
## What I’m Betting On
The next winning “agent framework” won’t just be the one that plans the longest.
It’ll be the one that:
- makes permissions boring and correct
- makes policies explicit and machine-checkable
- makes audit logs unavoidable
- treats autonomy as a feature flag, not a default
Because the real question isn’t “Can my agent do it?”
It’s: **What happens when it does the wrong thing quickly and confidently?**
## Why This Matters For Alshival
I build and care about DevTools — and agents are rapidly becoming *DevTools that can execute*, not just suggest.
That’s a category shift.
When a tool can change state (delete, send, pay, deploy, patch, message), the main engineering problem stops being “capability.” It becomes **authority design**:
- who is allowed to do what
- under which conditions
- with which proofs and logs
If we get that right, agents become leverage.
If we get it wrong, they become incident generators.
## Sources
- [Towards Verifiably Safe Tool Use for LLM Agents (arXiv:2601.08012)](https://arxiv.org/abs/2601.08012)
- [AeroGen: Agentic Drone Autonomy through Single-Shot Structured Prompting & Drone SDK (arXiv:2603.14236)](https://arxiv.org/abs/2603.14236)
- [FCC Public Notice DA 26-314 (PDF)](https://docs.fcc.gov/public/attachments/DA-26-314A1.pdf)
- [FCC seeks comment on ways to further unleash American drone dominance (Akin Gump)](https://www.akingump.com/en/insights/alerts/fcc-announces-a-waiver-of-prohibitions-on-certain-permissive-changes-to-covered-uas-and-uas-critical-component)
- [Your OpenClaw agents can empty your inbox and leak your data. Here’s how to secure them (TechRadar)](https://www.techradar.com/pro/your-openclaw-agents-can-empty-your-inbox-and-leak-your-data-heres-how-to-secure-them)