Agentic AI can move beyond chatbots to complete government workflows. Learn what Stateâs rollout teaches defense teams about autonomy, control, and adoption.

Agentic AI for Government Work: Lessons for Defense
State Department employees went from 3,000 beta testers to roughly 45,000â50,000 active users of an internal generative AI chatbotâout of a workforce of about 80,000. That adoption curve is the real headline, not the chatbot itself.
Because once people trust an AI assistant to answer questionsââHow much leave do I have?ââthe next demand is obvious: âGreat, now do the thing for me.â Thatâs exactly the jump the State Department CIO, Kelly Fletcher, described when she talked about moving beyond StateChat into agentic AI: systems that can take actions across multiple tools and workflows on an employeeâs behalf.
This post is part of our AI in Government & Public Sector series, and Iâm going to take a firm stance: agentic AI will matter more for national security operations than most flashy model demos. Not because it writes better prose, but because it can reduce âadministrative dragâ that quietly steals hours from mission planning, intelligence analysis, cyber defense, logistics, and diplomatic operations.
Agentic AI is about outcomes, not answers
Agentic AI is most valuable when itâs measured by tasks completed, not prompts answered. A chatbot that explains a policy is helpful; an agent that files the form, routes the approval, updates the record, and confirms completion is operational.
In the State Department example, Fletcherâs point was simple:
âI want it to not only tell me âHow much leave do I haveâ⌠but then I want it to put in my leave slip, which is in a different system.â
That âdifferent systemâ part is where government and defense workflows usually break. Policy knowledge is centralized enough to search. Execution is scattered across HR platforms, case management systems, ticketing tools, legacy apps, and role-based portals.
Why this matters for defense and national security
Defense and national security organizations have the same frictionâjust with higher stakes:
- A planner pulls data from one system, drafts a brief in another, and coordinates approvals in a third.
- A cyber analyst triages alerts in one console, submits an incident ticket in another, and documents steps in a knowledge base.
- A logistics officer checks status across multiple dashboards, then emails five people to reconcile mismatches.
Agentic AI is the first mainstream technology that can plausibly span those seamsâif itâs engineered with security, auditing, and guardrails from day one.
The boring win: reducing administrative toil at scale
The State Departmentâs best early use cases werenât âAI does diplomacy.â They were practical, like helping staff navigate internal rules. Fletcherâs example about moving a pet to an overseas post is funny because itâs true: employees waste time hunting through manuals for niche but urgent requirements.
Administrative toil is a national security issue for one reason: it competes directly with time for mission work.
Hereâs how Iâve seen this play out in real programs: leaders spend months arguing about model accuracy, while the workforce burns thousands of hours each week on tasks that donât require judgmentâjust persistence across systems.
Where agentic AI actually pays off
Agentic AI tends to deliver value fastest in workflows that are:
- High-volume (lots of repetitions across the workforce)
- Rules-based (clear policy or procedural constraints)
- Cross-system (the work breaks because tools donât talk)
- Low-risk or controllable-risk (you can add approvals and audit trails)
For government and defense environments, strong candidates include:
- Leave, travel, training requests, and status checks
- Policy-guided routing (who needs to approve what, and when)
- Drafting and packaging routine artifacts (memos, summaries, tickets)
- Cyber alert triage and prioritization (with human confirmation)
- Case intake and document classification (with strict access controls)
If you want a simple north star: start where the agent can save 15 minutes per person per week across 10,000 people. Thatâs 2,500 hours per weekâmore than 60 full-time equivalents of time returned to mission work.
Adoption is the hard partâand Stateâs numbers prove it
One of Fletcherâs most practical observations was that State underestimated the effort required to build confidence and habits.
She described having to answer, even a month prior, whether it was allowed to use the tool. Thatâs not a technical problem; itâs a governance and change-management problem.
The adoption pattern you should expect
Most agencies and defense orgs will see a similar sequence:
- Phase 1: Permission â âAm I allowed to use this?â
- Phase 2: Safety â âWill I get in trouble if itâs wrong?â
- Phase 3: Fit â âWhen should I use it vs. do it myself?â
- Phase 4: Trust â âCan it touch my systems, or just advise me?â
Agentic AI ramps the stakes because it moves from content generation to action execution. The moment an AI can submit a form, trigger a workflow, or change a record, users (and auditors) will demand clear answers to:
- Who approved the action?
- What data was used?
- What policy constrained the decision?
- What changed, exactly?
If you canât answer those questions automatically, agentic AI wonât scale in a regulated national security environment.
What training looks like when it works
The training that sticks isnât âhereâs how to prompt.â Itâs:
- Role-based playbooks (what to use, when, and why)
- Three or four blessed workflows (repeatable wins)
- Clear escalation paths (what to do when the AI is wrong)
- A feedback loop that visibly improves the system every month
Stateâs jump to 45,000â50,000 users suggests a key truth: once employees see personal time returned to them, adoption becomes contagious. But you have to get them over the first hump.
Agentic AI changes the security modelâplan for it upfront
The reason agentic AI is both promising and risky is the same: it can operate across systems.
In defense and national security, the security model canât be âthe AI has access to everything.â It has to be least privilege, with explicit controls on what actions are allowed.
Practical guardrails for agentic AI in sensitive environments
If youâre designing or buying agentic AI for government workflows, these are non-negotiable:
- Action gating: the agent can propose actions; humans approve high-impact steps.
- Strong identity binding: every action is attributable to a user and role, with clear delegation rules.
- Tool-level permissions: the agent inherits scoped permissions per system, not blanket access.
- Audit-by-default logs: prompts, tool calls, retrieved documents, and final actions are recorded and reviewable.
- Policy-as-code constraints: the agent is restricted by explicit rules (what it can file, change, route, or disclose).
A useful mental model is: treat an AI agent like a junior staffer with access to your systems. You wouldnât give a new hire admin privileges on day one. Donât do it with an agent either.
Testing is harder than people admit
Traditional software testing checks whether a function returns the expected value. Agentic AI testing must validate whether the agent:
- followed the approved workflow,
- used authorized data,
- asked for approval at the right moment,
- and produced a correct, auditable action.
Thatâs why many early agent deployments will succeed in constrained domains (HR, IT helpdesk, policy navigation) before expanding into mission operations.
From State Department workflows to mission operations
The bridge from Stateâs administrative use cases to defense missions isnât theoretical. Itâs the same workflow pattern with different nouns.
Mission planning: agents as staff accelerators
Mission planning often involves collecting inputs, reconciling versions, routing approvals, and producing briefings. An agent can:
- pull the latest authoritative data from approved sources,
- assemble a draft product in the right format,
- generate a checklist of missing inputs,
- route to the correct approvers,
- track responses and update the package.
Humans still own decisions. The agent owns the glue work.
Intelligence analysis: agents as context managers
Analysts drown in documents, not because they canât read, but because time is limited. Properly governed AI agents can:
- watch for new items in assigned collections,
- summarize changes since the last brief,
- extract entities and link them to existing case files,
- recommend what needs a human look.
The win isnât âAI replaces analysis.â The win is AI protects analyst time for the hard calls.
Cybersecurity: agents as alert triage partners
Fletcher mentioned embedding AI to prioritize cybersecurity alerts. Thatâs where agentic patterns shine:
- cluster similar alerts,
- enrich with context (asset criticality, recent changes, known IOCs),
- open a ticket with a pre-filled narrative,
- propose containment steps with a required human approval step.
This is how you scale cyber operations without pretending staffing doesnât matter.
A practical roadmap for deploying agentic AI in government
Most organizations get stuck because they start with a giant vision and no execution path. A better approach is staged autonomy.
Stage 1: âRead-onlyâ copilots (weeks)
Start with assistants that can search, summarize, and draftâusing approved internal knowledge bases.
Success metric: weekly active users and time saved per task.
Stage 2: âSuggest-onlyâ agents (1â3 months)
Let agents propose workflows and fill out forms, but require users to click âsubmit.â
Success metric: reduction in cycle time (how long approvals and submissions take).
Stage 3: âBounded-actionâ agents (3â9 months)
Grant limited action permissions for low-risk tasks (e.g., create tickets, schedule meetings, route documents) with full audit logging.
Success metric: task completion rate and exception rate (how often humans had to intervene).
Stage 4: Mission-aligned agents (ongoing)
Expand into mission workflows only after governance, security, and testing maturity is proven.
Success metric: mission throughput (more analyses produced, faster planning cycles, improved triage).
What leaders should do before buying anything
If youâre responsible for AI adoption in defense, national security, or the broader public sector, youâll save time by forcing clarity early:
- Pick one painful workflow with a measurable baseline (cycle time, backlog size, labor hours).
- Decide the autonomy ceiling upfront (read-only, suggest-only, or bounded-action).
- Define the audit requirement in plain language (who needs to see what, and when).
- Plan training as a product, not a one-off webinar.
- Budget for integration, because cross-system action is where value lives.
Agentic AI succeeds when itâs treated as an operational system, not an innovation lab experiment.
Where this is headed in 2026âand why you should care now
AI vendors are openly targeting agents that can run multi-hour or multi-day tasks with limited supervision. That timeline isnât distant. Procurement cycles, accreditation, and workforce training alone can eat a year.
The State Departmentâs experience offers a clean lesson for our AI in Government & Public Sector series: the fastest path to mission impact is often through the âunsexyâ work. Automate the administrative friction first, earn trust, and then expand autonomy into mission workflows.
If youâre building toward agentic AI for mission planning and operational efficiency, the next step is straightforward: identify two workflowsâone administrative and one mission-adjacentâwhere an agent can reduce cycle time without expanding risk. Then prove it with logs, controls, and measurable outcomes.
The question Iâd leave you with is the one leaders should be asking heading into 2026 budgeting: Which mission-critical teams are doing âglue workâ that an agent could handleâand what would they do with that time back?