All Insights
7 min read

The Case for Graduated Agent Execution

Why starting solo and escalating to teams beats launching full swarms. A framework from building AI-augmented operations at enterprise scale.

AI OperationsFrameworkMulti-Agent Systems
JW

Jason Walker

State CISO, Florida

The conventional wisdom in multi-agent AI system design is binary: either you run everything solo, bottlenecking at your own capacity, or you spin up a full swarm and pay the coordination tax upfront. I've learned through building real operational systems that this framing is wrong. The answer isn't one or the other. It's graduated execution—starting lean and escalating only when thresholds demand it.

This matters because coordination overhead is real. Every agent you add doesn't multiply your throughput linearly. It adds handoff latency, decision-making overhead, and the cognitive load of managing parallel work. At enterprise scale—managing 35 Florida state agencies while running doctoral research and building trading infrastructure—I discovered that most work doesn't need teams. But some work does, and the question becomes: how do you know the difference?

The Binary Trap

I spent the first eight months of building AI-augmented operations thinking in absolutes. Either a task was simple enough that a single agent could handle it end-to-end, or it was complex enough to justify spinning up a team. This led to two bad patterns. First, I overloaded solo agents with tasks that should have been distributed—literature reviews, dissertation data processing, FAIR simulations—which created bottlenecks and degraded output quality. Second, I launched full teams for work that didn't need coordination, wasting token budget and creating unnecessary decision trees.

The inflection point came when I was managing twelve concurrent projects with overlapping stakeholders. I had a solo agent reviewing literature, a solo agent doing threat analysis, a solo agent tracking agency engagement. Each was running at capacity. Adding more solo agents didn't scale. Launching dedicated teams for each function added friction. What I needed was a middle ground.

The Four-Tier Model

Graduated execution uses four tiers, each with clear entry criteria and escalation signals:

Tier 1: Solo Agent. The default. One agent, working through a task sequentially. Typical work: research tasks, draft writing, data analysis, fact-checking, single-domain problems. Entry threshold: under 50 output items, single domain, deliverable fits in one session. This is 80% of work. Coordination overhead: zero. Token cost: lowest.

Tier 2: Clone (Parallel Solo). When a single agent can't keep pace with bulk work, you spawn identical agents running the same task across different input sets. Useful for: processing meeting transcripts in parallel, analyzing multiple agencies simultaneously, running backtests on different asset universes, extracting data from 30 research papers. Entry threshold: 50-200 input items, single task type, deterministic work. Coordination: minimal (aggregation phase at the end). Token cost: linear scaling.

Tier 3: Lightweight Team. Two or three agents with complementary roles, working on the same deliverable but from different angles. Example: one agent drafts a briefing, another fact-checks it, a third optimizes for the audience. Or: one agent runs FAIR simulations while another prepares the statistical analysis. Entry threshold: multi-phase deliverable, 3-5 day timeline, 2-3 domain expertise required, high stakes (executive visibility, publication). Coordination: moderate (3-4 async handoffs). Token cost: 40-60% premium for overlap.

Tier 4: Full Team. Four or more agents with distinct roles, working a complex multi-month project with stakeholder management, iterative feedback, and cross-domain dependencies. Example: a doctoral dissertation team with chapter writer, literature reviewer, methodology analyst, and advisor bot. Entry threshold: 10+ week project, 4+ domains, multiple stakeholders, feedback loops, publication or leadership deliverable. Coordination: high (structured, daily). Token cost: 70-100% premium.

Threshold-Based Routing

The trick is knowing when to escalate. This isn't subjective. I use measurable thresholds:

File or item count. Under 10 items? Definitely solo. 10-50? Solo plus maybe a clone run. 50-200? Clone tier or lightweight team. 200+? Full team or batch processing.

Deliverable phases. Single output? Solo. Two phases (draft + review)? Clone or lightweight team. Three or more phases with feedback loops? Full team.

Domain count. Tasks confined to one domain (e.g., "analyze this agency's risk posture")? Solo. Two or three domains (e.g., "threat intelligence + policy + organizational capacity")? Lightweight team. Four or more? Full team.

Stakeholder count. One stakeholder (or none, self-directed)? Solo. Two? Lightweight team. Three or more? Full team.

Timeline. Same-session work (4-8 hours)? Solo. Single week? Solo or lightweight team. Multiple weeks? Lightweight or full team. Monthly or longer? Full team.

Signal-Word Routing

Natural language maps to execution tiers automatically if you train yourself to listen for it:

"Review this" → Solo. "Review these 30 papers" → Clone. "Draft a briefing incorporating multiple agencies' risk postures and external benchmarks" → Lightweight team. "Build the full FAIR simulation infrastructure, write the methodology chapter, prepare the statistical analysis, and coordinate with the advisor" → Full team.

The pattern: single imperative verb + singular object = solo. Plural objects or multiple verbs = escalate. Stakeholders mentioned = escalate further.

The Escalation Window

Don't pre-commit to a tier. Start solo. If signals emerge mid-task—the work is harder than expected, output quality is degrading, you're running out of context window—escalate up a tier. This is the critical advantage over binary thinking. You don't need to know upfront whether a task needs a team. You can discover it.

I built this into my dispatch logic: start with a solo agent, but include escalation criteria in the prompt. If the agent detects it's running out of capacity, it flags the work for promotion to a clone run or lightweight team. This reduces the guesswork and makes escalation data-driven.

Evidence from Operations

I tracked token spend and delivery quality across 120 tasks over three months:

Solo agents: 89 tasks, 2.1 tokens per output word, 94% delivery rate. Fastest time-to-first-output: 6 minutes.

Clone runs: 22 tasks, 1.8 tokens per output word (token efficiency improves with parallelization), 98% delivery rate. Higher throughput on bulk work, but coordination time adds 15-30 minutes.

Lightweight teams: 8 tasks, 2.4 tokens per output word (overlap increases token cost slightly), 100% delivery rate. Output quality 8-12% higher on complex deliverables.

Full teams: 1 ongoing task (the crypto trading platform), 3.1 tokens per output word (significant overlap and coordination), but enabled a $2.1M portfolio backtest that no individual agent could deliver in one sprint.

The data supports the model: solo is most efficient for most work, but there's a threshold—around 50-100 input items or 2-3 week timelines—where escalation becomes justified.

The Coordination Tax

Here's what people miss about multi-agent systems: you don't just add agents and expect linear scaling. Every agent you add increases:

Decision latency. With one agent, decisions happen in-session. With four agents on a team, you need handoff decisions (what does Agent A pass to Agent B?), merge decisions (how do we combine outputs?), and conflict resolution (what if two agents reach different conclusions?).

Token overhead. Agents need to explain their work to each other. A solo agent's implicit context becomes explicit instruction between agents.

Communication surface. The number of potential miscommunications grows with every agent pair. With 4 agents, you have 6 potential communication paths. With 8 agents, you have 28.

These costs are real and measurable. They're worth it for genuinely complex work, but they're not worth it for straightforward tasks. Graduated execution keeps you on the efficient frontier: using the lightest execution tier that delivers your required quality within your time budget.

The Closing Question

The standard framing asks: "Should I use multi-agent systems?" I'd reframe it: "At what threshold of complexity does my work require escalation?"

For most professionals, that threshold is high. Most problems aren't multi-domain or multi-stakeholder. Most work doesn't hit the file count or timeline thresholds. For those problems, a well-configured solo agent will outperform a team on token efficiency, speed, and quality.

But the problems that do cross those thresholds—long-term projects, multi-domain research, outputs consumed by multiple stakeholders—those problems demand coordination. The graduated execution model gives you a framework for recognizing when you're at that threshold and a protocol for escalating cleanly when you cross it.

Start solo. Scale as the work demands. Don't pre-architect for a team size you don't need yet. This approach has cut my average time-to-delivery by 23% and improved output consistency across four operational domains.