← Back to Blog

Model Routing for AAO: How to Cut Agent Cost Without Cutting Quality

TL;DR: Model routing is the AAO practice of sending each agent step to the cheapest model and workflow that can do it reliably. The goal is not to use weaker AI. The goal is to stop wasting premium reasoning on work that only needed classification, retrieval, formatting, or a deterministic rule.

The short answer

Model routing cuts agent cost by matching task difficulty to capability. A support triage step, a source fetch, a JSON clean-up, a brand-tone rewrite, and a legal-risk decision do not need the same model, context window, tool access, or verification depth.

Good routing creates a small operating system around AI agents. It decides when to use a fast model, when to call a specialist agent, when to escalate to a premium reasoning model, and when to stop the workflow for human review.

Quotable nugget: The most expensive agent architecture is not the one with the best model. It is the one that uses the best model for work a cheaper path could have handled safely.

Why model routing belongs inside AAO

Assistive Agent Optimisation exists because businesses are no longer asking AI to answer occasional questions. They are using agents as operational staff: drafting content, checking code, triaging inboxes, auditing websites, preparing quotes, and watching dashboards. Once agents touch repeatable workflows, cost and reliability become design problems rather than procurement footnotes.

That is where routing matters. A prompt can improve a single response. A routing policy improves the whole system. It determines which agent works first, which evidence is gathered, which model is allowed to take action, and which outputs need a second check.

The serious teams are moving from “which model is best?” to “which model is best for this step, under this risk level, with this evidence, at this price?” That question is the heart of AAO.

The wrong way: one premium model for everything

The simplest agent system sends every request to the strongest available model with every useful tool attached. It feels sensible because the quality ceiling is high and the architecture is easy to explain. In production, it usually creates three problems.

  • Cost leakage: routine classification, formatting, and extraction jobs consume the same budget as complex reasoning.
  • Latency drag: easy tasks wait behind heavy workflows because everything takes the premium path.
  • Risk concentration: one agent accumulates too many permissions, too much context, and too many opportunities to make a consequential mistake.

That design can be acceptable for a prototype. It is rarely the right operating model for an AI workforce.

The better question: what capability does this step actually need?

Every agent step has a capability profile. Before choosing a model, define what the step actually requires. Is it classifying intent? Summarising a known source? Searching for evidence? Planning a multi-step action? Writing in brand voice? Executing a file edit? Verifying a claim against live data?

Those jobs differ along five dimensions:

Routing dimensions for agent workflows
DimensionLow-demand exampleHigh-demand example
Reasoning depthClassify a ticket categoryDiagnose conflicting analytics evidence
Context sizeRead one short form responseCompare a full codebase, brief, and live logs
RiskDraft an internal noteModify production content or advise on regulated claims
CreativityNormalise data fieldsWrite a strategic narrative for a founder audience
Verification needFormat valid JSONProve a live fix, source claim, or compliance decision

Routing becomes much easier when each step is labelled this way. The model choice stops being ideological and starts being operational.

A practical routing ladder

Most teams do not need a complex model marketplace on day one. They need a clear ladder. Start with the cheapest safe path and climb only when the task proves it needs more capability.

  1. Deterministic rule: use code, templates, regex, validation schemas, or database lookups when the task does not require judgement.
  2. Small/fast model: use it for classification, extraction, tagging, basic summarisation, and routine transformations.
  3. Specialist agent: use a role-specific prompt, memory slice, and tool set when the job has clear domain boundaries.
  4. Premium reasoning model: use it for ambiguous planning, multi-source synthesis, high-value strategy, debugging, or decisions with material downside.
  5. Human escalation: use it when the risk, uncertainty, or authority requirement exceeds what the agent system should own.

The ladder is not about being cheap for its own sake. It is about spending intelligence where intelligence changes the outcome.

Route by risk, not just by difficulty

A task can be technically easy and still risky. For example, changing a product title, sending an email, or updating metadata may require little reasoning but can have commercial or reputational consequences. A good router therefore considers both difficulty and consequence.

Low-risk, low-difficulty tasks can often run automatically. Low-risk, high-difficulty tasks may justify a stronger model but not a human. High-risk, low-difficulty tasks need permission gates. High-risk, high-difficulty tasks need both strong reasoning and verification.

Quotable nugget: Difficulty chooses the model. Risk chooses the gate.

Use verification to make cheaper routing safe

Cheap routes become viable when verification is designed well. A small model can extract facts if a validator checks the output shape. A drafting agent can write fast if another step checks sources, links, and brand constraints. A code assistant can make a small edit if tests and static checks run before release.

This is why AAO treats verification as part of the workflow, not as an afterthought. The verifier may be deterministic, model-based, human, or a combination. What matters is that the route has a defined failure detector.

Anthropic’s public guidance on effective agents makes the same broad point from another angle: agent systems work best when workflows are composable, tool use is deliberate, and evaluation is built around the task rather than the demo. That is the spirit of routing. You do not pay for magic. You design the path.

Context is a cost centre

Model spend is not only about which model answers. It is also about how much context you feed it. Agents often become expensive because every step receives the whole conversation, every document, every brand rule, and every historical note.

A routing-aware system trims context aggressively. The classifier does not need the whole knowledge base. The writer needs the brief, the audience, the evidence, and the target format. The verifier needs the requirements and the final output. The compliance reviewer needs policy-relevant passages, not the entire brainstorm.

Prompt caching, retrieval design, summaries, and durable memory all matter here. But the first principle is simpler: give each agent the smallest context that still lets it succeed.

The routing policy should be written down

If routing lives only in one engineer’s head, it will drift. Write the policy like an operations document. Define task classes, default paths, escalation triggers, and QA requirements.

A simple routing policy might say:

  • classification and tagging go to the fast model unless confidence is below a threshold
  • source-backed content must run retrieval first and citation QA last
  • production edits require a planning step, an execution step, and a verification step
  • regulated, legal, medical, financial, or brand-risk claims require human approval
  • unknown task types go to a planner, not directly to an executor

This turns routing from improvisation into governance.

What to measure

Routing only matters if it improves the operating metrics. Track the numbers that show whether the system is cheaper, faster, and still trustworthy.

  • Cost per completed task: not just tokens per call, but total cost after retries and QA.
  • First-pass success rate: how often the selected route completes without escalation.
  • Escalation rate: how often cheap paths correctly recognise that they need help.
  • Rework rate: how often outputs pass the route but fail downstream review.
  • Latency by task class: whether routing makes routine work noticeably faster.
  • Incident rate: whether automation creates unsafe sends, bad edits, unsupported claims, or broken handoffs.

Do not optimise for the lowest token bill if the result is more rework. The real metric is cost per trusted outcome.

A starter model-routing architecture

For many businesses, the first useful architecture is straightforward:

  1. Intake classifier: identifies task type, risk level, required tools, and confidence.
  2. Router: chooses deterministic path, small model, specialist agent, premium model, or human handoff.
  3. Specialist executor: performs the task with a narrow prompt and only the tools it needs.
  4. Verifier: checks schema, source grounding, tests, brand rules, or live evidence.
  5. Logger: records route, model, cost, outcome, errors, and escalation reason.

That architecture is boring in the best possible way. It gives you levers. You can see which tasks are expensive, which route fails, which verifier catches problems, and where human review is still necessary.

Where routing goes wrong

The common failure is false economy. Teams route to weaker models without changing the workflow around them. Then the output quality drops, trust collapses, and everyone decides routing was the problem. It was not. The problem was treating routing as model swapping rather than workflow design.

Other failure modes include:

  • using confidence scores that are not calibrated against real outcomes
  • forgetting to count retry cost
  • routing based on department politics rather than task properties
  • letting cheap models take irreversible actions without gates
  • failing to update the policy after new task patterns emerge

Routing is a living system. It should change as models improve, prices change, workflows mature, and new risks appear.

The AAO view: intelligence is an allocation problem

The next wave of AI operations will not be won by companies that simply buy more model capacity. It will be won by companies that allocate intelligence well. That means knowing when to automate, when to retrieve, when to reason, when to verify, and when to stop.

SAGEO makes a business visible to search engines, answer engines, and generative systems. AAO makes the business effective when AI agents become part of the operating team. Model routing is one of the bridge disciplines between those worlds because it translates abstract AI capability into repeatable commercial performance.

Quotable nugget: In agent operations, intelligence is not a subscription tier. It is a resource allocation discipline.

FAQ

What is model routing in AI agent systems?

Model routing is the process of choosing which model, agent, tool path, or human gate should handle each step in a workflow. It is based on task type, difficulty, risk, context size, latency needs, and verification requirements.

Does model routing mean using cheaper models?

Sometimes, but not always. Model routing means using the cheapest safe route. Many steps can use deterministic logic or smaller models, while ambiguous or high-risk steps should still use stronger models and stricter QA.

How does routing reduce AI agent cost?

Routing reduces cost by keeping routine classification, extraction, formatting, and low-risk drafting away from premium reasoning paths. It also reduces retry waste by escalating uncertain tasks earlier.

What is the biggest risk of model routing?

The biggest risk is routing to a cheaper path without adding appropriate verification. If the system cannot detect failures, lower model cost can turn into higher operational risk and more human rework.

How should a company start with model routing?

Start by labelling common tasks by difficulty and risk, then define a simple routing ladder: deterministic rule, fast model, specialist agent, premium reasoning model, and human escalation. Measure first-pass success, cost per trusted outcome, and rework rate.

About the author: Firdaus Nagree writes about SAGEO and AAO — the operating disciplines for being found, cited, and used in search and agent-led workflows.

Next: connect this model-routing layer to Assistive Agent Optimisation, agent memory architecture, and multi-agent architectures.