# Automatic Model Routing

Different parts of an agent want different models. Long reasoning wants a frontier model. Quick "fix this typo" calls want a fast cheap one. Vision wants a vision model. OpenHuman handles this with a built-in **router provider** so you never have to think about it.

## How a request gets routed

The model parameter on any chat call can take one of two shapes:

* **Concrete model name**. e.g. `anthropic/claude-sonnet-4`. Routes to the default provider with that exact model.
* **Hint prefix**. e.g. `hint:reasoning`. Looks the hint up in the route table and resolves to a `(provider, model)` pair.

```rust
// src/openhuman/providers/router.rs
fn resolve(&self, model: &str) -> (usize, String) {
    if let Some(hint) = model.strip_prefix("hint:") {
        if let Some((idx, resolved_model)) = self.routes.get(hint) {
            return (*idx, resolved_model.clone());
        }
    }
    (self.default_index, model.to_string())
}
```

The router wraps several pre-created providers (Anthropic, OpenAI, Google, Groq, etc.) and picks the right one per request. Hints can be remapped at runtime without restarting the core.

## Common hints

| Hint             | Typical target              | When it's used                                        |
| ---------------- | --------------------------- | ----------------------------------------------------- |
| `hint:reasoning` | A strong reasoning model    | Multi-step planning, math, code-heavy turns           |
| `hint:fast`      | A fast/cheap model          | UI helpers, autocompletes, small classification calls |
| `hint:vision`    | A vision-capable model      | Screenshots, image attachments, OCR                   |
| `hint:summarize` | A model good at compression | Memory tree summary builders                          |
| `hint:code`      | A code-tuned model          | Native coder turns                                    |

The exact mappings are configurable; the defaults ship sensible per-provider routes.

## One subscription

Routing happens behind a single OpenHuman subscription. You don't hold separate API keys for Anthropic, OpenAI, Google etc., the backend brokers access, and the router picks the right one per task. That's the "one subscription, many providers" promise from the README, made concrete.

## Overriding routes

* **Globally**. config TOML (`Config` struct in `src/openhuman/config/schema/types.rs`) can supply a custom route table at startup.
* **Per call**. pass a concrete model name (no `hint:` prefix) and the router falls through to the default provider with that exact model.
* **For a skill**. skills can pin a hint or a model in their manifest.

## Why this isn't just "model switcher"

Routing isn't a UI dropdown. The agent loop itself emits hints based on what it's about to do. You don't pick the model; the *task* does. That's the difference between "multi-model" and "smart routing".

## See also

* [Smart Token Compression](/openhuman/features/token-compression.md). what makes large reasoning calls affordable.
* [Native Tools](/openhuman/features/native-tools.md). different tool calls hint at different routes.
* [Local AI (optional)](/openhuman/features/model-routing/local-ai.md). lightweight chat hints can run on-device.


---

# Agent Instructions: Querying This Documentation

If you need additional information that is not directly available in this page, you can query the documentation dynamically by asking a question.

Perform an HTTP GET request on the current page URL with the `ask` query parameter:

```
GET https://tinyhumans.gitbook.io/openhuman/features/model-routing.md?ask=<question>
```

The question should be specific, self-contained, and written in natural language.
The response will contain a direct answer to the question and relevant excerpts and sources from the documentation.

Use this mechanism when the answer is not explicitly present in the current page, you need clarification or additional context, or you want to retrieve related documentation sections.
