Three-Axis Model

Omegon organizes inference along three independent axes. The agent adjusts these autonomously based on task complexity, or operators can override explicitly. This gives fine-grained control over capability, cost, and context without coupling them.

Three-axis routing model: capability tier, thinking level, and context class feed into the routing engine

Axis 1: Capability Tier

Controls which model family is used. The agent switches tiers with set_model_tier.

TierUse CaseTypical Models
localOn-device, zero API costOllama (qwen3, devstral, llama)
retributionSimple lookups, boilerplateClaude Haiku, GPT-4o-mini, Groq
victoryRoutine coding, executionClaude Sonnet, GPT-4o, Codestral
glorianaArchitecture, deep reasoningClaude Opus, o1-pro
/model                   # Opens selector with all available models
set_model_tier(victory)  # Agent self-selects (automatic)

Axis 2: Thinking Level

Controls extended reasoning budget — how many tokens the model spends "thinking" before responding.

LevelUse Case
offNo extended thinking. Fastest, cheapest.
minimalBrief sanity check.
lowLight reasoning for straightforward tasks.
mediumDefault. Balanced reasoning for general work.
highDeep reasoning for architecture, debugging, multi-step problems.
/think high     # Manual override
/think off      # Disable for speed

Axis 3: Context Class

Controls context window capacity. Larger windows allow more conversation history and memory injection.

ClassWindowUse Case
Squad128k tokensQuick tasks, simple edits
Maniple272k tokensStandard development work
Clan400k tokensLarge refactors, multi-file analysis
Legion1M tokensMassive codebases, full-project comprehension
/context legion   # Pin to maximum context
/context squad    # Minimize for cost

Orthogonality

The three axes are independent. You can run gloriana + off + squad (powerful model, no thinking, small window) or retribution + high + legion (cheap model, deep reasoning, huge context). The agent tunes these combinations based on what it's doing:

Provider Matrix

Which providers serve which tiers depends on available credentials and model capabilities:

Providerlocalretributionvictorygloriana
AnthropicHaikuSonnetOpus
OpenAI4o-mini4o / o3o1-pro
Codexcodex-mini
OpenRoutervariesvariesvaries
Groq
xAIGrok
MistralCodestral
Cerebras
Ollama

The routing engine (routing.rs) scores available providers against the requested tier and selects the best match. See Providers for details.