File size: 5,563 Bytes
bc1c255 5309389 bc1c255 5309389 bc1c255 8ac636e 5309389 52e5f26 82903a0 52e5f26 a43b342 5309389 52e5f26 bc1c255 52e5f26 47a8f8e 5309389 bc1c255 5309389 230798d 5309389 bc1c255 5309389 bc1c255 52e5f26 bc1c255 5309389 52e5f26 bc1c255 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 | ---
license: apache-2.0
tags:
- llm-routing
- model-selection
- budget-optimization
- knn
language:
- en
library_name: sklearn
pipeline_tag: text-classification
---
# R2-Router: LLM Router with Joint Model-Budget Optimization
**R2-Router** intelligently routes each query to the optimal (LLM, token budget) pair, jointly optimizing accuracy and inference cost. Ranked **#1** on the [RouterArena](https://routerarena.github.io/) leaderboard.
**Paper**: [R2-Router (arxiv)](https://arxiv.org/abs/TODO)
## RouterArena Performance
Official leaderboard results on 8,400 queries:
| Metric | Value |
|--------|-------|
| Accuracy | 71.23% |
| Cost per 1K Queries | $0.061 |
| Arena Score (beta=0.1) | **71.60** |
| Robustness Score | 45.71% |
| Rank | **#1** |
## Quick Start
### Installation
We recommend using [uv](https://docs.astral.sh/uv/) for fast, reliable environment setup:
```bash
# Install uv (if not already installed)
curl -LsSf https://astral.sh/uv/install.sh | sh
# Create environment and install dependencies
uv venv .venv && source .venv/bin/activate
uv pip install scikit-learn numpy joblib huggingface_hub vllm
```
### With vLLM Server (Recommended)
Start the embedding server once, then route from any process without reloading the model:
```bash
# Terminal 1: Start vLLM embedding server (runs once, stays alive)
uv pip install vllm
vllm serve Qwen/Qwen3-0.6B --runner pooling --port 8000
```
```python
# Terminal 2: Route queries (connects to the running server)
from huggingface_hub import snapshot_download
import sys
path = snapshot_download("JiaqiXue/r2-router")
sys.path.insert(0, path)
from router import R2Router
router = R2Router.from_pretrained(path, embed_url="http://localhost:8000")
result = router.route_text("What is the capital of France?")
print(f"Model: {result['model_full_name']}, Budget: {result['token_limit']}")
```
### Adjusting Lambda (Cost-Accuracy Tradeoff)
The `lambda` parameter controls the tradeoff between accuracy and cost:
- **lambda → 1.0**: Minimize cost (routes to cheaper models)
- **lambda → 0.0**: Maximize accuracy (routes to the best model regardless of cost)
- **Default: 0.999** (strongly cost-sensitive, as used in our RouterArena submission)
```python
# Cost-sensitive (default, as submitted to RouterArena)
router = R2Router.from_pretrained(path, lambda_val=0.999)
# Balanced accuracy vs cost
router = R2Router.from_pretrained(path, lambda_val=0.5)
# Accuracy-first (ignores cost, always picks highest quality)
router = R2Router.from_pretrained(path, lambda_val=0.0)
# Override lambda per query
result = router.route_text("Solve this integral", lambda_val=0.5)
```
### Train from Scratch
```python
from huggingface_hub import snapshot_download
import sys
path = snapshot_download("JiaqiXue/r2-router")
sys.path.insert(0, path)
from router import R2Router
# Train KNN with custom hyperparameters
router = R2Router.from_training_data(path, k=80, lambda_val=0.999)
```
## Architecture
R2-Router jointly optimizes **which model** to use and **how many tokens** to allocate per query.
### Routing Formula
```
risk(M, b) = (1 - lambda) * predicted_quality(query, M, b) - lambda * predicted_tokens(query, M) * price_M / 1e6
(M*, b*) = argmax risk
```
### Pipeline
```
Input Query
|
[1] Embed with Qwen3-0.6B -> 1024-dim vector
|
[2] For each (model, budget) pair:
- KNN predicts quality (accuracy)
- KNN predicts output token count
- Compute risk = (1-lambda) * quality - lambda * cost
|
[3] Select (model, budget) with highest risk
|
Output: (model_name, token_budget)
```
### Model Pool (6 LLMs)
| Model | Output $/M tokens |
|-------|------------------|
| Qwen3-235B-A22B | $0.463 |
| Qwen3-Next-80B-A3B | $1.10 |
| Qwen3-30B-A3B | $0.33 |
| Qwen3-Coder-Next | $0.30 |
| Gemini 2.5 Flash | $2.50 |
| Claude 3 Haiku | $1.25 |
### Token Budgets
4 output token limits: **100, 200, 400, 800** tokens.
### Key Parameters
| Parameter | Value |
|-----------|-------|
| KNN K | 80 |
| Lambda | 0.999 |
| Distance Metric | Cosine |
| KNN Weights | Distance-weighted |
| Embedding Dim | 1024 |
## Repository Contents
```
config.json # Router configuration (models, budgets, prices, hyperparams)
router.py # Self-contained inference code (embed + route)
training_data/
embeddings.npy # Sub_10 training embeddings (809 x 1024)
labels.json # Per-(model, budget) accuracy & token labels
checkpoints/
quality_knn_*.joblib # Pre-fitted KNN quality predictors (18 total)
token_knn_*.joblib # Pre-fitted KNN token predictors (6 total)
```
### Ways to Use
| Method | GPU? | Description |
|--------|------|-------------|
| `route_text()` + vLLM server | Yes (server) | Start `vllm serve` once, route from anywhere via HTTP |
| `route_text()` + local vLLM | Yes (local) | Auto-loads Qwen3-0.6B on first call, caches it |
| `route(embedding)` | No | Route from pre-computed 1024-dim embedding |
| `from_training_data(path)` | No | Train your own KNN with custom hyperparameters |
## Training Details
- **Training Data**: RouterArena sub_10 split (809 queries, 10% of full 8,400)
- **Method**: KNeighborsRegressor with cosine distance, distance-weighted
- **Evaluation**: Full 8,400 RouterArena queries (no data leakage)
- **Training Time**: < 1 second (KNN fitting)
## Citation
```bibtex
@article{r2router2026,
title={R2-Router: A New Paradigm for LLM Routing with Reasoning},
author={TODO},
year={2026},
url={https://arxiv.org/abs/TODO}
}
```
## License
Apache 2.0
|