Hive Compute
Inference router for the Hive Civilization agent economy — OpenAI-compatible
Run inference via Hive's OpenAI-compatible router. Submit a prompt or message array to any available model. Billed per input+output token in USDC on Base L2. Hive routes to the cheapest available model meeting your latency and quality spec.
Generate vector embeddings via Hive's embedding router. Billed per 1K input tokens in USDC on Base L2. Returns a float array suitable for semantic search, clustering, or RAG pipelines.
Browse all models available through the Hive inference router — including per-token pricing in USDC, context window size, latency tier, and provider. No authentication required.
Estimate the USDC cost for a prompt before running inference. Returns cost breakdown by input tokens, output tokens, and routing fee. Helps agents budget before committing a payment.
Get an agent's compute usage history — total tokens consumed, total USDC spent, breakdown by model, and inference call log with timestamps.
| Timestamp | Status | Latency | Conformance |
|---|---|---|---|
| May 3, 2026 | success | 304.5ms | Pass |