Continuum AI Releases OrcaRouter and OrcaRouter Lite for Unified Model Inference
Continuum AI has released OrcaRouter and OrcaRouter Lite, announced in a press release. The unified inference layer connects over 200 frontier and open source language models with zero markup for bring-your-own-key traffic. Developers pay providers directly, while Continuum generates revenue through features such as caching, governance, and policy management.
OrcaRouter Lite is open source under the MIT license and can be self-hosted on a laptop, virtual private server, or cluster. It uses SQLite by default and does not require Postgres, Redis, or Kubernetes. The hosted OrcaRouter version offers accelerated inference, failover under 50 milliseconds, and adaptive routing that learns from live traffic.
The system provides unified billing across providers including OpenAI, Anthropic, Google, Mistral, and DeepSeek. User keys remain encrypted at rest with AES-256-GCM. Continuum is also offering free credits to AI developers and independent builders, accessible at OrcaRouter.ai, with the open source code available on GitHub.
We hope you enjoyed this article.
Consider subscribing to one of our newsletters like Daily AI Brief.
Also, consider following us on social media:
Subscribe to Daily AI Brief
Daily report covering major AI developments and industry news, with both top stories and complete market updates
Market report
AI’s Time-to-Market Quagmire: Why Enterprises Struggle to Scale AI Innovation
The 2025 AI Governance Benchmark Report by ModelOp provides insights from 100 senior AI and data leaders across various industries, highlighting the challenges enterprises face in scaling AI initiatives. The report emphasizes the importance of AI governance and automation in overcoming fragmented systems and inconsistent practices, showcasing how early adoption correlates with faster deployment and stronger ROI.
Read more