Skip to main content

LLM Routing

Genie Nexus provides powerful LLM routing capabilities that allow you to seamlessly work with multiple AI providers through a unified OpenAI-compatible API. This guide covers all aspects of LLM routing, from basic setup to advanced features.

Overview

LLM routing in Genie Nexus enables you to:

  • Unified API: Use OpenAI-compatible endpoints for all providers
  • Provider Abstraction: Switch between providers without code changes
  • Intelligent Fallback: Automatically failover when providers are unavailable
  • Request Optimization: Batch requests and optimize token usage
  • Cost Management: Track and control spending across providers
  • Response Streaming: Support both streaming and non-streaming responses

Supported Providers

  • Anthropic
  • CloudRift
  • Google Studio AI
  • Google Vertex AI
  • Groq
  • Inference Net
  • Kluster AI
  • Mistral
  • OpenAI
  • Together AI
  • xAI
note

Any other OpenAI-compatible provider is supported, just configure it as a provider and use it in a deployment.

Static Responses

A static response can be configured to test your deployments or for development purposes. Your request will not leave the Genie Nexus platform and no costs will be incurred.