Supported LLM Providers

9 min read

Overview

Archestra Platform acts as a security proxy between your AI applications and LLM providers. It currently supports the following LLM providers.

OpenAI

Supported OpenAI APIs

  • Chat Completions API (/chat/completions) - ✅ Fully supported
  • Responses API (/responses) - ⚠️ Not yet supported (GitHub Issue #720)

OpenAI Connection Details

  • Base URL: http://localhost:9000/v1/openai/{profile-id}
  • Authentication: Pass your OpenAI API key in the Authorization header as Bearer <your-api-key>

Important Notes

  • Use Chat Completions API: Ensure your application uses the /chat/completions endpoint (not /responses). Many frameworks default to this, but some like Vercel AI SDK require explicit configuration (add .chat to the provider instance).
  • Streaming: OpenAI streaming responses require your cloud provider's load balancer to support long-lived connections. See Cloud Provider Configuration for more details.

Anthropic

Supported Anthropic APIs

  • Messages API (/messages) - ✅ Fully supported

Anthropic Connection Details

  • Base URL: http://localhost:9000/v1/anthropic/{profile-id}
  • Authentication: Pass your Anthropic API key in the x-api-key header

Google Gemini

Archestra supports both the Google AI Studio (Gemini Developer API) and Vertex AI implementations of the Gemini API.

Supported Gemini APIs

  • Generate Content API (:generateContent) - ✅ Fully supported
  • Stream Generate Content API (:streamGenerateContent) - ✅ Fully supported

Gemini Connection Details

  • Base URL: http://localhost:9000/v1/gemini/{profile-id}/v1beta
  • Authentication:

Using Vertex AI

To use Vertex AI instead of Google AI Studio, configure these environment variables:

VariableRequiredDescription
ARCHESTRA_GEMINI_VERTEX_AI_ENABLEDYesSet to true to enable Vertex AI mode
ARCHESTRA_GEMINI_VERTEX_AI_PROJECTYesYour GCP project ID
ARCHESTRA_GEMINI_VERTEX_AI_LOCATIONNoGCP region (default: us-central1)
ARCHESTRA_GEMINI_VERTEX_AI_CREDENTIALS_FILENoPath to service account JSON key file

For GKE deployments, we recommend using Workload Identity which provides secure, keyless authentication. This eliminates the need for service account JSON key files.

Setup steps:

  1. Create a GCP service account with Vertex AI permissions:
gcloud iam service-accounts create archestra-vertex-ai \
  --display-name="Archestra Vertex AI"

gcloud projects add-iam-policy-binding PROJECT_ID \
  --member="serviceAccount:archestra-vertex-ai@PROJECT_ID.iam.gserviceaccount.com" \
  --role="roles/aiplatform.user"
  1. Bind the GCP service account to the Kubernetes service account:
gcloud iam service-accounts add-iam-policy-binding \
  archestra-vertex-ai@PROJECT_ID.iam.gserviceaccount.com \
  --role="roles/iam.workloadIdentityUser" \
  --member="serviceAccount:PROJECT_ID.svc.id.goog[NAMESPACE/KSA_NAME]"

Replace NAMESPACE with your Helm release namespace and KSA_NAME with the Kubernetes service account name (defaults to archestra-platform).

  1. Configure Helm values to annotate the service account:
archestra:
  orchestrator:
    kubernetes:
      serviceAccount:
        annotations:
          iam.gke.io/gcp-service-account: archestra-vertex-ai@PROJECT_ID.iam.gserviceaccount.com
  env:
    ARCHESTRA_GEMINI_VERTEX_AI_ENABLED: "true"
    ARCHESTRA_GEMINI_VERTEX_AI_PROJECT: "PROJECT_ID"
    ARCHESTRA_GEMINI_VERTEX_AI_LOCATION: "us-central1"

With this configuration, Application Default Credentials (ADC) will automatically use the bound GCP service account—no credentials file needed.

Other Environments

For non-GKE environments, Vertex AI supports several authentication methods through Application Default Credentials (ADC):

  • Service account key file: Set ARCHESTRA_GEMINI_VERTEX_AI_CREDENTIALS_FILE to the path of a service account JSON key file
  • Local development: Use gcloud auth application-default login to authenticate with your user account
  • Cloud environments: Attached service accounts on Compute Engine, Cloud Run, and Cloud Functions are automatically detected
  • AWS/Azure: Use workload identity federation to authenticate without service account keys

See the Vertex AI authentication guide for detailed setup instructions for each environment.

Cerebras

Cerebras provides fast inference for open-source AI models through an OpenAI-compatible API.

Supported Cerebras APIs

  • Chat Completions API (/chat/completions) - ✅ Fully supported

Cerebras Connection Details

  • Base URL: http://localhost:9000/v1/cerebras/{agent-id}
  • Authentication: Pass your Cerebras API key in the Authorization header as Bearer <your-api-key>

Important Notes

Cohere

Cohere provides enterprise-grade LLMs designed for safe, controllable, and efficient AI applications. The platform offers features like safety guardrails, function calling, and both synchronous and streaming APIs.

Supported Cohere APIs

  • Chat API (/chat) - ✅ Fully supported
  • Streaming: ✅ Fully supported

Cohere Connection Details

  • Base URL: http://localhost:9000/v1/cohere/{profile-id}
  • Authentication: Pass your Cohere API key in the Authorization header as Bearer <your-api-key>

Environment Variables

VariableRequiredDescription
ARCHESTRA_COHERE_BASE_URLNoCohere API base URL (default: https://api.cohere.ai)
ARCHESTRA_CHAT_COHERE_API_KEYNoDefault API key for Cohere (can be overridden per conversation/team/org)

Important Notes

Mistral AI

Mistral AI provides state-of-the-art open and commercial AI models through an OpenAI-compatible API.

Supported Mistral APIs

  • Chat Completions API (/chat/completions) - ✅ Fully supported

Mistral Connection Details

  • Base URL: http://localhost:9000/v1/mistral/{agent-id}
  • Authentication: Pass your Mistral API key in the Authorization header as Bearer <your-api-key>

Getting an API Key

You can get an API key from the Mistral AI Console.

vLLM

vLLM is a high-throughput and memory-efficient inference and serving engine for LLMs. It's ideal for self-hosted deployments where you want to run open-source models on your own infrastructure.

Supported vLLM APIs

  • Chat Completions API (/chat/completions) - ✅ Fully supported (OpenAI-compatible)

vLLM Connection Details

  • Base URL: http://localhost:9000/v1/vllm/{profile-id}
  • Authentication: Pass your vLLM API key (if configured) in the Authorization header as Bearer <your-api-key>. Many vLLM deployments don't require authentication.

Environment Variables

VariableRequiredDescription
ARCHESTRA_VLLM_BASE_URLYesvLLM server base URL (e.g., http://localhost:8000/v1 or your vLLM endpoint)
ARCHESTRA_CHAT_VLLM_API_KEYNoAPI key for vLLM server (optional, many deployments don't require auth)

Important Notes

  • Configure base URL to enable vLLM: The vLLM provider is only available when ARCHESTRA_VLLM_BASE_URL is set. Without it, vLLM won't appear as an option in the platform.
  • No API key required for most deployments: Unlike cloud providers, self-hosted vLLM typically doesn't require authentication. The ARCHESTRA_CHAT_VLLM_API_KEY is only needed if your vLLM deployment has authentication enabled.

Ollama

Ollama is a local LLM runner that makes it easy to run open-source large language models on your machine. It's perfect for local development, testing, and privacy-conscious deployments.

Supported Ollama APIs

  • Chat Completions API (/chat/completions) - ✅ Fully supported (OpenAI-compatible)

Ollama Connection Details

  • Base URL: http://localhost:9000/v1/ollama/{profile-id}
  • Authentication: Pass your Ollama API key (if configured) in the Authorization header as Bearer <your-api-key>. Ollama typically doesn't require authentication.

Environment Variables

VariableRequiredDescription
ARCHESTRA_OLLAMA_BASE_URLYesOllama server base URL (e.g., http://localhost:11434/v1 for default Ollama)
ARCHESTRA_CHAT_OLLAMA_API_KEYNoAPI key for Ollama server (optional, Ollama typically doesn't require auth)

Important Notes

  • Configure base URL to enable Ollama: The Ollama provider is only available when ARCHESTRA_OLLAMA_BASE_URL is set. Without it, Ollama won't appear as an option in the platform.
  • Default Ollama port: Ollama runs on port 11434 by default. The OpenAI-compatible API is available at http://localhost:11434/v1.
  • No API key required: Ollama typically doesn't require authentication for local deployments.
  • Model availability: Models must be pulled first using ollama pull <model-name> before they can be used through Archestra.

Zhipu AI

Zhipu AI (Z.ai) is a Chinese AI company offering the GLM (General Language Model) series of large language models. The platform provides both free and commercial models with strong performance in Chinese and English language tasks.

Supported Zhipu AI APIs

  • Chat Completions API (/chat/completions) - ✅ Fully supported (OpenAI-compatible)

Zhipu AI Connection Details

  • Base URL: http://localhost:9000/v1/zhipuai/{profile-id}
  • Authentication: Pass your Zhipu AI API key in the Authorization header as Bearer <your-api-key>

Environment Variables

VariableRequiredDescription
ARCHESTRA_ZHIPUAI_BASE_URLNoZhipu AI API base URL (default: https://api.z.ai/api/paas/v4)
ARCHESTRA_CHAT_ZHIPUAI_API_KEYNoDefault API key for Zhipu AI (can be overridden per conversation/team/org)
  • GLM-4.5-Flash (Free tier) - Fast inference model with good performance
  • GLM-4.5 - Balanced model for general use
  • GLM-4.5-Air - Lightweight model optimized for speed
  • GLM-4.6 - Enhanced version with improved capabilities
  • GLM-4.7 - Latest model with advanced features

Important Notes

  • OpenAI-compatible API: Zhipu AI's API follows the OpenAI Chat Completions format, making it easy to switch between providers
  • API Key format: Obtain your API key from the Zhipu AI Platform
  • Free tier available: The GLM-4.5-Flash model is available on the free tier for testing and development
  • Chinese language support: GLM models excel at Chinese language understanding and generation, while maintaining strong English capabilities

Amazon Bedrock

Supported Bedrock APIs

  • Converse API (/converse) - ✅ Fully supported (AWS Docs)
  • Converse Stream API (/converse-stream) - ✅ Fully supported (AWS Docs)
  • InvokeModel API (/invoke) - ⚠️ Not yet supported (AWS Docs)
  • OpenAI-compatible API (Mantle) - ⚠️ Not yet supported (AWS Docs)

Bedrock Connection Details

  • Base URL: http://localhost:9000/v1/bedrock/{profile-id}
  • Authentication: Pass your Amazon Bedrock API key in the Authorization header as Bearer <your-api-key>

Environment Variables

VariableRequiredDescription
ARCHESTRA_BEDROCK_BASE_URLYesBedrock runtime endpoint URL (e.g., https://bedrock-runtime.us-east-1.amazonaws.com)
ARCHESTRA_BEDROCK_INFERENCE_PROFILE_PREFIXNoRegion prefix for cross-region inference profiles (e.g., us or eu)
ARCHESTRA_CHAT_BEDROCK_API_KEYNoDefault API key for Bedrock (can be overridden per conversation/team/org)

ARCHESTRA_BEDROCK_BASE_URL

This variable is required to enable the Bedrock provider. It specifies the regional endpoint for the Bedrock Runtime API. The URL format follows AWS regional endpoints:

https://bedrock-runtime.{region}.amazonaws.com

ARCHESTRA_BEDROCK_INFERENCE_PROFILE_PREFIX

Some Bedrock models, such as Anthropic's Claude, require cross-region inference profiles. Set this variable to enable those models. If not set, only models with on-demand inference support will be available.

For more details, see how inference works in Amazon Bedrock.