Skip to main content
Cube Cloud includes a set of predefined LLM models for powering your AI agents. If you need to use a specific model version, a fine-tuned model, or route requests through a particular inference provider, you can bring your own LLM (BYOLLM) by registering a custom model in Cube Cloud. BYOLLM supports OpenAI and Anthropic model families. You can use models hosted directly by these providers or through compatible inference endpoints.

Supported providers

Cube Cloud supports the following LLM providers for custom models:
ProviderDescription
OpenAIModels accessed via the OpenAI API, including GPT-4o, GPT-4.1, and custom fine-tuned models
AnthropicModels accessed via the Anthropic API, including Claude 3.5, Claude 3.7, Claude 4, and newer releases
Different inference providers that expose OpenAI-compatible or Anthropic-compatible APIs are also supported. This lets you route requests through providers such as Azure OpenAI Service, Amazon Bedrock, or any endpoint that implements the same API contract.

Register a custom model

To register a custom model, navigate to AI → Models in the Cube Cloud settings and add a new model entry. You will need to provide:
  • Model name — a display name used to identify the model across your deployment
  • Provider — select either OpenAI or Anthropic
  • Model identifier — the model ID as recognized by the provider API (e.g., gpt-4o-2024-11-20, claude-sonnet-4-20250514)
  • API key — the secret key for authenticating with the provider
  • API base URL (optional) — a custom endpoint URL if you are using a third-party inference provider or a private deployment instead of the default provider endpoint
API keys are stored securely in Cube Cloud and are never exposed in logs or API responses. Rotate your keys regularly and use keys scoped to the minimum required permissions.

Assign a custom model to an agent

After registering a custom model, you can assign it to any agent.

Using the UI

In the Cube Cloud console, navigate to the agent configuration page and select your custom model from the LLM dropdown. Custom models appear alongside predefined models.

Using YAML configuration

Reference a custom model in your YAML agent configuration using the byom key:
agents:
  - name: my-agent
    space: my-space
    llm:
      byom:
        name: my-custom-model
You can also reference a custom model by its numeric ID:
agents:
  - name: my-agent
    space: my-space
    llm:
      byom:
        id: 123
Custom embedding models are also supported using the same syntax on the embedding_llm field:
agents:
  - name: my-agent
    space: my-space
    embedding_llm:
      byom:
        name: my-custom-embedding-model

Predefined models

Cube Cloud provides the following predefined models that you can use without any additional configuration: Anthropic models:
Model IDDescription
claude_3_5_sonnetv2Claude 3.5 Sonnet v2
claude_3_7_sonnetClaude 3.7 Sonnet
claude_3_7_sonnet_thinkingClaude 3.7 Sonnet with extended thinking
claude_4_sonnetClaude 4 Sonnet
claude_4_5_sonnetClaude 4.5 Sonnet
claude_4_5_haikuClaude 4.5 Haiku
claude_4_5_opusClaude 4.5 Opus
OpenAI models:
Model IDDescription
gpt_4oGPT-4o
gpt_4_1GPT-4.1
gpt_4_1_miniGPT-4.1 Mini
gpt_5GPT-5
gpt_5_miniGPT-5 Mini
o3o3
o4_minio4-mini
Embedding models:
Model IDDescription
text-embedding-3-largeOpenAI text-embedding-3-large
text-embedding-3-smallOpenAI text-embedding-3-small
If none of the predefined models meet your requirements, use the BYOLLM feature to register a custom model.