Skip to Content

Model Catalog

OfoxAI provides unified access to mainstream LLMs. You can browse all available models in the OfoxAI Model Catalog , or programmatically retrieve complete model information via the Models API.

Model Naming Convention

All models follow the provider/model-name format:

anthropic/claude-sonnet-4.5 google/gemini-3-flash-preview moonshotai/kimi-k2.5

Models API Standard

OfoxAI’s Models API follows the OpenRouter standard, returning complete metadata for each model in JSON format.

API Response Structure

Root Response Object

{ "object": "list", "data": [ /* Array of Model objects */ ] }

Model Object

Each model contains the following standardized fields:

FieldTypeDescription
idstringModel identifier used in API requests, e.g. "anthropic/claude-sonnet-4.5"
canonical_slugstringPermanent model identifier, never changes
namestringModel display name
creatednumberTime model was added (Unix timestamp)
descriptionstringDetailed description of model capabilities and features
context_lengthnumberMaximum context window size (tokens)
architectureArchitectureModel technical architecture information
pricingPricingModel pricing information
top_providerTopProviderPrimary provider configuration
supported_parametersstring[]List of supported API parameters

Architecture Object

Describes the model’s input/output modalities and tokenizer information:

{ "modality": "text+image+file->text", "input_modalities": ["text", "image", "file"], "output_modalities": ["text"], "tokenizer": "claude", "instruct_type": null }
FieldDescription
modalityShorthand for input/output modalities, e.g. text+image->text
input_modalitiesSupported input types: text, image, audio, file
output_modalitiesSupported output types: text
tokenizerTokenizer type
instruct_typeInstruction format type (null for some models)

Pricing Object

All prices are in USD per token. A value of "0" indicates free.

{ "prompt": "0.000001", "completion": "0.000005", "input_cache_read": "0.0000001", "input_cache_write_5m": "0.00000125", "input_cache_write_1h": "0.000002" }
FieldDescription
promptInput token price
completionOutput token price
input_cache_readCache read token price
input_cache_write_5m5-minute cache write token price
input_cache_write_1h1-hour cache write token price

Different models use different tokenizers, so even with identical input and output text, the token count (and cost) may vary. Use the usage field in the response to get the actual token consumption.

TopProvider Object

{ "context_length": 200000, "max_completion_tokens": 8192, "is_moderated": false }
FieldDescription
context_lengthProvider-level context limit
max_completion_tokensMaximum tokens per response
is_moderatedWhether content moderation is enabled

Supported Parameters

The supported_parameters array indicates which OpenAI-compatible parameters the model supports:

ParameterDescription
temperatureSampling temperature control
top_pNucleus sampling parameter
max_tokensMaximum response length
stopCustom stop sequences
toolsFunction Calling / Tool Use
tool_choiceTool selection strategy
response_formatOutput format specification (JSON Mode)
reasoningDeep reasoning mode

Fetching the Model List

The Models API is a public endpoint that does not require an API Key.

Terminal
curl https://api.ofox.ai/v1/models

For the complete live model list and pricing, visit the OfoxAI Model Catalog . For API endpoint details, see the Models API Reference.

Last updated on