Skip to content
SolidRusT.ai

Models

Retrieve information about available models.

GET /v1/models
Terminal window
curl https://api.solidrust.ai/v1/models \
-H "Authorization: Bearer YOUR_API_KEY"
{
"object": "list",
"data": [
{
"id": "vllm-primary",
"object": "model",
"created": 1704067200,
"owned_by": "solidrust",
"permission": [],
"root": "vllm-primary",
"parent": null
},
{
"id": "bge-m3",
"object": "model",
"created": 1704067200,
"owned_by": "solidrust",
"permission": [],
"root": "bge-m3",
"parent": null
}
]
}
Model IDDescriptionContext LengthBest For
vllm-primaryRecommended alias (currently Qwen3-4B)8192All chat tasks - use this
qwen3-4bDirect model reference8192When you need a specific model
Model IDDimensionsMax InputBest For
bge-m310248192 tokensSemantic search, RAG
  • Chat completions: Use vllm-primary
  • Embeddings: Use bge-m3

When local GPU infrastructure is unavailable, requests automatically route to cloud providers:

PrimaryFailover Chain
vllm-primaryOpenAI GPT-4o-mini → Claude Haiku

This ensures high availability while maintaining API compatibility. You can detect failover by checking the model field in responses - it will indicate which model actually served the request.