Developer Docs

VIP Multi-Model API Gateway

Private capacity for priority AI workloads.

Ling.AI VIP runs as an isolated English instance with its own API endpoint, Redis runtime, database schema, and outbox workers for priority workloads.

Three compatible entry points

All examples on this VIP site target api.vip.lingapi.ai, keeping SDK configuration separate from the standard English instance.

OpenAI SDK(Chat) POST /v1/chat/completions
from openai import OpenAI

client = OpenAI(
    base_url="https://api.vip.lingapi.ai/v1",
    api_key="<YOUR_API_KEY>"
)

response = client.chat.completions.create(
    model="<model id from /v1/models>",
    messages=[{"role": "user", "content": "Introduce the Ling.AI API Gateway in three sentences."}],
    stream=False
)
OpenAI SDK(Responses) POST /v1/responses
from openai import OpenAI

client = OpenAI(
    base_url="https://api.vip.lingapi.ai/v1",
    api_key="<YOUR_API_KEY>"
)

response = client.responses.create(
    model="<model id from /v1/models>",
    input="Summarize the public endpoints and tool capabilities available in this system."
)
HTTP(x-api-key) GET /v1/models
curl -sS https://api.vip.lingapi.ai/v1/models \
  -H "x-api-key: <YOUR_API_KEY>"

Explore the docs

Browse public APIs, model capabilities, tool integrations, account usage, and async task flows based on the capabilities enabled in this instance.

Unified API surface OpenAI /v1、Anthropic Messages、Gemini generateContent
Models and media Chat reasoning, Responses tools, image generation/editing, video generation, embeddings, and audio transcription
Accounts and observability API keys, IP allowlists, wallet, packages, usage analytics, cost trends, and request history
Gateway operations Dedicated routing, isolated runtime cache, private outbox workers, provider failover, prompt caching, async tasks, and task status lookup