Skip to content

Chat Completions

Compatible with the OpenAI /v1/chat/completions endpoint. Supports Claude and GPT model families.

Endpoint

  • Method: POST
  • URL: https://1688token.ai/v1/chat/completions
  • Auth: Bearer Token

Request Parameters

ParameterTypeRequiredDescription
modelstringModel name — see supported models below
messagesarrayList of conversation messages
streambooleanEnable streaming output. Default: false
temperaturenumberRandomness, range 0~2. Default: 1
max_tokensintegerMaximum output tokens
top_pnumberNucleus sampling probability, range 0~1
toolsarrayTool/function call definitions
tool_choicestring/objectTool selection strategy

messages Format

Each message is an object with:

FieldTypeDescription
rolestringsystem / user / assistant / tool
contentstring | arrayMessage content; supports text or multimodal

Request Examples

python
from openai import OpenAI

client = OpenAI(
    api_key="YOUR_API_KEY",
    base_url="https://1688token.ai/v1"
)

response = client.chat.completions.create(
    model="claude-sonnet-4-6",
    messages=[
        {"role": "system", "content": "You are a professional AI assistant."},
        {"role": "user", "content": "Explain large language models in one sentence."}
    ],
    temperature=0.7,
    max_tokens=512
)

print(response.choices[0].message.content)
javascript
import OpenAI from "openai";

const client = new OpenAI({
  apiKey: "YOUR_API_KEY",
  baseURL: "https://1688token.ai/v1"
});

const response = await client.chat.completions.create({
  model: "claude-sonnet-4-6",
  messages: [
    { role: "system", content: "You are a professional AI assistant." },
    { role: "user", content: "Explain large language models in one sentence." }
  ],
  temperature: 0.7,
  max_tokens: 512
});

console.log(response.choices[0].message.content);
bash
curl https://1688token.ai/v1/chat/completions \
  -H "Content-Type: application/json" \
  -H "Authorization: Bearer YOUR_API_KEY" \
  -d '{
    "model": "claude-sonnet-4-6",
    "messages": [
      {"role": "system", "content": "You are a professional AI assistant."},
      {"role": "user", "content": "Explain large language models in one sentence."}
    ],
    "temperature": 0.7,
    "max_tokens": 512
  }'

Response Example

json
{
  "id": "chatcmpl-abc123",
  "object": "chat.completion",
  "created": 1715000000,
  "model": "claude-sonnet-4-6",
  "choices": [
    {
      "index": 0,
      "message": {
        "role": "assistant",
        "content": "A large language model is an AI system trained on massive text datasets that can understand and generate natural language."
      },
      "finish_reason": "stop"
    }
  ],
  "usage": {
    "prompt_tokens": 32,
    "completion_tokens": 28,
    "total_tokens": 60
  }
}

Streaming

Set stream: true to receive Server-Sent Events (SSE):

python
stream = client.chat.completions.create(
    model="gpt-4o",
    messages=[{"role": "user", "content": "Write a short poem"}],
    stream=True
)

for chunk in stream:
    delta = chunk.choices[0].delta.content
    if delta:
        print(delta, end="", flush=True)
javascript
const stream = await client.chat.completions.create({
  model: "gpt-4o",
  messages: [{ role: "user", content: "Write a short poem" }],
  stream: true
});

for await (const chunk of stream) {
  const delta = chunk.choices[0]?.delta?.content;
  if (delta) process.stdout.write(delta);
}

Supported Models

Claude

ModelContextDescription
claude-opus-4-7200KStrongest reasoning, best for complex tasks
claude-sonnet-4-6200KBalanced performance & speed — recommended
claude-haiku-4-5-20251001200KFast & lightweight for high-concurrency

GPT

ModelContextDescription
gpt-4o128KMultimodal flagship
gpt-4o-mini128KCost-effective everyday model
gpt-4.11MUltra-long context for large documents
gpt-4.1-mini1MLong-context lightweight version