Chat Completions
Compatible with the OpenAI /v1/chat/completions endpoint. Supports Claude and GPT model families.
Endpoint
- Method:
POST - URL:
https://1688token.ai/v1/chat/completions - Auth: Bearer Token
Request Parameters
| Parameter | Type | Required | Description |
|---|---|---|---|
model | string | ✅ | Model name — see supported models below |
messages | array | ✅ | List of conversation messages |
stream | boolean | — | Enable streaming output. Default: false |
temperature | number | — | Randomness, range 0~2. Default: 1 |
max_tokens | integer | — | Maximum output tokens |
top_p | number | — | Nucleus sampling probability, range 0~1 |
tools | array | — | Tool/function call definitions |
tool_choice | string/object | — | Tool selection strategy |
messages Format
Each message is an object with:
| Field | Type | Description |
|---|---|---|
role | string | system / user / assistant / tool |
content | string | array | Message content; supports text or multimodal |
Request Examples
python
from openai import OpenAI
client = OpenAI(
api_key="YOUR_API_KEY",
base_url="https://1688token.ai/v1"
)
response = client.chat.completions.create(
model="claude-sonnet-4-6",
messages=[
{"role": "system", "content": "You are a professional AI assistant."},
{"role": "user", "content": "Explain large language models in one sentence."}
],
temperature=0.7,
max_tokens=512
)
print(response.choices[0].message.content)javascript
import OpenAI from "openai";
const client = new OpenAI({
apiKey: "YOUR_API_KEY",
baseURL: "https://1688token.ai/v1"
});
const response = await client.chat.completions.create({
model: "claude-sonnet-4-6",
messages: [
{ role: "system", content: "You are a professional AI assistant." },
{ role: "user", content: "Explain large language models in one sentence." }
],
temperature: 0.7,
max_tokens: 512
});
console.log(response.choices[0].message.content);bash
curl https://1688token.ai/v1/chat/completions \
-H "Content-Type: application/json" \
-H "Authorization: Bearer YOUR_API_KEY" \
-d '{
"model": "claude-sonnet-4-6",
"messages": [
{"role": "system", "content": "You are a professional AI assistant."},
{"role": "user", "content": "Explain large language models in one sentence."}
],
"temperature": 0.7,
"max_tokens": 512
}'Response Example
json
{
"id": "chatcmpl-abc123",
"object": "chat.completion",
"created": 1715000000,
"model": "claude-sonnet-4-6",
"choices": [
{
"index": 0,
"message": {
"role": "assistant",
"content": "A large language model is an AI system trained on massive text datasets that can understand and generate natural language."
},
"finish_reason": "stop"
}
],
"usage": {
"prompt_tokens": 32,
"completion_tokens": 28,
"total_tokens": 60
}
}Streaming
Set stream: true to receive Server-Sent Events (SSE):
python
stream = client.chat.completions.create(
model="gpt-4o",
messages=[{"role": "user", "content": "Write a short poem"}],
stream=True
)
for chunk in stream:
delta = chunk.choices[0].delta.content
if delta:
print(delta, end="", flush=True)javascript
const stream = await client.chat.completions.create({
model: "gpt-4o",
messages: [{ role: "user", content: "Write a short poem" }],
stream: true
});
for await (const chunk of stream) {
const delta = chunk.choices[0]?.delta?.content;
if (delta) process.stdout.write(delta);
}Supported Models
Claude
| Model | Context | Description |
|---|---|---|
claude-opus-4-7 | 200K | Strongest reasoning, best for complex tasks |
claude-sonnet-4-6 | 200K | Balanced performance & speed — recommended |
claude-haiku-4-5-20251001 | 200K | Fast & lightweight for high-concurrency |
GPT
| Model | Context | Description |
|---|---|---|
gpt-4o | 128K | Multimodal flagship |
gpt-4o-mini | 128K | Cost-effective everyday model |
gpt-4.1 | 1M | Ultra-long context for large documents |
gpt-4.1-mini | 1M | Long-context lightweight version |
