Skip to main content
POST
/
v1
/
chat
/
completions
curl --request POST \
  --url https://api.evolink.ai/v1/chat/completions \
  --header 'Authorization: Bearer <token>' \
  --header 'Content-Type: application/json' \
  --data '
{
  "model": "MiniMax-M2.5",
  "messages": [
    {
      "role": "user",
      "content": "Please introduce yourself"
    }
  ]
}
'
{
  "id": "cmpl-04ea926191a14749b7f2c7a48a68abc6",
  "model": "MiniMax-M2.5",
  "object": "chat.completion",
  "created": 1698999496,
  "choices": [
    {
      "index": 0,
      "message": {
        "role": "assistant",
        "content": "Hi there! How can I help you?"
      },
      "finish_reason": "stop"
    }
  ],
  "usage": {
    "prompt_tokens": 8,
    "completion_tokens": 292,
    "total_tokens": 300
  }
}

Authorizations

Authorization
string
header
required

##All APIs require Bearer Token authentication##

Get API Key:

Visit API Key Management Page to get your API Key

Add to request header:

Authorization: Bearer YOUR_API_KEY

Body

application/json
model
enum<string>
required

Chat model name

Available options:
MiniMax-M2.5
Example:

"MiniMax-M2.5"

messages
object[]
required

List of conversation messages, supports multi-turn dialogue

Minimum array length: 1
max_tokens
integer

Maximum number of tokens for generated content, upper limit is 2048

Note:

  • Content exceeding the limit will be truncated
  • If generation is interrupted due to length, try increasing this value
Required range: 1 <= x <= 2048
Example:

2048

stream
boolean

Whether to return the response in streaming mode

  • true: Stream response, return content in real-time chunks
  • false: Wait for complete response before returning
Example:

false

temperature
number

Sampling temperature, controls output randomness

Note:

  • Lower values (e.g. 0.1): More deterministic, focused output
  • Higher values (e.g. 0.9): More random, creative output
  • Range: (0, 1], excluding 0
Required range: x <= 1
Example:

0.7

top_p
number

Nucleus Sampling parameter

Note:

  • Controls sampling from tokens with cumulative probability
  • e.g. 0.9 means selecting from tokens reaching 90% cumulative probability
  • Range: (0, 1], excluding 0

Recommendation: Do not adjust temperature and top_p simultaneously

Required range: x <= 1
Example:

0.9

Whether to enable web search

  • true: Enable web search, the model will search the internet for latest information as needed
  • false: Disable web search
Example:

true

search_options
object

Web search options, requires enable_search: true

Response

Chat completion successful

id
string

Unique identifier for the chat completion

Example:

"cmpl-04ea926191a14749b7f2c7a48a68abc6"

model
string

Model name actually used

Example:

"MiniMax-M2.5"

object
enum<string>

Response type

Available options:
chat.completion
Example:

"chat.completion"

created
integer

Creation timestamp

Example:

1698999496

choices
object[]

List of chat completion choices

usage
object

Token usage statistics