Passer au contenu principal
POST
/
v1
/
chat
/
completions
curl --request POST \ --url https://direct.evolink.ai/v1/chat/completions \ --header 'Authorization: Bearer <token>' \ --header 'Content-Type: application/json' \ --data ' { "model": "gemini-2.5-flash-lite", "messages": [ { "role": "user", "content": "Please introduce yourself" } ] } '
{
  "id": "chatcmpl-20251010015944503180122WJNB8Eid",
  "model": "gemini-2.5-flash-lite",
  "object": "chat.completion",
  "created": 1760032810,
  "choices": [
    {
      "index": 0,
      "message": {
        "role": "assistant",
        "content": "Hello! I'm pleased to introduce myself.\n\nI'm a Large Language Model, trained and developed by Google.\n\nSimply put, you can think of me as a \"smart brain\" that has been trained on massive amounts of text data and is able to understand and generate human language. My core capability is processing and generating text. Specifically, I can do the following:\n\n**1. Information Query & Knowledge Answering**\nI can act like a \"talking encyclopedia,\" answering various questions, whether they're about scientific knowledge, historical events, or everyday facts.\n\n**2. Creative Writing & Text Generation**\nI can create various types of text based on your requirements, such as:\n*   **Writing**: Poetry, stories, scripts, emails, speeches, advertising copy, etc.\n*   **Planning**: Travel plans, study outlines, event proposals, etc.\n*   **Brainstorming**: Working with you to generate new ideas and spark creativity.\n\n**3. Translation & Language Processing**\nI'm proficient in multiple languages and can provide fast, fluent translation services. I can also help you polish, proofread, summarize, or rewrite text to make your expression clearer and more professional.\n\n**4. Programming & Code Assistance**\nI can write code snippets, explain code logic, debug errors, or \"translate\" code from one programming language to another, making me a helpful companion for programmers.\n\n**5. Logical Analysis & Reasoning**\nI can help you analyze complex problems, organize logical chains, and make inferences and summaries based on the information you provide.\n\n---\n\n**In summary**, my goal is to be a powerful and useful tool that helps you obtain information more efficiently, complete tasks, and spark creativity through natural language communication.\n\n**Remember:** I'm an artificial intelligence, my knowledge comes from the data I've learned, and it may not be the most up-to-date. Sometimes I may also make mistakes, so for very important information, I recommend you verify it again.",
        "tool_calls": [
          {
            "id": "<string>",
            "type": "function",
            "function": {
              "name": "<string>",
              "arguments": "<string>"
            }
          }
        ]
      },
      "logprobs": {
        "content": [
          {
            "token": "<string>",
            "logprob": 123,
            "bytes": [
              123
            ],
            "top_logprobs": [
              {
                "token": "<string>",
                "logprob": 123,
                "bytes": [
                  123
                ]
              }
            ]
          }
        ]
      },
      "finish_reason": "stop"
    }
  ],
  "usage": {
    "prompt_tokens": 13,
    "completion_tokens": 1891,
    "total_tokens": 1904,
    "prompt_tokens_details": {
      "cached_tokens": 0,
      "text_tokens": 13,
      "audio_tokens": 0,
      "image_tokens": 0
    },
    "completion_tokens_details": {
      "text_tokens": 0,
      "audio_tokens": 0,
      "reasoning_tokens": 1480
    },
    "input_tokens": 0,
    "output_tokens": 0,
    "input_tokens_details": null
  }
}

Documentation Index

Fetch the complete documentation index at: https://docs.evolink.ai/llms.txt

Use this file to discover all available pages before exploring further.

BaseURL : La BaseURL par défaut est https://direct.evolink.ai, qui offre une meilleure prise en charge des modèles de texte et des connexions persistantes. https://api.evolink.ai est le point d’accès principal pour les services multimodaux et sert d’adresse de secours pour les modèles de texte.

Autorisations

Authorization
string
header
requis

##Toutes les API nécessitent une authentification Bearer Token##

Obtenir une clé API :

Visitez la Page de gestion des clés API pour obtenir votre clé API

Ajouter à l'en-tête de requête :

Authorization: Bearer YOUR_API_KEY

Corps

application/json
model
enum<string>
défaut:gemini-2.5-flash-lite
requis

Nom du modèle de chat

Options disponibles:
gemini-2.5-flash-lite
Exemple:

"gemini-2.5-flash-lite"

messages
object[]
requis

Liste des messages de chat, prend en charge le dialogue multi-tours et l'entrée multimodale

Minimum array length: 1
stream
boolean
défaut:false

Indique si la réponse doit être retournée en mode streaming

  • true : Retour en streaming, reçoit le contenu en fragments en temps réel
  • false : Retourne la réponse complète en une seule fois
Exemple:

false

max_completion_tokens
integer | null

Maximum number of completion tokens for the generated response, corresponding to Gemini's maxOutputTokens.

Plage requise: 1 <= x <= 65536
Exemple:

2000

max_tokens
integer

Maximum number of tokens for the generated response, compatible with the legacy OpenAI parameter.

Plage requise: 1 <= x <= 65536
Exemple:

2000

temperature
number
défaut:1

Température d'échantillonnage, contrôle le caractère aléatoire de la sortie

Description :

  • Valeurs basses (ex. 0.2) : Sortie plus déterministe et ciblée
  • Valeurs élevées (ex. 1.5) : Sortie plus aléatoire et créative
Plage requise: 0 <= x <= 2
Exemple:

0.7

top_p
number
défaut:1

Paramètre d'échantillonnage Nucleus

Description :

  • Contrôle l'échantillonnage à partir des tokens avec une probabilité cumulative
  • Par exemple, 0.9 signifie sélectionner parmi les tokens dont la probabilité cumulative atteint 90 %
  • Par défaut : 1.0 (considère tous les tokens)

Recommandation : Ne pas ajuster temperature et top_p simultanément

Plage requise: 0 <= x <= 1
Exemple:

0.9

frequency_penalty
number | null
défaut:0

Frequency penalty coefficient. Range: -2.0 to 2.0. Corresponds to Gemini's frequencyPenalty.

Plage requise: -2 <= x <= 2
Exemple:

0

presence_penalty
number | null
défaut:0

Presence penalty coefficient. Range: -2.0 to 2.0. Corresponds to Gemini's presencePenalty.

Plage requise: -2 <= x <= 2
Exemple:

0

stop

Stop sequences. Supports a string or string array, corresponding to Gemini's stopSequences.

n
integer | null
défaut:1

Number of generated candidates.

Plage requise: x >= 1
Exemple:

1

reasoning_effort
enum<string> | null
défaut:medium

Limite l'effort de raisonnement. Gemini 2.5 Flash et Flash Lite prennent en charge none pour désactiver le raisonnement ; low/medium/high correspondent à différents budgets.

Options disponibles:
none,
low,
medium,
high
Exemple:

"medium"

seed
integer | null

Random seed used to make output as reproducible as possible, corresponding to Gemini's seed.

Exemple:

12345

logprobs
boolean | null
défaut:false

Whether to return token logprob information, corresponding to Gemini's responseLogprobs.

Exemple:

true

top_logprobs
integer | null

Number of top logprob values returned for each token, corresponding to Gemini's logprobs.

Plage requise: 0 <= x <= 20
Exemple:

5

response_format
object

Response format settings, supporting JSON mode and JSON Schema, corresponding to Gemini's responseMimeType, responseSchema and responseJsonSchema.

stream_options
object

Streaming response options. Can be set when stream is true.

tools
object[] | null

List of tool definitions for Function Calling.

tool_choice

Controls tool-calling behavior.

Options disponibles:
none,
auto,
required
extra_body
object

Gemini extension parameters.

Réponse

Complétion de chat générée avec succès

id
string

Identifiant unique pour la complétion de chat

Exemple:

"chatcmpl-20251010015944503180122WJNB8Eid"

model
string

Nom du modèle réellement utilisé

Exemple:

"gemini-2.5-flash-lite"

object
enum<string>

Type de réponse

Options disponibles:
chat.completion
Exemple:

"chat.completion"

created
integer

Horodatage de création

Exemple:

1760032810

choices
object[]

Liste des choix de complétion de chat

usage
object

Statistiques d'utilisation des jetons