DeepSeek V3 API

The DeepSeek V3 API wraps the DeepSeek V3 model and lets you submit structured prompts to receive conversational responses. Ideal for natural dialogue, character-based prompting, or general-purpose language generation.

Base URL: https://api.inferenceapis.com

Endpoints

Chat Completion

POST /

Submit a conversation and get a response from the model.

Examples


import requests

payload = {
    "model": "deepseek-ai/DeepSeek-V3",
    "messages": [
        {"role": "system", "content": "Act like you're a cowboy."},
        {"role": "user", "content": "What did you do today?"}
    ],
}

headers = {
    'Authorization': 'Bearer YOUR_API_KEY',
    'Content-Type': 'application/json'
}
response = requests.post(
    'https://api.inferenceapis.com',
    json=payload,
    headers=headers
)

print(response.json())
        

Parameters

Parameter Type Required Description
model string Yes Must be deepseek-ai/DeepSeek-V3
messages array Yes List of objects with role (system/user/assistant) and content.
max_tokens integer No Max tokens to generate (default is model-defined).
temperature float No Controls randomness. Higher = more creative.
top_p float No Nucleus sampling probability cutoff.
min_p float No Minimum cumulative probability during sampling.
top_k integer No Limits token sampling to top-k tokens.
repetition_penalty float No Discourages repeating tokens.
presence_penalty float No Discourages previously mentioned topics.
frequency_penalty float No Penalizes frequent tokens.
seed integer No Random seed for deterministic outputs.