large language
 DeepSeek V3 0324
DeepSeek V3 0324
 DeepSeek V3 0324
DeepSeek V3 0324A state-of-the-art 671B-parameter MoE LLM licensed for commercial use
Model details
View repositoryExample usage
DeepSeek V3 0324 runs using the Baseten Inference Stack and is accessible via an OpenAI-compatible API endpoint.
Input
1# You can use this model with any of the OpenAI clients in any language!
2# Simply change the API Key to get started
3
4from openai import OpenAI
5
6client = OpenAI(
7    api_key="YOUR_API_KEY",
8    base_url="https://inference.baseten.co/v1"
9)
10
11response = client.chat.completions.create(
12    model="deepseek-ai/DeepSeek-V3-0324",
13    messages=[
14        {
15            "role": "user",
16            "content": "Implement Hello World in Python"
17        }
18    ],
19    stop=[],
20    stream=True,
21    stream_options={
22        "include_usage": True,
23        "continuous_usage_stats": True
24    },
25    top_p=1,
26    max_tokens=1000,
27    temperature=1,
28    presence_penalty=0,
29    frequency_penalty=0
30)
31
32for chunk in response:
33    if chunk.choices and chunk.choices[0].delta.content is not None:
34        print(chunk.choices[0].delta.content, end="", flush=True)JSON output
1{
2    "id": "8456fe51db3548789f199cfb8c8efd35",
3    "object": "text_completion",
4    "created": 1735236968,
5    "model": "deepseek-ai/DeepSeek-V3-0324",
6    "choices": [
7        {
8            "index": 0,
9            "text": "FizzBuzz is a classic programming problem where you print numbers from 1 to 100...",
10            "logprobs": null,
11            "finish_reason": "stop",
12            "matched_stop": 1
13        }
14    ],
15    "usage": {
16        "prompt_tokens": 14,
17        "total_tokens": 240,
18        "completion_tokens": 226,
19        "prompt_tokens_details": null
20    }
21}




