Anthropic: Claude 3.5 Sonnet
anthropic/claude-3.5-sonnet
New Claude 3.5 Sonnet delivers better-than-Opus capabilities, faster-than-Sonnet speeds, at the same Sonnet prices. Sonnet is particularly good at: - Coding: Scores ~49% on SWE-Bench Verified, higher than the last best score, and without any fancy prompt scaffolding - Data science: Augments human data science expertise; navigates unstructured data while using multiple tools for insights - Visual processing: excelling at interpreting charts, graphs, and images, accurately transcribing text to derive insights beyond just the text alone - Agentic tasks: exceptional tool use, making it great at agentic tasks (i.e. complex, multi-step problem solving tasks that require engaging with other systems)
ByanthropicInput typeOutput type
Recent activity on Claude 3.5 Sonnet
Tokens processed per day
Thoughput
(tokens/s)
ProvidersMin (tokens/s)Max (tokens/s)Avg (tokens/s)
Anthropic4.1939.9710.64
Vertex AI2.5940.7616.93
Amazon Bedrock23.4635.7329.64
First Token Latency
(ms)
ProvidersMin (ms)Max (ms)Avg (ms)
Anthropic66231991081.69
Vertex AI2032353878714.00
Amazon Bedrock165328982077.50
Providers for Claude 3.5 Sonnet
ZenMux Provider to the best providers that are able to handle your prompt size and parameters, with fallbacks to maximize uptime.
Latency
0.88
s
Throughput
8.31
tps
Uptime
100.00
%
Recent uptime
Oct 10,2025 - 3 PM100.00%
Price
Input
$ 3
/ M tokens
Output
$ 15
/ M tokens
Cache read
$ 0.3
/ M tokens
Cache write 5m
$ 3.75
/ M tokens
Cache write 1h
$ 6
/ M tokens
Cache write
-
Web search
$ 0.01
/ request
Model limitation
Context
200.00K
Max output
8.19K
Supported Parameters
max_completion_tokens
temperature
top_p
frequency_penalty
-
presence_penalty
-
seed
-
logit_bias
-
logprobs
-
top_logprobs
-
response_format
-
stop
tools
tool_choice
parallel_tool_calls
-
Model Protocol Compatibility
openai
anthropic
Data policy
Prompt training
false
Prompt Logging
Zero retention
Moderation
Responsibility of developer
Status Page
status page
Latency
-
Throughput
-
Uptime
100.00
%
Recent uptime
Oct 10,2025 - 3 PM100.00%
Price
Input
$ 3
/ M tokens
Output
$ 15
/ M tokens
Cache read
$ 0.3
/ M tokens
Cache write 5m
$ 3.75
/ M tokens
Cache write 1h
-
Cache write
-
Web search
-
Model limitation
Context
200.00K
Max output
8.19K
Supported Parameters
max_completion_tokens
temperature
top_p
frequency_penalty
-
presence_penalty
-
seed
-
logit_bias
-
logprobs
-
top_logprobs
-
response_format
-
stop
tools
tool_choice
parallel_tool_calls
-
Model Protocol Compatibility
openai
anthropic
Data policy
Prompt training
false
Prompt Logging
Zero retention
Moderation
Responsibility of developer
Status Page
status page
Latency
-
Throughput
-
Uptime
100.00
%
Recent uptime
Oct 10,2025 - 3 PM100.00%
Price
Input
$ 3
/ M tokens
Output
$ 15
/ M tokens
Cache read
$ 0.3
/ M tokens
Cache write 5m
$ 3.75
/ M tokens
Cache write 1h
-
Cache write
-
Web search
-
Model limitation
Context
200.00K
Max output
8.19K
Supported Parameters
max_completion_tokens
temperature
top_p
frequency_penalty
-
presence_penalty
-
seed
-
logit_bias
-
logprobs
-
top_logprobs
-
response_format
-
stop
tools
tool_choice
parallel_tool_calls
-
Model Protocol Compatibility
openai
anthropic
Data policy
Prompt training
false
Prompt Logging
Zero retention
Moderation
Responsibility of developer
Status Page
status page
Sample code and API for Claude 3.5 Sonnet
ZenMux normalizes requests and responses across providers for you.
OpenAI-PythonPythonTypeScriptOpenAI-TypeScriptcURL
python
from openai import OpenAI

client = OpenAI(
  base_url="https://zenmux.ai/api/v1",
  api_key="<ZenMux_API_KEY>",
)

completion = client.chat.completions.create(
  model="anthropic/claude-3.5-sonnet",
  messages=[
    {
      "role": "user",
      "content": [
        {
          "type": "text",
          "text": "What is in this image?"
        }
      ]
    }
  ]
)
print(completion.choices[0].message.content)