Anthropic: Claude 3.5 Haiku
anthropic/claude-3.5-haiku
Claude 3.5 Haiku features offers enhanced capabilities in speed, coding accuracy, and tool use. Engineered to excel in real-time applications, it delivers quick response times that are essential for dynamic tasks such as chat interactions and immediate coding suggestions. This makes it highly suitable for environments that demand both speed and precision, such as software development, customer service bots, and data management systems.
ByanthropicInput typeOutput type
Recent activity on Claude 3.5 Haiku
Tokens processed per day
Thoughput
(tokens/s)
ProvidersMin (tokens/s)Max (tokens/s)Avg (tokens/s)
Anthropic9.3545.3212.55
Amazon Bedrock3140.1336.23
Vertex AI5.1140.7625.68
First Token Latency
(ms)
ProvidersMin (ms)Max (ms)Avg (ms)
Anthropic672780724.24
Amazon Bedrock141934442056.50
Vertex AI1766312226114.00
Providers for Claude 3.5 Haiku
ZenMux Provider to the best providers that are able to handle your prompt size and parameters, with fallbacks to maximize uptime.
Latency
0.76
s
Throughput
10.62
tps
Uptime
100.00
%
Recent uptime
Oct 10,2025 - 3 PM100.00%
Price
Input
$ 0.8
/ M tokens
Output
$ 4
/ M tokens
Cache read
$ 0.08
/ M tokens
Cache write 5m
$ 1
/ M tokens
Cache write 1h
$ 1.6
/ M tokens
Cache write
-
Web search
$ 0.01
/ request
Model limitation
Context
200.00K
Max output
8.19K
Supported Parameters
max_completion_tokens
temperature
top_p
frequency_penalty
-
presence_penalty
-
seed
-
logit_bias
-
logprobs
-
top_logprobs
-
response_format
-
stop
tools
tool_choice
parallel_tool_calls
-
Model Protocol Compatibility
openai
anthropic
Data policy
Prompt training
false
Prompt Logging
Zero retention
Moderation
Responsibility of developer
Status Page
status page
Latency
-
Throughput
-
Uptime
100.00
%
Recent uptime
Oct 10,2025 - 3 PM100.00%
Price
Input
$ 0.8
/ M tokens
Output
$ 4
/ M tokens
Cache read
$ 0.08
/ M tokens
Cache write 5m
$ 1
/ M tokens
Cache write 1h
-
Cache write
-
Web search
-
Model limitation
Context
200.00K
Max output
8.19K
Supported Parameters
max_completion_tokens
temperature
top_p
frequency_penalty
-
presence_penalty
-
seed
-
logit_bias
-
logprobs
-
top_logprobs
-
response_format
-
stop
tools
tool_choice
parallel_tool_calls
-
Model Protocol Compatibility
openai
anthropic
Data policy
Prompt training
false
Prompt Logging
Zero retention
Moderation
Responsibility of developer
Status Page
status page
Latency
-
Throughput
-
Uptime
100.00
%
Recent uptime
Oct 10,2025 - 3 PM100.00%
Price
Input
$ 0.8
/ M tokens
Output
$ 4
/ M tokens
Cache read
$ 0.08
/ M tokens
Cache write 5m
$ 1
/ M tokens
Cache write 1h
-
Cache write
-
Web search
-
Model limitation
Context
200.00K
Max output
8.19K
Supported Parameters
max_completion_tokens
temperature
top_p
frequency_penalty
-
presence_penalty
-
seed
-
logit_bias
-
logprobs
-
top_logprobs
-
response_format
-
stop
tools
tool_choice
parallel_tool_calls
-
Model Protocol Compatibility
openai
anthropic
Data policy
Prompt training
false
Prompt Logging
Zero retention
Moderation
Responsibility of developer
Status Page
status page
Sample code and API for Claude 3.5 Haiku
ZenMux normalizes requests and responses across providers for you.
OpenAI-PythonPythonTypeScriptOpenAI-TypeScriptcURL
python
from openai import OpenAI

client = OpenAI(
  base_url="https://zenmux.ai/api/v1",
  api_key="<ZenMux_API_KEY>",
)

completion = client.chat.completions.create(
  model="anthropic/claude-3.5-haiku",
  messages=[
    {
      "role": "user",
      "content": [
        {
          "type": "text",
          "text": "What is in this image?"
        }
      ]
    }
  ]
)
print(completion.choices[0].message.content)