Skip to content

Commit 391ad77

Browse files
mazyu36iamjoel
authored andcommitted
feat: support Claude 3.5 Haiku on Amazon Bedrock (#10265)
1 parent c5422af commit 391ad77

File tree

2 files changed

+122
-0
lines changed

2 files changed

+122
-0
lines changed
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,61 @@
1+
model: anthropic.claude-3-5-haiku-20241022-v1:0
2+
label:
3+
en_US: Claude 3.5 Haiku
4+
model_type: llm
5+
features:
6+
- agent-thought
7+
- vision
8+
- tool-call
9+
- stream-tool-call
10+
model_properties:
11+
mode: chat
12+
context_size: 200000
13+
# docs: https://docs.aws.amazon.com/bedrock/latest/userguide/model-parameters-anthropic-claude-messages.html
14+
parameter_rules:
15+
- name: max_tokens
16+
use_template: max_tokens
17+
required: true
18+
type: int
19+
default: 4096
20+
min: 1
21+
max: 4096
22+
help:
23+
zh_Hans: 停止前生成的最大令牌数。请注意,Anthropic Claude 模型可能会在达到 max_tokens 的值之前停止生成令牌。不同的 Anthropic Claude 模型对此参数具有不同的最大值。
24+
en_US: The maximum number of tokens to generate before stopping. Note that Anthropic Claude models might stop generating tokens before reaching the value of max_tokens. Different Anthropic Claude models have different maximum values for this parameter.
25+
# docs: https://docs.anthropic.com/claude/docs/system-prompts
26+
- name: temperature
27+
use_template: temperature
28+
required: false
29+
type: float
30+
default: 1
31+
min: 0.0
32+
max: 1.0
33+
help:
34+
zh_Hans: 生成内容的随机性。
35+
en_US: The amount of randomness injected into the response.
36+
- name: top_p
37+
required: false
38+
type: float
39+
default: 0.999
40+
min: 0.000
41+
max: 1.000
42+
help:
43+
zh_Hans: 在核采样中,Anthropic Claude 按概率递减顺序计算每个后续标记的所有选项的累积分布,并在达到 top_p 指定的特定概率时将其切断。您应该更改温度或top_p,但不能同时更改两者。
44+
en_US: In nucleus sampling, Anthropic Claude computes the cumulative distribution over all the options for each subsequent token in decreasing probability order and cuts it off once it reaches a particular probability specified by top_p. You should alter either temperature or top_p, but not both.
45+
- name: top_k
46+
required: false
47+
type: int
48+
default: 0
49+
min: 0
50+
# tip docs from aws has error, max value is 500
51+
max: 500
52+
help:
53+
zh_Hans: 对于每个后续标记,仅从前 K 个选项中进行采样。使用 top_k 删除长尾低概率响应。
54+
en_US: Only sample from the top K options for each subsequent token. Use top_k to remove long tail low probability responses.
55+
- name: response_format
56+
use_template: response_format
57+
pricing:
58+
input: '0.001'
59+
output: '0.005'
60+
unit: '0.001'
61+
currency: USD
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,61 @@
1+
model: us.anthropic.claude-3-5-haiku-20241022-v1:0
2+
label:
3+
en_US: Claude 3.5 Haiku(US.Cross Region Inference)
4+
model_type: llm
5+
features:
6+
- agent-thought
7+
- vision
8+
- tool-call
9+
- stream-tool-call
10+
model_properties:
11+
mode: chat
12+
context_size: 200000
13+
# docs: https://docs.aws.amazon.com/bedrock/latest/userguide/model-parameters-anthropic-claude-messages.html
14+
parameter_rules:
15+
- name: max_tokens
16+
use_template: max_tokens
17+
required: true
18+
type: int
19+
default: 4096
20+
min: 1
21+
max: 4096
22+
help:
23+
zh_Hans: 停止前生成的最大令牌数。请注意,Anthropic Claude 模型可能会在达到 max_tokens 的值之前停止生成令牌。不同的 Anthropic Claude 模型对此参数具有不同的最大值。
24+
en_US: The maximum number of tokens to generate before stopping. Note that Anthropic Claude models might stop generating tokens before reaching the value of max_tokens. Different Anthropic Claude models have different maximum values for this parameter.
25+
# docs: https://docs.anthropic.com/claude/docs/system-prompts
26+
- name: temperature
27+
use_template: temperature
28+
required: false
29+
type: float
30+
default: 1
31+
min: 0.0
32+
max: 1.0
33+
help:
34+
zh_Hans: 生成内容的随机性。
35+
en_US: The amount of randomness injected into the response.
36+
- name: top_p
37+
required: false
38+
type: float
39+
default: 0.999
40+
min: 0.000
41+
max: 1.000
42+
help:
43+
zh_Hans: 在核采样中,Anthropic Claude 按概率递减顺序计算每个后续标记的所有选项的累积分布,并在达到 top_p 指定的特定概率时将其切断。您应该更改温度或top_p,但不能同时更改两者。
44+
en_US: In nucleus sampling, Anthropic Claude computes the cumulative distribution over all the options for each subsequent token in decreasing probability order and cuts it off once it reaches a particular probability specified by top_p. You should alter either temperature or top_p, but not both.
45+
- name: top_k
46+
required: false
47+
type: int
48+
default: 0
49+
min: 0
50+
# tip docs from aws has error, max value is 500
51+
max: 500
52+
help:
53+
zh_Hans: 对于每个后续标记,仅从前 K 个选项中进行采样。使用 top_k 删除长尾低概率响应。
54+
en_US: Only sample from the top K options for each subsequent token. Use top_k to remove long tail low probability responses.
55+
- name: response_format
56+
use_template: response_format
57+
pricing:
58+
input: '0.001'
59+
output: '0.005'
60+
unit: '0.001'
61+
currency: USD

0 commit comments

Comments
 (0)