You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
workspace_id: 0# In the future, there will be the concept of public/private workspaces. Public models are managed by the public workspace, private models by the private workspace. Currently, all models belong to the public workspace, and the public workspace id is temporarily set to 0.
5
-
name: "deepseek-r1-distill-qwen-32b-250120"
3
+
name: "doubao"
6
4
frame: "eino"
7
5
protocol: "ark"
8
6
protocol_config:
9
-
api_key: "a715a14b-3b93-47da-8bc9-844c12fecff7"
10
-
model: "ep-20250304143659-5bcjt"
7
+
api_key: "***"
8
+
model: "***"
11
9
param_config:
12
10
param_schemas:
13
11
- name: "temperature"
14
12
label: "temperature"
15
-
desc: "Increasing temperature will make the model output more diverse and creative. Conversely, lowering the temperature will make the output more compliant with instructions but reduce diversity. It is recommended not to adjust together with 'Top p'."
13
+
desc: "Increasing temperature makes model output more diverse and creative, while decreasing it makes output more focused on instructions but less diverse. It's recommended not to adjust this simultaneously with 'Top p'."
16
14
type: "float"
17
15
min: "0"
18
16
max: "1.0"
19
17
default_val: "0.7"
20
18
- name: "max_tokens"
21
19
label: "max_tokens"
22
-
desc: "Controls the maximum length of model output tokens. Typically, 100 tokens are about 150 Chinese characters."
20
+
desc: "Controls the maximum number of tokens in model output. Typically, 100 tokens equals about 150 Chinese characters."
23
21
type: "int"
24
22
min: "1"
25
-
max: "8192"
23
+
max: "4096"
26
24
default_val: "2048"
27
-
# multimodal model
28
-
- id: 2
29
-
workspace_id: 0# In the future, there will be the concept of public/private workspaces. Public models are managed by the public workspace, private models by the private workspace. Currently, all models belong to the public workspace, and the public workspace id is temporarily set to 0.
desc: "Increasing temperature will make the model output more diverse and creative. Conversely, lowering the temperature will make the output more compliant with instructions but reduce diversity. It is recommended not to adjust together with 'Top p'."
72
-
type: "float"
73
-
min: "0"
74
-
max: "1.0"
75
-
default_val: "0.7"
76
-
- name: "max_tokens"
77
-
label: "max_tokens"
78
-
desc: "Controls the maximum length of model output tokens. Typically, 100 tokens are about 150 Chinese characters."
79
-
type: "int"
80
-
min: "1"
81
-
max: "8192"
82
-
default_val: "2048"
83
-
# fc model
84
-
- id: 3
85
-
workspace_id: 0# In the future, there will be the concept of public/private workspaces. Public models are managed by the public workspace, private models by the private workspace. Currently, all models belong to the public workspace, and the public workspace id is temporarily set to 0.
desc: "Increasing temperature will make the model output more diverse and creative. Conversely, lowering the temperature will make the output more compliant with instructions but reduce diversity. It is recommended not to adjust together with 'Top p'."
121
-
type: "float"
122
-
min: "0"
123
-
max: "1.0"
124
-
default_val: "0.1"
125
25
- name: "top_p"
126
26
label: "top_p"
127
-
desc: "The model will consider token results within top_p probability mass."
27
+
desc: "Selects the minimum token set with cumulative probability reaching top_p during generation, excluding tokens outside the set, balancing diversity and reasonableness."
128
28
type: "float"
129
-
min: "0"
29
+
min: "0.001"
130
30
max: "1.0"
131
-
default_val: "0.1"
132
-
- name: "max_tokens"
133
-
label: "max_tokens"
134
-
desc: "Controls the maximum length of model output tokens. Typically, 100 tokens are about 150 Chinese characters."
135
-
type: "int"
136
-
min: "1"
137
-
max: "8192"
138
-
default_val: "2048"
139
-
- name: "top_k"
140
-
label: "top_k"# Displayed as a name on the front end
141
-
desc: "Only sample from the top k tokens with the highest probability to limit the candidate range and improve generation stability."# Displayed as a description on the front end
142
-
type: "int"# Required. Must be float, int, bool, string
143
-
min: "1"
144
-
max: "100"
145
-
default_val: "50"
146
-
- name: "frequency_penalty"
147
-
label: "frequency_penalty"# Displayed as a name on the front end
148
-
desc: "Penalizes generated tokens, with higher frequency resulting in higher penalties, suppressing repetitive content."# Displayed as a description on the front end
149
-
type: "float"# Required. Must be float, int, bool, string
150
-
min: "0"
151
-
max: "2.0"
152
-
default_val: "0"
153
-
- name: "presence_penalty"
154
-
label: "presence_penalty"# Displayed as a name on the front end
155
-
desc: "Penalizes all tokens that have appeared, preventing the same content from appearing repeatedly, increasing content diversity."# Displayed as a description on the front end
156
-
type: "float"# Required. Must be float, int, bool, string
0 commit comments