You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: apps/site/docs/en/model-common-config.mdx
+14-27Lines changed: 14 additions & 27 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -74,37 +74,21 @@ MIDSCENE_MODEL_FAMILY="doubao-seed" # "doubao-vision" is also supported
74
74
# Optional: control reasoning effort (low, medium, high)
75
75
# MIDSCENE_MODEL_REASONING_EFFORT="medium"
76
76
```
77
-
### Qwen3.6 {#qwen36}
77
+
### Qwen3.X Series {#qwen3x}
78
78
79
-
Using Alibaba Cloud's `qwen3.6-plus` as an example. It is recommended to disable the platform's default thinking mode to improve execution speed, the environment variable configuration is as follows:
79
+
Qwen3.5 and Qwen3.6 are currently supported in the Qwen3.X series.
To enable thinking mode, remove the `MIDSCENE_MODEL_REASONING_ENABLED="false"` line and add `MIDSCENE_MODEL_REASONING_BUDGET="500"` to control thinking cost.
90
-
91
-
You can also use Qwen3.6 from [OpenRouter](https://openrouter.ai/qwen).
92
-
93
-
### Qwen3.5 {#qwen35}
94
-
95
-
Using Alibaba Cloud's `qwen3.5-plus` as an example. It is recommended to disable the platform's default thinking mode to improve execution speed, the environment variable configuration is as follows:
81
+
Using Alibaba Cloud's `qwen3.5-plus` as an example. It is recommended to disable the platform's default thinking mode to improve execution speed:
MIDSCENE_MODEL_NAME="qwen3.5-plus"# For Qwen3.6, use "qwen3.6-plus"
87
+
MIDSCENE_MODEL_FAMILY="qwen3.5"# For Qwen3.6, use "qwen3.6"
102
88
MIDSCENE_MODEL_REASONING_ENABLED="false"
103
89
```
104
90
105
-
To enable thinking mode, remove the `MIDSCENE_MODEL_REASONING_ENABLED="false"` line and add `MIDSCENE_MODEL_REASONING_BUDGET="500"` to control thinking cost.
106
-
107
-
You can also use Qwen3.5 from [OpenRouter](https://openrouter.ai/qwen).
91
+
To enable thinking mode, set `MIDSCENE_MODEL_REASONING_ENABLED="true"` and add `MIDSCENE_MODEL_REASONING_BUDGET="500"` to control thinking cost.
You can also use Qwen3-VL from [OpenRouter](https://openrouter.ai/qwen).
121
105
122
-
### Zhipu GLM-V {#glm-v}
106
+
### Zhipu GLM-V Series {#glm-v}
123
107
124
-
Zhipu GLM-V is an open-source vision model from Zhipu AI. Using `GLM-4.6V` as an example:
108
+
Zhipu GLM-V is a vision understanding model from Zhipu AI. The latest versions include `GLM-4.6V` (open-source) and `GLM-5V-Turbo`.
109
+
110
+
It is recommended to disable the platform's default thinking mode to improve execution speed. Using `GLM-4.6V` as an example:
125
111
126
112
Obtain an API key from [Z.AI (Global)](https://z.ai/manage-apikey/apikey-list) or [BigModel (CN)](https://bigmodel.cn/usercenter/proj-mgmt/apikeys), and set:
127
113
128
114
```bash
129
-
MIDSCENE_MODEL_BASE_URL="https://api.z.ai/api/paas/v4"# Or https://open.bigmodel.cn/api/paas/v4
115
+
MIDSCENE_MODEL_BASE_URL="https://open.bigmodel.cn/api/paas/v4"# Or https://api.z.ai/api/paas/v4
130
116
MIDSCENE_MODEL_API_KEY="......"
131
-
MIDSCENE_MODEL_NAME="glm-4.6v"
117
+
MIDSCENE_MODEL_NAME="glm-4.6v"# For GLM-5V-Turbo, use "glm-5v-turbo"
132
118
MIDSCENE_MODEL_FAMILY="glm-v"
119
+
MIDSCENE_MODEL_REASONING_ENABLED="false"
133
120
```
134
121
135
-
**Learn more about Zhipu GLM-V**
122
+
**Learn more about the open-source Zhipu GLM-V model**
@@ -53,7 +53,7 @@ If you are unsure where to start, pick whichever model is easiest to access toda
53
53
| Model family | Deployment | Midscene notes |
54
54
| --- | --- | --- |
55
55
| Doubao Seed Model<br />[Quick setup](./model-common-config#doubao-seed-model)| Volcano Engine:<br />[Doubao-Seed-1.6-Vision](https://www.volcengine.com/docs/82379/1799865)<br />[Doubao-Seed-2.0-Lite](https://www.volcengine.com/docs/82379/1799865)| ⭐⭐⭐⭐<br />Strong at UI planning and targeting<br />Slightly slower |
56
-
| Qwen3.5<br />[Quick setup](./model-common-config#qwen35)|[Alibaba Cloud](https://help.aliyun.com/zh/model-studio/vision)<br/>[OpenRouter](https://openrouter.ai/qwen)| ⭐⭐⭐⭐<br />Stronger than Qwen3-VL and Qwen2.5-VL |
56
+
| Qwen3.5<br />[Quick setup](./model-common-config#qwen3x)|[Alibaba Cloud](https://help.aliyun.com/zh/model-studio/vision)<br/>[OpenRouter](https://openrouter.ai/qwen)| ⭐⭐⭐⭐<br />Stronger than Qwen3-VL and Qwen2.5-VL |
57
57
| Zhipu GLM-4.6V<br />[Quick setup](./model-common-config#glm-v)|[Z.AI (Global)](https://docs.z.ai/guides/vlm/glm-4.6v)<br/>[BigModel (CN)](https://docs.bigmodel.cn/cn/guide/models/vlm/glm-4.6v)| Newly integrated, welcome to try it out<br />Weights open-sourced on [HuggingFace](https://huggingface.co/zai-org/GLM-4.6V)|
58
58
| Gemini-3-Pro / Gemini-3-Flash<br />[Quick setup](./model-common-config#gemini-3-pro)|[Google Cloud](https://ai.google.dev/gemini-api/docs/models/gemini)| ⭐⭐⭐<br />Gemini-3-Flash is supported<br />Price is higher than Doubao and Qwen |
59
59
| UI-TARS<br />[Quick setup](./model-common-config#ui-tars)|[Volcano Engine](https://www.volcengine.com/docs/82379/1536429)| ⭐⭐<br />Strong exploratory ability but results vary by scenario<br />Open-source versions available ([HuggingFace](https://huggingface.co/bytedance-research/UI-TARS-72B-SFT) / [GitHub](https://github.com/bytedance/ui-tars)) |
0 commit comments