Skip to content

Commit 81553d9

Browse files
committed
feat(docs): supports zhipu glm-5v-turbo
1 parent d2e4653 commit 81553d9

File tree

4 files changed

+29
-51
lines changed

4 files changed

+29
-51
lines changed

apps/site/docs/en/model-common-config.mdx

Lines changed: 14 additions & 27 deletions
Original file line numberDiff line numberDiff line change
@@ -74,37 +74,21 @@ MIDSCENE_MODEL_FAMILY="doubao-seed" # "doubao-vision" is also supported
7474
# Optional: control reasoning effort (low, medium, high)
7575
# MIDSCENE_MODEL_REASONING_EFFORT="medium"
7676
```
77-
### Qwen3.6 {#qwen36}
77+
### Qwen3.X Series {#qwen3x}
7878

79-
Using Alibaba Cloud's `qwen3.6-plus` as an example. It is recommended to disable the platform's default thinking mode to improve execution speed, the environment variable configuration is as follows:
79+
Qwen3.5 and Qwen3.6 are currently supported in the Qwen3.X series.
8080

81-
```bash
82-
MIDSCENE_MODEL_BASE_URL="https://dashscope.aliyuncs.com/compatible-mode/v1"
83-
MIDSCENE_MODEL_API_KEY="......"
84-
MIDSCENE_MODEL_NAME="qwen3.6-plus"
85-
MIDSCENE_MODEL_FAMILY="qwen3.6"
86-
MIDSCENE_MODEL_REASONING_ENABLED="false"
87-
```
88-
89-
To enable thinking mode, remove the `MIDSCENE_MODEL_REASONING_ENABLED="false"` line and add `MIDSCENE_MODEL_REASONING_BUDGET="500"` to control thinking cost.
90-
91-
You can also use Qwen3.6 from [OpenRouter](https://openrouter.ai/qwen).
92-
93-
### Qwen3.5 {#qwen35}
94-
95-
Using Alibaba Cloud's `qwen3.5-plus` as an example. It is recommended to disable the platform's default thinking mode to improve execution speed, the environment variable configuration is as follows:
81+
Using Alibaba Cloud's `qwen3.5-plus` as an example. It is recommended to disable the platform's default thinking mode to improve execution speed:
9682

9783
```bash
9884
MIDSCENE_MODEL_BASE_URL="https://dashscope.aliyuncs.com/compatible-mode/v1"
9985
MIDSCENE_MODEL_API_KEY="......"
100-
MIDSCENE_MODEL_NAME="qwen3.5-plus"
101-
MIDSCENE_MODEL_FAMILY="qwen3.5"
86+
MIDSCENE_MODEL_NAME="qwen3.5-plus" # For Qwen3.6, use "qwen3.6-plus"
87+
MIDSCENE_MODEL_FAMILY="qwen3.5" # For Qwen3.6, use "qwen3.6"
10288
MIDSCENE_MODEL_REASONING_ENABLED="false"
10389
```
10490

105-
To enable thinking mode, remove the `MIDSCENE_MODEL_REASONING_ENABLED="false"` line and add `MIDSCENE_MODEL_REASONING_BUDGET="500"` to control thinking cost.
106-
107-
You can also use Qwen3.5 from [OpenRouter](https://openrouter.ai/qwen).
91+
To enable thinking mode, set `MIDSCENE_MODEL_REASONING_ENABLED="true"` and add `MIDSCENE_MODEL_REASONING_BUDGET="500"` to control thinking cost.
10892

10993
### Qwen3-VL {#qwen3-vl}
11094

@@ -119,20 +103,23 @@ MIDSCENE_MODEL_FAMILY="qwen3-vl"
119103

120104
You can also use Qwen3-VL from [OpenRouter](https://openrouter.ai/qwen).
121105

122-
### Zhipu GLM-V {#glm-v}
106+
### Zhipu GLM-V Series {#glm-v}
123107

124-
Zhipu GLM-V is an open-source vision model from Zhipu AI. Using `GLM-4.6V` as an example:
108+
Zhipu GLM-V is a vision understanding model from Zhipu AI. The latest versions include `GLM-4.6V` (open-source) and `GLM-5V-Turbo`.
109+
110+
It is recommended to disable the platform's default thinking mode to improve execution speed. Using `GLM-4.6V` as an example:
125111

126112
Obtain an API key from [Z.AI (Global)](https://z.ai/manage-apikey/apikey-list) or [BigModel (CN)](https://bigmodel.cn/usercenter/proj-mgmt/apikeys), and set:
127113

128114
```bash
129-
MIDSCENE_MODEL_BASE_URL="https://api.z.ai/api/paas/v4" # Or https://open.bigmodel.cn/api/paas/v4
115+
MIDSCENE_MODEL_BASE_URL="https://open.bigmodel.cn/api/paas/v4" # Or https://api.z.ai/api/paas/v4
130116
MIDSCENE_MODEL_API_KEY="......"
131-
MIDSCENE_MODEL_NAME="glm-4.6v"
117+
MIDSCENE_MODEL_NAME="glm-4.6v" # For GLM-5V-Turbo, use "glm-5v-turbo"
132118
MIDSCENE_MODEL_FAMILY="glm-v"
119+
MIDSCENE_MODEL_REASONING_ENABLED="false"
133120
```
134121

135-
**Learn more about Zhipu GLM-V**
122+
**Learn more about the open-source Zhipu GLM-V model**
136123

137124
- Github: [https://github.com/zai-org/GLM-V](https://github.com/zai-org/GLM-V)
138125
- Hugging Face: [https://huggingface.co/zai-org/GLM-4.6V](https://huggingface.co/zai-org/GLM-4.6V)

apps/site/docs/en/model-strategy.mdx

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,7 @@ import TroubleshootingLLMConnectivity from './common/troubleshooting-llm-connect
66

77
If you want to try Midscene right away, pick a model and follow its configuration guide:
88
* [Doubao Seed Model](./model-common-config#doubao-seed-model)
9-
* Qwen Models: [Qwen3.6](./model-common-config#qwen36), [Qwen3.5](./model-common-config#qwen35), [Qwen3-VL](./model-common-config#qwen3-vl)
9+
* Qwen Models: [Qwen3.X Series](./model-common-config#qwen3x), [Qwen3-VL](./model-common-config#qwen3-vl)
1010
* [Zhipu GLM-V](./model-common-config#glm-v)
1111
* [Zhipu AutoGLM](./model-common-config#auto-glm)
1212
* [Gemini-3-Pro / Gemini-3-Flash](./model-common-config#gemini-3-pro)
@@ -53,7 +53,7 @@ If you are unsure where to start, pick whichever model is easiest to access toda
5353
| Model family | Deployment | Midscene notes |
5454
| --- | --- | --- |
5555
| Doubao Seed Model<br />[Quick setup](./model-common-config#doubao-seed-model) | Volcano Engine:<br />[Doubao-Seed-1.6-Vision](https://www.volcengine.com/docs/82379/1799865)<br />[Doubao-Seed-2.0-Lite](https://www.volcengine.com/docs/82379/1799865) | ⭐⭐⭐⭐<br />Strong at UI planning and targeting<br />Slightly slower |
56-
| Qwen3.5<br />[Quick setup](./model-common-config#qwen35) | [Alibaba Cloud](https://help.aliyun.com/zh/model-studio/vision)<br/>[OpenRouter](https://openrouter.ai/qwen) | ⭐⭐⭐⭐<br />Stronger than Qwen3-VL and Qwen2.5-VL |
56+
| Qwen3.5<br />[Quick setup](./model-common-config#qwen3x) | [Alibaba Cloud](https://help.aliyun.com/zh/model-studio/vision)<br/>[OpenRouter](https://openrouter.ai/qwen) | ⭐⭐⭐⭐<br />Stronger than Qwen3-VL and Qwen2.5-VL |
5757
| Zhipu GLM-4.6V<br />[Quick setup](./model-common-config#glm-v) | [Z.AI (Global)](https://docs.z.ai/guides/vlm/glm-4.6v)<br/>[BigModel (CN)](https://docs.bigmodel.cn/cn/guide/models/vlm/glm-4.6v) | Newly integrated, welcome to try it out<br />Weights open-sourced on [HuggingFace](https://huggingface.co/zai-org/GLM-4.6V) |
5858
| Gemini-3-Pro / Gemini-3-Flash<br />[Quick setup](./model-common-config#gemini-3-pro) | [Google Cloud](https://ai.google.dev/gemini-api/docs/models/gemini) | ⭐⭐⭐<br />Gemini-3-Flash is supported<br />Price is higher than Doubao and Qwen |
5959
| UI-TARS<br />[Quick setup](./model-common-config#ui-tars) | [Volcano Engine](https://www.volcengine.com/docs/82379/1536429) | ⭐⭐<br />Strong exploratory ability but results vary by scenario<br />Open-source versions available ([HuggingFace](https://huggingface.co/bytedance-research/UI-TARS-72B-SFT) / [GitHub](https://github.com/bytedance/ui-tars)) |

apps/site/docs/zh/model-common-config.mdx

Lines changed: 11 additions & 20 deletions
Original file line numberDiff line numberDiff line change
@@ -78,29 +78,17 @@ MIDSCENE_MODEL_FAMILY="doubao-seed" # 也兼容 "doubao-vision"
7878
# MIDSCENE_MODEL_REASONING_EFFORT="medium"
7979
```
8080

81-
### 千问 Qwen3.6 {#qwen36}
81+
### 千问 Qwen3.X 系列 {#qwen3x}
8282

83-
[阿里云](https://www.aliyun.com/)`qwen3.6-plus` 模型为例,推荐关闭平台默认的思考模式以提升执行速度,其环境变量配置如下:
84-
85-
```bash
86-
MIDSCENE_MODEL_BASE_URL="https://dashscope.aliyuncs.com/compatible-mode/v1"
87-
MIDSCENE_MODEL_API_KEY="......"
88-
MIDSCENE_MODEL_NAME="qwen3.6-plus"
89-
MIDSCENE_MODEL_FAMILY="qwen3.6"
90-
MIDSCENE_MODEL_REASONING_ENABLED="false"
91-
```
92-
93-
如果需要开启思考模式,可以配置 `MIDSCENE_MODEL_REASONING_ENABLED="true"` 并增加 `MIDSCENE_MODEL_REASONING_BUDGET="500"` 配置以控制思考耗时。
94-
95-
### 千问 Qwen3.5 {#qwen35}
83+
目前已支持 Qwen3.X 系列中的 Qwen3.5、Qwen3.6。
9684

9785
[阿里云](https://www.aliyun.com/)`qwen3.5-plus` 模型为例,推荐关闭平台默认的思考模式以提升执行速度,其环境变量配置如下:
9886

9987
```bash
10088
MIDSCENE_MODEL_BASE_URL="https://dashscope.aliyuncs.com/compatible-mode/v1"
10189
MIDSCENE_MODEL_API_KEY="......"
102-
MIDSCENE_MODEL_NAME="qwen3.5-plus"
103-
MIDSCENE_MODEL_FAMILY="qwen3.5"
90+
MIDSCENE_MODEL_NAME="qwen3.5-plus" # Qwen3.6 系列的 plus 模型是 "qwen3.6-plus"
91+
MIDSCENE_MODEL_FAMILY="qwen3.5" # Qwen3.6 系列对应的 family 是 "qwen3.6"
10492
MIDSCENE_MODEL_REASONING_ENABLED="false"
10593
```
10694

@@ -117,20 +105,23 @@ MIDSCENE_MODEL_NAME="qwen3-vl-plus"
117105
MIDSCENE_MODEL_FAMILY="qwen3-vl"
118106
```
119107

120-
### 智谱 GLM-V {#glm-v}
108+
### 智谱 GLM-V 系列 {#glm-v}
121109

122-
智谱 GLM-V 是智谱 AI 推出的开源视觉模型。以 `GLM-4.6V` 为例:
110+
智谱 GLM-V 是智谱 AI 推出的视觉理解模型。最新版本有 `GLM-4.6V`(开源)、`GLM-5V-Turbo`
111+
112+
推荐关闭平台默认的思考模式以提升执行速度,以 `GLM-4.6V` 为例:
123113

124114
[Z.AI(国际)](https://z.ai/manage-apikey/apikey-list)[BigModel(国内)](https://bigmodel.cn/usercenter/proj-mgmt/apikeys)获取 API 密钥,然后设置:
125115

126116
```bash
127117
MIDSCENE_MODEL_BASE_URL="https://open.bigmodel.cn/api/paas/v4" # 或 https://api.z.ai/api/paas/v4
128118
MIDSCENE_MODEL_API_KEY="......"
129-
MIDSCENE_MODEL_NAME="glm-4.6v"
119+
MIDSCENE_MODEL_NAME="glm-4.6v" # GLM-5V-Turbo 的对应模型名为 "glm-5v-turbo"
130120
MIDSCENE_MODEL_FAMILY="glm-v"
121+
MIDSCENE_MODEL_REASONING_ENABLED="false"
131122
```
132123

133-
**了解更多关于智谱 GLM-V**
124+
**了解更多关于智谱 GLM-V 开源模型**
134125

135126
- Github: [https://github.com/zai-org/GLM-V](https://github.com/zai-org/GLM-V)
136127
- Hugging Face: [https://huggingface.co/zai-org/GLM-4.6V](https://huggingface.co/zai-org/GLM-4.6V)

apps/site/docs/zh/model-strategy.mdx

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,7 @@ import TroubleshootingLLMConnectivity from './common/troubleshooting-llm-connect
66

77
如果你想快速开始体验 Midscene,请选择模型并参考配置文档:
88
* [豆包 Seed 模型](./model-common-config#doubao-seed-model)
9-
* 千问系列: [Qwen3.6](./model-common-config#qwen36), [Qwen3.5](./model-common-config#qwen35), [Qwen3-VL](./model-common-config#qwen3-vl)
9+
* 千问系列: [Qwen3.X 系列](./model-common-config#qwen3x), [Qwen3-VL](./model-common-config#qwen3-vl)
1010
* [智谱 GLM-V](./model-common-config#glm-v)
1111
* [智谱 AutoGLM](./model-common-config#auto-glm)
1212
* [Gemini-3-Pro / Gemini-3-Flash](./model-common-config#gemini-3-pro)
@@ -53,7 +53,7 @@ DOM 定位方案的稳定性不足预期,它常在 Canvas 元素、CSS backgro
5353
|模型系列|部署|Midscene 评价|
5454
|---|---|---|
5555
|豆包 Seed 模型<br />[快速配置](./model-common-config#doubao-seed-model)|火山引擎版本:<br />[Doubao-Seed-1.6-Vision](https://www.volcengine.com/docs/82379/1799865)<br />[Doubao-Seed-2.0-Lite](https://www.volcengine.com/docs/82379/1799865)|⭐⭐⭐⭐<br/>UI 操作规划、定位能力较强<br />速度略慢|
56-
|千问 Qwen3.5<br />[快速配置](./model-common-config#qwen35)|[阿里云](https://help.aliyun.com/zh/model-studio/vision)<br/>[OpenRouter](https://openrouter.ai/qwen)|⭐⭐⭐⭐<br/>综合效果优于 Qwen3-VL 和 Qwen2.5-VL |
56+
|千问 Qwen3.5<br />[快速配置](./model-common-config#qwen3x)|[阿里云](https://help.aliyun.com/zh/model-studio/vision)<br/>[OpenRouter](https://openrouter.ai/qwen)|⭐⭐⭐⭐<br/>综合效果优于 Qwen3-VL 和 Qwen2.5-VL |
5757
|智谱 GLM-4.6V<br />[快速配置](./model-common-config#glm-v)|[Z.AI (Global)](https://docs.z.ai/guides/vlm/glm-4.6v)<br/>[BigModel (CN)](https://docs.bigmodel.cn/cn/guide/models/vlm/glm-4.6v)|全新接入,欢迎体验<br />模型权重开源[HuggingFace](https://huggingface.co/zai-org/GLM-4.6V) |
5858
|Gemini-3-Pro / Gemini-3-Flash<br />[快速配置](./model-common-config#gemini-3-pro)|[Google Cloud](https://ai.google.dev/gemini-api/docs/models/gemini)|⭐⭐⭐<br />支持 Gemini-3-Flash<br />价格高于豆包和千问|
5959
|UI-TARS <br />[快速配置](./model-common-config#ui-tars)|[火山引擎](https://www.volcengine.com/docs/82379/1536429)|⭐⭐<br /> 有探索能力,但在不同场景表现可能差异较大<br />有开源版本([HuggingFace](https://huggingface.co/bytedance-research/UI-TARS-72B-SFT) / [Github](https://github.com/bytedance/ui-tars)|

0 commit comments

Comments
 (0)