Skip to content

Commit 5a7ff02

Browse files
authored
Merge pull request #6123 from hiyouga/hiyouga/fix_qwen2vl_vllm
[inference] fix qwen2vl vllm infer Former-commit-id: 0ce173e
2 parents 7ed5a71 + dcc67ac commit 5a7ff02

File tree

1 file changed

+9
-1
lines changed

1 file changed

+9
-1
lines changed

src/llamafactory/chat/vllm_engine.py

+9-1
Original file line numberDiff line numberDiff line change
@@ -112,7 +112,15 @@ async def _generate(
112112
if not any(IMAGE_PLACEHOLDER in message["content"] for message in messages):
113113
messages[0]["content"] = IMAGE_PLACEHOLDER * len(images) + messages[0]["content"]
114114

115-
paired_messages = messages + [{"role": "assistant", "content": ""}]
115+
if self.template.mm_plugin.__class__.__name__ == "Qwen2vlPlugin": # temporary solution
116+
image_str = "<|vision_start|>" + self.template.mm_plugin.image_token + "<|vision_end|>"
117+
else:
118+
image_str = self.template.mm_plugin.image_token
119+
120+
paired_messages = [
121+
{"role": message["role"], "content": message["content"].replace(IMAGE_PLACEHOLDER, image_str)}
122+
for message in messages
123+
] + [{"role": "assistant", "content": ""}]
116124
system = system or self.generating_args["default_system"]
117125
prompt_ids, _ = self.template.encode_oneturn(self.tokenizer, paired_messages, system, tools)
118126
prompt_length = len(prompt_ids)

0 commit comments

Comments
 (0)