How to test a multimodal model? #520
Unanswered
xiaozcy
asked this question in
User Support
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Hi team, I just learned about GuideLLM recently, and I tried to use a custom dataset to test a multimodal model(Qwen2.5-VL-3B-Instruct), but I found that the request sent was Incorrect.
My dataset looks like:
{"image": "/path/text.jpeg", "text": "What do you see happening in this image?\n<image>"}And the generated request body looks like:
{ "messages": [ { "role": "user", "content": [{"type": "text", "text": "What do you see happening in this image?\\n<image>'"}] }, { "role": "user", "content": [{"type": "image_url", "image_url": "data:image/jpeg;basexxxxxx"}] }] }But the request body expected to be generated looks like:
{ "messages": [ { "role": "user", "content": [ {"type": "text", "text": "What do you see happening in this image?\\n<image>'"}, {"type": "image_url", "image_url": "data:image/jpeg;basexxxxxx"}] }] }My command looks like:
I'm not sure if I'm using it the wrong way, and I'm really looking forward to your support, Thanks.
Beta Was this translation helpful? Give feedback.
All reactions