-
Notifications
You must be signed in to change notification settings - Fork 3k
Open
Description
Hi,
I followed these guides to export microsoft/Phi-3.5-vision-instruct to the openvino format so I can run it on my NPU:
- https://docs.openvino.ai/2025/model-server/ovms_demos_vlm_npu.html?utm_source=chatgpt.com#references
- https://docs.openvino.ai/2025/openvino-workflow-generative/inference-with-genai/inference-with-genai-on-npu.html
My goal is to use VLMPipeline on the NPU with MAX_PROMPT_LEN set to higher than 1024 (for example 2048).
However, when I pass the MAX_PROMPT_LEN parameter to VLMPipeline, same as @asteroidcow in issue #33385 , I am getting "NotFound: Unsupported property MAX_PROMPT_LEN by CPU plugin.". It seems like for some reason VLMPipeline does not support MAX_PROMPT_LEN even though LLMPipeline does.
I would appreciate some help.
(openvino version - 2025.4.0)
Thanks.
Reactions are currently unavailable
Metadata
Metadata
Assignees
Labels
No labels