Skip to content

phi3.5-vision fails on CPU #1146

Open
Open
@suyash-narain

Description

Hi,

I am using a linux aarch64 device using ORT and onnxruntime-genai v0.5.2

On executing phi3.5-vision model on CPU following the steps mentioned: https://onnxruntime.ai/docs/genai/tutorials/phi3-v.html#run-on-cpu

The program gets killed with oom-error. My device has 16GB memory. I can easily execute phi3.5-mini models on my device, but phi3.5-vision is failing due to oom-kill error.

my error log is as follows:

python3 phi3-v.py -m /tmp/cpu_and_mobile/cpu-int4-rtn-block-32-acc-level-4/ -p cpu
Loading model...
Model loaded
Image Path (comma separated; leave empty if no image): car.jpg
Using image: car.jpg
Prompt: describe image
Processing images and prompt...
Generating response...
Killed

are there any specific formats of image the model takes in?

I have faced this 'killed' issue with ORT before. With ORT, i have to set the flag enable_cpu_mem_arena to False

How do i execute the same using the provided python script https://github.com/microsoft/onnxruntime-genai/blob/rel-0.5.2/examples/python/phi3v.py

Do ort-genai also have such flags while executing generator models?

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions