Description
Hi,
I am using a linux aarch64 device using ORT and onnxruntime-genai v0.5.2
On executing phi3.5-vision model on CPU following the steps mentioned: https://onnxruntime.ai/docs/genai/tutorials/phi3-v.html#run-on-cpu
The program gets killed with oom-error. My device has 16GB memory. I can easily execute phi3.5-mini models on my device, but phi3.5-vision is failing due to oom-kill error.
my error log is as follows:
python3 phi3-v.py -m /tmp/cpu_and_mobile/cpu-int4-rtn-block-32-acc-level-4/ -p cpu
Loading model...
Model loaded
Image Path (comma separated; leave empty if no image): car.jpg
Using image: car.jpg
Prompt: describe image
Processing images and prompt...
Generating response...
Killed
are there any specific formats of image the model takes in?
I have faced this 'killed' issue with ORT before. With ORT, i have to set the flag enable_cpu_mem_arena
to False
How do i execute the same using the provided python script https://github.com/microsoft/onnxruntime-genai/blob/rel-0.5.2/examples/python/phi3v.py
Do ort-genai also have such flags while executing generator models?