@@ -269,42 +269,6 @@ jobs:
269269 diff pred2.txt ref.txt
270270 echo "Chat sample python" passed
271271
272- benchmark_genai-ubuntu :
273- runs-on : ubuntu-24.04
274- defaults :
275- run :
276- shell : bash
277- steps :
278- - uses : actions/checkout@v4
279- with :
280- submodules : recursive
281- - uses : actions/setup-python@v4
282- with :
283- python-version : 3.11
284- - name : Install OpenVINO
285- run : |
286- mkdir ./ov/
287- curl ${{ env.l_ov_link }} | tar --directory ./ov/ --strip-components 1 -xz
288- sudo ./ov/install_dependencies/install_openvino_dependencies.sh
289- - name : Build app
290- run : |
291- source ./ov/setupvars.sh
292- cmake -DCMAKE_BUILD_TYPE=Release -S ./ -B ./build/
293- cmake --build ./build/ --config Release -j
294- - name : Download and convert and model
295- run : |
296- source ./ov/setupvars.sh
297- python -m pip install ./thirdparty/openvino_tokenizers/[transformers] --extra-index-url https://storage.openvinotoolkit.org/simple/wheels/nightly
298- python -m pip install -r ./samples/requirements.txt
299- optimum-cli export openvino --trust-remote-code --weight-format fp16 --model TinyLlama/TinyLlama-1.1B-Chat-v1.0 TinyLlama-1.1B-Chat-v1.0
300- - name : Run
301- env :
302- PYTHONPATH : " ./build"
303- run : |
304- source ./ov/setupvars.sh
305- timeout 60s ./build/samples/cpp/text_generation/benchmark_genai -m ./TinyLlama-1.1B-Chat-v1.0/ -p "Why is the sun yellow?" --nw 2 -n 3 --mt 50 -d CPU
306- timeout 60s python ./samples/python/text_generation/benchmark_genai.py -m ./TinyLlama-1.1B-Chat-v1.0/ -p "Why is the sun yellow?" -nw 2 -n 3 -mt 50 -d CPU
307-
308272 visual_language_chat_sample-ubuntu-minicpm_v2_6 :
309273 runs-on : ubuntu-22.04-16-cores
310274 steps :
@@ -399,18 +363,6 @@ jobs:
399363 f.write(content.encode("utf-8"))
400364 - run : diff cpp2.txt py2.txt
401365
402- visual_language_chat_sample-ubuntu-llava_1_5 :
403- uses : ./.github/workflows/job_vlm_sample_llava.yml
404- with :
405- model_id : llava-hf/llava-1.5-7b-hf
406- model_dir : llava_1_5_7b_ov
407-
408- visual_language_chat_sample-ubuntu-llava_next :
409- uses : ./.github/workflows/job_vlm_sample_llava.yml
410- with :
411- model_id : llava-hf/llava-v1.6-mistral-7b-hf
412- model_dir : llava_v1_6_mistral_7b_ov
413-
414366 visual_language_chat_sample-ubuntu-internvl2 :
415367 runs-on : ubuntu-22.04-16-cores
416368 steps :
@@ -611,7 +563,7 @@ jobs:
611563 Overall_Status :
612564 name : ci/gha_overall_status_causal_lm
613565 needs : [cpp-greedy_causal_lm-windows, cpp-Phi-1_5, cpp-greedy_causal_lm-redpajama-3b-chat, cpp-chat_sample-ubuntu, cpp-continuous-batching-ubuntu,
614- visual_language_chat_sample-ubuntu-minicpm_v2_6, visual_language_chat_sample-ubuntu-llava_1_5, visual_language_chat_sample-ubuntu-llava_next, visual_language_chat_sample-ubuntu- internvl2,
566+ visual_language_chat_sample-ubuntu-minicpm_v2_6, visual_language_chat_sample-ubuntu-internvl2,
615567 cpp-continuous-batching-windows, cpp-continuous-batching-macos]
616568 if : ${{ always() }}
617569 runs-on : ubuntu-latest
0 commit comments