File tree Expand file tree Collapse file tree
samples/cpp/module_genai/config_yaml/Qwen3-Omni
src/cpp/src/module_genai/modules/md_text_to_speech/models Expand file tree Collapse file tree Original file line number Diff line number Diff line change @@ -34,7 +34,7 @@ pipeline_modules:
3434 code_predictor_single_codec_embedding_model_path : " ./tests/module_genai/cpp/test_models/Qwen3-Omni-4B-Instruct-multilingual-int4/qwen3_omni_code_predictor_codec_embed_model.xml"
3535 speech_decoder_model_path : " ./tests/module_genai/cpp/test_models/Qwen3-Omni-4B-Instruct-multilingual-int4/qwen3_omni_speech_decoder_model.xml"
3636 sample_codec_token_greedy_search : true
37- merge_ar_and_sce_ov_models : false
37+ merge_ar_and_sce_ov_models : true
3838 force_ar_model_inference_precision_f32 : false
3939
4040 pipeline_result :
Original file line number Diff line number Diff line change @@ -209,7 +209,7 @@ void TextToSpeechImpl_Qwen3Omni::load_code_predictor_models(const ov::AnyMap& tt
209209 if (m_sample_codec_token_greedy_search && m_merge_ar_and_sce_ov_models) {
210210 GENAI_INFO (" TextToSpeechModule[" + module_desc->name +
211211 " ]: sample_codec_token_greedy_search is enabled, will use greedy decoding in sample_codec_token" );
212- merge_code_predictor_ov_models (ar_models, sce_models);
212+ merge_code_predictor_ov_models (ar_models, sce_models, tts_props );
213213 if (m_enable_merge_ov_models) {
214214 // Release original infer requests to save memory since they won't be used anymore.
215215 m_code_predictor_ar_infers.clear ();
@@ -387,7 +387,8 @@ std::shared_ptr<ov::Model> merge_neighbor_models(std::shared_ptr<ov::Model>& mod
387387};
388388
389389void TextToSpeechImpl_Qwen3Omni::merge_code_predictor_ov_models (std::vector<std::shared_ptr<ov::Model>>& ar_models,
390- std::vector<std::shared_ptr<ov::Model>>& sce_models) {
390+ std::vector<std::shared_ptr<ov::Model>>& sce_models,
391+ const ov::AnyMap& tts_props) {
391392 if (ar_models.size () < 2 ) {
392393 GENAI_WARN (" TextToSpeechModule[" + module_desc->name + " ]: Not enough AR models to merge (found " +
393394 std::to_string (ar_models.size ()) + " ), will skip merging and use separate AR/SCE infer requests" );
Original file line number Diff line number Diff line change @@ -69,7 +69,8 @@ class TextToSpeechImpl_Qwen3Omni : public TextToSpeechModule {
6969
7070 void load_code_predictor_models (const ov::AnyMap& tts_props);
7171 void merge_code_predictor_ov_models (std::vector<std::shared_ptr<ov::Model>>& ar_models,
72- std::vector<std::shared_ptr<ov::Model>>& sce_models);
72+ std::vector<std::shared_ptr<ov::Model>>& sce_models,
73+ const ov::AnyMap& tts_props);
7374 std::vector<int64_t > code_predictor_ar_infers_merged_ov (int cp_steps,
7475 std::vector<float >& autoregressive_sequence,
7576 size_t batch,
You can’t perform that action at this time.
0 commit comments