1212 SpeechT5HifiGan ,
1313 AutoModelForSequenceClassification
1414)
15- from diffusers .pipelines import DiffusionPipeline , LDMSuperResolutionPipeline
15+ from diffusers .pipelines import DiffusionPipeline , LDMSuperResolutionPipeline , LTXPipeline
1616from optimum .intel .openvino import (
1717 OVModelForCausalLM ,
1818 OVModelForSeq2SeqLM ,
2323 OVPipelineForImage2Image ,
2424 OVModelForFeatureExtraction ,
2525 OVModelForTextToSpeechSeq2Seq ,
26- OVModelForSequenceClassification
26+ OVModelForSequenceClassification ,
27+ OVLTXPipeline ,
2728)
2829from llm_bench_utils .ov_model_classes import OVMPTModel , OVLDMSuperResolutionPipeline , OVChatGLMModel
2930from dataclasses import dataclass , field
3031
3132
3233@dataclass
3334class UseCase :
34- task = ''
35+ task = ""
3536 model_types : list [str ] = field (default_factory = list )
3637 ov_cls : type | None = None
3738 pt_cls : type | None = AutoModel
@@ -45,17 +46,17 @@ class UseCaseImageGen(UseCase):
4546 pt_cls : type | None = DiffusionPipeline
4647
4748 TASK = {
48- "text2img" : {"name" : ' text-to-image' , "ov_cls" : OVDiffusionPipeline },
49- "img2img" : {"name" : ' image-to-image' , "ov_cls" : OVPipelineForImage2Image },
50- "inpainting" : {"name" : ' inpainting' , "ov_cls" : OVPipelineForInpainting }
49+ "text2img" : {"name" : " text-to-image" , "ov_cls" : OVDiffusionPipeline },
50+ "img2img" : {"name" : " image-to-image" , "ov_cls" : OVPipelineForImage2Image },
51+ "inpainting" : {"name" : " inpainting" , "ov_cls" : OVPipelineForInpainting }
5152 }
5253
5354
5455@dataclass
5556class UseCaseVideoGen (UseCase ):
5657 task = "video_gen"
57- ov_cls : type | None = OVDiffusionPipeline
58- pt_cls : type | None = DiffusionPipeline
58+ ov_cls : type | None = OVLTXPipeline
59+ pt_cls : type | None = LTXPipeline
5960
6061
6162@dataclass
@@ -81,35 +82,35 @@ class UseCaseTextGen(UseCase):
8182
8283@dataclass
8384class UseCaseCodeGen (UseCase ):
84- task = ' code_gen'
85+ task = " code_gen"
8586 ov_cls : type | None = OVModelForCausalLM
8687 pt_cls : type | None = AutoModelForCausalLM
8788
8889
8990@dataclass
9091class UseCaseImageCls (UseCase ):
91- task = ' image_cls'
92+ task = " image_cls"
9293 ov_cls : type | None = OVModelForCausalLM
9394 pt_cls : type | None = AutoModelForCausalLM
9495
9596
9697@dataclass
9798class UseCaseLDMSuperResolution (UseCase ):
98- task = ' ldm_super_resolution'
99+ task = " ldm_super_resolution"
99100 ov_cls : type | None = OVLDMSuperResolutionPipeline
100101 pt_cls : type | None = LDMSuperResolutionPipeline
101102
102103
103104@dataclass
104105class UseCaseTextEmbeddings (UseCase ):
105- task = ' text_embed'
106+ task = " text_embed"
106107 ov_cls : type | None = OVModelForFeatureExtraction
107108 pt_cls : type | None = AutoModel
108109
109110
110111@dataclass
111112class UseCaseTextReranker (UseCase ):
112- task = ' text_rerank'
113+ task = " text_rerank"
113114 ov_cls : type | None = OVModelForSequenceClassification
114115 pt_cls : type | None = AutoModelForSequenceClassification
115116
@@ -125,36 +126,36 @@ def is_qwen_causallm_arch(config):
125126
126127@dataclass
127128class UseCaseTextToSpeech (UseCase ):
128- task = ' text_to_speech'
129+ task = " text_to_speech"
129130 ov_cls : type | None = OVModelForTextToSpeechSeq2Seq
130131 pt_cls : type | None = SpeechT5ForTextToSpeech
131132 tokenizer_cls : type = SpeechT5Processor
132133 vocoder_cls : type = SpeechT5HifiGan
133134
134135
135136USE_CASES = {
136- ' image_gen' : [UseCaseImageGen ([' stable-diffusion-' , ' ssd-' , ' tiny-sd' , ' small-sd' , ' lcm-' , ' sdxl' , ' dreamlike' , "flux" ])],
137- ' video_gen' : [UseCaseVideoGen (' ltx' )],
137+ " image_gen" : [UseCaseImageGen ([" stable-diffusion-" , " ssd-" , " tiny-sd" , " small-sd" , " lcm-" , " sdxl" , " dreamlike" , "flux" ])],
138+ " video_gen" : [UseCaseVideoGen ([ " ltx" ] )],
138139 "visual_text_gen" : [UseCaseVLM (["llava" , "llava-next" , "qwen2-vl" , "llava-qwen2" , "internvl-chat" , "minicpmv" , "phi3-v" ,
139140 "minicpm-v" , "minicpmo" , "maira2" , "qwen2-5-vl" , "smolvlm" ])],
140- ' speech_to_text' : [UseCaseSpeech2Text ([' whisper' ])],
141- ' image_cls' : [UseCaseImageCls ([' vit' ])],
142- ' code_gen' : [UseCaseCodeGen (["codegen" , "codegen2" , "stable-code" ]),
143- UseCaseCodeGen ([' replit' ], ov_cls = OVMPTModel ),
144- UseCaseCodeGen ([' codet5' ], ov_cls = OVModelForSeq2SeqLM )],
145- ' text_gen' : [UseCaseTextGen ([' arcee' , "decoder" , "falcon" , "glm" , "aquila" , "gpt" , "gpt-" , "gpt2" , "open-llama" , "openchat" , "neural-chat" , "llama" ,
141+ " speech_to_text" : [UseCaseSpeech2Text ([" whisper" ])],
142+ " image_cls" : [UseCaseImageCls ([" vit" ])],
143+ " code_gen" : [UseCaseCodeGen (["codegen" , "codegen2" , "stable-code" ]),
144+ UseCaseCodeGen ([" replit" ], ov_cls = OVMPTModel ),
145+ UseCaseCodeGen ([" codet5" ], ov_cls = OVModelForSeq2SeqLM )],
146+ " text_gen" : [UseCaseTextGen ([" arcee" , "decoder" , "falcon" , "glm" , "aquila" , "gpt" , "gpt-" , "gpt2" , "open-llama" , "openchat" , "neural-chat" , "llama" ,
146147 "tiny-llama" , "tinyllama" , "opt" , "opt-" , "pythia" , "pythia-" , "stablelm" , "stablelm-" , "stable-zephyr-" , "rocket-" ,
147148 "vicuna" , "dolly" , "bloom" , "red-pajama" , "xgen" , "longchat" , "jais" , "orca-mini" , "baichuan" , "qwen" , "zephyr" ,
148149 "mistral" , "mixtral" , "phi" , "phi2-" , "minicpm" , "gemma" , "deci" , "phi3" , "internlm" , "olmo" , "starcoder" , "instruct-gpt" ,
149150 "granite" , "granitemoe" , "gptj" , "yi-" ]),
150- UseCaseTextGen (['t5' ], ov_cls = OVModelForSeq2SeqLM , pt_cls = T5ForConditionalGeneration ),
151- UseCaseTextGen ([' mpt' ], OVMPTModel ),
152- UseCaseTextGen ([' blenderbot' ], ov_cls = OVModelForSeq2SeqLM , pt_cls = BlenderbotForConditionalGeneration ),
153- UseCaseTextGen ([' chatglm' ], ov_cls = OVChatGLMModel , pt_cls = AutoModel )],
154- ' ldm_super_resolution' : [UseCaseLDMSuperResolution ([' ldm-super-resolution' ])],
155- ' text_embed' : [UseCaseTextEmbeddings (["qwen3" , "bge" , "bert" , "albert" , "roberta" , "xlm-roberta" ])],
156- ' text_rerank' : [UseCaseTextReranker (["qwen3" , "bge" , "bert" , "albert" , "roberta" , "xlm-roberta" ])],
157- ' text_to_speech' : [UseCaseTextToSpeech ([' speecht5' ])],
151+ UseCaseTextGen (["t5" ], ov_cls = OVModelForSeq2SeqLM , pt_cls = T5ForConditionalGeneration ),
152+ UseCaseTextGen ([" mpt" ], OVMPTModel ),
153+ UseCaseTextGen ([" blenderbot" ], ov_cls = OVModelForSeq2SeqLM , pt_cls = BlenderbotForConditionalGeneration ),
154+ UseCaseTextGen ([" chatglm" ], ov_cls = OVChatGLMModel , pt_cls = AutoModel )],
155+ " ldm_super_resolution" : [UseCaseLDMSuperResolution ([" ldm-super-resolution" ])],
156+ " text_embed" : [UseCaseTextEmbeddings (["qwen3" , "bge" , "bert" , "albert" , "roberta" , "xlm-roberta" ])],
157+ " text_rerank" : [UseCaseTextReranker (["qwen3" , "bge" , "bert" , "albert" , "roberta" , "xlm-roberta" ])],
158+ " text_to_speech" : [UseCaseTextToSpeech ([" speecht5" ])],
158159}
159160
160161PA_ATTENTION_BACKEND = "PA"
0 commit comments