import ImageGenerationModelsTable from './_components/image-generation-models-table'; import VideoGenerationModelsTable from './_components/video-generation-models-table'; import LLMModelsTable from './_components/llm-models-table'; import VLMModelsTable from './_components/vlm-models-table'; import WhisperModelsTable from './_components/whisper-models-table'; import TextEmbeddingsModelsTable from './_components/text-embeddings-models-table'; import SpeechGenerationModelsTable from './_components/speech-generation-models-table'; import TextRerankModelsTable from './_components/text-rerank-models-table';
:::info Models Compatibility Other models with similar architectures may also work successfully even if not explicitly validated. Consider testing any unlisted models to verify compatibility with your specific use case. :::
:::tip LoRA Support LLM pipeline supports LoRA adapters. :::
::::info
The LLM pipeline can work with other similar topologies produced by optimum-intel with the same model signature.
The model is required to have the following inputs after the conversion:
input_idscontains the tokens.attention_maskis filled with1.beam_idxselects beams.position_ids(optional) encodes a position of currently generating token in the sequence and a singlelogitsoutput.
:::note
Models should belong to the same family and have the same tokenizers.
:::
::::
:::info LoRA Support VLM pipeline does not support LoRA adapters. :::
:::warning VLM Models Notes
To convert InternVL2 models, timm and einops are required:
pip install timm einopsopenbmb/MiniCPM-o-2_6doesn't supporttransformers>=4.52which is required foroptimum-cliexport.--task image-text-to-textis required foroptimum-cli export openvino --trust-remote-codebecauseimage-text-to-textisn'tMiniCPM-o-2_6's native task.
Models' configs aren't consistent. It's required to override the default eos_token_id with the one from a tokenizer:
generation_config.set_eos_token_id(pipe.get_tokenizer().get_eos_token_id())Apply https://huggingface.co/microsoft/Phi-4-multimodal-instruct/discussions/78/files to fix the model export for transformers>=4.50
The model requires transformers>=4.57 for the export with optimum-cli.
The model requires transformers>=4.48 for the export with optimum-cli.
:::
:::info LoRA Support Speech recognition pipeline does not support LoRA adapters. :::
:::info LoRA Support Speech generation pipeline does not support LoRA adapters. :::
:::info LoRA Support Text embeddings pipeline does not support LoRA adapters. :::
:::warning Text Embeddings Models Notes
Qwen3 Embedding models require --task feature-extraction during the conversion with optimum-cli.
:::
:::info LoRA Support Text rerank pipeline does not support LoRA adapters. :::
:::warning Text Rerank Models Notes
Text Rerank models require appropriate --task provided during the conversion with optimum-cli. Task can be found in the table above.
:::
:::info Hugging Face Notes Some models may require access request submission on the Hugging Face page to be downloaded.
If https://huggingface.co/ is down, the conversion step won't be able to download the models. :::