Supported Models
Other models with similar architectures may also work successfully even if not explicitly validated. Consider testing any unlisted models to verify compatibility with your specific use case.
Large Language Models (LLMs)
LLM pipeline supports LoRA adapters.
The LLM pipeline can work with other similar topologies produced by optimum-intel with the same model signature.
The model is required to have the following inputs after the conversion:
input_idscontains the tokens.attention_maskis filled with1.beam_idxselects beams.position_ids(optional) encodes a position of currently generating token in the sequence and a singlelogitsoutput.
Models should belong to the same family and have the same tokenizers.
Image Generation Models
Visual Language Models (VLMs)
VLM pipeline does not support LoRA adapters.
| Architecture | Models | Example HuggingFace Models |
|---|---|---|
InternVLChat | InternVLChatModel (Notes) | |
LLaVA | LLaVA-v1.5 | |
nanoLLaVA | nanoLLaVA | |
| nanoLLaVA-1.5 | ||
LLaVA-NeXT | LLaVA-v1.6 | |
LLaVA-NeXT-Video | LLaVA-Next-Video | |
MiniCPMO | MiniCPM-o-2_6 (Notes) | |
MiniCPMV | MiniCPM-V-2_6 | |
Phi3VForCausalLM | phi3_v (Notes) | |
Phi4MMForCausalLM | phi4mm (Notes) | |
Qwen2-VL | Qwen2-VL | |
Qwen2.5-VL | Qwen2.5-VL | |
Gemma3ForConditionalGeneration | gemma3 |
InternVL2
To convert InternVL2 models, timm and einops are required:
pip install timm einops
MiniCPMO
openbmb/MiniCPM-o-2_6doesn't supporttransformers>=4.52which is required foroptimum-cliexport.--task image-text-to-textis required foroptimum-cli export openvino --trust-remote-codebecauseimage-text-to-textisn'tMiniCPM-o-2_6's native task.
phi3_v
Models' configs aren't consistent. It's required to override the default eos_token_id with the one from a tokenizer:
generation_config.set_eos_token_id(pipe.get_tokenizer().get_eos_token_id())
phi4mm
Apply https://huggingface.co/microsoft/Phi-4-multimodal-instruct/discussions/78/files to fix the model export for transformers>=4.50
Speech Recognition Models (Whisper-based)
Speech recognition pipeline does not support LoRA adapters.
| Architecture | Models | Example HuggingFace Models |
|---|---|---|
WhisperForConditionalGeneration | Whisper | |
| Distil-Whisper |
Speech Generation Models
Speech generation pipeline does not support LoRA adapters.
| Architecture | Models | Example HuggingFace Models |
|---|---|---|
SpeechT5ForTextToSpeech | SpeechT5 TTS |
Text Embeddings Models
Text embeddings pipeline does not support LoRA adapters.
| Architecture | Example HuggingFace Models |
|---|---|
BertModel | |
MPNetForMaskedLM | |
RobertaForMaskedLM | |
XLMRobertaModel | |
Qwen3ForCausalLM |
Qwen3 Embedding models require --task feature-extraction during the conversion with optimum-cli.
Text Rerank Models
Text rerank pipeline does not support LoRA adapters.
| Architecture | `optimum-cli` task | Example HuggingFace Models |
|---|---|---|
BertForSequenceClassification | text-classification | |
XLMRobertaForSequenceClassification | text-classification | |
ModernBertForSequenceClassification | text-classification | |
Qwen3ForCausalLM | text-generation-with-past |
Text Rerank models require appropriate --task provided during the conversion with optimum-cli. Task can be found in the table above.
Some models may require access request submission on the Hugging Face page to be downloaded.
If https://huggingface.co/ is down, the conversion step won't be able to download the models.