Skip to content

Commit e34fccb

Browse files
natkebaijumeswani
andauthored
Update README.md (#1121)
Co-authored-by: Baiju Meswani <[email protected]>
1 parent 66e8817 commit e34fccb

File tree

1 file changed

+6
-6
lines changed

1 file changed

+6
-6
lines changed

README.md

+6-6
Original file line numberDiff line numberDiff line change
@@ -13,14 +13,14 @@ It implements the generative AI loop for ONNX models, including pre and post pro
1313

1414
See documentation at https://onnxruntime.ai/docs/genai.
1515

16-
| Support matrix | Supported now | Under development | On the roadmap |
16+
|Support matrix|Supported now|Under development|On the roadmap|
1717
| -------------- | ------------- | ----------------- | -------------- |
1818
| Model architectures | Gemma <br/> Llama * <br/> Mistral + <br/> Phi (language + vision) <br/> Qwen <br/> Nemotron <br/> Granite <br/> AMD OLMo | Whisper | Stable diffusion |
19-
| API | Python <br/> C# <br/> C/C++ <br/> Java ^ | Objective-C | |
20-
| Platform | Linux <br/> Windows <br/> Mac ^ <br/> Android ^ | | iOS |
21-
| Architecture | x86 <br/> x64 <br/> Arm64 ~ | | |
22-
| Hardware Acceleration | CUDA <br/> DirectML <br/> | QNN <br/> OpenVINO <br/> ROCm | |
23-
| Features | | Interactive decoding <br/> Customization (fine-tuning) | Speculative decoding |
19+
|API| Python <br/>C# <br/>C/C++ <br/> Java ^ |Objective-C||
20+
|Platform| Linux <br/> Windows <br/>Mac ^ <br/>Android ^ ||iOS |||
21+
|Architecture|x86 <br/> x64 <br/> Arm64 ~ ||||
22+
|Hardware Acceleration|CUDA<br/>DirectML<br/>|QNN <br/> OpenVINO <br/> ROCm ||
23+
|Features|MultiLoRA <br/> Continuous decoding (session continuation)^ | Constrained decoding | Speculative decoding |
2424

2525
\* The Llama model architecture supports similar model families such as CodeLlama, Vicuna, Yi, and more.
2626

0 commit comments

Comments
 (0)