You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardexpand all lines: .github/workflows/reminder_comment.yml
+6-2
Original file line number
Diff line number
Diff line change
@@ -2,7 +2,6 @@ name: PR Reminder Comment Bot
2
2
on:
3
3
pull_request_target:
4
4
types: [opened]
5
-
6
5
jobs:
7
6
pr_reminder:
8
7
runs-on: ubuntu-latest
@@ -15,7 +14,12 @@ jobs:
15
14
owner: context.repo.owner,
16
15
repo: context.repo.repo,
17
16
issue_number: context.issue.number,
18
-
body: '👋 Hi! Thank you for contributing to the vLLM project.\n Just a reminder: PRs would not trigger full CI run by default. Instead, it would only run `fastcheck` CI which starts running only a small and essential subset of CI tests to quickly catch errors. You can run other CI tests on top of those by going to your `fastcheck` build on Buildkite UI (linked in the PR checks section) and unblock them. If you do not have permission to unblock, ping `simon-mo` or `khluu` to add you in our Buildkite org. \n\nOnce the PR is approved and ready to go, your PR reviewer(s) can run CI to test the changes comprehensively before merging.\n\n To run CI, PR reviewers can do one of these:\n- Add `ready` label to the PR\n- Enable auto-merge.\n\n🚀'
17
+
body: '👋 Hi! Thank you for contributing to the vLLM project.\n\n' +
18
+
'💬 Join our developer Slack at https://slack.vllm.ai to discuss your PR in #pr-reviews, coordinate on features in #feat- channels, or join special interest groups in #sig- channels.\n\n' +
19
+
'Just a reminder: PRs would not trigger full CI run by default. Instead, it would only run `fastcheck` CI which starts running only a small and essential subset of CI tests to quickly catch errors. You can run other CI tests on top of those by going to your `fastcheck` build on Buildkite UI (linked in the PR checks section) and unblock them. If you do not have permission to unblock, ping `simon-mo` or `khluu` to add you in our Buildkite org.\n\n' +
20
+
'Once the PR is approved and ready to go, your PR reviewer(s) can run CI to test the changes comprehensively before merging.\n\n' +
21
+
'To run CI, PR reviewers can either: Add `ready` label to the PR or enable auto-merge.\n\n' +
Copy file name to clipboardexpand all lines: docs/source/models/supported_models.md
+19-4
Original file line number
Diff line number
Diff line change
@@ -726,14 +726,14 @@ See [this page](#generative-models) for more information on how to use generativ
726
726
*`h2oai/h2ovl-mississippi-800m`, `h2oai/h2ovl-mississippi-2b`, etc.
727
727
*
728
728
* ✅︎
729
-
*
729
+
*\*
730
730
-*`Idefics3ForConditionalGeneration`
731
731
* Idefics3
732
732
* T + I
733
733
*`HuggingFaceM4/Idefics3-8B-Llama3` etc.
734
734
* ✅︎
735
735
*
736
-
*
736
+
* ✅︎
737
737
-*`InternVLChatModel`
738
738
* InternVL 2.5, Mono-InternVL, InternVL 2.0
739
739
* T + I<sup>E+</sup>
@@ -799,7 +799,7 @@ See [this page](#generative-models) for more information on how to use generativ
799
799
* ✅︎
800
800
-*`NVLM_D_Model`
801
801
* NVLM-D 1.0
802
-
* T + I<sup>E+</sup>
802
+
* T + I<sup>+</sup>
803
803
*`nvidia/NVLM-D-72B`, etc.
804
804
*
805
805
* ✅︎
@@ -846,6 +846,13 @@ See [this page](#generative-models) for more information on how to use generativ
846
846
* ✅︎
847
847
* ✅︎
848
848
* ✅︎
849
+
-*`Qwen2_5_VLForConditionalGeneration`
850
+
* Qwen2.5-VL
851
+
* T + I<sup>E+</sup> + V<sup>E+</sup>
852
+
*`Qwen/Qwen2.5-VL-3B-Instruct`, `Qwen/Qwen2.5-VL-72B-Instruct`, etc.
853
+
*
854
+
* ✅︎
855
+
* ✅︎
849
856
-*`UltravoxModel`
850
857
* Ultravox
851
858
* T + A<sup>E+</sup>
@@ -859,7 +866,11 @@ See [this page](#generative-models) for more information on how to use generativ
859
866
<sup>+</sup> Multiple items can be inputted per text prompt for this modality.
860
867
861
868
:::{note}
862
-
To use `DeepSeek-VL2` series models, you have to pass `--hf_overrides '{"architectures": ["DeepseekVLV2ForCausalLM"]}'` when running vLLM.
869
+
To use DeepSeek-VL2 series models, you have to pass `--hf_overrides '{"architectures": ["DeepseekVLV2ForCausalLM"]}'` when running vLLM.
870
+
:::
871
+
872
+
:::{note}
873
+
H2O-VL series models will be available in V1 once we support backends other than FlashAttention.
863
874
:::
864
875
865
876
:::{note}
@@ -876,6 +887,10 @@ The chat template for Pixtral-HF is incorrect (see [discussion](https://huggingf
876
887
A corrected version is available at <gh-file:examples/template_pixtral_hf.jinja>.
877
888
:::
878
889
890
+
:::{note}
891
+
To use Qwen2.5-VL series models, you have to install Huggingface `transformers` library from source via `pip install git+https://github.com/huggingface/transformers`.
892
+
:::
893
+
879
894
### Pooling Models
880
895
881
896
See [this page](pooling-models) for more information on how to use pooling models.
0 commit comments