Skip to content

Commit 7a0a9da

Browse files
varun-sundar-rabindranathvarun sundar rabindranath
and
varun sundar rabindranath
authored
[Doc] V1 : Update LoRA status (#17133)
Signed-off-by: varun sundar rabindranath <vsundarr@redhat.com> Co-authored-by: varun sundar rabindranath <vsundarr@redhat.com>
1 parent 69bff9b commit 7a0a9da

File tree

1 file changed

+1
-6
lines changed

1 file changed

+1
-6
lines changed

docs/source/getting_started/v1_user_guide.md

+1-6
Original file line numberDiff line numberDiff line change
@@ -44,8 +44,8 @@ This living user guide outlines a few known **important changes and limitations*
4444
|-----------------|-----------------------------------------------------------------------------------|
4545
| **Prefix Caching** | <nobr>🚀 Optimized</nobr> |
4646
| **Chunked Prefill** | <nobr>🚀 Optimized</nobr> |
47+
| **LoRA** | <nobr>🚀 Optimized</nobr> |
4748
| **Logprobs Calculation** | <nobr>🟢 Functional</nobr> |
48-
| **LoRA** | <nobr>🟢 Functional ([PR #13096](https://github.com/vllm-project/vllm/pull/13096))</nobr>|
4949
| **Multimodal Models** | <nobr>🟢 Functional</nobr> |
5050
| **FP8 KV Cache** | <nobr>🟢 Functional on Hopper devices ([PR #15191](https://github.com/vllm-project/vllm/pull/15191))</nobr>|
5151
| **Spec Decode** | <nobr>🚧 WIP ([PR #13933](https://github.com/vllm-project/vllm/pull/13933))</nobr>|
@@ -121,11 +121,6 @@ Although we have re-implemented and partially optimized many features and models
121121
These features are already supported in vLLM V1, but their optimization is still
122122
in progress.
123123

124-
- **LoRA**: LoRA is functionally working on vLLM V1 but its performance is
125-
inferior to that of V0. The team is actively working on improving its
126-
performance
127-
(e.g., see [PR #13096](https://github.com/vllm-project/vllm/pull/13096)).
128-
129124
- **Spec Decode**: Currently, only ngram-based spec decode is supported in V1. There
130125
will be follow-up work to support other types of spec decode (e.g., see [PR #13933](https://github.com/vllm-project/vllm/pull/13933)). We will prioritize the support for Eagle, MTP compared to draft model based spec decode.
131126

0 commit comments

Comments
 (0)