fix(deps): update dependency vllm to ^0.11.0 #257
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
This PR contains the following updates:
| Package | Change | Age | Confidence |
|
Generated Summary
This summary was generated with ❤️ by rigging
| vllm |
|
|
^0.5.0
->^0.11.0
|Release Notes
vllm-project/vllm (vllm)
v0.11.0
Compare Source
Highlights
This release features 538 commits, 207 contributors (65 new contributors)!
Model Support
Engine Core
Hardware & Performance
Large Scale Serving & Performance
Quantization
API & Frontend
Security
Dependencies
xm.mark_step
in favor oftorch_xla.sync
(#25254).V0 Deprecation
What's Changed
cpu_attn.py:_run_sdpa_forward
for better memory access by @ignaciosica in #24701--enable-log-outputs
does not match the documentation by @kebe7jun in #24626_validate_and_reshape_mm_tensor
by @lgeiger in #24742supports_kw
by @lgeiger in #24773s3_utils
type hints withBaseClient
by @Zerohertz in #24825stop
in reasoning content by @gaocegege in #14550kv_output_aggregator
support heterogeneous by @LCAIZJ in #23917MultiModalConfig
fromconfig/__init__.py
toconfig/multimodal.py
by @hmellor in #24659HuggingFace
->Hugging Face
inIntegration with Hugging Face
docs by @sergiopaniego in #24889is_flashmla_supported
Check Error by @yewentao256 in #24774n_groups % tp_size == 0
by @tomeras91 in #24593SpeculativeConfig
fromconfig/__init__.py
toconfig/speculative.py
by @hmellor in #24904EngineCoreRequest
arguments in tests and fix extra kwargs by @qthequartermasterman in #24987CpuGpuBuffer
for block table tensors by @njhill in #24795AutoModelForVision2Seq
by @DarkLight1337 in #25065cutlass_mla
hang by @alexm-redhat in #24966MultiModalCache
by @lgeiger in #25006sliding_window
from text config in Gemma3 MM by @hmellor in #25085Configuration
📅 Schedule: Branch creation - At any time (no schedule defined), Automerge - At any time (no schedule defined).
🚦 Automerge: Enabled.
♻ Rebasing: Whenever PR becomes conflicted, or you tick the rebase/retry checkbox.
🔕 Ignore: Close this PR and you won't be reminded about this update again.
This PR has been generated by Renovate Bot.