- EPSS 1.7%
- Veröffentlicht 19.03.2025 15:33:28
- Zuletzt bearbeitet 01.07.2025 20:52:17
vLLM is a high-throughput and memory-efficient inference and serving engine for LLMs. When vLLM is configured to use Mooncake, unsafe deserialization exposed directly over ZMQ/TCP on all network interfaces will allow attackers to execute remote code ...
CVE-2025-29770
- EPSS 0.32%
- Veröffentlicht 19.03.2025 15:31:00
- Zuletzt bearbeitet 31.07.2025 15:58:58
vLLM is a high-throughput and memory-efficient inference and serving engine for LLMs. The outlines library is one of the backends used by vLLM to support structured output (a.k.a. guided decoding). Outlines provides an optional cache for its compiled...
CVE-2025-25183
- EPSS 0.32%
- Veröffentlicht 07.02.2025 20:15:34
- Zuletzt bearbeitet 01.07.2025 20:58:00
vLLM is a high-throughput and memory-efficient inference and serving engine for LLMs. Maliciously constructed statements can lead to hash collisions, resulting in cache reuse, which can interfere with subsequent responses and cause unintended behavio...
CVE-2025-24357
- EPSS 1.01%
- Veröffentlicht 27.01.2025 18:15:41
- Zuletzt bearbeitet 27.06.2025 19:30:59
vLLM is a library for LLM inference and serving. vllm/model_executor/weight_utils.py implements hf_model_weights_iterator to load the model checkpoint, which is downloaded from huggingface. It uses the torch.load function and the weights_only paramet...