Vllm-project

Vllm

22 Schwachstellen gefunden.

Hinweis: Diese Liste kann unvollständig sein. Daten werden ohne Gewähr im Ursprungsformat bereitgestellt.
Exploit
  • EPSS 0.14%
  • Veröffentlicht 30.05.2025 18:38:45
  • Zuletzt bearbeitet 01.07.2025 20:42:13

vLLM is an inference and serving engine for large language models (LLMs). In version 0.8.0 up to but excluding 0.9.0, the vLLM backend used with the /v1/chat/completions OpenAPI endpoint fails to validate unexpected or malformed input in the "pattern...

  • EPSS 0.08%
  • Veröffentlicht 30.05.2025 18:36:01
  • Zuletzt bearbeitet 02.06.2025 17:32:17

vLLM is an inference and serving engine for large language models (LLMs). Version 0.8.0 up to but excluding 0.9.0 have a Denial of Service (ReDoS) that causes the vLLM server to crash if an invalid regex was provided while using structured output. Th...

  • EPSS 0.07%
  • Veröffentlicht 30.05.2025 18:33:40
  • Zuletzt bearbeitet 02.06.2025 17:32:17

vLLM is an inference and serving engine for large language models (LLMs). In versions 0.8.0 up to but excluding 0.9.0, hitting the /v1/completions API with a invalid json_schema as a Guided Param kills the vllm server. This vulnerability is similar ...

Exploit
  • EPSS 0.12%
  • Veröffentlicht 30.05.2025 17:36:16
  • Zuletzt bearbeitet 19.06.2025 00:55:27

vLLM, an inference and serving engine for large language models (LLMs), has a Regular Expression Denial of Service (ReDoS) vulnerability in the file `vllm/entrypoints/openai/tool_parsers/pythonic_tool_parser.py` of versions 0.6.4 up to but excluding ...

  • EPSS 0.1%
  • Veröffentlicht 29.05.2025 16:36:12
  • Zuletzt bearbeitet 30.05.2025 16:31:03

vLLM is an inference and serving engine for large language models (LLMs). In versions starting from 0.7.0 to before 0.9.0, in the file vllm/multimodal/hasher.py, the MultiModalHasher class has a security and data integrity issue in its image hashing ...

  • EPSS 0.06%
  • Veröffentlicht 29.05.2025 16:32:42
  • Zuletzt bearbeitet 30.05.2025 16:31:03

vLLM is an inference and serving engine for large language models (LLMs). Prior to version 0.9.0, when a new prompt is processed, if the PageAttention mechanism finds a matching prefix chunk, the prefill process speeds up, which is reflected in the T...

Exploit
  • EPSS 0.37%
  • Veröffentlicht 20.05.2025 17:32:27
  • Zuletzt bearbeitet 13.08.2025 16:35:57

vLLM, an inference and serving engine for large language models (LLMs), has an issue in versions 0.6.5 through 0.8.4 that ONLY impacts environments using the `PyNcclPipe` KV cache transfer integration with the V0 engine. No other configurations are a...

  • EPSS 1.31%
  • Veröffentlicht 06.05.2025 16:53:52
  • Zuletzt bearbeitet 31.07.2025 18:05:30

vLLM is an inference and serving engine for large language models. In a multi-node vLLM deployment using the V0 engine, vLLM uses ZeroMQ for some multi-node communication purposes. The secondary vLLM hosts open a `SUB` ZeroMQ socket and connect to an...

Exploit
  • EPSS 2.48%
  • Veröffentlicht 30.04.2025 00:25:00
  • Zuletzt bearbeitet 28.05.2025 19:12:58

vLLM is a high-throughput and memory-efficient inference and serving engine for LLMs. Versions starting from 0.6.5 and prior to 0.8.5, having vLLM integration with mooncake, are vulnerable to remote code execution due to using pickle based serializat...

Exploit
  • EPSS 0.57%
  • Veröffentlicht 30.04.2025 00:24:53
  • Zuletzt bearbeitet 28.05.2025 19:15:56

vLLM is a high-throughput and memory-efficient inference and serving engine for LLMs. Versions starting from 0.8.0 and prior to 0.8.5 are affected by a critical performance vulnerability in the input preprocessing logic of the multimodal tokenizer. T...