Vllm

Vllm

18 Schwachstellen gefunden.

Hinweis: Diese Liste kann unvollständig sein. Daten werden ohne Gewähr im Ursprungsformat bereitgestellt.
  • EPSS 0.15%
  • Veröffentlicht 01.12.2025 22:45:42
  • Zuletzt bearbeitet 03.12.2025 17:52:26

vLLM is an inference and serving engine for large language models (LLMs). Prior to 0.11.1, vllm has a critical remote code execution vector in a config class named Nemotron_Nano_VL_Config. When vllm loads a model config that contains an auto_map entr...

  • EPSS 0.05%
  • Veröffentlicht 21.11.2025 01:22:37
  • Zuletzt bearbeitet 04.12.2025 17:40:47

vLLM is an inference and serving engine for large language models (LLMs). From version 0.5.5 to before 0.11.1, users can crash the vLLM engine serving multimodal models by passing multimodal embedding inputs with correct ndim but incorrect shape (e.g...

  • EPSS 0.05%
  • Veröffentlicht 21.11.2025 01:21:29
  • Zuletzt bearbeitet 04.12.2025 17:42:10

vLLM is an inference and serving engine for large language models (LLMs). From version 0.5.5 to before 0.11.1, the /v1/chat/completions and /tokenize endpoints allow a chat_template_kwargs request parameter that is used in the code before it is prope...

  • EPSS 0.3%
  • Veröffentlicht 21.11.2025 01:18:38
  • Zuletzt bearbeitet 04.12.2025 17:14:20

vLLM is an inference and serving engine for large language models (LLMs). From versions 0.10.2 to before 0.11.1, a memory corruption vulnerability could lead to a crash (denial-of-service) and potentially remote code execution (RCE), exists in the Co...

Exploit
  • EPSS 0.46%
  • Veröffentlicht 07.10.2025 14:15:38
  • Zuletzt bearbeitet 16.10.2025 18:02:09

vLLM is an inference and serving engine for large language models (LLMs). Before version 0.11.0rc2, the API key support in vLLM performs validation using a method that was vulnerable to a timing attack. API key validation uses a string comparison tha...

  • EPSS 0.37%
  • Veröffentlicht 21.08.2025 14:41:03
  • Zuletzt bearbeitet 09.10.2025 18:04:53

vLLM is an inference and serving engine for large language models (LLMs). From 0.1.0 to before 0.10.1.1, a Denial of Service (DoS) vulnerability can be triggered by sending a single HTTP GET request with an extremely large header to an HTTP endpoint....

Exploit
  • EPSS 0.09%
  • Veröffentlicht 30.05.2025 18:38:45
  • Zuletzt bearbeitet 01.07.2025 20:42:13

vLLM is an inference and serving engine for large language models (LLMs). In version 0.8.0 up to but excluding 0.9.0, the vLLM backend used with the /v1/chat/completions OpenAPI endpoint fails to validate unexpected or malformed input in the "pattern...

Exploit
  • EPSS 0.12%
  • Veröffentlicht 30.05.2025 17:36:16
  • Zuletzt bearbeitet 19.06.2025 00:55:27

vLLM, an inference and serving engine for large language models (LLMs), has a Regular Expression Denial of Service (ReDoS) vulnerability in the file `vllm/entrypoints/openai/tool_parsers/pythonic_tool_parser.py` of versions 0.6.4 up to but excluding ...

Exploit
  • EPSS 0.35%
  • Veröffentlicht 20.05.2025 17:32:27
  • Zuletzt bearbeitet 13.08.2025 16:35:57

vLLM, an inference and serving engine for large language models (LLMs), has an issue in versions 0.6.5 through 0.8.4 that ONLY impacts environments using the `PyNcclPipe` KV cache transfer integration with the V0 engine. No other configurations are a...

  • EPSS 0.48%
  • Veröffentlicht 06.05.2025 16:53:52
  • Zuletzt bearbeitet 31.07.2025 18:05:30

vLLM is an inference and serving engine for large language models. In a multi-node vLLM deployment using the V0 engine, vLLM uses ZeroMQ for some multi-node communication purposes. The secondary vLLM hosts open a `SUB` ZeroMQ socket and connect to an...