Lucene search

K
Vllm-projectVllm

16 matches found

CVE
CVE
added 2025/02/07 8:15 p.m.200 views

CVE-2025-25183

vLLM is a high-throughput and memory-efficient inference and serving engine for LLMs. Maliciously constructed statements can lead to hash collisions, resulting in cache reuse, which can interfere with subsequent responses and cause unintended behavior. Prefix caching makes use of Python's built-in ...

2.6CVSS6.8AI score0.00028EPSS
CVE
CVE
added 2025/01/27 6:15 p.m.198 views

CVE-2025-24357

vLLM is a library for LLM inference and serving. vllm/model_executor/weight_utils.py implements hf_model_weights_iterator to load the model checkpoint, which is downloaded from huggingface. It uses the torch.load function and the weights_only parameter defaults to False. When torch.load loads malic...

8.8CVSS7.7AI score0.00079EPSS
CVE
CVE
added 2025/03/19 4:15 p.m.197 views

CVE-2025-29770

vLLM is a high-throughput and memory-efficient inference and serving engine for LLMs. The outlines library is one of the backends used by vLLM to support structured output (a.k.a. guided decoding). Outlines provides an optional cache for its compiled grammars on the local filesystem. This cache has...

6.5CVSS6.4AI score0.00195EPSS
CVE
CVE
added 2025/03/19 4:15 p.m.196 views

CVE-2025-29783

vLLM is a high-throughput and memory-efficient inference and serving engine for LLMs. When vLLM is configured to use Mooncake, unsafe deserialization exposed directly over ZMQ/TCP on all network interfaces will allow attackers to execute remote code on distributed hosts. This is a remote code execu...

9CVSS9.8AI score0.00745EPSS
CVE
CVE
added 2025/04/30 1:15 a.m.152 views

CVE-2025-32444

vLLM is a high-throughput and memory-efficient inference and serving engine for LLMs. Versions starting from 0.6.5 and prior to 0.8.5, having vLLM integration with mooncake, are vulnerable to remote code execution due to using pickle based serialization over unsecured ZeroMQ sockets. The vulnerable...

10CVSS9.8AI score0.0077EPSS
CVE
CVE
added 2025/05/20 6:15 p.m.148 views

CVE-2025-47277

vLLM, an inference and serving engine for large language models (LLMs), has an issue in versions 0.6.5 through 0.8.4 that ONLY impacts environments using the PyNcclPipe KV cache transfer integration with the V0 engine. No other configurations are affected. vLLM supports the use of the PyNcclPipe cl...

9.8CVSS9.5AI score0.00067EPSS
CVE
CVE
added 2025/04/30 1:15 a.m.140 views

CVE-2025-46560

vLLM is a high-throughput and memory-efficient inference and serving engine for LLMs. Versions starting from 0.8.0 and prior to 0.8.5 are affected by a critical performance vulnerability in the input preprocessing logic of the multimodal tokenizer. The code dynamically replaces placeholder tokens (...

7.5CVSS6.6AI score0.00062EPSS
CVE
CVE
added 2025/04/30 1:15 a.m.132 views

CVE-2025-30202

vLLM is a high-throughput and memory-efficient inference and serving engine for LLMs. Versions starting from 0.5.2 and prior to 0.8.5 are vulnerable to denial of service and data exposure via ZeroMQ on multi-node vLLM deployment. In a multi-node vLLM deployment, vLLM uses ZeroMQ for some multi-node...

7.5CVSS7.7AI score0.00071EPSS
CVE
CVE
added 2025/05/30 7:15 p.m.102 views

CVE-2025-48943

vLLM is an inference and serving engine for large language models (LLMs). Version 0.8.0 up to but excluding 0.9.0 have a Denial of Service (ReDoS) that causes the vLLM server to crash if an invalid regex was provided while using structured output. This vulnerability is similar to GHSA-6qc9-v4r8-22x...

6.5CVSS7AI score0.00052EPSS
CVE
CVE
added 2025/03/20 10:15 a.m.100 views

CVE-2024-9053

vllm-project vllm version 0.6.0 contains a vulnerability in the AsyncEngineRPCServer() RPC server entrypoints. The core functionality run_server_loop() calls the function _make_handler_coro(), which directly uses cloudpickle.loads() on received messages without any sanitization. This can result in ...

9.8CVSS9.7AI score0.00641EPSS
CVE
CVE
added 2025/05/06 5:16 p.m.95 views

CVE-2025-30165

vLLM is an inference and serving engine for large language models. In a multi-node vLLM deployment using the V0 engine, vLLM uses ZeroMQ for some multi-node communication purposes. The secondary vLLM hosts open a SUB ZeroMQ socket and connect to an XPUB socket on the primary vLLM host. When data is...

8CVSS8.2AI score0.00277EPSS
CVE
CVE
added 2025/05/29 5:15 p.m.92 views

CVE-2025-46722

vLLM is an inference and serving engine for large language models (LLMs). In versions starting from 0.7.0 to before 0.9.0, in the file vllm/multimodal/hasher.py, the MultiModalHasher class has a security and data integrity issue in its image hashing method. Currently, it serializes PIL.Image.Image ...

7.3CVSS4.6AI score0.00093EPSS
CVE
CVE
added 2025/05/29 5:15 p.m.91 views

CVE-2025-46570

vLLM is an inference and serving engine for large language models (LLMs). Prior to version 0.9.0, when a new prompt is processed, if the PageAttention mechanism finds a matching prefix chunk, the prefill process speeds up, which is reflected in the TTFT (Time to First Token). These timing differenc...

2.6CVSS3.6AI score0.00031EPSS
CVE
CVE
added 2025/05/30 7:15 p.m.90 views

CVE-2025-48942

vLLM is an inference and serving engine for large language models (LLMs). In versions 0.8.0 up to but excluding 0.9.0, hitting the /v1/completions API with a invalid json_schema as a Guided Param kills the vllm server. This vulnerability is similar GHSA-9hcf-v7m4-6m2j/CVE-2025-48943, but for regex ...

6.5CVSS6.9AI score0.00052EPSS
CVE
CVE
added 2025/05/30 7:15 p.m.90 views

CVE-2025-48944

vLLM is an inference and serving engine for large language models (LLMs). In version 0.8.0 up to but excluding 0.9.0, the vLLM backend used with the /v1/chat/completions OpenAPI endpoint fails to validate unexpected or malformed input in the "pattern" and "type" fields when the tools functionality ...

6.5CVSS7AI score0.00066EPSS
CVE
CVE
added 2025/05/30 6:15 p.m.89 views

CVE-2025-48887

vLLM, an inference and serving engine for large language models (LLMs), has a Regular Expression Denial of Service (ReDoS) vulnerability in the file vllm/entrypoints/openai/tool_parsers/pythonic_tool_parser.py of versions 0.6.4 up to but excluding 0.9.0. The root cause is the use of a highly comple...

6.5CVSS6.9AI score0.00047EPSS