Lucene search

K
githubGitHub Advisory DatabaseGHSA-WC36-9694-F9RF
HistorySep 17, 2024 - 6:33 p.m.

vLLM Denial of Service via the best_of parameter

2024-09-1718:33:26
CWE-400
GitHub Advisory Database
github.com
7
vllm dos vulnerability
ilab model serve
json web api
best_of parameter
resource exhaustion
system unresponsive

CVSS3

6.2

Attack Vector

LOCAL

Attack Complexity

LOW

Privileges Required

NONE

User Interaction

NONE

Scope

UNCHANGED

Confidentiality Impact

NONE

Integrity Impact

NONE

Availability Impact

HIGH

CVSS:3.1/AV:L/AC:L/PR:N/UI:N/S:U/C:N/I:N/A:H

AI Score

6.8

Confidence

High

EPSS

0

Percentile

9.6%

A vulnerability was found in the ilab model serve component, where improper handling of the best_of parameter in the vllm JSON web API can lead to a Denial of Service (DoS). The API used for LLM-based sentence or chat completion accepts a best_of parameter to return the best completion from several options. When this parameter is set to a large value, the API does not handle timeouts or resource exhaustion properly, allowing an attacker to cause a DoS by consuming excessive system resources. This leads to the API becoming unresponsive, preventing legitimate users from accessing the service.

Affected configurations

Vulners
Node
vllmRange0.5.0.post1
VendorProductVersionCPE
*vllm*cpe:2.3:a:*:vllm:*:*:*:*:*:*:*:*

CVSS3

6.2

Attack Vector

LOCAL

Attack Complexity

LOW

Privileges Required

NONE

User Interaction

NONE

Scope

UNCHANGED

Confidentiality Impact

NONE

Integrity Impact

NONE

Availability Impact

HIGH

CVSS:3.1/AV:L/AC:L/PR:N/UI:N/S:U/C:N/I:N/A:H

AI Score

6.8

Confidence

High

EPSS

0

Percentile

9.6%

Related for GHSA-WC36-9694-F9RF