Skip to main content
Press slash or control plus K to focus the search. Use the arrow keys to navigate results and press enter to open a threat.
Reconnecting to live updates…

CVE-2024-8768: Reachable Assertion

0
High
VulnerabilityCVE-2024-8768cvecve-2024-8768
Published: Tue Sep 17 2024 (09/17/2024, 16:20:42 UTC)
Source: CVE Database V5

Description

CVE-2024-8768 is a high-severity vulnerability in the vLLM library where sending a completions API request with an empty prompt causes the API server to crash, resulting in a denial of service. The flaw requires no authentication or user interaction and can be exploited remotely over the network. Although it does not impact confidentiality or integrity, the availability impact is significant as it disrupts service. No known exploits are currently reported in the wild. European organizations using vLLM in production environments could face service outages if targeted. Mitigation involves input validation to reject empty prompts and applying any vendor patches once available. Countries with strong AI and machine learning sectors, such as Germany, France, and the UK, are more likely to be affected due to higher adoption of such libraries. The vulnerability has a CVSS score of 7. 5, reflecting its high severity due to ease of exploitation and impact on availability.

AI-Powered Analysis

AILast updated: 11/27/2025, 08:12:33 UTC

Technical Analysis

CVE-2024-8768 is a reachable assertion vulnerability identified in the vLLM library, which is used to provide language model completions via an API. The vulnerability occurs when a completions API request is made with an empty prompt string. This triggers an assertion failure within the server code, causing the vLLM API server to crash and become unavailable. The flaw requires no privileges or user interaction, and can be exploited remotely by sending a crafted API request over the network. The impact is a denial of service (DoS) condition, where the service becomes unresponsive until restarted. The vulnerability does not affect confidentiality or integrity, as it does not allow data leakage or modification. The CVSS 3.1 base score is 7.5 (high), with vector AV:N/AC:L/PR:N/UI:N/S:U/C:N/I:N/A:H, indicating network attack vector, low attack complexity, no privileges or user interaction needed, unchanged scope, no confidentiality or integrity impact, and high availability impact. No patches or known exploits are currently reported, but the vulnerability is publicly disclosed and could be weaponized. The root cause is insufficient input validation on the prompt parameter, allowing an empty string to reach an assertion that is not designed to handle it. This type of flaw is critical in API services that require high availability, especially in production environments relying on vLLM for AI-driven applications.

Potential Impact

For European organizations, the primary impact is service disruption due to denial of service attacks exploiting this vulnerability. Organizations using vLLM for AI completions in customer-facing or internal applications could experience downtime, affecting business continuity and user experience. This could lead to operational delays, loss of productivity, and potential reputational damage if services are unavailable. While no data breach or integrity compromise is involved, the availability impact can be severe for sectors relying on real-time AI services such as finance, healthcare, and telecommunications. Additionally, attackers could use this vulnerability as a vector to cause cascading failures in complex AI-driven workflows. The lack of authentication or user interaction requirements means attackers can easily trigger the DoS remotely, increasing the threat surface. European entities with critical AI infrastructure should prioritize addressing this vulnerability to maintain service reliability.

Mitigation Recommendations

Immediate mitigation should focus on implementing input validation to reject empty prompt strings before processing API requests. Organizations should audit their vLLM API usage to ensure that all client inputs are sanitized and validated. Monitoring API logs for anomalous requests with empty prompts can help detect exploitation attempts. Since no official patches are currently available, consider deploying rate limiting or API request filtering to reduce the risk of automated exploitation. Once vendor patches or updates are released, apply them promptly. Additionally, implement robust service monitoring and automated restart mechanisms to minimize downtime in case of crashes. For critical deployments, consider isolating the vLLM service in containerized environments with resource limits to contain potential crashes. Finally, maintain incident response plans that include DoS scenarios targeting AI services.

Need more detailed analysis?Upgrade to Pro Console

Technical Details

Data Version
5.2
Assigner Short Name
redhat
Date Reserved
2024-09-12T21:29:58.462Z
Cvss Version
3.1
State
PUBLISHED

Threat ID: 691ec609337afffbc0f72899

Added to database: 11/20/2025, 7:40:57 AM

Last enriched: 11/27/2025, 8:12:33 AM

Last updated: 1/7/2026, 5:20:35 AM

Views: 71

Community Reviews

0 reviews

Crowdsource mitigation strategies, share intel context, and vote on the most helpful responses. Sign in to add your voice and help keep defenders ahead.

Sort by
Loading community insights…

Want to contribute mitigation steps or threat intel context? Sign in or create an account to join the community discussion.

Actions

PRO

Updates to AI analysis require Pro Console access. Upgrade inside Console → Billing.

Please log in to the Console to use AI analysis features.

Need more coverage?

Upgrade to Pro Console in Console -> Billing for AI refresh and higher limits.

For incident response and remediation, OffSeq services can help resolve threats faster.

Latest Threats