Skip to main content
Press slash or control plus K to focus the search. Use the arrow keys to navigate results and press enter to open a threat.
Reconnecting to live updates…

CVE-2024-8768: Reachable Assertion

0
High
VulnerabilityCVE-2024-8768cvecve-2024-8768
Published: Tue Sep 17 2024 (09/17/2024, 16:20:42 UTC)
Source: CVE Database V5

Description

CVE-2024-8768 is a high-severity vulnerability in the vLLM library where sending a completions API request with an empty prompt causes the API server to crash, resulting in a denial of service. This flaw does not affect confidentiality or integrity but severely impacts availability. The vulnerability requires no authentication or user interaction and can be exploited remotely over the network. There are no known exploits in the wild yet, and no patches have been published as of now. European organizations using the vLLM library in production environments, especially those providing AI-based services or APIs, are at risk of service disruption. Mitigation involves input validation to prevent empty prompts and implementing robust error handling to avoid server crashes. Countries with significant AI technology adoption and cloud service usage, such as Germany, France, the UK, and the Netherlands, are more likely to be affected. Due to the ease of exploitation and high impact on availability, this vulnerability is rated as high severity.

AI-Powered Analysis

AILast updated: 11/20/2025, 07:49:11 UTC

Technical Analysis

CVE-2024-8768 is a vulnerability identified in the vLLM library, a tool used for language model serving and completions. The flaw arises when the completions API receives a request with an empty prompt string. Instead of handling this input gracefully, the server triggers a reachable assertion failure, causing the vLLM API server process to crash. This results in a denial of service (DoS) condition, where legitimate users cannot access the service until it is restarted or recovered. The vulnerability has a CVSS 3.1 base score of 7.5, reflecting its high severity due to network exploitability (AV:N), low attack complexity (AC:L), no privileges required (PR:N), no user interaction (UI:N), and impact limited to availability (A:H) without affecting confidentiality or integrity. The vulnerability affects version 0 of the vLLM library, with no patches currently available and no known exploits in the wild. The root cause is insufficient input validation and error handling for empty prompt inputs, which should be addressed by adding checks to reject or safely handle such requests. Since the vLLM library is used in AI-driven applications and services, this vulnerability can disrupt operations relying on language model completions, impacting service availability and user experience.

Potential Impact

For European organizations, the primary impact is service disruption due to denial of service attacks exploiting this vulnerability. Organizations providing AI-based services, chatbots, or any applications leveraging the vLLM library for language model completions may experience downtime or degraded service availability. This can lead to operational interruptions, loss of customer trust, and potential financial losses, especially for businesses relying on real-time AI interactions. Since the vulnerability does not affect confidentiality or integrity, data breaches or manipulation are not direct concerns. However, prolonged outages could indirectly affect business continuity and reputation. The ease of exploitation without authentication means attackers can launch DoS attacks remotely, increasing the risk of widespread disruption. European sectors with high AI adoption, such as finance, healthcare, and technology, could be particularly vulnerable if they integrate vLLM-based services into their infrastructure.

Mitigation Recommendations

To mitigate this vulnerability, organizations should implement strict input validation on the completions API to reject or safely handle empty prompt requests before they reach the vLLM server. Developers should update the vLLM library or apply patches once available from the maintainers. In the interim, deploying web application firewalls (WAFs) or API gateways that filter out malformed or empty prompt requests can reduce exposure. Monitoring and alerting on server crashes or unusual API request patterns can help detect exploitation attempts early. Additionally, implementing redundancy and failover mechanisms for the vLLM service can minimize downtime in case of crashes. Organizations should also conduct code reviews and fuzz testing on AI service inputs to identify similar vulnerabilities proactively. Finally, maintaining up-to-date backups and incident response plans will aid in rapid recovery if a denial of service occurs.

Need more detailed analysis?Get Pro

Technical Details

Data Version
5.2
Assigner Short Name
redhat
Date Reserved
2024-09-12T21:29:58.462Z
Cvss Version
3.1
State
PUBLISHED

Threat ID: 691ec609337afffbc0f72899

Added to database: 11/20/2025, 7:40:57 AM

Last enriched: 11/20/2025, 7:49:11 AM

Last updated: 11/20/2025, 8:42:12 AM

Views: 3

Community Reviews

0 reviews

Crowdsource mitigation strategies, share intel context, and vote on the most helpful responses. Sign in to add your voice and help keep defenders ahead.

Sort by
Loading community insights…

Want to contribute mitigation steps or threat intel context? Sign in or create an account to join the community discussion.

Actions

PRO

Updates to AI analysis require Pro Console access. Upgrade inside Console → Billing.

Please log in to the Console to use AI analysis features.

Need enhanced features?

Contact root@offseq.com for Pro access with improved analysis and higher rate limits.

Latest Threats