Skip to main content
Press slash or control plus K to focus the search. Use the arrow keys to navigate results and press enter to open a threat.
Reconnecting to live updates…

CVE-2025-62372: CWE-129: Improper Validation of Array Index in vllm-project vllm

0
High
VulnerabilityCVE-2025-62372cvecve-2025-62372cwe-129
Published: Fri Nov 21 2025 (11/21/2025, 01:22:37 UTC)
Source: CVE Database V5
Vendor/Project: vllm-project
Product: vllm

Description

vLLM is an inference and serving engine for large language models (LLMs). From version 0.5.5 to before 0.11.1, users can crash the vLLM engine serving multimodal models by passing multimodal embedding inputs with correct ndim but incorrect shape (e.g. hidden dimension is wrong), regardless of whether the model is intended to support such inputs (as defined in the Supported Models page). This issue has been patched in version 0.11.1.

AI-Powered Analysis

AILast updated: 11/21/2025, 01:59:56 UTC

Technical Analysis

CVE-2025-62372 is a vulnerability classified under CWE-129 (Improper Validation of Array Index) found in the vLLM project, an inference and serving engine for large language models (LLMs). The vulnerability exists in versions from 0.5.5 to before 0.11.1. It allows an attacker to crash the vLLM engine by sending multimodal embedding inputs that have the correct number of dimensions (ndim) but an incorrect shape, specifically an invalid hidden dimension size. This malformed input triggers improper array index validation within the engine, leading to a denial-of-service (DoS) condition by crashing the service. The vulnerability does not require authentication or user interaction, and the attack vector is network-based, making it remotely exploitable. The flaw affects the availability of the service but does not impact confidentiality or integrity. The vulnerability was publicly disclosed on November 21, 2025, with a CVSS v4.0 score of 8.3 (high severity), reflecting its ease of exploitation and significant impact on service availability. The issue has been addressed and patched in vLLM version 0.11.1. No known exploits have been reported in the wild as of the publication date. The vulnerability is particularly relevant for organizations deploying multimodal LLMs using vLLM, as the attack exploits input shapes that the model may not be designed to handle, regardless of the model's intended support for such inputs.

Potential Impact

For European organizations, the primary impact of CVE-2025-62372 is the potential for denial-of-service attacks against AI inference services running vLLM versions 0.5.5 to before 0.11.1. This can disrupt critical AI-driven applications, including natural language processing, multimodal data analysis, and other AI services that rely on vLLM for inference. Service outages could affect customer-facing applications, internal AI workflows, and research activities, leading to operational downtime and potential financial losses. Since the vulnerability does not affect confidentiality or integrity, data breaches are unlikely; however, the availability impact can degrade trust in AI services and delay business processes. Organizations heavily invested in AI and machine learning infrastructure, especially those offering AI-as-a-service or integrating LLMs into their products, face increased risk. The lack of required authentication lowers the barrier for attackers, increasing the likelihood of exploitation if vulnerable versions are exposed to untrusted networks. Additionally, the growing adoption of multimodal models in Europe heightens the relevance of this vulnerability.

Mitigation Recommendations

1. Immediate upgrade to vLLM version 0.11.1 or later, which contains the patch for this vulnerability. 2. Implement strict input validation on all inference endpoints to verify not only the number of dimensions but also the expected shape and size of embedding inputs before processing. 3. Deploy network-level protections such as Web Application Firewalls (WAFs) or API gateways to detect and block malformed or anomalous requests targeting inference services. 4. Apply rate limiting and anomaly detection to reduce the risk of denial-of-service attempts exploiting this vulnerability. 5. Isolate AI inference services within secure network segments and restrict access to trusted clients where possible. 6. Monitor logs and telemetry for unusual crashes or malformed input patterns indicative of exploitation attempts. 7. Engage in regular vulnerability scanning and penetration testing focused on AI infrastructure to identify similar issues proactively. 8. Educate development and operations teams about the importance of validating input shapes and dimensions in AI model serving environments.

Need more detailed analysis?Get Pro

Technical Details

Data Version
5.2
Assigner Short Name
GitHub_M
Date Reserved
2025-10-10T14:22:48.204Z
Cvss Version
4.0
State
PUBLISHED

Threat ID: 691fc3ff70da09562fa7fc94

Added to database: 11/21/2025, 1:44:31 AM

Last enriched: 11/21/2025, 1:59:56 AM

Last updated: 11/21/2025, 2:11:56 PM

Views: 12

Community Reviews

0 reviews

Crowdsource mitigation strategies, share intel context, and vote on the most helpful responses. Sign in to add your voice and help keep defenders ahead.

Sort by
Loading community insights…

Want to contribute mitigation steps or threat intel context? Sign in or create an account to join the community discussion.

Actions

PRO

Updates to AI analysis require Pro Console access. Upgrade inside Console → Billing.

Please log in to the Console to use AI analysis features.

Need enhanced features?

Contact root@offseq.com for Pro access with improved analysis and higher rate limits.

Latest Threats