Skip to main content
Press slash or control plus K to focus the search. Use the arrow keys to navigate results and press enter to open a threat.
Reconnecting to live updates…

CVE-2026-22773: CWE-770: Allocation of Resources Without Limits or Throttling in vllm-project vllm

0
Medium
VulnerabilityCVE-2026-22773cvecve-2026-22773cwe-770
Published: Sat Jan 10 2026 (01/10/2026, 06:39:02 UTC)
Source: CVE Database V5
Vendor/Project: vllm-project
Product: vllm

Description

vLLM is an inference and serving engine for large language models (LLMs). In versions from 0.6.4 to before 0.12.0, users can crash the vLLM engine serving multimodal models that use the Idefics3 vision model implementation by sending a specially crafted 1x1 pixel image. This causes a tensor dimension mismatch that results in an unhandled runtime error, leading to complete server termination. This issue has been patched in version 0.12.0.

AI-Powered Analysis

AILast updated: 01/10/2026, 07:09:27 UTC

Technical Analysis

CVE-2026-22773 is a vulnerability classified under CWE-770 (Allocation of Resources Without Limits or Throttling) affecting the vLLM project, an inference and serving engine for large language models. Specifically, versions from 0.6.4 up to but not including 0.12.0 are vulnerable. The flaw arises when the engine processes multimodal models that incorporate the Idefics3 vision model implementation. An attacker can send a specially crafted 1x1 pixel image that triggers a tensor dimension mismatch during inference. This mismatch leads to an unhandled runtime error, causing the vLLM server process to crash and terminate unexpectedly. The vulnerability does not require user interaction but does require low-level privileges to send the crafted input. The impact is a denial of service (DoS) due to server termination, affecting availability but not confidentiality or integrity. The CVSS v3.1 score is 6.5, reflecting medium severity with network attack vector, low attack complexity, and no user interaction needed. No known exploits are currently reported in the wild. The issue was publicly disclosed and patched in vLLM version 0.12.0, which includes proper input validation and resource management to prevent tensor dimension mismatches and unhandled errors.

Potential Impact

For European organizations deploying vLLM for AI inference, particularly those using multimodal models with vision components, this vulnerability poses a risk of denial of service. An attacker could remotely crash inference servers, causing service outages and disruption of AI-driven applications such as automated customer support, image recognition, or data analysis. This could lead to operational downtime, loss of productivity, and potential reputational damage. While the vulnerability does not expose sensitive data or allow unauthorized data modification, the availability impact can be significant in environments relying heavily on AI services. Organizations in sectors like finance, healthcare, automotive, and research institutions that integrate large language models with vision capabilities are especially vulnerable. The lack of known exploits reduces immediate risk, but the ease of exploitation and network accessibility means attackers could weaponize this flaw if discovered. The medium severity rating suggests prioritizing patching to maintain service continuity and prevent potential denial of service attacks.

Mitigation Recommendations

To mitigate CVE-2026-22773, European organizations should immediately upgrade all vLLM deployments to version 0.12.0 or later, where the vulnerability is patched. In addition to upgrading, organizations should implement strict input validation and sanitization for all data fed into AI inference engines, especially for multimodal inputs involving images. Deploy rate limiting and resource throttling mechanisms to prevent resource exhaustion from malformed or malicious inputs. Monitoring and alerting on unexpected server crashes or abnormal inference errors can help detect exploitation attempts early. Isolate AI inference services in segmented network zones to limit exposure to untrusted inputs. Conduct regular security assessments and code reviews of AI serving infrastructure to identify similar resource allocation or input handling issues. Finally, maintain an inventory of AI model components and their versions to ensure timely patch management and vulnerability tracking.

Need more detailed analysis?Upgrade to Pro Console

Technical Details

Data Version
5.2
Assigner Short Name
GitHub_M
Date Reserved
2026-01-09T18:27:19.387Z
Cvss Version
3.1
State
PUBLISHED

Threat ID: 6961f7c0c540fa4b5456583e

Added to database: 1/10/2026, 6:54:56 AM

Last enriched: 1/10/2026, 7:09:27 AM

Last updated: 1/10/2026, 10:30:10 PM

Views: 17

Community Reviews

0 reviews

Crowdsource mitigation strategies, share intel context, and vote on the most helpful responses. Sign in to add your voice and help keep defenders ahead.

Sort by
Loading community insights…

Want to contribute mitigation steps or threat intel context? Sign in or create an account to join the community discussion.

Actions

PRO

Updates to AI analysis require Pro Console access. Upgrade inside Console → Billing.

Please log in to the Console to use AI analysis features.

Need more coverage?

Upgrade to Pro Console in Console -> Billing for AI refresh and higher limits.

For incident response and remediation, OffSeq services can help resolve threats faster.

Latest Threats