Skip to main content
Press slash or control plus K to focus the search. Use the arrow keys to navigate results and press enter to open a threat.
Reconnecting to live updates…

CVE-2026-22807: CWE-94: Improper Control of Generation of Code ('Code Injection') in vllm-project vllm

0
High
VulnerabilityCVE-2026-22807cvecve-2026-22807cwe-94
Published: Wed Jan 21 2026 (01/21/2026, 21:13:11 UTC)
Source: CVE Database V5
Vendor/Project: vllm-project
Product: vllm

Description

CVE-2026-22807 is a high-severity code injection vulnerability in vLLM versions 0. 10. 1 up to but not including 0. 14. 0. The flaw arises because vLLM loads Hugging Face `auto_map` dynamic modules during model resolution without properly restricting execution of remote code, allowing attacker-controlled Python code to run at server startup. Exploitation requires the attacker to influence the model repository or path, which can be local or remote, enabling arbitrary code execution before any API request handling and without authentication. This vulnerability impacts confidentiality, integrity, and availability of the host system. The issue is fixed in vLLM version 0. 14.

AI-Powered Analysis

AILast updated: 01/21/2026, 21:50:17 UTC

Technical Analysis

CVE-2026-22807 is a critical code injection vulnerability classified under CWE-94 affecting the vLLM project, an inference and serving engine for large language models. The vulnerability exists in versions 0.10.1 through 0.13.x, where vLLM loads Hugging Face `auto_map` dynamic modules during model resolution without enforcing the `trust_remote_code` flag. This lack of gating allows an attacker who can control the model repository or path—either locally or remotely—to inject and execute arbitrary Python code on the host system at server startup. Notably, this code execution occurs before any API request handling and does not require any authentication or user interaction, significantly lowering the barrier for exploitation. The attack vector involves influencing the model source, which could be a malicious or compromised Hugging Face repository or a local directory under attacker control. The impact includes full compromise of the host system, enabling attackers to exfiltrate data, disrupt services, or pivot within the network. The vulnerability has a CVSS 3.1 score of 8.8, reflecting its high severity with network attack vector, low complexity, no privileges required, and high impact on confidentiality, integrity, and availability. The issue was publicly disclosed on January 21, 2026, and fixed in vLLM version 0.14.0 by enforcing proper controls on dynamic code loading. No known exploits in the wild have been reported yet, but the potential for damage is significant given the widespread use of vLLM in AI model serving.

Potential Impact

For European organizations, this vulnerability poses a significant risk to the security of AI infrastructure, particularly those deploying large language models using vLLM. Successful exploitation can lead to full system compromise, allowing attackers to execute arbitrary code, steal sensitive data, disrupt AI services, or use the compromised host as a foothold for further attacks. Given the increasing reliance on AI-driven applications in sectors such as finance, healthcare, and government, the impact could extend to critical services and intellectual property theft. The fact that exploitation does not require authentication or user interaction increases the threat level. Additionally, organizations using models from external or untrusted repositories are at higher risk. The vulnerability could also undermine trust in AI deployments and cause operational downtime, regulatory compliance issues, and reputational damage.

Mitigation Recommendations

The primary mitigation is to upgrade all vLLM deployments to version 0.14.0 or later, where the vulnerability is fixed by enforcing the `trust_remote_code` gating mechanism. Organizations should audit their AI model sources and restrict model loading to trusted and verified repositories only. Implement strict access controls and monitoring on directories used for model storage to prevent unauthorized modifications. Employ network segmentation to isolate AI serving infrastructure from general user networks. Use runtime application self-protection (RASP) or endpoint detection and response (EDR) solutions to detect anomalous code execution during model loading. Regularly review and update security policies related to AI model deployment and supply chain security. Finally, conduct penetration testing and code reviews focusing on dynamic code loading mechanisms in AI frameworks.

Need more detailed analysis?Upgrade to Pro Console

Technical Details

Data Version
5.2
Assigner Short Name
GitHub_M
Date Reserved
2026-01-09T22:50:10.288Z
Cvss Version
3.1
State
PUBLISHED

Threat ID: 697146bc4623b1157ced22d7

Added to database: 1/21/2026, 9:35:56 PM

Last enriched: 1/21/2026, 9:50:17 PM

Last updated: 1/21/2026, 10:54:55 PM

Views: 9

Community Reviews

0 reviews

Crowdsource mitigation strategies, share intel context, and vote on the most helpful responses. Sign in to add your voice and help keep defenders ahead.

Sort by
Loading community insights…

Want to contribute mitigation steps or threat intel context? Sign in or create an account to join the community discussion.

Actions

PRO

Updates to AI analysis require Pro Console access. Upgrade inside Console → Billing.

Please log in to the Console to use AI analysis features.

Need more coverage?

Upgrade to Pro Console in Console -> Billing for AI refresh and higher limits.

For incident response and remediation, OffSeq services can help resolve threats faster.

Latest Threats