Skip to main content
Press slash or control plus K to focus the search. Use the arrow keys to navigate results and press enter to open a threat.
Reconnecting to live updates…

CVE-2026-22807: CWE-94: Improper Control of Generation of Code ('Code Injection') in vllm-project vllm

0
High
VulnerabilityCVE-2026-22807cvecve-2026-22807cwe-94
Published: Wed Jan 21 2026 (01/21/2026, 21:13:11 UTC)
Source: CVE Database V5
Vendor/Project: vllm-project
Product: vllm

Description

CVE-2026-22807 is a high-severity code injection vulnerability in the vLLM inference engine for large language models, affecting versions 0. 10. 1 up to but not including 0. 14. 0. The flaw arises because vLLM loads Hugging Face auto_map dynamic modules during model resolution without verifying the trustworthiness of remote code, allowing attacker-controlled Python code execution at server startup. Exploitation requires the attacker to influence the model repository or path, either locally or remotely, and does not require API access or prior authentication. This vulnerability enables arbitrary code execution with full privileges on the host system before any request handling occurs. Although no known exploits are currently reported in the wild, the CVSS score of 8. 8 reflects the critical impact on confidentiality, integrity, and availability.

AI-Powered Analysis

AILast updated: 01/29/2026, 08:46:36 UTC

Technical Analysis

CVE-2026-22807 is a critical code injection vulnerability identified in the vLLM project, an inference and serving engine for large language models (LLMs). The vulnerability exists in versions 0.10.1 through 0.13.x, where vLLM loads Hugging Face's `auto_map` dynamic modules during the model resolution phase without enforcing the `trust_remote_code` flag. This lack of gating means that if an attacker can control the model repository or path—whether a local directory or a remote Hugging Face repository—they can inject malicious Python code that executes on the server at startup. This execution occurs before the server begins handling any API requests, and crucially, does not require any authentication or user interaction, significantly lowering the barrier to exploitation. The impact is severe, as arbitrary code execution on the host can lead to full system compromise, data theft, service disruption, or use of the host as a pivot point for further attacks. The vulnerability is classified under CWE-94 (Improper Control of Generation of Code) and has a CVSS v3.1 score of 8.8, indicating high severity with network attack vector, no privileges required, low attack complexity, and high impact on confidentiality, integrity, and availability. The issue was publicly disclosed on January 21, 2026, and fixed in vLLM version 0.14.0. No known exploits have been reported in the wild yet, but the potential for damage is significant given the widespread use of vLLM in AI model serving environments. Organizations using affected versions should prioritize upgrading to 0.14.0 or later and restrict model sources to trusted repositories to prevent injection of malicious code.

Potential Impact

For European organizations, the impact of CVE-2026-22807 can be substantial, particularly for those leveraging vLLM to serve large language models in production or research environments. Successful exploitation allows attackers to execute arbitrary code with the privileges of the vLLM process, potentially leading to full system compromise. This can result in unauthorized access to sensitive data, disruption of AI services, and lateral movement within corporate networks. Given the increasing reliance on AI and machine learning infrastructure in sectors such as finance, healthcare, telecommunications, and government, the compromise of vLLM hosts could lead to significant operational and reputational damage. Additionally, the vulnerability's exploitation does not require authentication or user interaction, increasing the risk of automated or opportunistic attacks. The ability to inject code at server startup means that even before any legitimate requests are processed, the attacker can establish persistence or backdoors. This threat is particularly critical for organizations integrating third-party or community models from Hugging Face repositories without strict validation, as these sources could be manipulated by adversaries. The lack of known exploits in the wild provides a window for proactive defense, but the high severity score underscores the urgency of mitigation.

Mitigation Recommendations

To mitigate CVE-2026-22807, organizations should immediately upgrade all vLLM instances to version 0.14.0 or later, where the vulnerability is patched. Until upgrades can be applied, restrict the model loading process to trusted, vetted local directories and repositories, avoiding dynamic loading of models from untrusted or external Hugging Face sources. Implement strict access controls and monitoring on the directories and repositories used for model storage to detect unauthorized changes. Employ network segmentation to isolate vLLM servers from critical infrastructure and limit exposure to external networks. Use application whitelisting and runtime application self-protection (RASP) tools to detect and prevent unauthorized code execution. Additionally, audit and validate all third-party models before deployment to ensure they do not contain malicious code. Incorporate security scanning tools that analyze model repositories for suspicious or unexpected Python code. Finally, maintain comprehensive logging and alerting on vLLM startup processes to quickly identify anomalous behavior indicative of exploitation attempts.

Need more detailed analysis?Upgrade to Pro Console

Technical Details

Data Version
5.2
Assigner Short Name
GitHub_M
Date Reserved
2026-01-09T22:50:10.288Z
Cvss Version
3.1
State
PUBLISHED

Threat ID: 697146bc4623b1157ced22d7

Added to database: 1/21/2026, 9:35:56 PM

Last enriched: 1/29/2026, 8:46:36 AM

Last updated: 2/7/2026, 6:39:10 AM

Views: 145

Community Reviews

0 reviews

Crowdsource mitigation strategies, share intel context, and vote on the most helpful responses. Sign in to add your voice and help keep defenders ahead.

Sort by
Loading community insights…

Want to contribute mitigation steps or threat intel context? Sign in or create an account to join the community discussion.

Actions

PRO

Updates to AI analysis require Pro Console access. Upgrade inside Console → Billing.

Please log in to the Console to use AI analysis features.

Need more coverage?

Upgrade to Pro Console in Console -> Billing for AI refresh and higher limits.

For incident response and remediation, OffSeq services can help resolve threats faster.

Latest Threats