Skip to main content
Press slash or control plus K to focus the search. Use the arrow keys to navigate results and press enter to open a threat.
Reconnecting to live updates…

CVE-2026-24779: CWE-918: Server-Side Request Forgery (SSRF) in vllm-project vllm

0
High
VulnerabilityCVE-2026-24779cvecve-2026-24779cwe-918
Published: Tue Jan 27 2026 (01/27/2026, 22:01:13 UTC)
Source: CVE Database V5
Vendor/Project: vllm-project
Product: vllm

Description

CVE-2026-24779 is a Server-Side Request Forgery (SSRF) vulnerability in the vLLM inference engine for large language models, affecting versions prior to 0. 14. 1. The flaw exists in the MediaConnector class's load_from_url and load_from_url_async methods, where inconsistent parsing of backslashes by different Python libraries allows bypassing hostname restrictions. This enables attackers to coerce the server into making arbitrary internal network requests, potentially targeting internal services or other containerized pods. The vulnerability is especially critical in containerized environments like llm-d, where exploitation could lead to internal network scanning, denial of service, or unauthorized access to sensitive data. The CVSS v3. 1 score is 7. 1 (high severity), reflecting network attack vector, low complexity, and high confidentiality impact. No known exploits are reported in the wild yet.

AI-Powered Analysis

AILast updated: 02/04/2026, 09:15:37 UTC

Technical Analysis

CVE-2026-24779 is a high-severity Server-Side Request Forgery (SSRF) vulnerability identified in the vLLM project, an inference and serving engine for large language models. The vulnerability resides in the MediaConnector class, specifically in the load_from_url and load_from_url_async methods that fetch and process media from user-supplied URLs. These methods attempt to restrict requests to allowed hosts by parsing URLs using two different Python libraries, each interpreting backslashes differently. This discrepancy allows attackers to craft URLs that bypass hostname restrictions, coercing the vLLM server into making arbitrary HTTP requests to internal network resources. In containerized environments such as llm-d, where vLLM pods operate, this can be exploited to scan internal networks, interact with other pods, or send malicious requests to internal management endpoints. For example, attackers might send falsified metrics to destabilize the system or cause denial of service. The vulnerability requires low privileges (PR:L) but no user interaction, and the attack vector is network-based, making exploitation feasible remotely. The impact primarily affects confidentiality by exposing internal resources, with limited availability impact. The vendor patched the vulnerability in version 0.14.1 by addressing the inconsistent URL parsing and enforcing proper hostname restrictions. No public exploits have been reported yet, but the nature of SSRF in containerized AI serving environments makes this a significant risk.

Potential Impact

For European organizations deploying vLLM versions prior to 0.14.1, especially within containerized or cloud-native AI infrastructures, this SSRF vulnerability poses a substantial risk. Attackers could leverage it to pivot within internal networks, accessing sensitive internal services or data not intended to be exposed externally. This could lead to unauthorized data disclosure, internal reconnaissance, or denial of service conditions affecting AI service availability. Given the increasing adoption of AI inference engines in sectors like finance, healthcare, and critical infrastructure across Europe, exploitation could disrupt essential services or compromise confidential information. Container orchestration platforms commonly used in Europe, such as Kubernetes, often run pods with network segmentation assumptions that SSRF can undermine. The vulnerability's ability to bypass hostname restrictions means traditional network controls might be insufficient. Although no known exploits exist yet, the high severity and ease of exploitation warrant immediate attention to prevent potential lateral movement and internal attacks within European enterprise environments.

Mitigation Recommendations

European organizations should immediately upgrade all vLLM deployments to version 0.14.1 or later to apply the official patch. Until upgrades are completed, implement strict network egress filtering at the container and host levels to restrict outbound HTTP requests from vLLM pods to only trusted external endpoints. Employ runtime security tools to monitor and alert on unusual outbound network activity from AI inference containers. Review and harden internal service authentication and authorization to mitigate risks if SSRF is exploited to reach internal endpoints. Consider deploying web application firewalls (WAFs) or API gateways that can detect and block SSRF patterns in incoming requests. Conduct thorough security assessments of containerized AI environments, focusing on inter-pod communication and internal API exposure. Finally, establish incident response procedures to quickly isolate affected pods and networks if suspicious activity is detected.

Need more detailed analysis?Upgrade to Pro Console

Technical Details

Data Version
5.2
Assigner Short Name
GitHub_M
Date Reserved
2026-01-26T21:06:47.869Z
Cvss Version
3.1
State
PUBLISHED

Threat ID: 697936c84623b1157c4a64f3

Added to database: 1/27/2026, 10:06:00 PM

Last enriched: 2/4/2026, 9:15:37 AM

Last updated: 2/7/2026, 4:28:21 AM

Views: 177

Community Reviews

0 reviews

Crowdsource mitigation strategies, share intel context, and vote on the most helpful responses. Sign in to add your voice and help keep defenders ahead.

Sort by
Loading community insights…

Want to contribute mitigation steps or threat intel context? Sign in or create an account to join the community discussion.

Actions

PRO

Updates to AI analysis require Pro Console access. Upgrade inside Console → Billing.

Please log in to the Console to use AI analysis features.

Need more coverage?

Upgrade to Pro Console in Console -> Billing for AI refresh and higher limits.

For incident response and remediation, OffSeq services can help resolve threats faster.

Latest Threats