CVE-2026-24779: CWE-918: Server-Side Request Forgery (SSRF) in vllm-project vllm
CVE-2026-24779 is a Server-Side Request Forgery (SSRF) vulnerability in the vLLM inference engine for large language models, affecting versions prior to 0. 14. 1. The flaw exists in the MediaConnector class's URL loading methods, where inconsistent parsing of backslashes by different Python libraries allows bypassing hostname restrictions. This enables attackers to coerce the vLLM server into making arbitrary requests to internal network resources. The vulnerability is especially critical in containerized environments like llm-d, where compromised pods can scan internal networks, interact with other pods, or cause denial of service by sending malicious requests to management endpoints. The CVSS score is 7. 1 (high severity), reflecting network attack vector, low complexity, and partial confidentiality impact without user interaction. No known exploits are reported yet, and version 0. 14.
AI Analysis
Technical Summary
The vulnerability CVE-2026-24779 affects the vLLM project, an inference and serving engine for large language models, specifically versions prior to 0.14.1. The issue lies in the MediaConnector class, which provides methods load_from_url and load_from_url_async to fetch media from user-supplied URLs. These methods use two different Python parsing libraries to enforce hostname restrictions. However, these libraries interpret backslashes differently, allowing attackers to craft URLs that bypass hostname restrictions and cause the server to make arbitrary HTTP requests to internal network resources. This SSRF vulnerability enables attackers to interact with internal services that are otherwise inaccessible externally. In containerized environments such as llm-d, where vLLM runs in pods, exploitation could allow an attacker to scan internal networks, communicate with other pods, or send malicious requests to internal management endpoints. For example, sending false metrics to the llm-d management interface could destabilize the system or cause denial of service. The vulnerability has a CVSS 3.1 score of 7.1, indicating high severity due to network attack vector, low attack complexity, and high confidentiality impact, though integrity and availability impacts are limited. No authentication or user interaction is required, but some privileges on the vLLM instance are necessary (PR:L). The vulnerability was published on January 27, 2026, and fixed in version 0.14.1. No known exploits have been reported in the wild to date.
Potential Impact
For European organizations deploying vLLM versions prior to 0.14.1, especially in containerized AI inference environments, this SSRF vulnerability poses significant risks. Attackers could leverage the flaw to access internal network resources that are not exposed externally, potentially extracting sensitive data or reconnaissance information. In multi-tenant or microservices architectures common in European cloud and enterprise environments, compromised vLLM pods could interact with other pods or services, leading to lateral movement or privilege escalation. The ability to send malicious requests to internal management endpoints could disrupt AI service availability, causing denial of service or system instability. Confidentiality is the primary concern, as internal data or metadata could be exposed. Given the increasing adoption of AI and container orchestration platforms in Europe, this vulnerability could impact sectors such as finance, healthcare, and government, where sensitive data and AI workloads coexist. The lack of known exploits reduces immediate risk but does not eliminate the threat, especially as attackers often target AI infrastructure. Organizations failing to patch or isolate vulnerable vLLM instances risk operational disruption and data breaches.
Mitigation Recommendations
European organizations should immediately upgrade all vLLM deployments to version 0.14.1 or later to apply the official patch. For environments where immediate upgrade is not feasible, implement strict network segmentation and firewall rules to restrict outbound requests from vLLM pods to only trusted endpoints. Employ egress filtering at the container or orchestration platform level to prevent unauthorized internal network scanning or access. Monitor network traffic from vLLM instances for unusual or unexpected requests, especially to internal management endpoints. Review and harden internal management interfaces to require strong authentication and limit exposure. Conduct regular security audits of containerized AI workloads and apply runtime security tools to detect anomalous behavior. Additionally, validate and sanitize all user-supplied URLs before processing to prevent malformed inputs exploiting parsing inconsistencies. Finally, maintain an incident response plan tailored to AI infrastructure compromise scenarios.
Affected Countries
Germany, France, United Kingdom, Netherlands, Sweden, Finland, Denmark, Belgium, Italy, Spain
CVE-2026-24779: CWE-918: Server-Side Request Forgery (SSRF) in vllm-project vllm
Description
CVE-2026-24779 is a Server-Side Request Forgery (SSRF) vulnerability in the vLLM inference engine for large language models, affecting versions prior to 0. 14. 1. The flaw exists in the MediaConnector class's URL loading methods, where inconsistent parsing of backslashes by different Python libraries allows bypassing hostname restrictions. This enables attackers to coerce the vLLM server into making arbitrary requests to internal network resources. The vulnerability is especially critical in containerized environments like llm-d, where compromised pods can scan internal networks, interact with other pods, or cause denial of service by sending malicious requests to management endpoints. The CVSS score is 7. 1 (high severity), reflecting network attack vector, low complexity, and partial confidentiality impact without user interaction. No known exploits are reported yet, and version 0. 14.
AI-Powered Analysis
Technical Analysis
The vulnerability CVE-2026-24779 affects the vLLM project, an inference and serving engine for large language models, specifically versions prior to 0.14.1. The issue lies in the MediaConnector class, which provides methods load_from_url and load_from_url_async to fetch media from user-supplied URLs. These methods use two different Python parsing libraries to enforce hostname restrictions. However, these libraries interpret backslashes differently, allowing attackers to craft URLs that bypass hostname restrictions and cause the server to make arbitrary HTTP requests to internal network resources. This SSRF vulnerability enables attackers to interact with internal services that are otherwise inaccessible externally. In containerized environments such as llm-d, where vLLM runs in pods, exploitation could allow an attacker to scan internal networks, communicate with other pods, or send malicious requests to internal management endpoints. For example, sending false metrics to the llm-d management interface could destabilize the system or cause denial of service. The vulnerability has a CVSS 3.1 score of 7.1, indicating high severity due to network attack vector, low attack complexity, and high confidentiality impact, though integrity and availability impacts are limited. No authentication or user interaction is required, but some privileges on the vLLM instance are necessary (PR:L). The vulnerability was published on January 27, 2026, and fixed in version 0.14.1. No known exploits have been reported in the wild to date.
Potential Impact
For European organizations deploying vLLM versions prior to 0.14.1, especially in containerized AI inference environments, this SSRF vulnerability poses significant risks. Attackers could leverage the flaw to access internal network resources that are not exposed externally, potentially extracting sensitive data or reconnaissance information. In multi-tenant or microservices architectures common in European cloud and enterprise environments, compromised vLLM pods could interact with other pods or services, leading to lateral movement or privilege escalation. The ability to send malicious requests to internal management endpoints could disrupt AI service availability, causing denial of service or system instability. Confidentiality is the primary concern, as internal data or metadata could be exposed. Given the increasing adoption of AI and container orchestration platforms in Europe, this vulnerability could impact sectors such as finance, healthcare, and government, where sensitive data and AI workloads coexist. The lack of known exploits reduces immediate risk but does not eliminate the threat, especially as attackers often target AI infrastructure. Organizations failing to patch or isolate vulnerable vLLM instances risk operational disruption and data breaches.
Mitigation Recommendations
European organizations should immediately upgrade all vLLM deployments to version 0.14.1 or later to apply the official patch. For environments where immediate upgrade is not feasible, implement strict network segmentation and firewall rules to restrict outbound requests from vLLM pods to only trusted endpoints. Employ egress filtering at the container or orchestration platform level to prevent unauthorized internal network scanning or access. Monitor network traffic from vLLM instances for unusual or unexpected requests, especially to internal management endpoints. Review and harden internal management interfaces to require strong authentication and limit exposure. Conduct regular security audits of containerized AI workloads and apply runtime security tools to detect anomalous behavior. Additionally, validate and sanitize all user-supplied URLs before processing to prevent malformed inputs exploiting parsing inconsistencies. Finally, maintain an incident response plan tailored to AI infrastructure compromise scenarios.
Technical Details
- Data Version
- 5.2
- Assigner Short Name
- GitHub_M
- Date Reserved
- 2026-01-26T21:06:47.869Z
- Cvss Version
- 3.1
- State
- PUBLISHED
Threat ID: 697936c84623b1157c4a64f3
Added to database: 1/27/2026, 10:06:00 PM
Last enriched: 1/27/2026, 10:20:34 PM
Last updated: 1/27/2026, 11:28:31 PM
Views: 3
Community Reviews
0 reviewsCrowdsource mitigation strategies, share intel context, and vote on the most helpful responses. Sign in to add your voice and help keep defenders ahead.
Want to contribute mitigation steps or threat intel context? Sign in or create an account to join the community discussion.
Related Threats
CVE-2026-24783: CWE-682: Incorrect Calculation in script3 soroban-fixed-point-math
HighCVE-2026-24770: CWE-22: Improper Limitation of a Pathname to a Restricted Directory ('Path Traversal') in infiniflow ragflow
CriticalCVE-2024-1545: CWE-1256: Improper Restriction of Software Interfaces to Hardware Features in WolfSSL wolfCrypt
MediumCVE-2026-24778: CWE-79: Improper Neutralization of Input During Web Page Generation ('Cross-site Scripting') in TryGhost Ghost
HighCVE-2026-24765: CWE-502: Deserialization of Untrusted Data in sebastianbergmann phpunit
HighActions
Updates to AI analysis require Pro Console access. Upgrade inside Console → Billing.
Need more coverage?
Upgrade to Pro Console in Console -> Billing for AI refresh and higher limits.
For incident response and remediation, OffSeq services can help resolve threats faster.