Skip to main content
Press slash or control plus K to focus the search. Use the arrow keys to navigate results and press enter to open a threat.
Reconnecting to live updates…

CVE-2024-9053: CWE-502 Deserialization of Untrusted Data in vllm-project vllm-project/vllm

0
Critical
VulnerabilityCVE-2024-9053cvecve-2024-9053cwe-502
Published: Thu Mar 20 2025 (03/20/2025, 10:09:33 UTC)
Source: CVE Database V5
Vendor/Project: vllm-project
Product: vllm-project/vllm

Description

vllm-project vllm version 0.6.0 contains a vulnerability in the AsyncEngineRPCServer() RPC server entrypoints. The core functionality run_server_loop() calls the function _make_handler_coro(), which directly uses cloudpickle.loads() on received messages without any sanitization. This can result in remote code execution by deserializing malicious pickle data.

AI-Powered Analysis

AILast updated: 10/15/2025, 13:08:14 UTC

Technical Analysis

CVE-2024-9053 is a critical vulnerability identified in the vllm-project's vllm software, version 0.6.0. The vulnerability stems from the unsafe deserialization of untrusted data within the AsyncEngineRPCServer() RPC server entrypoints. Specifically, the run_server_loop() function calls _make_handler_coro(), which uses cloudpickle.loads() directly on incoming messages without any sanitization or validation. Cloudpickle is a Python library that extends the standard pickle module to serialize and deserialize Python objects, but it is inherently unsafe when handling untrusted input because it can execute arbitrary code during deserialization. An attacker can exploit this by sending maliciously crafted pickle data to the RPC server, leading to remote code execution (RCE) on the server hosting the vllm instance. This vulnerability does not require any authentication or user interaction, making it highly exploitable remotely over the network. The CVSS v3.0 score of 9.8 reflects the critical nature of this flaw, with attack vector network (AV:N), low attack complexity (AC:L), no privileges required (PR:N), no user interaction (UI:N), and high impact on confidentiality (C:H), integrity (I:H), and availability (A:H). No patches or fixes have been linked yet, and no known exploits are reported in the wild, but the vulnerability's nature suggests it could be weaponized quickly. The vulnerability is categorized under CWE-502 (Deserialization of Untrusted Data), a well-known class of security issues that frequently lead to RCE in Python applications that use pickle or cloudpickle without proper safeguards.

Potential Impact

For European organizations, the impact of CVE-2024-9053 can be severe. vllm is used in AI and machine learning workloads, often processing sensitive data or running critical inference services. Exploitation could lead to full system compromise, data theft, service disruption, or use of compromised systems as a foothold for lateral movement within networks. Confidentiality breaches could expose proprietary AI models or sensitive input data, while integrity violations could allow attackers to manipulate AI outputs or training data, undermining trust in AI-driven decisions. Availability impacts could disrupt AI services critical to business operations. Given the critical CVSS score and the lack of required authentication, attackers can remotely exploit this vulnerability at scale. European organizations in sectors such as finance, healthcare, manufacturing, and research that leverage AI technologies are particularly at risk. The absence of known exploits currently provides a window for proactive mitigation, but the threat landscape could evolve rapidly.

Mitigation Recommendations

1. Immediate mitigation should focus on restricting network access to the vllm RPC server, limiting exposure to trusted internal networks only. 2. Implement strict input validation and filtering to prevent untrusted or malformed pickle data from reaching the deserialization function. 3. Replace or modify the use of cloudpickle.loads() with safer deserialization alternatives or sandboxed environments that limit code execution capabilities. 4. Monitor network traffic and logs for anomalous or unexpected RPC calls that could indicate exploitation attempts. 5. Apply any forthcoming patches or updates from the vllm-project as soon as they become available. 6. Consider deploying runtime application self-protection (RASP) or endpoint detection and response (EDR) tools that can detect and block suspicious deserialization activities. 7. Educate developers and security teams about the risks of unsafe deserialization and enforce secure coding practices in AI/ML projects. 8. If patching is delayed, consider isolating vllm instances in containerized or virtualized environments with minimal privileges to limit potential damage.

Need more detailed analysis?Get Pro

Technical Details

Data Version
5.1
Assigner Short Name
@huntr_ai
Date Reserved
2024-09-20T18:43:46.911Z
Cvss Version
3.0
State
PUBLISHED

Threat ID: 68ef9b2e178f764e1f470ea1

Added to database: 10/15/2025, 1:01:34 PM

Last enriched: 10/15/2025, 1:08:14 PM

Last updated: 10/15/2025, 6:03:22 PM

Views: 3

Community Reviews

0 reviews

Crowdsource mitigation strategies, share intel context, and vote on the most helpful responses. Sign in to add your voice and help keep defenders ahead.

Sort by
Loading community insights…

Want to contribute mitigation steps or threat intel context? Sign in or create an account to join the community discussion.

Actions

PRO

Updates to AI analysis require Pro Console access. Upgrade inside Console → Billing.

Please log in to the Console to use AI analysis features.

Need enhanced features?

Contact root@offseq.com for Pro access with improved analysis and higher rate limits.

Latest Threats