Skip to main content
Press slash or control plus K to focus the search. Use the arrow keys to navigate results and press enter to open a threat.
Reconnecting to live updates…

Researchers Find Serious AI Bugs Exposing Meta, Nvidia, and Microsoft Inference Frameworks

0
Critical
Vulnerabilityremotercepython
Published: Fri Nov 14 2025 (11/14/2025, 15:20:00 UTC)
Source: The Hacker News

Description

Cybersecurity researchers have uncovered critical remote code execution vulnerabilities impacting major artificial intelligence (AI) inference engines, including those from Meta, Nvidia, Microsoft, and open-source PyTorch projects such as vLLM and SGLang. "These vulnerabilities all traced back to the same root cause: the overlooked unsafe use of ZeroMQ (ZMQ) and Python's pickle deserialization,"

AI-Powered Analysis

AILast updated: 11/14/2025, 23:36:50 UTC

Technical Analysis

Researchers have identified critical remote code execution (RCE) vulnerabilities affecting several prominent AI inference engines, including Meta's Llama framework, Nvidia's TensorRT-LLM, Microsoft's Sarathi-Serve, and open-source PyTorch-based projects like vLLM and SGLang. The vulnerabilities stem from an unsafe deserialization pattern involving Python's pickle module used in conjunction with ZeroMQ (ZMQ) sockets exposed over the network without authentication. Specifically, the use of ZeroMQ's recv_pyobj() method deserializes incoming data using pickle, which is inherently unsafe when processing untrusted input. This pattern, termed ShadowMQ, originated in Meta's Llama framework (CVE-2024-50050) and propagated through code reuse and copy-pasting across multiple projects. Some vulnerabilities have been assigned CVEs with high CVSS scores (6.3 to 8.8), while others remain unpatched or partially fixed. Exploitation allows attackers to execute arbitrary code remotely on AI inference nodes, potentially leading to privilege escalation, model theft, and deployment of malicious payloads such as cryptocurrency miners. The widespread reuse of vulnerable code amplifies the risk, as multiple AI frameworks share the same insecure deserialization logic. The issue highlights the dangers of rapid AI development and code sharing without adequate security review. Additionally, related threats include JavaScript injection attacks in AI-powered IDEs, emphasizing the broader attack surface in AI ecosystems. The vulnerabilities underscore the need for secure deserialization practices, network authentication, and rigorous code audits in AI infrastructure.

Potential Impact

For European organizations, the impact of these vulnerabilities is significant due to the increasing adoption of AI inference frameworks in research, cloud services, and enterprise AI deployments. Successful exploitation could lead to full compromise of AI inference nodes, enabling attackers to execute arbitrary code, steal proprietary AI models, and disrupt AI services. This could result in intellectual property loss, service downtime, and reputational damage. In cloud environments, compromised nodes could serve as footholds for lateral movement, escalating the risk to broader IT infrastructure. Financially motivated attackers might deploy cryptocurrency miners or ransomware payloads, increasing operational costs and recovery efforts. The vulnerabilities also threaten AI supply chain integrity, as compromised inference engines could propagate malicious code or corrupted models. Given Europe's strong AI research hubs and cloud adoption, the threat could affect sectors such as finance, healthcare, manufacturing, and government services relying on AI. The lack of authentication on ZMQ sockets exacerbates the risk, especially in multi-tenant or hybrid cloud environments common in Europe. Overall, the vulnerabilities pose a critical risk to confidentiality, integrity, and availability of AI-powered systems in Europe.

Mitigation Recommendations

European organizations should immediately audit their AI inference deployments for use of vulnerable frameworks and unsafe deserialization patterns involving ZeroMQ and Python pickle. Specific mitigations include: 1) Applying all available patches and updates from Meta, Nvidia, Microsoft, and open-source projects like vLLM and SGLang. 2) Disabling or restricting network exposure of ZeroMQ sockets, enforcing strong authentication and encryption (e.g., CURVE/ZAP mechanisms) to prevent unauthorized access. 3) Replacing Python pickle deserialization with safer serialization methods such as JSON or protobuf where feasible. 4) Conducting thorough code reviews to identify and eliminate copy-pasted insecure deserialization logic across AI projects. 5) Implementing network segmentation and strict access controls around AI inference clusters to limit attack surface. 6) Monitoring AI infrastructure for anomalous activity indicative of exploitation attempts, including unexpected process executions or network connections. 7) Educating developers and DevOps teams on secure coding practices related to deserialization and third-party library usage. 8) For AI-powered IDEs and related tools, disable auto-run features, vet extensions carefully, and audit any third-party integrations like MCP servers. These targeted actions go beyond generic advice and address the root causes and propagation vectors of the vulnerabilities.

Need more detailed analysis?Get Pro

Technical Details

Article Source
{"url":"https://thehackernews.com/2025/11/researchers-find-serious-ai-bugs.html","fetched":true,"fetchedAt":"2025-11-14T23:36:25.770Z","wordCount":1365}

Threat ID: 6917bcfbed594783724528cf

Added to database: 11/14/2025, 11:36:27 PM

Last enriched: 11/14/2025, 11:36:50 PM

Last updated: 11/16/2025, 4:09:14 AM

Views: 34

Community Reviews

0 reviews

Crowdsource mitigation strategies, share intel context, and vote on the most helpful responses. Sign in to add your voice and help keep defenders ahead.

Sort by
Loading community insights…

Want to contribute mitigation steps or threat intel context? Sign in or create an account to join the community discussion.

Actions

PRO

Updates to AI analysis require Pro Console access. Upgrade inside Console → Billing.

Please log in to the Console to use AI analysis features.

Need enhanced features?

Contact root@offseq.com for Pro access with improved analysis and higher rate limits.

Latest Threats