Skip to main content

CVE-2025-8747: CWE-502 Deserialization of Untrusted Data in Google Keras

High
VulnerabilityCVE-2025-8747cvecve-2025-8747cwe-502
Published: Mon Aug 11 2025 (08/11/2025, 07:21:16 UTC)
Source: CVE Database V5
Vendor/Project: Google
Product: Keras

Description

A safe mode bypass vulnerability in the `Model.load_model` method in Keras versions 3.0.0 through 3.10.0 allows an attacker to achieve arbitrary code execution by convincing a user to load a specially crafted `.keras` model archive.

AI-Powered Analysis

AILast updated: 08/11/2025, 07:47:57 UTC

Technical Analysis

CVE-2025-8747 is a high-severity vulnerability affecting Google Keras versions 3.0.0 through 3.10.0, specifically in the `Model.load_model` method. This vulnerability is categorized under CWE-502, which involves deserialization of untrusted data. The core issue is a safe mode bypass that allows an attacker to execute arbitrary code by tricking a user into loading a maliciously crafted `.keras` model archive. The deserialization process in Keras's model loading functionality does not adequately validate or sanitize the contents of the model archive, enabling exploitation. When a user loads such a crafted model, the deserialization process can execute embedded malicious payloads, leading to arbitrary code execution within the context of the user running the Keras environment. The CVSS 4.0 base score of 8.6 reflects the high impact and exploitability of this vulnerability, with a vector indicating local attack with low complexity, no attacker privileges required, but user interaction is necessary. The vulnerability impacts confidentiality, integrity, and availability, as arbitrary code execution can lead to data theft, manipulation, or system compromise. No known exploits are currently reported in the wild, and no official patches have been linked yet, indicating that mitigation relies on cautious handling of model files and potential workarounds until a fix is released.

Potential Impact

For European organizations, this vulnerability poses significant risks, especially for entities relying on machine learning workflows that incorporate Keras for model development, deployment, or inference. Sectors such as finance, healthcare, automotive, and research institutions frequently use Keras for AI and ML applications. An attacker exploiting this vulnerability could execute arbitrary code on systems running vulnerable Keras versions, potentially leading to data breaches, intellectual property theft, disruption of AI services, or lateral movement within networks. Given the widespread adoption of Python and Keras in European academia and industry, the threat could affect both production environments and development pipelines. The requirement for user interaction (loading a malicious model) suggests that social engineering or supply chain attacks (e.g., tampered model files shared via collaboration platforms) are likely attack vectors. The impact is heightened in environments where model files are shared across teams or downloaded from untrusted sources without verification. Additionally, the compromise of AI models can undermine trust in automated decision-making systems, which is critical in regulated sectors prevalent in Europe.

Mitigation Recommendations

European organizations should implement several specific mitigations: 1) Enforce strict validation and integrity checks on all `.keras` model files before loading, including cryptographic signatures or hashes from trusted sources. 2) Restrict the use of `Model.load_model` to only trusted and verified model files, avoiding loading models from untrusted or unknown origins. 3) Employ sandboxing or containerization techniques when loading models to limit the potential impact of arbitrary code execution. 4) Monitor and audit model loading activities and related system calls to detect anomalous behavior indicative of exploitation attempts. 5) Educate data scientists and developers about the risks of loading untrusted models and encourage secure collaboration practices. 6) Track updates from Google and apply patches promptly once available. 7) Consider alternative model serialization formats or loading mechanisms that do not rely on vulnerable deserialization processes. 8) Implement network segmentation and least privilege principles to contain potential breaches resulting from exploitation.

Need more detailed analysis?Get Pro

Technical Details

Data Version
5.1
Assigner Short Name
Google
Date Reserved
2025-08-08T09:37:17.811Z
Cvss Version
4.0
State
PUBLISHED

Threat ID: 68999c95ad5a09ad00224b5d

Added to database: 8/11/2025, 7:32:37 AM

Last enriched: 8/11/2025, 7:47:57 AM

Last updated: 8/11/2025, 9:53:01 AM

Views: 4

Actions

PRO

Updates to AI analysis are available only with a Pro account. Contact root@offseq.com for access.

Please log in to the Console to use AI analysis features.

Need enhanced features?

Contact root@offseq.com for Pro access with improved analysis and higher rate limits.

Latest Threats