CVE-2026-34159: CWE-119: Improper Restriction of Operations within the Bounds of a Memory Buffer in ggml-org llama.cpp
llama.cpp is an inference of several LLM models in C/C++. Prior to version b8492, the RPC backend's deserialize_tensor() skips all bounds validation when a tensor's buffer field is 0. An unauthenticated attacker can read and write arbitrary process memory via crafted GRAPH_COMPUTE messages. Combined with pointer leaks from ALLOC_BUFFER/BUFFER_GET_BASE, this gives full ASLR bypass and remote code execution. No authentication required, just TCP access to the RPC server port. This issue has been patched in version b8492.
AI Analysis
Technical Summary
CVE-2026-34159 is a severe memory corruption vulnerability classified under CWE-119 (Improper Restriction of Operations within the Bounds of a Memory Buffer) affecting the llama.cpp project by ggml-org, which implements inference for several large language models (LLMs) in C/C++. The vulnerability exists in the RPC backend's deserialize_tensor() function prior to version b8492, where bounds validation is entirely skipped if a tensor's buffer field is set to zero. This logic flaw permits an unauthenticated attacker to send specially crafted GRAPH_COMPUTE messages over TCP to the RPC server, enabling arbitrary read and write operations on the process memory. Furthermore, the attacker can leverage pointer leaks exposed via ALLOC_BUFFER and BUFFER_GET_BASE commands to bypass Address Space Layout Randomization (ASLR), a critical memory protection mechanism. This combination allows the attacker to execute arbitrary code remotely with no authentication or user interaction required. The vulnerability affects all versions before b8492 and is rated with a CVSS 3.1 score of 9.8 (critical), reflecting its high impact on confidentiality, integrity, and availability. Although no known exploits have been reported in the wild yet, the ease of exploitation and severity make it a significant threat. The issue has been addressed in version b8492 by adding proper bounds checking and validation in the deserialize_tensor() function.
Potential Impact
The vulnerability enables unauthenticated remote attackers to execute arbitrary code on systems running vulnerable versions of llama.cpp, potentially leading to full system compromise. This can result in unauthorized data access, data manipulation, service disruption, and lateral movement within affected networks. Since llama.cpp is used for LLM inference, exploitation could allow attackers to manipulate AI model outputs or exfiltrate sensitive data processed by these models. The lack of authentication and requirement of only TCP access to the RPC port significantly broadens the attack surface, especially in environments where the RPC service is exposed to untrusted networks. Organizations relying on llama.cpp for AI workloads may face severe operational and reputational damage if exploited. The ability to bypass ASLR further increases the reliability of exploitation, making mitigation urgent.
Mitigation Recommendations
1. Immediately upgrade all deployments of llama.cpp to version b8492 or later, which contains the patch for this vulnerability. 2. Restrict network access to the RPC server port by implementing strict firewall rules and network segmentation to limit exposure only to trusted hosts. 3. Employ network-level authentication or VPNs to protect RPC endpoints from unauthorized access. 4. Monitor network traffic for unusual GRAPH_COMPUTE messages or unexpected RPC activity that could indicate exploitation attempts. 5. Conduct regular code audits and fuzz testing on RPC deserialization routines to detect similar memory safety issues proactively. 6. If upgrading is not immediately possible, consider disabling or restricting the RPC backend functionality temporarily to reduce risk. 7. Implement runtime protections such as AddressSanitizer or Control Flow Integrity (CFI) where feasible to detect exploitation attempts. 8. Maintain up-to-date intrusion detection and prevention systems tuned to detect anomalous memory manipulation patterns.
Affected Countries
United States, China, Germany, United Kingdom, Japan, South Korea, France, Canada, Australia, India
CVE-2026-34159: CWE-119: Improper Restriction of Operations within the Bounds of a Memory Buffer in ggml-org llama.cpp
Description
llama.cpp is an inference of several LLM models in C/C++. Prior to version b8492, the RPC backend's deserialize_tensor() skips all bounds validation when a tensor's buffer field is 0. An unauthenticated attacker can read and write arbitrary process memory via crafted GRAPH_COMPUTE messages. Combined with pointer leaks from ALLOC_BUFFER/BUFFER_GET_BASE, this gives full ASLR bypass and remote code execution. No authentication required, just TCP access to the RPC server port. This issue has been patched in version b8492.
AI-Powered Analysis
Machine-generated threat intelligence
Technical Analysis
CVE-2026-34159 is a severe memory corruption vulnerability classified under CWE-119 (Improper Restriction of Operations within the Bounds of a Memory Buffer) affecting the llama.cpp project by ggml-org, which implements inference for several large language models (LLMs) in C/C++. The vulnerability exists in the RPC backend's deserialize_tensor() function prior to version b8492, where bounds validation is entirely skipped if a tensor's buffer field is set to zero. This logic flaw permits an unauthenticated attacker to send specially crafted GRAPH_COMPUTE messages over TCP to the RPC server, enabling arbitrary read and write operations on the process memory. Furthermore, the attacker can leverage pointer leaks exposed via ALLOC_BUFFER and BUFFER_GET_BASE commands to bypass Address Space Layout Randomization (ASLR), a critical memory protection mechanism. This combination allows the attacker to execute arbitrary code remotely with no authentication or user interaction required. The vulnerability affects all versions before b8492 and is rated with a CVSS 3.1 score of 9.8 (critical), reflecting its high impact on confidentiality, integrity, and availability. Although no known exploits have been reported in the wild yet, the ease of exploitation and severity make it a significant threat. The issue has been addressed in version b8492 by adding proper bounds checking and validation in the deserialize_tensor() function.
Potential Impact
The vulnerability enables unauthenticated remote attackers to execute arbitrary code on systems running vulnerable versions of llama.cpp, potentially leading to full system compromise. This can result in unauthorized data access, data manipulation, service disruption, and lateral movement within affected networks. Since llama.cpp is used for LLM inference, exploitation could allow attackers to manipulate AI model outputs or exfiltrate sensitive data processed by these models. The lack of authentication and requirement of only TCP access to the RPC port significantly broadens the attack surface, especially in environments where the RPC service is exposed to untrusted networks. Organizations relying on llama.cpp for AI workloads may face severe operational and reputational damage if exploited. The ability to bypass ASLR further increases the reliability of exploitation, making mitigation urgent.
Mitigation Recommendations
1. Immediately upgrade all deployments of llama.cpp to version b8492 or later, which contains the patch for this vulnerability. 2. Restrict network access to the RPC server port by implementing strict firewall rules and network segmentation to limit exposure only to trusted hosts. 3. Employ network-level authentication or VPNs to protect RPC endpoints from unauthorized access. 4. Monitor network traffic for unusual GRAPH_COMPUTE messages or unexpected RPC activity that could indicate exploitation attempts. 5. Conduct regular code audits and fuzz testing on RPC deserialization routines to detect similar memory safety issues proactively. 6. If upgrading is not immediately possible, consider disabling or restricting the RPC backend functionality temporarily to reduce risk. 7. Implement runtime protections such as AddressSanitizer or Control Flow Integrity (CFI) where feasible to detect exploitation attempts. 8. Maintain up-to-date intrusion detection and prevention systems tuned to detect anomalous memory manipulation patterns.
Technical Details
- Data Version
- 5.2
- Assigner Short Name
- GitHub_M
- Date Reserved
- 2026-03-25T20:12:04.197Z
- Cvss Version
- 3.1
- State
- PUBLISHED
Threat ID: 69cd7224e6bfc5ba1dee83e5
Added to database: 4/1/2026, 7:29:40 PM
Last enriched: 4/1/2026, 7:48:06 PM
Last updated: 4/4/2026, 6:55:28 AM
Views: 37
Community Reviews
0 reviewsCrowdsource mitigation strategies, share intel context, and vote on the most helpful responses. Sign in to add your voice and help keep defenders ahead.
Want to contribute mitigation steps or threat intel context? Sign in or create an account to join the community discussion.
Actions
Updates to AI analysis require Pro Console access. Upgrade inside Console → Billing.
Need more coverage?
Upgrade to Pro Console for AI refresh and higher limits.
For incident response and remediation, OffSeq services can help resolve threats faster.
Latest Threats
Check if your credentials are on the dark web
Instant breach scanning across billions of leaked records. Free tier available.