Skip to main content
Press slash or control plus K to focus the search. Use the arrow keys to navigate results and press enter to open a threat.
Reconnecting to live updates…

CVE-2026-2069: Stack-based Buffer Overflow in ggml-org llama.cpp

0
Medium
VulnerabilityCVE-2026-2069cvecve-2026-2069
Published: Fri Feb 06 2026 (02/06/2026, 22:02:08 UTC)
Source: CVE Database V5
Vendor/Project: ggml-org
Product: llama.cpp

Description

A flaw has been found in ggml-org llama.cpp up to 55abc39. Impacted is the function llama_grammar_advance_stack of the file llama.cpp/src/llama-grammar.cpp of the component GBNF Grammar Handler. This manipulation causes stack-based buffer overflow. The attack needs to be launched locally. The exploit has been published and may be used. Patch name: 18993. To fix this issue, it is recommended to deploy a patch.

AI-Powered Analysis

Machine-generated threat intelligence

AILast updated: 02/23/2026, 22:11:01 UTC

Technical Analysis

CVE-2026-2069 identifies a stack-based buffer overflow vulnerability in the open-source ggml-org llama.cpp project, specifically within the llama_grammar_advance_stack function of the GBNF Grammar Handler component. This vulnerability arises from improper handling of stack memory during grammar advancement operations, leading to a buffer overflow condition. The affected code version is up to commit 55abc39. The flaw requires local access to exploit, meaning an attacker must have some level of local system access with limited privileges to trigger the overflow. No user interaction or elevated privileges are necessary, but the attack vector is limited to local execution. The vulnerability could allow an attacker to corrupt memory on the stack, potentially causing application crashes or enabling code execution under certain conditions. Although no known exploits are currently observed in the wild, proof-of-concept exploits have been published, increasing the risk of future exploitation. The vendor has released a patch (identified as patch 18993) to fix the issue by correcting the buffer handling logic. The CVSS 4.0 base score is 4.8, reflecting a medium severity due to the local attack vector and limited scope of impact. The vulnerability affects systems using llama.cpp, a lightweight C++ implementation of Facebook's LLaMA language model, commonly used in AI research and development environments.

Potential Impact

The primary impact of CVE-2026-2069 is the potential for local attackers to cause denial of service through application crashes or, in more severe cases, execute arbitrary code by exploiting the stack-based buffer overflow. This could compromise the integrity and availability of systems running vulnerable versions of llama.cpp. Since llama.cpp is often used in AI and machine learning research environments, exploitation could disrupt AI model training or inference processes, leading to operational downtime or corrupted outputs. The confidentiality impact is limited as the vulnerability does not directly expose sensitive data. However, if exploited for code execution, attackers could potentially escalate privileges or move laterally within the local environment. The requirement for local access limits the attack surface, but insider threats or compromised user accounts could leverage this vulnerability. Organizations relying on llama.cpp for AI workloads, especially in research labs, universities, and companies developing AI applications, face operational risks if the vulnerability is not patched promptly.

Mitigation Recommendations

To mitigate CVE-2026-2069, organizations should immediately apply the vendor-supplied patch (patch 18993) that corrects the buffer handling in the llama_grammar_advance_stack function. Beyond patching, restrict local access to systems running llama.cpp to trusted users only, employing strict access controls and monitoring for unusual local activity. Implement application whitelisting and endpoint detection and response (EDR) solutions to detect anomalous behavior indicative of exploitation attempts. Regularly audit and update all AI-related software dependencies to ensure vulnerabilities are addressed promptly. For environments where patching is delayed, consider running llama.cpp processes with the least privileges possible and within containerized or sandboxed environments to limit potential damage from exploitation. Additionally, maintain comprehensive logging and alerting on local process crashes or memory corruption events to enable rapid incident response.

Pro Console: star threats, build custom feeds, automate alerts via Slack, email & webhooks.Upgrade to Pro

Technical Details

Data Version
5.2
Assigner Short Name
VulDB
Date Reserved
2026-02-06T07:41:07.150Z
Cvss Version
4.0
State
PUBLISHED

Threat ID: 698667edf9fa50a62f37ce3d

Added to database: 2/6/2026, 10:15:09 PM

Last enriched: 2/23/2026, 10:11:01 PM

Last updated: 3/23/2026, 9:14:29 PM

Views: 520

Community Reviews

0 reviews

Crowdsource mitigation strategies, share intel context, and vote on the most helpful responses. Sign in to add your voice and help keep defenders ahead.

Sort by
Loading community insights…

Want to contribute mitigation steps or threat intel context? Sign in or create an account to join the community discussion.

Actions

PRO

Updates to AI analysis require Pro Console access. Upgrade inside Console → Billing.

Please log in to the Console to use AI analysis features.

Need more coverage?

Upgrade to Pro Console for AI refresh and higher limits.

For incident response and remediation, OffSeq services can help resolve threats faster.

Latest Threats

Breach by OffSeqOFFSEQFRIENDS — 25% OFF

Check if your credentials are on the dark web

Instant breach scanning across billions of leaked records. Free tier available.

Scan now
OffSeq TrainingCredly Certified

Lead Pen Test Professional

Technical5-day eLearningPECB Accredited
View courses