๐Ÿ” CVE Alert

CVE-2026-33298

HIGH 7.8

llama.cpp has a Heap Buffer Overflow via Integer Overflow in GGUF Tensor Parsing

CVSS Score
7.8
EPSS Score
0.0%
EPSS Percentile
12th

llama.cpp is an inference of several LLM models in C/C++. Prior to b7824, an integer overflow vulnerability in the `ggml_nbytes` function allows an attacker to bypass memory validation by crafting a GGUF file with specific tensor dimensions. This causes `ggml_nbytes` to return a significantly smaller size than required (e.g., 4MB instead of Exabytes), leading to a heap-based buffer overflow when the application subsequently processes the tensor. This vulnerability allows potential Remote Code Execution (RCE) via memory corruption. b7824 contains a fix.

CWE CWE-122 CWE-190
Vendor ggml-org
Product llama.cpp
Published Mar 24, 2026
Last Updated Mar 25, 2026
Stay Ahead of the Next One

Get instant alerts for ggml-org llama.cpp

Be the first to know when new high vulnerabilities affecting ggml-org llama.cpp are published โ€” delivered to Slack, Telegram or Discord.

Get Free Alerts โ†’ Free ยท No credit card ยท 60 sec setup

CVSS v3 Breakdown

CVSS:3.1/AV:L/AC:L/PR:N/UI:R/S:U/C:H/I:H/A:H
Attack Vector
Local
Attack Complexity
Low
Privileges Required
None
User Interaction
Required
Scope
Unchanged
Confidentiality
High
Integrity
High
Availability
High

Affected Versions

ggml-org / llama.cpp
< b7824

References

NVD โ†— CVE.org โ†— EPSS Data โ†—
github.com: https://github.com/ggml-org/llama.cpp/security/advisories/GHSA-96jg-mvhq-q7q7 github.com: https://github.com/ggml-org/llama.cpp/releases/tag/b7824