CVE-2025-46570

vLLM’s Chunk-Based Prefix Caching Vulnerable to Potential Timing Side-Channel

Description

vLLM is an inference and serving engine for large language models (LLMs). Prior to version 0.9.0, when a new prompt is processed, if the PageAttention mechanism finds a matching prefix chunk, the prefill process speeds up, which is reflected in the TTFT (Time to First Token). These timing differences caused by matching chunks are significant enough to be recognized and exploited. This issue has been patched in version 0.9.0.

Categories

2.6
CVSS
Severity: Low
CVSS 3.1 •
EPSS 0.03%
Vendor Advisory github.com Vendor Advisory github.com
Affected: vllm-project vllm
Published at:
Updated at:

References

Frequently Asked Questions

What is the severity of CVE-2025-46570?
CVE-2025-46570 has been scored as a low severity vulnerability.
How to fix CVE-2025-46570?
To fix CVE-2025-46570, make sure you are using an up-to-date version of the affected component(s) by checking the vendor release notes. As for now, there are no other specific guidelines available.
Is CVE-2025-46570 being actively exploited in the wild?
As for now, there are no information to confirm that CVE-2025-46570 is being actively exploited. According to its EPSS score, there is a ~0% probability that this vulnerability will be exploited by malicious actors in the next 30 days.
What software or system is affected by CVE-2025-46570?
CVE-2025-46570 affects vllm-project vllm.
This platform uses data from the NIST NVD, MITRE CVE, MITRE CWE, First.org and CISA KEV but is not endorsed or certified by these entities. CVE is a registred trademark of the MITRE Corporation and the authoritative source of CVE content is MITRE's CVE web site. CWE is a registred trademark of the MITRE Corporation and the authoritative source of CWE content is MITRE's CWE web site.
© 2025 Under My Watch. All Rights Reserved.