CVE-2025-47277

vLLM, an inference and serving engine for large language models (LLMs), has an issue in versions 0.6.5 through 0.8.4 that ONLY impacts environments using the `PyNcclPipe` KV cache transfer integration with the V0 engine. No other configurations are affected. vLLM supports the use of the `PyNcclPipe` class to establish a peer-to-peer communication domain for data transmission between distributed nodes. The GPU-side KV-Cache transmission is implemented through the `PyNcclCommunicator` class, while CPU-side control message passing is handled via the `send_obj` and `recv_obj` methods on the CPU side.​ The intention was that this interface should only be exposed to a private network using the IP address specified by the `--kv-ip` CLI parameter. The vLLM documentation covers how this must be limited to a secured network. The default and intentional behavior from PyTorch is that the `TCPStore` interface listens on ALL interfaces, regardless of what IP address is provided. The IP address given was only used as a client-side address to use. vLLM was fixed to use a workaround to force the `TCPStore` instance to bind its socket to a specified private interface. As of version 0.8.5, vLLM limits the `TCPStore` socket to the private interface as configured.
Configurations

Configuration 1 (hide)

cpe:2.3:a:vllm:vllm:*:*:*:*:*:*:*:*

History

13 Aug 2025, 16:35

Type Values Removed Values Added
First Time Vllm vllm
Vllm
CPE cpe:2.3:a:vllm:vllm:*:*:*:*:*:*:*:*
References () https://docs.vllm.ai/en/latest/deployment/security.html - () https://docs.vllm.ai/en/latest/deployment/security.html - Technical Description
References () https://github.com/vllm-project/vllm/commit/0d6e187e88874c39cda7409cf673f9e6546893e7 - () https://github.com/vllm-project/vllm/commit/0d6e187e88874c39cda7409cf673f9e6546893e7 - Patch
References () https://github.com/vllm-project/vllm/pull/15988 - () https://github.com/vllm-project/vllm/pull/15988 - Issue Tracking, Patch
References () https://github.com/vllm-project/vllm/security/advisories/GHSA-hjq4-87xh-g4fv - () https://github.com/vllm-project/vllm/security/advisories/GHSA-hjq4-87xh-g4fv - Exploit, Vendor Advisory
Summary
  • (es) vLLM, un motor de inferencia y servicio para modelos de lenguaje grandes (LLM), presenta un problema en las versiones 0.6.5 a 0.8.4 que SOLO afecta a entornos que utilizan la integración de transferencia de caché KV `PyNcclPipe` con el motor V0. Ninguna otra configuración se ve afectada. vLLM admite el uso de la clase `PyNcclPipe` para establecer un dominio de comunicación punto a punto para la transmisión de datos entre nodos distribuidos. La transmisión de caché KV del lado de la GPU se implementa mediante la clase `PyNcclCommunicator`, mientras que el paso de mensajes de control del lado de la CPU se gestiona mediante los métodos `send_obj` y `recv_obj` en el lado de la CPU. El objetivo era que esta interfaz solo se expusiera a una red privada utilizando la dirección IP especificada por el parámetro de CLI `--kv-ip`. La documentación de vLLM explica cómo esto debe limitarse a una red segura. El comportamiento predeterminado e intencional de PyTorch es que la interfaz `TCPStore` escucha en TODAS las interfaces, independientemente de la dirección IP proporcionada. La dirección IP proporcionada solo se usaba como dirección del cliente. vLLM se corrigió para usar una solución alternativa que obligaba a la instancia `TCPStore` a vincular su socket a una interfaz privada específica. A partir de la versión 0.8.5, vLLM limita el socket `TCPStore` a la interfaz privada configurada.

20 May 2025, 18:15

Type Values Removed Values Added
New CVE

Information

Published : 2025-05-20 18:15

Updated : 2025-08-13 16:35


NVD link : CVE-2025-47277

Mitre link : CVE-2025-47277

CVE.ORG link : CVE-2025-47277


JSON object : View

Products Affected

vllm

  • vllm
CWE
CWE-502

Deserialization of Untrusted Data