CVE-2025-66448
📋 TL;DR
This vulnerability in vLLM allows remote code execution when loading model configurations containing auto_map entries. Attackers can create a seemingly benign repository that points to a malicious backend, executing arbitrary code on victim systems even when trust_remote_code is explicitly disabled. Users of vLLM versions before 0.11.1 who load external model configurations are affected.
💻 Affected Systems
- vLLM
📦 What is this software?
Vllm by Vllm
⚠️ Risk & Real-World Impact
Worst Case
Complete system compromise allowing attackers to execute arbitrary code, steal data, deploy ransomware, or pivot to other systems.
Likely Case
Data exfiltration, cryptocurrency mining, or backdoor installation on vulnerable vLLM instances.
If Mitigated
Limited impact if proper network segmentation and least privilege principles are implemented.
🎯 Exploit Status
Exploitation requires the victim to load a malicious configuration file, which could be achieved through social engineering or supply chain attacks.
🛠️ Fix & Mitigation
✅ Official Fix
Patch Version: 0.11.1
Vendor Advisory: https://github.com/vllm-project/vllm/security/advisories/GHSA-8fr4-5q9j-m8gm
Restart Required: Yes
Instructions:
1. Update vLLM to version 0.11.1 or later using pip: pip install --upgrade vllm>=0.11.1
2. Restart all vLLM services and processes
3. Verify the fix by checking the version
🔧 Temporary Workarounds
Disable remote code loading
allAvoid loading any model configurations from untrusted sources or repositories containing auto_map entries.
Network restrictions
linuxBlock outbound connections from vLLM instances to prevent fetching remote code.
iptables -A OUTPUT -p tcp --dport 443 -j DROP
iptables -A OUTPUT -p tcp --dport 80 -j DROP
🧯 If You Can't Patch
- Only load model configurations from trusted, verified sources with known integrity.
- Implement strict network segmentation to isolate vLLM instances from untrusted networks.
🔍 How to Verify
Check if Vulnerable:
Check vLLM version: python -c "import vllm; print(vllm.__version__)" - if version is less than 0.11.1, system is vulnerable.
Check Version:
python -c "import vllm; print(vllm.__version__)"
Verify Fix Applied:
After updating, verify version is 0.11.1 or higher using the same command.
📡 Detection & Monitoring
Log Indicators:
- Unexpected network connections to external repositories
- Unusual Python module imports or executions
- Errors related to Nemotron_Nano_VL_Config class
Network Indicators:
- Outbound HTTP/HTTPS connections to unfamiliar repositories during model loading
SIEM Query:
source="vllm" AND ("auto_map" OR "Nemotron_Nano_VL_Config" OR "get_class_from_dynamic_module")