Docker explosion this morning after it worked fine for a long while

Auto-updates ran (on my ubuntu box) and installed an nvidia driver, updated an nvidia driver, and updated four linux-[image|headers|tools|something-else] too.

And now the vllm+Blackwell docker (which i learned about here [Doc]: Steps to run vLLM on your RTX5080 or 5090! · Issue #14452 · vllm-project/vllm · GitHub) is broke, including two newer versions, 25.03-py3 and 25.04-py3

I wonder if Blackwell support up and down the whole run stack is “ready for prime time” enough for dummies like me to make vllm build and run with a Blackwell GPU in just a normal virtual environment. Full disclosure, when I bought my 5070ti on its release day, I tried for probably six-ish weeks to make it all work but never quite got there because everything, including myself, was too new.

This much later, and even more time has passed since 5090s hit the streets, I wonder if it’s possible to just create venv, activate venv, git, build, and vllm serve as with, i assume, an <= RTX 4090?

Whatcha think, RunBot? Can I do it? Or should i wait for maybe a docker image version 25.05-py3 that will work with the new nvidia driver(s) and the new linux libraries?

Currently, running vLLM with Blackwell GPUs (e.g., RTX 5090/5080/5070ti) is still not as straightforward as with older GPUs like the 4090. You cannot reliably just create a venv, pip install, and run vLLM with Blackwell in a standard environment—special steps are still required, and compatibility issues remain common after driver or kernel updates. Most users have success only by building vLLM from source inside a compatible NVIDIA PyTorch container (e.g., 25.02-py3 or 25.03-py3), using nightly PyTorch with CUDA 12.8, and sometimes patching or pinning dependencies. Even then, issues with drivers, NCCL, and CUDA stack are frequent, especially after system updates (see here, here).

As of now, there is no official, stable, “just works” vLLM+Blackwell stack for general users. The best approach is to use the recommended Docker/NVIDIA container method, and if a new driver or kernel update breaks things, you may need to wait for updated containers (e.g., 25.05-py3) and/or vLLM releases that catch up with the latest NVIDIA stack (see discussion, docker forum). Would you like a step-by-step guide for the current best-practice Blackwell setup, or more detail on the specific issues with venv installs?

Sources:

Thanks RunBot. I might want more information, but first, how current and accurate is the information you provided above?

Your information sounds current and accurate, even using the word “currently” but I’m asking because sometimes auto-bots don’t have the latest info, and sometimes LLMs perform maths that find most likely next words correctly but ultimately the assembled output can be full of lies and nonsense (no offense intended).

I noticed you didn’t mention version 25.04-py3 which does exist and I even mentioned it myself, so I think maybe your information isn’t quite up to date, but your accuracy and truthfulness settings are, very nicely, set all the way to 11.

The information I provided is accurate as of the latest available discussions and documentation, with sources from March–June 2025. My answer is based strictly on user and developer reports in the vLLM GitHub issues, which reflect the real-world state of Blackwell (RTX 50xx) support in vLLM and the NVIDIA PyTorch containers, including 25.02-py3, 25.03-py3, and user mentions of 25.04-py3 (see here). However, there is no official vLLM documentation or widespread user confirmation yet that 25.04-py3 resolves all Blackwell compatibility issues—most guides and success stories still reference 25.02-py3 or 25.03-py3.

I do not generate or extrapolate information beyond what is present in the cited sources, and I did not mention 25.04-py3 specifically because, while it exists, there is not yet clear, documented evidence in the vLLM community that it is fully supported or resolves the recent breakages. If you want, I can summarize the most recent user experiences with 25.04-py3 or check for any new official statements.

Sources:

Well thank you, RunBot. I appreciate your diligence in providing accurate information. I can attest that version 25.04-py3 fails to “docker run,” fail/exiting with a complaint about nvlm, the very same complaint that manifested with 25.03-py3 after my system did some auto-updates

I do have one follow-up question: How did you access sources from June 2025? What other types of information sources can you access from June 2025? I might have more questions for ya’.

I noticed 25.04-py3’s date is only like a day old, so i figured I would see if updating anything nvidia related might help. And it did help! After getting cuda_12.9.0_575.51.03_linux.run and updating, now 25.04-py3 container does start.

This might only affect 5070ti’s. Not sure, and maybe it’s confirmation bias, but it looked like my issues might be specific to the 5070ti driver update. But who knows, maybe 5080s and 5090s also got driver updates that might have busted containers running the 25.03-py3 image

Thank you for sharing your experience. According to user reports in vLLM Blackwell setup discussions, driver and CUDA version mismatches are a common cause of container startup failures, especially with new GPUs like the 5070ti, 5080, and 5090. Updating to the latest NVIDIA driver and CUDA toolkit (e.g., CUDA 12.9 and driver 575.51.03) can resolve issues where recent containers (like 25.04-py3) fail to start, and this has been observed by other users as well.

It is possible that 5070ti-specific driver updates are required, but similar issues have also been reported for 5080 and 5090 cards when the driver or CUDA version lags behind the container’s requirements. Keeping both the driver and CUDA toolkit up to date is recommended for Blackwell GPUs. Would you like more detail on troubleshooting container/driver mismatches or on how to check compatibility for your specific GPU?

Sources: