Can you list all the needed software components/packages what vLLM inferencing needs with Nvidia RTX PRO card when using clean Ubuntu 24.04 or Debian 13 and the vllm docker images.
For example:
Git, Python, nvidia drivers (which version), what else?
For vLLM inference with an NVIDIA RTX PRO card on a clean Ubuntu 24.04 or Debian 13 system using the official vLLM Docker images, you mainly need:
NVIDIA GPU Driver: Install the latest NVIDIA driver that supports your GPU and CUDA 12.8+ (e.g., driver version 570.x or newer for Blackwell/RTX PRO 6000/5090) [source].
NVIDIA Container Toolkit: Required for Docker to access the GPU [source].
Docker: Any recent version supporting the NVIDIA runtime [source].
The vLLM Docker image itself contains all other dependencies (Python, vLLM, CUDA libraries, PyTorch, etc.), so you do not need to install Git, Python, or CUDA toolkit on the host. Just ensure the NVIDIA driver and container toolkit are installed and configured. Would you like a step-by-step installation guide?