Your current environment
Ubuntun
How would you like to use vllm
I currently need to investigate a question: Does the LoRa adapter for vllm have to run on the same GPU as the base model?
For example, if a basic model and a LoRa adapter are running on a GPU, and the video memory is almost full, can we run another LoRa adapter on another GPU of the same node?