vLLM Load balancing

Does vLLM come with its own load balancer. Whats the recommended way of load balancing an inference endpoint with vllm

The load balancers currently lie outside of the vLLM repo and are more like a component on top of it. Available options AFAIK are:

1 Like