|
About the Model Support category
|
|
0
|
128
|
March 20, 2025
|
|
[Field Report] AWQ on RTX 5060 Ti (SM_120 / Blackwell) — awq_marlin + TRITON_ATTN working
|
|
1
|
51
|
March 18, 2026
|
|
Trying to run Qwen3.5-397B-A17B-GPTQ-Int4
|
|
10
|
129
|
March 13, 2026
|
|
Suggestion to improve inferencing speed
|
|
17
|
327
|
March 11, 2026
|
|
Critique my vLLM configuration for qwen3-coder-next
|
|
3
|
77
|
March 10, 2026
|
|
Hosting Qwen 3.5 35B-A3B model
|
|
1
|
827
|
February 25, 2026
|
|
Mistral Small 3.2 finetune errors out: There is no module or parameter named 'language_model' in LlamaForCausalLM
|
|
3
|
410
|
February 18, 2026
|
|
Mistral-small-3.2: Unable to locate consolidated.safetensors.index.json
|
|
1
|
38
|
February 8, 2026
|
|
How to run Deep Seek OCR 2 in vllm
|
|
1
|
1018
|
January 27, 2026
|
|
Vllm-omni cannot load z-image-turbo
|
|
3
|
235
|
December 27, 2025
|
|
Llama 3.3 70B very slow
|
|
5
|
569
|
December 11, 2025
|
|
Text to speech support with /v1/audio/speech route
|
|
1
|
481
|
November 28, 2025
|
|
Using InternVL3 to perform OCR tasks yields worse results in vLLM than in LMDeploy
|
|
2
|
66
|
November 27, 2025
|
|
Serving minimax-m2
|
|
3
|
371
|
November 8, 2025
|
|
Disabling reasoning of Qwen3-VL-8B-Thinking per request
|
|
1
|
2701
|
October 29, 2025
|
|
Zerank - deploying using vllm
|
|
3
|
188
|
October 29, 2025
|
|
Vllm-ascend是否支持deepseek-ocr
|
|
2
|
316
|
October 21, 2025
|
|
Custom edit on the embedding out
|
|
2
|
55
|
October 15, 2025
|
|
Which ATTENTION BACKEND for gpt-oss in version 0.11.0?
|
|
1
|
468
|
October 4, 2025
|
|
Loading non-consolidated Mistral safetensors
|
|
3
|
354
|
September 30, 2025
|
|
Issue serving gemma3-27b-it
|
|
1
|
476
|
September 19, 2025
|
|
Progress bar to browser
|
|
0
|
41
|
September 11, 2025
|
|
Intermittent Service Downtime Issue with Magistral-Small-2506 Model on GPU VM
|
|
1
|
234
|
September 3, 2025
|
|
GPT OSS in docker container
|
|
1
|
309
|
August 19, 2025
|
|
Why does prefill use normal attention, while decode uses weight absorption in MLA?
|
|
1
|
185
|
August 5, 2025
|
|
Using vLLM on a HF model architecture modified locally
|
|
1
|
182
|
July 7, 2025
|
|
The vllm/vllm-openai version 0.9.1 is nearly 30% faster compared to lmsysorg/sglang:v0.4.7.post, but it stops running every two to three hours
|
|
0
|
167
|
June 23, 2025
|
|
Gemma 3 prefix caching in case of multimodal prompts
|
|
4
|
261
|
May 22, 2025
|
|
Will vLLM follow-up DeepSeek's inference system
|
|
3
|
549
|
May 13, 2025
|
|
Add Support for GLM-4 series model
|
|
1
|
178
|
April 16, 2025
|