Model Support
Llama DeepSeek
Topic | Replies | Views | Activity | |
---|---|---|---|---|
About the Model Support category |
![]() |
0 | 42 | March 20, 2025 |
Using vLLM on a HF model architecture modified locally |
![]() ![]() |
1 | 22 | July 7, 2025 |
The vllm/vllm-openai version 0.9.1 is nearly 30% faster compared to lmsysorg/sglang:v0.4.7.post, but it stops running every two to three hours |
![]() |
0 | 102 | June 23, 2025 |
Gemma 3 prefix caching in case of multimodal prompts |
![]() ![]() |
4 | 70 | May 22, 2025 |
Will vLLM follow-up DeepSeek's inference system |
![]() ![]() ![]() |
3 | 319 | May 13, 2025 |
Add Support for GLM-4 series model |
![]() ![]() |
1 | 84 | April 16, 2025 |
Why does phi3 implementation in vLLM inherit from llama? |
![]() ![]() |
1 | 24 | April 14, 2025 |
Does the latest version support deepseek-v3 tool call |
![]() |
0 | 75 | April 12, 2025 |