I want to use vllm to deploy gpt-oss-20b on H20 and want to use the parameters --tool parser and --reasoning-parser. Is there a docker image that can be run?
Yes, there is a dedicated Docker image for GPT-OSS models: vllm/vllm-openai:gptoss. You can run it on H100/H200 GPUs and pass additional parameters such as --tool-call-parser and --reasoning-parser. However, the available tool parsers and reasoning parsers are limited; for GPT-OSS, you should use --tool-call-parser openai and --reasoning-parser openai, as other parser names (e.g., GptOss) may not be available in all versions yet. Example command:
docker run --gpus all -p 8000:8000 --ipc=host vllm/vllm-openai:gptoss \
--model openai/gpt-oss-20b \
--async-scheduling \
--tool-call-parser openai \
--reasoning-parser openai
For more details and up-to-date options, see the official vLLM GPT-OSS recipe and Docker deployment guide. Would you like more detail on tool calling or parser compatibility?
Sources:
Is this vllm/vllm-openai:gptoss image deprecated or is the current plan to have it maintained and regularly updated?
I ask because:
- at the time of writing,
vllm/vllm-openai:gptossimage is behind the:latesttag - the current recipe for GPT-OSS in vLLM docs - GPT OSS - vLLM Recipes - doesn’t mention this image at all and instead references
vllm/vllm-openai:v0.11.0