I want to use vllm to deploy gpt-oss-20b on H20 and want to use the parameters --tool parser and --reasoning-parser. Is there a docker image that can be run?
Yes, there is a dedicated Docker image for GPT-OSS models: vllm/vllm-openai:gptoss. You can run it on H100/H200 GPUs and pass additional parameters such as --tool-call-parser and --reasoning-parser. However, the available tool parsers and reasoning parsers are limited; for GPT-OSS, you should use --tool-call-parser openai and --reasoning-parser openai, as other parser names (e.g., GptOss) may not be available in all versions yet. Example command:
docker run --gpus all -p 8000:8000 --ipc=host vllm/vllm-openai:gptoss \
--model openai/gpt-oss-20b \
--async-scheduling \
--tool-call-parser openai \
--reasoning-parser openai
For more details and up-to-date options, see the official vLLM GPT-OSS recipe and Docker deployment guide. Would you like more detail on tool calling or parser compatibility?
Sources: