GPT OSS vllm server

I want to use vllm to deploy gpt-oss-20b on H20 and want to use the parameters --tool parser and --reasoning-parser. Is there a docker image that can be run?

Yes, there is a dedicated Docker image for GPT-OSS models: vllm/vllm-openai:gptoss. You can run it on H100/H200 GPUs and pass additional parameters such as --tool-call-parser and --reasoning-parser. However, the available tool parsers and reasoning parsers are limited; for GPT-OSS, you should use --tool-call-parser openai and --reasoning-parser openai, as other parser names (e.g., GptOss) may not be available in all versions yet. Example command:

docker run --gpus all -p 8000:8000 --ipc=host vllm/vllm-openai:gptoss \
  --model openai/gpt-oss-20b \
  --async-scheduling \
  --tool-call-parser openai \
  --reasoning-parser openai

For more details and up-to-date options, see the official vLLM GPT-OSS recipe and Docker deployment guide. Would you like more detail on tool calling or parser compatibility?

Sources: