启动命令:
source /usr/local/lib/miniconda3/bin/activate SenseRL && \
vllm serve /mnt/afs/share_models/git_models/Qwen/Qwen3-32B \
--port 6366 \
--tensor-parallel-size 8 \
日志:
[2025-07-30 17:13:26] reset container password ...
[2025-07-30 17:13:26] check and install necesary softwares for tione ...
[2025-07-30 17:13:27] tikit installed.
[2025-07-30 17:13:28] sshd started.
[2025-07-30 17:13:28] config password-free login from launcher node to all worker nodes ...
[2025-07-30 17:13:30] set environment variables ...
[2025-07-30 17:13:30] start training ...
[2025-07-30 17:13:38] INFO 07-30 17:13:37 [__init__.py:243] Automatically detected platform cuda.
[2025-07-30 17:13:42] INFO 07-30 17:13:41 [__init__.py:31] Available plugins for group vllm.general_plugins:
[2025-07-30 17:13:42] INFO 07-30 17:13:41 [__init__.py:33] - lora_filesystem_resolver -> vllm.plugins.lora_resolvers.filesystem_resolver:register_filesystem_resolver
[2025-07-30 17:13:42] INFO 07-30 17:13:41 [__init__.py:36] All plugins in this group will be loaded. Set `VLLM_PLUGINS` to control which plugins to load.
[2025-07-30 17:13:44] usage: vllm [-h] [-v] {chat,complete,serve,bench,collect-env} ...
[2025-07-30 17:13:44] vllm: error: unrecognized arguments: \
[2025-07-30 17:13:45] train failed: exit status 2