A high-throughput and memory-efficient inference and serving engine for LLMs
Here are some things you can do to get started:
Introduce yourself by adding your picture and information about yourself and your interests to your profile. What is one thing youβd like to be asked about?
Get to know the community by browsing discussions that are already happening here. When you find a post interesting, informative, or entertaining, use the to show your appreciation or support!
Contribute by commenting, sharing your own perspective, asking questions, or offering feedback in the discussion. Before replying or starting new topics, please review the Community Guidelines.
Welcome! I would like explain more about the purpose of this forum, courtesy of @comaniac:
If you are confused about where to put their questions (forum vs. Github issues), especially GitHub issues are more accessible than the forum. A simple principle could be: all questions should go to forum, and Github issues should be something the vLLM developers want to track.
If you creates an GH issue for questions, we will reply saying that please ask this question in the forum, and close the issue.
If a topic in the forum ends up being a bug report or feature request, we should encourage the author to create an issue in the Github.
We encourage all community members to answer questions in the forum, whatever youβre a committer or not. Answering questions could also be an important metric for nominating committers (we attached forum activity link in the committer nomination letter).
We also encourage beginners to answer simple questions. This is a very good starting point for beginners to participate the community.