vLLM#

Description#
vLLM is a fast and easy-to-use library for LLM inference and serving. Originally developed in the Sky Computing Lab at UC Berkeley, vLLM has evolved into a community-driven project with contributions from both academia and industry. See for background info: https://blog.vllm.ai/2023/06/20/vllm.html
Home page for this solution: https://docs.vllm.ai/en/latest/
Overview#
Key |
Value |
---|---|
Name |
vllm |
Description |
A high-throughput and memory-efficient inference and serving engine for LLMs |
License |
Apache License 2.0 |
Programming Language |
Python |
Created |
2023-02-09 |
Last update |
2025-03-31 |
Github Stars |
43148 |
Project Home Page |
|
Code Repository |
|
OpenSSF Scorecard |
Note:
Created date is date that repro is created on Github.com.
Last update is only the last date I run an automatic check.
Do not attach a wrong value to github stars. Its a vanity metric! Stars count are misleading and don’t indicate if the SBB is high-quality or very popular.