vLLM

vLLM#

https://raw.githubusercontent.com/vllm-project/vllm/main/docs/source/assets/logos/vllm-logo-text-light.png

Description#

vLLM is a fast and easy-to-use library for LLM inference and serving. Originally developed in the Sky Computing Lab at UC Berkeley, vLLM has evolved into a community-driven project with contributions from both academia and industry. See for background info: https://blog.vllm.ai/2023/06/20/vllm.html

Home page for this solution: https://docs.vllm.ai/en/latest/

Overview#

Key

Value

Name

vllm

Description

A high-throughput and memory-efficient inference and serving engine for LLMs

License

Apache License 2.0

Programming Language

Python

Created

2023-02-09

Last update

2025-03-31

Github Stars

43148

Project Home Page

https://docs.vllm.ai

Code Repository

vllm-project/vllm

OpenSSF Scorecard

Report

Note:

  • Created date is date that repro is created on Github.com.

  • Last update is only the last date I run an automatic check.

  • Do not attach a wrong value to github stars. Its a vanity metric! Stars count are misleading and don’t indicate if the SBB is high-quality or very popular.