Skip to content

vLLM

vLLM is a fast and easy-to-use library for LLM inference and serving.

The provider is vllm and the model name is ignored.