摘要: vLLM is a fast and easy-to-use library for LLM inference and serving. github地址: https://github.com/vllm-project/vllm 官方文档: https://docs.vllm.ai/ 阅读全文
posted @ 2025-03-08 08:53 rvy 阅读(146) 评论(0) 推荐(0)