
natecompiles/vllmThis repository provides vLLM images compiled to work with older NVIDIA CUDA versions.
This project aims to provide pre-compiled vLLM Docker images for systems that can't use the latest CUDA toolkits. Currently, the repository offers images specifically compiled for CUDA 12.4.
vLLM is a high-throughput, easy-to-use LLM serving engine that leverages paged attention to achieve significantly higher throughput than traditional serving methods.




manifest unknown 错误
TLS 证书验证失败
DNS 解析超时
410 错误:版本过低
402 错误:流量耗尽
身份认证失败错误
429 限流错误
凭证保存错误
来自真实用户的反馈,见证轩辕镜像的优质服务