
longkeyy/llamafactorylongkeyy/llamafactory是一个多功能Docker镜像,支持多种硬件后端的模型训练与推理,包括NVIDIA CUDA、华为Ascend NPU和AMD ROCm。它提供Web UI和OpenAI兼容的API访问方式。
| 标签 | 后端 | 描述 | 推荐使用场景 |
|---|---|---|---|
cuda, latest | NVIDIA CUDA | NVIDIA GPU标准版本 | 配备CUDA GPU的机器 |
npu | Ascend NPU | 为华为Ascend NPU构建 | Ascend 310/910设备 |
rocm | AMD ROCm | 兼容AMD ROCm GPU | 配备AMD GPU的机器 |
注意:
latest标签始终指向最新的cuda镜像版本。
bashdocker run -it --gpus all \ --shm-size 16g \ --ipc host \ -p 7860:7860 -p 8000:8000 \ -v $PWD/hf_cache:/root/.cache/huggingface \ -v $PWD/ms_cache:/root/.cache/modelscope \ -v $PWD/om_cache:/root/.cache/openmind \ -v $PWD/data:/app/data \ -v $PWD/output:/app/output \ --restart unless-stopped \ longkeyy/llamafactory:cuda-0.9.2
确保已挂载必要的Ascend驱动和设备。
bashdocker run -it \ --shm-size 16g \ --ipc host \ --device /dev/davinci0 \ --device /dev/davinci_manager \ --device /dev/devmm_svm \ --device /dev/hisi_hdc \ -v /usr/local/dcmi:/usr/local/dcmi \ -v /usr/local/bin/npu-smi:/usr/local/bin/npu-smi \ -v /usr/local/Ascend/driver:/usr/local/Ascend/driver \ -v /etc/ascend_install.info:/etc/ascend_install.info \ -v $PWD/hf_cache:/root/.cache/huggingface \ -v $PWD/ms_cache:/root/.cache/modelscope \ -v $PWD/om_cache:/root/.cache/openmind \ -v $PWD/data:/app/data \ -v $PWD/output:/app/output \ -p 7860:7860 -p 8000:8000 \ longkeyy/llamafactory:npu-0.9.2
bashdocker run -it \ --shm-size 16g \ --ipc host \ --device /dev/kfd \ --device /dev/dri \ -v $PWD/hf_cache:/root/.cache/huggingface \ -v $PWD/ms_cache:/root/.cache/modelscope \ -v $PWD/om_cache:/root/.cache/openmind \ -v $PWD/data:/app/data \ -v $PWD/output:/app/output \ -v $PWD/saves:/app/saves \ -p 7860:7860 -p 8000:8000 \ --restart unless-stopped \ longkeyy/llamafactory:rocm-0.9.2
yamlversion: '3.8' services: llamafactory-cuda: image: longkeyy/llamafactory:cuda-0.9.2 container_name: llamafactory-cuda ports: - "7860:7860" - "8000:8000" volumes: - ./hf_cache:/root/.cache/huggingface - ./ms_cache:/root/.cache/modelscope - ./om_cache:/root/.cache/openmind - ./data:/app/data - ./output:/app/output deploy: resources: reservations: devices: - driver: nvidia count: all capabilities: [gpu] shm_size: "16gb" ipc: host tty: true restart: unless-stopped profiles: ["cuda"] llamafactory-npu: image: longkeyy/llamafactory:npu-0.9.2 container_name: llamafactory-npu ports: - "7860:7860" - "8000:8000" volumes: - ./hf_cache:/root/.cache/huggingface - ./ms_cache:/root/.cache/modelscope - ./om_cache:/root/.cache/openmind - ./data:/app/data - ./output:/app/output - /usr/local/dcmi:/usr/local/dcmi - /usr/local/bin/npu-smi:/usr/local/bin/npu-smi - /usr/local/Ascend/driver:/usr/local/Ascend/driver - /etc/ascend_install.info:/etc/ascend_install.info devices: - /dev/davinci0 - /dev/davinci_manager - /dev/devmm_svm - /dev/hisi_hdc shm_size: "16gb" ipc: host tty: true stdin_open: true restart: unless-stopped profiles: ["npu"] llamafactory-rocm: image: longkeyy/llamafactory:rocm-0.9.2 container_name: llamafactory-rocm ports: - "7860:7860" - "8000:8000" volumes: - ./hf_cache:/root/.cache/huggingface - ./ms_cache:/root/.cache/modelscope - ./om_cache:/root/.cache/openmind - ./data:/app/data - ./output:/app/output - ./saves:/app/saves devices: - /dev/kfd:/dev/kfd - /dev/dri:/dev/dri shm_size: "16gb" ipc: host tty: true stdin_open: true restart: unless-stopped profiles: ["rocm"]
bash# CUDA docker compose --profile cuda up -d # NPU docker compose --profile npu up -d # ROCm docker compose --profile rocm up -d
| 主机路径 | 容器路径 | 描述 |
|---|---|---|
./hf_cache | /root/.cache/huggingface | HuggingFace模型缓存 |
./ms_cache | /root/.cache/modelscope | ModelScope模型缓存 |
./om_cache | /root/.cache/openmind | OpenMind缓存 |
./data | /app/data | 数据集目录 |
./output | /app/output | 输出目录 |
./saves(仅ROCm) | /app/saves | 模型保存目录 |
--restart unless-stopped参数确保服务弹性。llamafactory-cli webui。ARG参数自定义构建。| 标签 | 大小 |
|---|---|
cuda | ~42.9 GB |
npu | ~34.6 GB |
rocm | ~76.6 GB |






manifest unknown 错误
TLS 证书验证失败
DNS 解析超时
410 错误:版本过低
402 错误:流量耗尽
身份认证失败错误
429 限流错误
凭证保存错误
来自真实用户的反馈,见证轩辕镜像的优质服务