openeuler/euler-copilot-fast-inferenceAn inference framework that can be used for large-scale CPU inference, based on openEuler basic image, containerized package after release, can be used for one-click deployment
| Tags | Currently | Architectures |
|---|---|---|
| 1.0.0-oe2203sp3 | euler-copilot-fast-inference 1.0.0 on openEuler 22.03-LTS-SP3 | arm64,amd64 |
Start the container based on the application image
Access the /home/euler-copilot-fast-inference directory to get the inference process
Specify the large model path and execute the inference process
Specify the large model path as the first argument, using -i to specify the input prompt word
View model inference results
example download: Due to the large size of euler-copilot-fast-inference image, it is recommended to pull it locally separately before starting the container:
docker pull openeuler/euler-copilot-fast-inference:{Tag}
usage:
download qwen1_5-7b-chat-q4_0.gguf
wget [***]
start docker,copy weight file into docker,start inference
# cd source code cd euler-copilot-fast-inference # start docker docker run --name **** -it -d --net=host --privileged=true --entrypoint=bash openeuler/fast-llm:1.0.0-oe2203sp3 # copy weight file docker cp qwen1_5-7b-chat-q4_0.gguf <CONTAINER ID>:/home/euler-copilot-fast-inference # start inference docker exec -it <name> bash cd /home/euler-copilot-fast-inference # example ./fast-llm qwen1_5-7b-chat-q4_0.gguf -t 0.0 -i "中国最高的山峰是?" -n 64
If you have any questions or want to use some special features, please submit an issue or a pull request on openeuler-docker-images.
manifest unknown 错误
TLS 证书验证失败
DNS 解析超时
410 错误:版本过低
402 错误:流量耗尽
身份认证失败错误
429 限流错误
凭证保存错误
来自真实用户的反馈,见证轩辕镜像的优质服务