openeuler/euler-copilot-fast-inference-qwenAn inference framework that can be used for large-scale CPU inference, based on the Kunpeng and openEuler basic image, containerized package after release, can be used for one-click deployment
The tag of each euler-copilot-fast-inference-qwen docker image is consist of the version of euler-copilot-fast-inference-qwen and the version of basic image. The details are as follows
| Tags | Currently | Architectures |
|---|---|---|
| 1.5-oe2203sp3 | euler-copilot-fast-inference-qwen 1.5 on openEuler 22.03-lts-sp3 | arm64 |
Start the container based on the application image
Access the /home/euler-copilot-fast-inference directory to get the inference process
Specify the large model path and execute the inference process
Specify the large model path as the first argument, using -i to specify the input prompt word
View model inference results
example download: | Due to the large size of euler-copilot-fast-inference image, it is recommended to pull it locally separately before starting the container:
docker pull openeuler/euler-copilot-fast-inference:{Tag}
usage: |
downloadqwen1_5-7b-chat-q4_0.gguf
wget [***]
start docker,copy weight file into docker,start inference
# cd source code cd euler-copilot-fast-inference # start docker docker run --name **** -it -d --net=host --privileged=true --entrypoint=bash openeuler/euler-copilot-fast-inference-qwen:1.5-oe2203sp3 # start inference docker exec -it <name> bash cd /home/euler-copilot-fast-inference # example ./fast-llm qwen1_5-7b-chat-q4_0.gguf -t 0.0 -i "中国最高的山峰是?" -n 64
If you have any questions or want to use some special features, please submit an issue or a pull request on openeuler-docker-images.



manifest unknown 错误
TLS 证书验证失败
DNS 解析超时
410 错误:版本过低
402 错误:流量耗尽
身份认证失败错误
429 限流错误
凭证保存错误
来自真实用户的反馈,见证轩辕镜像的优质服务