alpine/ollamaThis repository provides a minimal CPU-only Ollama Docker image, specifically designed to run on systems without GPU support. At just 70MB, this image is significantly smaller than the official Ollama image, which is around 4GB.
ollama latest b99944c07117 3 hours ago 69.3MB
[***]
[***]
[***]
Lightweight: The official Ollama image is over 4GB in size, which can be overkill for systems that only need CPU-based processing. This image is only 70MB, making it much faster to download and deploy.
CPU-only Support: This image is tailored for systems without GPUs. It ensures you can run Ollama efficiently, even on basic or resource-constrained environments, without needing specialized hardware.
Run Anywhere: Whether you're working on local servers, edge devices, or cloud environments that don’t offer GPU resources, this image allows you to run Ollama anywhere, focusing purely on CPU-based operations.
bashdocker pull alpine/ollama
docker rm -f ollama docker run -d -p ***:*** -v ~/.ollama/root/.ollama --name ollama alpine/ollama
llama3.2, only run once. It will save the model locally, you can re-use it later.docker exec -ti ollama ollama pull llama3.2
If you don't want to download, you can choice to use alpine/llama3.2 image directly. I create this with model "llama3.2" integrated already
docker run -d -p ***:*** --name llama3.2 alpine/llama3.2
$ curl http://localhost:***/api/generate -d '{ "model": "llama3.2", "prompt":"Why is the sky blue?" }' {"model":"llama3.2","created_at":"2024-10-16T00:25:58.59931201Z","response":"The","done":false} {"model":"llama3.2","created_at":"2024-10-16T00:25:58.695826838Z","response":" sky","done":false} {"model":"llama3.2","created_at":"2024-10-16T00:25:58.780917761Z","response":" appears","done":false} {"model":"llama3.2","created_at":"2024-10-16T00:25:58.992556209Z","response":" blue","done":false} {"model":"llama3.2","created_at":"2024-10-16T00:25:59.085970606Z","response":" because","done":false} {"model":"llama3.2","created_at":"2024-10-16T00:25:59.30869749Z","response":" of","done":false} ...
If you monitor the CPU usage, for example, with htop, you would see the high CPU usage
You can deploy the Ollama web UI to chat with it directly. There are many tools available, but I won't recommend any specific one.
this image could be deployed to any enviornment, for example, in kubernetes cluster, you can use it to analyze logs, streamlining logs with local LLMs, etc.
Announce — Minimal CPU-only Ollama Docker Image - [***]






manifest unknown 错误
TLS 证书验证失败
DNS 解析超时
410 错误:版本过低
402 错误:流量耗尽
身份认证失败错误
429 限流错误
凭证保存错误
来自真实用户的反馈,见证轩辕镜像的优质服务