text-generation-webui Docker 镜像下载 - 轩辕镜像
text-generation-webui 镜像详细信息和使用指南
text-generation-webui 镜像标签列表和版本信息
text-generation-webui 镜像拉取命令和加速下载
text-generation-webui 镜像使用说明和配置指南
Docker 镜像加速服务 - 轩辕镜像平台
国内开发者首选的 Docker 镜像加速平台
极速拉取 Docker 镜像服务
相关 Docker 镜像推荐
热门 Docker 镜像下载
text-generation-webui 镜像详细信息
text-generation-webui 镜像标签列表
text-generation-webui 镜像使用说明
text-generation-webui 镜像拉取命令
Docker 镜像加速服务
轩辕镜像平台优势
镜像下载指南
相关 Docker 镜像推荐
text-generation-webui 镜像详细说明
text-generation-webui 使用指南
text-generation-webui 配置说明
text-generation-webui 官方文档
This image's Dockerfile and documentation are available on github: [***]
Announcement
The updates to support version 3 of text-generation-webui require that the bind mount paths are updated in the container. Please review the new example docker-compose.yml file for details and check the releases page for more information. TLDR; *:/app/* must be updated to *:/app/user_data/* for most paths.
Introduction
This project dockerises the deployment of oobabooga/text-generation-webui and its variants. It provides a default configuration corresponding to a standard deployment of the application with all extensions enabled, and a base version without extensions. Versions are offered for Nvidia GPU nvidia, AMD GPU (unstable) rocm, Intel Arc (unstable) arc, and CPU-only cpu. Pre-built images are available on Docker Hub: [***]
The goal of this project is to be to oobabooga/text-generation-webui, what AbdBarho/stable-diffusion-webui-docker is to AUTOMATIC1111/stable-diffusion-webui.
Quick-Start
- Pull the repo:
git clone [***] - Point your terminal to the downloaded folder (e.g.,
cd text-generation-webui-docker) - (Optional) Edit
docker-compose.ymlto your requirements - Start the server (the image will be pulled automatically for the first run):
docker compose up - Navigate to
127.0.0.1:7860and enjoy your local instance of oobabooga's text-generation-webui!
Usage
This repo provides a template docker-compose.yml and a structured config folder to store the application files. The project officially targets Linux as the deployment platform, however the images will also work on Docker Desktop for Windows. There is no plan to support Apple Silicon because that platform runs docker inside a VM, and Apple has chosen not to support virtualised hardware access. Intel Macs should be fine to run with the variant suitable for their GPU.
Check the issues for hints and tips for your platform (and remember to search closed issues too!)
Pre-Requisites
- docker
- docker compose
- CUDA docker runtime (optional, for Nvidia GPU-powered inferencing)
Ask your favourite LLM how to install and configure docker, docker-compose, and the Nvidia CUDA docker runtime for your platform!
Docker Compose
This is the recommended deployment method (it is the easiest and quickest way to manage folders and settings through updates and reinstalls). The recommended variant is default (it is the full version of the standard application with all default bundled extensions installed, set up for Nvidia GPU accelerated inference).
Select variant
Tagged release versions are published on a regular basis - check hub.docker.com/r/atinoda/text-generation-webui for available tags. Pulling an untagged variant will pull the latest stable release. Unstable, latest versions are available via nightly builds.
Choose the desired variant by setting the image :tag in docker-compose.yml using the pattern {VARIANT}-{PLATFORM}, or {VARIANT}-{PLATFORM}-{VERSION} to specify a specific release.
| Variant | Description |
|---|---|
default-* | Standard deployment with all default bundled extensions installed. Normal image intended for everyday usage. |
base-* | Basic deployment with no extensions installed. Slimmer image intended for customisation or lightweight deployment. |
| Platform | Description |
|---|---|
*-nvidia | CUDA 12.4 inference acceleration. |
*-nvidia-noavx2 | CUDA 12.4 inference acceleration with no AVX2 CPU instructions. Typical use-case is legacy CPU with modern GPU. |
*-nvidia-tenssorrtllm | CUDA 12.4 inference acceleration with additional TensorRT-LLM library pre-installed. This has not been updated in a long time! |
*-cpu | CPU-only inference. Has become surprisingly fast since the early days! |
*-rocm | ROCM 5.6 inference acceleration. Experimental and unstable. |
*-arc | Intel Arc XPU and oneAPI inference acceleration. Not compatible with Intel integrated GPU (iGPU). Experimental and unstable. |
| Examples | Description |
|---|---|
default | Standard deployment with all extensions, configured for Nvidia GPU accelerated inferencing. Same as default-nvidia. This version is recommended for most users. |
default-cpu | Standard deployment with all extensions, set up for CPU-only inference. This version is useful if you don't have a supported GPU. |
{VARIANT}-{PLATFORM}-{VERSION} | Build of each {VARIANT}-{PLATFORM} tagged with the release {VERSION} of the text-generation-webui (e.g., default-nvidia-snapshot-2024-02-04). Visit obabooga/text-generation-webui/releases for release notes. Go to hub.docker.com/r/atinoda/text-generation-webui to see the available pre-built versions. |
{VARIANT}-{PLATFORM}-nightly | Automated nightly build of the variant. These images are built and pushed automatically - they are untested and may be unstable. Suitable when more frequent updates are required and instability is not an issue. |
Deploy
Deploy the service:
docker compose up
Remove
Remove the service:
docker compose down -v
Configuration
These configuration instructions describe the relevant details for this docker wrapper. Refer to oobabooga/text-generation-webui documentation for usage of the application itself.
Ports
Three commonly used ports are exposed:
| Port | Description | Configuration |
|---|---|---|
7860 | Web UI port | Pre-configured and enabled in docker-compose.yml |
5000 | API port | Enable by adding --api --extensions api to launch args then uncomment mapping in docker-compose.yml |
5005 | Streaming port | Enable by adding --api --extensions api to launch args then uncomment mapping in docker-compose.yml |
Extensions may use additional ports - check the application documentation for more details.
Volumes
The provided example docker compose maps several volumes from the local config directory into the container: loras, models, presets, prompts, training, extensions. If these folders are empty, they will be initialised when the container is run.
Extensions will persist their state between container launches if you use a mapped folder - but they will not automatically update when a new image is released, so this feature is disabled by default. The whole extensions folder can be mapped (all extensions are persisted) or individual extensions can be mapped one at a time. Examples are given in the docker-compose.yml.
If you are getting an error about missing files, try clearing these folders and letting the service re-populate them.
Extra launch arguments
Extra launch arguments can be defined in the environment variable EXTRA_LAUNCH_ARGS (e.g., "--model MODEL_NAME", to load a model at launch). The provided default extra arguments are --verbose and --listen (which makes the webui available on your local network) and these are set in the docker-compose.yml.
Launch arguments should be defined as a space-separated list, just like writing them on the command line. These arguments are passed to the server.py module.
Runtime extension build
Extensions which should be built during startup can be defined in the environment variable BUILD_EXTENSIONS_LIVE (e.g., "coqui_tts whisper_stt", will rebuild those extensions at launch). This feature may be useful if you are developing a third-party extension and need its dependencies to refresh at launch.
Startup times will be much slower if you use this feature, because it will rebuild the named extensions every time the container is started (i.e., don't use this feature unless you are certain that you need it.)
Extension names for runtime build should be defined as a space-separated list.
Updates
These projects are moving quickly! To update to the most recent version on Docker hub, pull the latest image:
docker compose pull
Then recreate the container:
docker compose up
When the container is launched, it will print out how many commits behind origin the current build is, so you can decide if you want to update it. Docker hub images will be periodically updated. The default-nightly image is built every day but it is not manually tested. If you need bleeding edge versions you must build locally.
Build (optional)
The provided docker-compose.yml.build shows how to build the image locally. You can use it as a reference to modify the original docker-compose.yml, or you can rename it and use it as-is. Choose the desired variant to build by setting the build target and then run:
docker compose build
To do a clean build and ensure the latest version:
docker compose build --no-cache
If you choose a different variant later, you must rebuild the image.
Developers / Advanced Users
The Dockerfile can be easily modified to compile and run the application from a local source folder. This is useful if you want to do some development or run a custom version. See the Dockerfile itself for instructions on how to do this.
Support is not provided for this deployment pathway. It is assumed that you are competent and willing to do your own debugging! Pro-tip: start by placing a text-generation-webui repo into the project folder.
Standalone Container
NOT recommended, instructions are included for completeness.
Run
Run a network accessible container (and destroy it upon completion):
docker run -it --rm -e EXTRA_LAUNCH_ARGS="--listen --verbose" --gpus all -p 7860:7860 atinoda/text-generation-webui:default-nvidia
Build and run (optional)
Build the image for the default target and tag it as local :
docker build --target default-nvidia -t text-generation-webui:local .
Run the local image with local network access (and destroy it upon completion):
docker run -it --rm -e EXTRA_LAUNCH_ARGS="--listen --verbose" --gpus all -p 7860:7860 text-generation-webui:local
Known Issues
AMD GPU ROCM
The rocm variant is reported to be working, but it is blind-built and not regularly tested due to a lack of hardware. User reports and insights are welcomed.
Thanks to @Alkali-V2 for confirming successful deployment with RX 6800 on Unraid.
Intel Arc GPU
The arc variant is blind-built and untested due to a lack of hardware. User reports and insights are welcomed.
Extensions
The following are known issues and they are planned to be investigated. Testing and insights are welcomed!
multimodal: Crashes because model is not loaded at start***: Requires an account, causes a crashsilero_tts: Does not work due to pydantic dependency problemsuperbooga/superboogav2: Crashes on startup
Contributions
Contributions are welcomed - please feel free to submit a PR. More variants (e.g., AMD/ROC-M support) and Windows support can help lower the barrier to entry, make this technology accessible to as many people as possible, and push towards democratising the severe impacts that AI is having on our society.
Also - it's fun to code and LLMs are cool.
DISCLAIMER
THE SOFTWARE IS PROVIDED 'AS IS', WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
用户好评
来自真实用户的反馈,见证轩辕镜像的优质服务
oldzhang
运维工程师
Linux服务器
"Docker加速体验非常流畅,大镜像也能快速完成下载。"
Nana
后端开发
Mac桌面
"配置Docker镜像源后,拉取速度快了数倍,开发体验提升明显。"
Qiang
平台研发
K8s集群
"轩辕镜像在K8s集群中表现很稳定,容器部署速度明显加快。"
小敏
测试工程师
Windows桌面
"Docker镜像下载不再超时,测试环境搭建更加高效。"
晨曦
DevOps工程师
宝塔面板
"配置简单,Docker镜像源稳定,适合快速部署环境。"
阿峰
资深开发
群晖NAS
"在群晖NAS上配置后,镜像下载速度飞快,非常适合家庭实验环境。"
俊仔
后端工程师
飞牛NAS
"Docker加速让容器搭建顺畅无比,再也不用等待漫长的下载。"
Lily
测试经理
Linux服务器
"镜像源覆盖面广,更新及时,团队一致反馈体验不错。"
浩子
云平台工程师
Podman容器
"使用轩辕镜像后,Podman拉取镜像稳定无比,生产环境可靠。"
Kai
运维主管
爱快路由
"爱快系统下配置加速服务,Docker镜像拉取速度提升非常大。"
翔子
安全工程师
Linux服务器
"镜像源稳定性高,安全合规,Docker拉取无忧。"
亮哥
架构师
K8s containerd
"大规模K8s集群下镜像加速效果显著,节省了大量时间。"
慧慧
平台开发
Docker Compose
"配置Compose镜像加速后,整体构建速度更快了。"
Tina
技术支持
Windows桌面
"配置简单,镜像拉取稳定,适合日常开发环境。"
宇哥
DevOps Leader
极空间NAS
"在极空间NAS上使用Docker加速,体验流畅稳定。"
小静
数据工程师
Linux服务器
"Docker镜像源下载速度快,大数据环境搭建轻松完成。"
磊子
SRE
宝塔面板
"使用轩辕镜像后,CI/CD流程整体快了很多,值得推荐。"
阿Yang
前端开发
Mac桌面
"国内网络环境下,Docker加速非常给力,前端环境轻松搭建。"
Docker迷
架构师
威联通NAS
"威联通NAS下配置镜像加速后,Docker体验比官方源好很多。"
方宇
系统工程师
绿联NAS
"绿联NAS支持加速配置,Docker镜像下载快且稳定。"
常见问题
免费版仅支持 Docker Hub 加速,不承诺可用性和速度;专业版支持更多镜像源,保证可用性和稳定速度,提供优先客服响应。
免费版仅支持 docker.io;专业版支持 docker.io、gcr.io、ghcr.io、registry.k8s.io、nvcr.io、quay.io、mcr.microsoft.com、docker.elastic.co 等。
当返回 402 Payment Required 错误时,表示流量已耗尽,需要充值流量包以恢复服务。
通常由 Docker 版本过低导致,需要升级到 20.x 或更高版本以支持 V2 协议。
先检查 Docker 版本,版本过低则升级;版本正常则验证镜像信息是否正确。
使用 docker tag 命令为镜像打上新标签,去掉域名前缀,使镜像名称更简洁。
轩辕镜像下载加速使用手册
探索更多轩辕镜像的使用方法,找到最适合您系统的配置方式
登录仓库拉取
通过 Docker 登录认证访问私有仓库
Linux
在 Linux 系统配置镜像加速服务
Windows/Mac
在 Docker Desktop 配置镜像加速
Docker Compose
Docker Compose 项目配置加速
K8s Containerd
Kubernetes 集群配置 Containerd
宝塔面板
在宝塔面板一键配置镜像加速
群晖
Synology 群晖 NAS 配置加速
飞牛
飞牛 fnOS 系统配置镜像加速
极空间
极空间 NAS 系统配置加速服务
爱快路由
爱快 iKuai 路由系统配置加速
绿联
绿联 NAS 系统配置镜像加速
威联通
QNAP 威联通 NAS 配置加速
Podman
Podman 容器引擎配置加速
其他仓库配置
ghcr、Quay、nvcr 等镜像仓库
专属域名拉取
无需登录使用专属域名加速
需要其他帮助?请查看我们的 常见问题 或 官方QQ群: 13763429