
dustynv/text-generation-webui
CONTAINERSIMAGESRUNBUILD
llama.cpp exllama2 AutoGPTQ transformers[!WARNING]
If you're using the llama.cpp loader, the model format has changed from GGML to GGUF. Existing GGML models can be converted using theconvert-llama-ggmlv3-to-gguf.pyscript inllama.cpp(or you can often find the GGUF conversions on HuggingFace Hub)
This container has a default run command that will automatically start the webserver like this:
bashcd /opt/text-generation-webui && python3 server.py \ --model-dir=/data/models/text-generation-webui \ --listen --verbose
To launch the container, run the command below, and then navigate your browser to [***]
bash./run.sh $(./autotag text-generation-webui)
While the server and models are dynamically configurable from within the webui at runtime, see here for optional command-line settings:
For example, after you've downloaded a model, you can load it directly at startup like so:
bash./run.sh $(./autotag text-generation-webui) /bin/bash -c \ "cd /opt/text-generation-webui && python3 server.py \ --model-dir=/data/models/text-generation-webui \ --model=llama-2-13b-chat.Q4_K_M.gguf \ --loader=llamacpp \ --n-gpu-layers=128 \ --listen --chat --verbose
See here for instructions for downloading models - you can do this from within the webui, or by running their download-model.py script:
bash./run.sh --workdir=/opt/text-generation-webui $(./autotag text-generation-webui) /bin/bash -c \ 'python3 download-model.py --output=/data/models/text-generation-webui TheBloke/Llama-2-7b-Chat-GPTQ'
This will download specified model from HuggingFace Hub and place it under the /data/models/text-generation-webui mounted directory (which is where you should store models so they aren't lost when the container exits)
n-gpu-layers to 128 in the loader settingsn_gqa to 8 (otherwise an error will occur)Q4_K_M model quantizationsDefault or Notebook tabInstruct chat mode and set the Instruction Template to Llama-v2 (in the Parameters tab)
Answer the questions. part)>> What games do you like to play? I'm a large language model, so I can play text-based games and answer questions on a wide variety of topics. Here are some game ideas: - 20 Questions: You or I can think of an object, and the other player tries to guess what it is by asking up to 20 yes-or-no questions. The goal is to guess the object in as few questions as possible. - Hangman: I can think of a word or phrase and you try to guess it by suggesting letters. Each incorrect guess will bring the hangman one step closer to being complete! - Word association: I'll give you a word and you have to come up with a word that is related to it. We can keep playing back and forth as long as you'd like! - Trivia: I can provide questions on a wide range of topics, from history to science to pop culture. See how many questions you can answer correctly! - Storytelling: I can start telling a story and then stop at a cliffhanger. You can then try to guess what happens next or even take over the storytelling and continue it in your own - Jokes: I love to tell jokes and make people laugh with my "Dad humor"! Knock knock! *giggles*
text-generation-webui:main | |
|---|---|
| Aliases | text-generation-webui |
| Requires | L4T ['>=34.1.0'] |
| Dependencies | build-essential cuda cudnn python numpy cmake onnx pytorch:2.2 torchvision huggingface_hub rust transformers auto_gptq exllama llama_cpp |
| Dockerfile | Dockerfile |
| Images | dustynv/text-generation-webui:main-r36.2.0 (2023-12-18, 8.1GB) |
text-generation-webui:1.7 | |
|---|---|
| Requires | L4T ['>=34.1.0'] |
| Dependencies | build-essential cuda cudnn python numpy cmake onnx pytorch:2.2 torchvision huggingface_hub rust transformers auto_gptq exllama llama_cpp |
| Dockerfile | Dockerfile |
| Images | dustynv/text-generation-webui:1.7-r35.4.1 (2023-12-05, 6.4GB) |
text-generation-webui:6a7cd01 | |
|---|---|
| Requires | L4T ['>=34.1.0'] |
| Dependencies | build-essential cuda cudnn python numpy cmake onnx pytorch:2.2 torchvision huggingface_hub rust transformers auto_gptq exllama llama_cpp |
| Dockerfile | Dockerfile |
| Repository/Tag | Date | Arch | Size |
|---|---|---|---|
dustynv/text-generation-webui:1.7-r35.4.1 | 2023-12-05 | arm64 | 6.4GB |
dustynv/text-generation-webui:main-r36.2.0 | 2023-12-18 | arm64 | 8.1GB |
dustynv/text-generation-webui:r35.2.1 | 2024-02-01 | arm64 | 6.6GB |
dustynv/text-generation-webui:r35.3.1 | 2024-02-03 | arm64 | 6.6GB |
dustynv/text-generation-webui:r35.4.1 | 2024-02-01 | arm64 | 6.6GB |
dustynv/text-generation-webui:r35.4.1-cp310 | 2024-04-12 | arm64 | 6.4GB |
dustynv/text-generation-webui:r36.2.0 | 2024-02-03 | arm64 | 8.3GB |
Container images are compatible with other minor versions of JetPack/L4T:
• L4T R32.7 containers can run on other versions of L4T R32.7 (JetPack 4.6+)
• L4T R35.x containers can run on other versions of L4T R35.x (JetPack 5.1+)
To start the container, you can use jetson-containers run and autotag, or manually put together a docker run command:
bash# automatically pull or build a compatible container image jetson-containers run $(autotag text-generation-webui) # or explicitly specify one of the container images above jetson-containers run dustynv/text-generation-webui:r35.4.1-cp310 # or if using 'docker run' (specify image and mounts/ect) sudo docker run --runtime nvidia -it --rm --network=host dustynv/text-generation-webui:r35.4.1-cp310
jetson-containers runforwards arguments todocker runwith some defaults added (like--runtime nvidia, mounts a/datacache, and detects devices)
autotagfinds a container image that's compatible with your version of JetPack/L4T - either locally, pulled from a registry, or by building it.
To mount your own directories into the container, use the -v or --volume flags:
bashjetson-containers run -v /path/on/host:/path/in/container $(autotag text-generation-webui)
To launch the container running a command, as opposed to an interactive shell:
bashjetson-containers run $(autotag text-generation-webui) my_app --abc xyz
You can pass any options to it that you would to docker run, and it'll print out the full command that it constructs before executing it.
If you use autotag as shown above, it'll ask to build the container for you if needed. To manually build it, first do the system setup, then run:
bashjetson-containers build text-generation-webui
The dependencies from above will be built into the container, and it'll be tested during. Run it with --help for build options.






manifest unknown 错误
TLS 证书验证失败
DNS 解析超时
410 错误:版本过低
402 错误:流量耗尽
身份认证失败错误
429 限流错误
凭证保存错误
来自真实用户的反馈,见证轩辕镜像的优质服务