Docker image for the Text Generation Web UI: A Gradio web UI for Large Language Models. Supports transformers, GPTQ, llama.cpp (GGUF), Llama models
Note
The legacy APIs no longer work with the latest version of the
Text Generation Web UI. They were deprecated in
November 2023 and have now been completely removed.
If you want to use the LEGACY APIs, please set the image tag
to 1.9.5
. You will also have to add port 6000 for the
legacy REST API and/or port 6005 for the legacy Websockets API.
- Ubuntu 22.04 LTS
- CUDA 12.1.1
- Python 3.10.12
- Text Generation Web UI
- Torch 2.2.0
- xformers 0.0.24
- Jupyter Lab
- runpodctl
- OhMyRunPod
- RunPod File Uploader
- croc
- rclone
- speedtest-cli
- screen
- tmux
This image is designed to work on RunPod. You can use my custom RunPod template to launch it on RunPod.
docker run -d \
--gpus all \
-v /workspace \
-p 3000:3001 \
-p 5000:5001 \
-p 8888:8888 \
-p 2999:2999 \
-e VENV_PATH="/workspace/venvs/text-generation-webui" \
ashleykza/oobabooga:latest
You can obviously substitute the image name and tag with your own.
Connect Port | Internal Port | Description |
---|---|---|
3000 | 3001 | Text Generation Web UI |
5000 | 5001 | Open AI Compatible API |
8888 | 8888 | Jupyter Lab |
2999 | 2999 | RunPod File Uploader |
Variable | Description | Default |
---|---|---|
VENV_PATH | Set the path for the Python venv for the app | /workspace/venvs/text-generation-webui |
DISABLE_AUTOLAUNCH | Disable Web UI from launching automatically | (not set) |
Text Generation Web UI creates a log file, and you can tail the log instead of killing the service to view the logs.
Application | Log file |
---|---|
Text Generation Web UI | /workspace/logs/textgen.log |
For example:
tail -f /workspace/logs/textgen.log
Pull requests and issues on GitHub are welcome. Bug fixes and new features are encouraged.
You can contact me and get help with deploying your container to RunPod on the RunPod Discord Server below, my username is ashleyk.