Text Generation Webui – v3.16
π Text Generation WebUI v3.16 just droppedβand itβs a game-changer for local LLM folks!
New portable build via symlink? Yes, please. Devs juggling multiple setups can now switch models and configs without reinstalling. Big shoutout to @reksar! π
macOS Apple Silicon usersβyour day is saved. Python deps now work flawlessly on Tahoe (thanks @drieschel)! π
Backend upgrades? Oh yeah:
- llama.cpp updated to latest GGML fork β now supports Llama-Mini-2.0 and Ring-Mini-2.0! Tiny but mighty models, unlocked.
- ExLlamaV3 v0.0.11 = faster inference, smoother text flow.
- Triton-Windows updated to 3.5.0.post21 β better CUDA perf on Windows rigs.
Portable builds are now even easier:
π₯ Download β π¦ Unzip β πΎ Copy your old `user_data` folder in β β All models, themes, and settings preserved. No pip. No venvs. Just AI magic.
Pick your build:
- NVIDIA? β `cuda12.4` (new) or `cuda11.7` (legacy)
- AMD/Intel? β Use `vulkan`
- CPU-only? β `cpu` build
- Mac? β `macos-arm64` (M-series) or `macos-x86_64`
No install. No fuss. Just drop-in, run, and chat with your LLMs like never before. π
