Text Generation Webui – v3.14
✨ Oobabooga’s Text Generation Web UI – v3.14 Update ✨
Hey AI Enthusiasts! 🤖 The popular local LLM interface, Text-Generation-WebUI, just dropped v3.14 with some solid improvements:
- Multi-GPU Performance: Enhanced `bitsandbytes` support for faster inference on multi-GPU setups (8 & 4-bit quantization).
- ExLlamaV3 Integration: New `/v1/internal/logits` endpoint added for advanced integrations using `exllamav3` and `exllamav3_hf`.
- Qwen Support: Now supports Qwen 3-Next models with ExLlamaV3 (requires `flato`).
- llama.cpp Update: Upgraded to the latest ggml-org/llama.cpp.
- Dependency Updates: `transformers` (v4.57), `exllamav3` (v0.0.7), and `bitsandbytes` (v0.48) all updated!
Bug Fixes: Chat history loading & macOS portable build issues resolved.
📦 Portable Builds Available: Grab the latest version for your setup: https://github.com/oobabooga/text-generation-webui/releases/tag/v3.14
- NVIDIA (cuda12.4, cuda11.7)
- AMD/Intel (vulkan)
- CPU Only
- Mac (Apple Silicon & Intel)
Happy generating! ✨
https://github.com/oobabooga/text-generation-webui/releases/tag/v3.14