Text Generation Webui – v3.14

Text Generation Webui – v3.14

✨ Oobabooga’s Text Generation Web UI – v3.14 Update ✨

Hey AI Enthusiasts! 🤖 The popular local LLM interface, Text-Generation-WebUI, just dropped v3.14 with some solid improvements:

  • Multi-GPU Performance: Enhanced `bitsandbytes` support for faster inference on multi-GPU setups (8 & 4-bit quantization).
  • ExLlamaV3 Integration: New `/v1/internal/logits` endpoint added for advanced integrations using `exllamav3` and `exllamav3_hf`.
  • Qwen Support: Now supports Qwen 3-Next models with ExLlamaV3 (requires `flato`).
  • llama.cpp Update: Upgraded to the latest ggml-org/llama.cpp.
  • Dependency Updates: `transformers` (v4.57), `exllamav3` (v0.0.7), and `bitsandbytes` (v0.48) all updated!

Bug Fixes: Chat history loading & macOS portable build issues resolved.

📦 Portable Builds Available: Grab the latest version for your setup: https://github.com/oobabooga/text-generation-webui/releases/tag/v3.14

  • NVIDIA (cuda12.4, cuda11.7)
  • AMD/Intel (vulkan)
  • CPU Only
  • Mac (Apple Silicon & Intel)

Happy generating! ✨

https://github.com/oobabooga/text-generation-webui/releases/tag/v3.14