Text Generation Webui – v3.19
🚀 Text Generation WebUI v3.19 just dropped—and it’s a game-changer for MoE lovers!
Qwen3-Next is now fully supported in llama.cpp, with massive speed gains on both full GPU and hybrid CPU/GPU setups. Say goodbye to slow MoE inference!
✨ New features:
- 🎛️ –ubatch-size slider — fine-tune batch performance like a pro
- 🚀 Optimized defaults for MoE efficiency out of the box
🔧 Backend upgrades:
- llama.cpp updated to latest ggml-org (ff55414) → Qwen3-Next ✅
- ExLlamaV3 bumped to v0.0.16
- coqui-tts now compatible with Transformers 4.55
📦 PORTABLE BUILDS ARE LIVE!
No install. No fuss. Just download, unzip, run:
- NVIDIA → `cuda12.4`
- AMD/Intel GPU → `vulkan`
- CPU only → `cpu`
- Apple Silicon Mac → `macos-arm64`
💡 Upgrading?
Grab the new zip → paste your old `user_data` folder in → all models, settings, and custom themes stay perfectly intact.
Go break some MoE speed records. 🤖💥
