Text Generation Webui – v3.19
π Text Generation WebUI v3.19 just droppedβand itβs a game-changer for MoE lovers!
Qwen3-Next is now fully supported in llama.cpp, with massive speed gains on both full GPU and hybrid CPU/GPU setups. Say goodbye to slow MoE inference!
β¨ New features:
- ποΈ –ubatch-size slider β fine-tune batch performance like a pro
- π Optimized defaults for MoE efficiency out of the box
π§ Backend upgrades:
- llama.cpp updated to latest ggml-org (ff55414) β Qwen3-Next β
- ExLlamaV3 bumped to v0.0.16
- coqui-tts now compatible with Transformers 4.55
π¦ PORTABLE BUILDS ARE LIVE!
No install. No fuss. Just download, unzip, run:
- NVIDIA β `cuda12.4`
- AMD/Intel GPU β `vulkan`
- CPU only β `cpu`
- Apple Silicon Mac β `macos-arm64`
π‘ Upgrading?
Grab the new zip β paste your old `user_data` folder in β all models, settings, and custom themes stay perfectly intact.
Go break some MoE speed records. π€π₯
