Text Generation Webui – v3.19

Text Generation Webui – v3.19

πŸš€ Text Generation WebUI v3.19 just droppedβ€”and it’s a game-changer for MoE lovers!

Qwen3-Next is now fully supported in llama.cpp, with massive speed gains on both full GPU and hybrid CPU/GPU setups. Say goodbye to slow MoE inference!

✨ New features:

  • πŸŽ›οΈ –ubatch-size slider β€” fine-tune batch performance like a pro
  • πŸš€ Optimized defaults for MoE efficiency out of the box

πŸ”§ Backend upgrades:

  • llama.cpp updated to latest ggml-org (ff55414) β†’ Qwen3-Next βœ…
  • ExLlamaV3 bumped to v0.0.16
  • coqui-tts now compatible with Transformers 4.55

πŸ“¦ PORTABLE BUILDS ARE LIVE!

No install. No fuss. Just download, unzip, run:

  • NVIDIA β†’ `cuda12.4`
  • AMD/Intel GPU β†’ `vulkan`
  • CPU only β†’ `cpu`
  • Apple Silicon Mac β†’ `macos-arm64`

πŸ’‘ Upgrading?

Grab the new zip β†’ paste your old `user_data` folder in β†’ all models, settings, and custom themes stay perfectly intact.

Go break some MoE speed records. πŸ€–πŸ’₯

πŸ”— View Release