Text Generation Webui – v4.3
๐จ Text-Generation-WebUI v4.3 is live! ๐จ
Hey AI tinkerers & devs โ fresh update dropped, and itโs packed with performance wins, new backends, and security upgrades. Hereโs the lowdown:
—
๐น ๐ฅ Brand-new backend: `ik_llama.cpp`
A high-octane fork by the imatrix creator, now baked into TGWU:
- โ
New quant formats (Q4_K_M, Q6_K, etc.)
- ๐ง Hadamard-based KV cache quantization โ way more accurate, on by default
- โก Built for MoE models & CPU inference (yes, really fast)
โ Grab it via `textgen-portable-ik` or `–ik` flag!
—
๐น ๐ง API upgrades (OpenAI-compatible!)
The `/v1/completions` endpoint now supports:
- `echo`: Returns prompt + completion in one go
- `logprobs`: Token-level log probabilities (prompt & generated)
- `top_logprobs_ids`: Top token IDs per position โ perfect for probing model confidence ๐ฏ
—
๐น ๐จ Gradio UX + Security Boost
- ๐ Custom Gradio fork = ~50ms faster UI interactions
- ๐ Fixed ACL bypass (Windows/macOS path quirks)
- โ
Server-side validation for Dropdown/Radio/CheckboxGroup
- ๐ก๏ธ SSRF fix in superbooga: blocks internal/private IPs
—
๐น ๐ง Bug fixes & polish
- `–idle-timeout` now works for encode/decode + parallel generations โ
- Stopping strings fixed (e.g., `<|return|>` vs `<|result|>`)
- Qwen3.5 MoE loads cleanly via ExLlamaV3_HF
- `ban_eos_token` finally works (EOS suppression at logit level)
—
๐น ๐ฆ Dependency upgrades
- ๐ฆ `llama.cpp` โ latest (`a1cfb64`) + Gemma-4 support
- ๐ `ExLlamaV3` โ v0.0.28
- ๐ฆ `transformers` โ 5.5
- โจ Auto-detects `torch_dtype` from model config (override with `–bf16`)
- ๐๏ธ Removed obsolete `models/config.yaml` โ templates pulled from model metadata now
—
๐น ๐ Terminology update
“Truncation length” โ now “context length” in logs (more accurate, less confusing!)
—
๐น ๐ฆ Portable builds โ GGUF-ready & zero-install
| Platform | Build to Use |
|———-|————–|
| NVIDIA (old driver) | `cuda12.4` |
| NVIDIA (new driver, CUDA >13) | `cuda13.1` |
| AMD/Intel GPU | `vulkan` |
| AMD (ROCm) | `rocm` |
| CPU-only | `cpu` |
| Apple Silicon | `macos-arm64` |
| Intel Mac | `macos-x86_64` |
๐ Updating? Just swap the folder โ keep `user_data/`, and now you can even move it one level up for shared use across versions ๐
—
Let me know if you want a quick-start walkthrough on `ik_llama.cpp` or portable builds! ๐ ๏ธ๐
๐ View Release