Text Generation Webui – v4.6
🚀 Major Update Alert: text-generation-webui v4.6 is here!
If you’re looking for the “AUTOMATIC1111” of LLMs, this is your playground. The latest update brings massive improvements for anyone building agentic workflows or running local models with precision.
🛠️ Precision Tool Calling
Tired of agents running commands without permission? You can now enable Tool call confirmation in the Chat tab. It adds inline approve/reject buttons, giving you total oversight over every command execution.
🔌 Expanded MCP Ecosystem
The Model Context Protocol (MCP) just got a huge boost!
- You can now configure local subprocess-based MCP servers via `user_data/mcp.json`.
- If you’re already using the configuration format from Claude Desktop or Cursor, it’s a seamless transition.
- Tool discovery is now cached to keep your interface snappy.
🧠 Advanced “Chain of Thought” Management
New controls for reasoning models are officially live:
- Use the new UI checkbox or `–preserve-thinking` CLI flag to decide if thinking blocks from previous turns stay in your context.
- The UI is now smarter—it only displays “Reasoning effort” and “Enable thinking” controls when you’re actually using a model that supports them.
🖥️ Smoother UI/UX
- Independent Sidebars: Sidebars now toggle independently and remember their state even after a page refresh.
- Visual Polish: Improved light mode borders, fixed code block copy buttons, and smoother scrolling during model loading.
⚙️ Under the Hood & Security
- Performance: `llama.cpp` now defaults to `–draft-min 48` for better speculative decoding performance.
- Security: Critical SSRF vulnerability fixes for URL fetching are included to keep your local environment safe.
- Bug Fixes: Resolved issues with Gemma 4 thinking tags and UI token leaks during tool calls.
Time to fire up those local models and test out these new agentic features! 🛠️✨
