Author: Tater Totterson

  • ComfyUI – v0.19.3

    ComfyUI – v0.19.3

    ComfyUI v0.19.3 is officially live! 🚀

    If you’re into building complex, node-based pipelines for Stable Diffusion, you know ComfyUI is the ultimate playground for granular control. Whether you’re working with SDXL, ControlNet, or custom LoRAs, this update keeps your engine running smoothly.

    This release is all about stability and fine-tuning the gears under the hood:

    • Bug Fixes & Refinements: The devs have implemented several under-the-hood tweaks to improve the reliability of node execution and workflow management. No more unexpected hiccups in your logic chains!
    • Performance Polishing: Small but mighty optimizations have been added to ensure that those massive, multi-node graphs run as smoothly as possible—especially during heavy, high-resolution renders.

    Pro-tip: When you pull this update, make sure to check in on your custom nodes too! Keeping your extensions synced with the core update is key to preventing broken workflows. 🛠️

    🔗 View Release

  • ComfyUI – v0.19.2

    ComfyUI – v0.19.2

    New update alert for the node-based wizards! 🛠️

    ComfyUI v0.19.2 is officially out. If you’re deep in the trenches of generative AI, you know ComfyUI is the powerhouse GUI that lets you stitch together complex Stable Diffusion pipelines with surgical precision using a modular, node-based interface.

    While this release focuses on a version bump, keeping your instance updated is critical for:

    • Node Compatibility: Ensuring those essential community custom nodes don’t break your existing workflows.
    • Backend Stability: Maintaining smooth integration with the latest underlying AI models and sampling methods.
    • Performance Tweaks: Staying aligned with the rapid-fire optimization updates coming from the Comfy-Org team.

    Time to pull those latest changes and keep those generations flowing! 🚀

    🔗 View Release

  • Ollama – v0.21.0

    Ollama – v0.21.0

    New update alert for Ollama! 🚨

    If you’re looking to run powerful LLMs like Llama 3, DeepSeek-R1, or Mistral locally on your own hardware, Ollama just made things a bit easier for the Windows crew. This tool is a total game-changer for anyone wanting privacy and control over their AI workflows without relying on the cloud.

    What’s new in v0.21.0:

    • Smoother Windows Setup: The launch command now includes proactive guidance for WSL (Windows Subsystem for Linux). Instead of leaving you to figure out the Linux subsystem on your own, it provides helpful instructions directly within the interface to streamline the installation process. 🛠️

    This update is all about reducing friction for fellow Windows devs and tinkerers getting their local environments up and running! 💻

    🔗 View Release

  • Ollama – v0.21.0-rc1

    Ollama – v0.21.0-rc1

    New update alert for Ollama! 🚨 The latest release candidate, v0.21.0-rc1, is officially out and ready for testing.

    If you’re running local LLMs like Llama 3, DeepSeek-R1, or Mistral, this update focuses on making the setup process much smoother for our Windows-based tinkering sessions. 🛠️

    What’s new in v0.21.0-rc1:

    • Improved Windows WSL Guidance: No more guessing games during setup! The launch command has been updated to provide direct, helpful guidance for the Windows Subsystem for Linux (WSL) right within the interface. This makes transitioning from a native Windows environment to a WSL configuration much more intuitive and less prone to errors.

    This is a great step toward making local AI deployment even more seamless across all platforms. Keep those models running! 🚀

    🔗 View Release

  • Ollama – v0.21.0-rc0

    Ollama – v0.21.0-rc0

    Ollama just dropped a new release candidate, v0.21.0-rc0, and it’s all about fine-tuning performance for some heavy hitters! 🚀

    If you’re looking to run powerful LLMs like Llama 3, DeepSeek-R1, or Gemma locally without a massive cloud budget, this is the tool to keep in your kit. This update specifically targets optimization for those of us tinkering on Apple Silicon.

    What’s new in this release:

    • MLX Optimization: A specific fix has been implemented for the MLX backend regarding Gemma 4 cache usage. It now utilizes a logical view, which streamlines how the model handles data during inference. 🛠️

    This is a big win for memory management! If you are experimenting with the latest Gemma models on Mac, this update ensures much smoother performance and more reliable data handling during your local runs. Happy tinkering! 💻✨

    🔗 View Release

  • Text Generation Webui – v4.5.2

    Text Generation Webui – v4.5.2

    Big news for all the local LLM enthusiasts! The project formerly known as text-generation-webui has officially been rebranded to TextGen! 🚀 Check out the new home at github.com/oobabooga/textgen.

    The latest update (v4.5.2) is packed with stability improvements and critical fixes for those of us playing with the newest models:

    • Gemma 4 Support: Major fixes for tool calling, handling special characters (like quotes and newlines), and improved rendering for thinking blocks in the UI.
    • VRAM Optimization: A much-needed reduction in VRAM peak usage during prompt logprobs forward passes—perfect for squeezing more performance out of your GPU. 🧠
    • UI Refinements: Added a sky-blue color for quoted text in light mode and improved logits display.
    • Bug Squashing: Fixed various issues including chat scroll freezing, tool icon shrinking, and BOS/EOS token overwriting for GGUF models.
    • Dependency Updates: Fresh updates for both `llama.cpp` and `ik_llama.cpp` (the fork with those sweet new quant types).

    Pro-tip for the tinkerers: If you use the portable builds, updating is a breeze! Just extract the new version and swap your `user_data` folder. You can even move `user_data` one level up to share it between different installation folders. Happy generating! 🛠️

    🔗 View Release

  • Text Generation Webui – v4.5.1

    Text Generation Webui – v4.5.1

    Big news for all the local LLM enthusiasts! The project formerly known as text-generation-webui has officially undergone a massive rebranding to TextGen! 🚀

    This latest update (v4.5.1) is all about stability, UI polish, and critical optimizations for the Gemma 4 model family. Whether you are running heavy quantizations or experimenting with complex tool calling, this release brings essential tweaks under the hood to keep your local inference smooth.

    What’s New in This Release:

    • Identity Shift: The project is now officially TextGen! You can find the updated repository at `github.com/oobabooga/textgen`.
    • Gemma 4 Optimization: Significant fixes for Gemma 4 tool calling, including much better handling of quotes and newlines, plus improved rendering for consecutive “thinking” blocks.
    • VRAM Efficiency: A huge win for GPU users! There is a much-needed reduction in VRAM peak usage during the prompt logprobs forward pass—perfect for squeezing more performance out of your hardware. 🧠
    • UI Enhancements: Added a fresh sky-blue color for quoted text in light mode and improved logits display to make debugging easier.
    • Bug Squashing:
    • Fixed chat scroll issues when interacting with “thinking” blocks.
    • Resolved tool icon SVG shrinking during long tool calls.
    • Fixed various BOS/EOS token issues for models lacking specific chat templates.
    • Dependency Updates: Includes fresh updates for both `llama.cpp` and the `ik_llama.cpp` fork, bringing those awesome new quantization types to your workflow.

    Pro-Tip for Tinkerers: 🛠️

    If you use the portable builds, updating is a breeze! Just download the latest version, extract it, and swap your existing `user_data` folder into the new directory. Even better, since version 4.0, you can place `user_data` one level up next to your install folder so multiple versions can share the same models and settings!

    🔗 View Release

  • Tater – Tater v71

    Tater – Tater v71

    🚀 Tater v71 — “Voice, Evolved” is here! 🎤

    Calling all local-LLM tinkerers and automation wizards! The latest update for Tater—your privacy-first, Hydra-powered AI assistant—has officially landed. This release is a massive leap forward in making voice interactions feel less like a command line and more like a natural conversation. If you’ve been running heavy pipelines for STT/TTS, get ready for a much smoother, unified experience.

    What’s new in v7el:

    • Unified Voice Core: Say goodbye to fragmented services! The Voice Core is now fully merged into the main Tater engine. This means significantly less overhead and a much more stable pipeline for your AI workflows. 🧠
    • Lightning-Fast Latency: We’ve optimized the STT (Speech-to-Text) and TTS (Text-to-Speech) loops. With improved end-of-speech detection and early-start TTS experiments, Tater responds much faster than ever before. ⚡
    • Deep Visibility & Debugging: No more mystery lags! New comprehensive voice metrics and enhanced logging allow you to see exactly what is happening under the hood during STT/TTS transitions. 📊
    • Smart Audio Routing: Take control of your soundscape! Tater is now “location aware,” allowing you to tie specific speakers to different assistants or ensure responses play back precisely on the device you used to speak. 🔊
    • System-Wide Speech Integration: Configuration is now unified via the new `Models` tab. Whether it’s a system announcement, an ESPHome voice device, or a Home Assistant media player, everything uses the same shared logic. One config to rule them all! 🛠️
    • Enhanced ESPHome Power: We’ve deepened the integration further, offering improved entity handling and direct control from the UI. Your ESPHome voice devices are now true system inputs. 🧩

    Everything feels tighter, faster, and more intelligent. If you’re looking to level up your local AI stack, it’s time to give v71 a spin! 🚀

    🔗 View Release

  • Text Generation Webui – v4.5

    Text Generation Webui – v4.5

    Big news for the local LLM crowd! The legendary text-generation-webui has officially undergone a rebrand and is now known as TextGen! 🚀 This update brings some much-needed stability and performance tweaks to your local inference workflows.

    Here is what’s new in this release:

    • VRAM & Performance Optimization: There is a reduction in peak VRAM usage during prompt logprobs forward passes. If you are running tight hardware setups or trying to squeeze maximum context into your GPU, this is a massive win! 🧠
    • Improved UI/UX:
    • Reading long conversations just got easier with a new sky-blue color for quoted text in light mode.
    • Significant bug fixes prevent chat scrolling from getting stuck on “thinking” blocks and stop tool icons from shrinking during long calls.
    • Critical Bug Fixes:
    • Gemma-4 Tool Calling: Fixed issues with handling double quotes and newline characters in arguments, ensuring much more reliable agentic behavior. 🛠️
    • Token Management: Resolved issues where BOS/EOS tokens weren’t being set correctly for models lacking chat templates, and fixed duplicate BOS token prepending in ExLlamav3.
    • Under-the-Hood Updates:
    • The project has moved! Find the new home at `github.com/oobabooga/textgen`.
    • Includes the latest versions of `llama.cpp` and `ik_llama.cpp` for better backend support.

    If you’ve been tinkering with tool-calling models or struggling with VRAM spikes, this is a must-have update for your local stack! 💻✨

    🔗 View Release

  • ComfyUI – v0.19.1

    ComfyUI – v0.19.1

    New update alert for the node-based wizards! 🛠️

    ComfyUI v0.19.1 is officially out. For those of you building complex pipelines, ComfyUI remains the powerhouse node-based GUI designed for advanced Stable Diffusion workflows and highly customized generative AI image generation.

    What’s new in this release:

    • Version Bump: The engine has been updated to v0.19.1.
    • Maintenance & Stability: This incremental update focuses on critical bug fixes and performance optimizations. These types of updates are essential for keeping your custom, heavy-duty workflows running smoothly without crashing mid-render.

    Keep those workflows experimental and your nodes organized! 🚀

    🔗 View Release