• ComfyUI – v0.20.2

    ComfyUI – v0.20.2

    ComfyUI v0.20.2 is officially live! ๐Ÿš€

    If youโ€™re a fan of node-based wizardry, you know ComfyUI is the ultimate playground for building complex Stable Diffusion pipelines without touching a line of code. It’s incredibly modular, making it a go-to for anyone wanting to orchestrate everything from SDXL generation to intricate ControlNet workflows.

    This latest minor update brings some sweet new compatibility to your node graphs:

    • OneTainer ERNIE LoRA Support: The big news here is the integration of OneTainer ERNIE LoRAs! This makes it much smoother to plug these specific fine-tuned models directly into your existing workflows. ๐Ÿ› ๏ธ

    Whether you’re upscaling, inpainting, or experimenting with new LCM models, this update keeps your toolkit expanding. Happy tinkering!

    ๐Ÿ”— View Release

  • Ollama – v0.23.0

    Ollama – v0.23.0

    Ollama v0.23.0 is officially live! ๐Ÿš€

    If you aren’t running Ollama yet, you are missing out on the gold standard for local LLM orchestration. Itโ€™s the ultimate toolkit for pulling and running heavy hitters like Llama 3, DeepSeek-R1, and Mistral directly on your hardwareโ€”no cloud subscriptions or API keys needed.

    The team is moving at lightning speed, and this latest update brings some great refinements to your local workflow:

    • Claude-style Integration: This release introduces significant backend work to support Claude-style application structures, making it even easier to integrate sophisticated prompting patterns into your local setups.
    • Enhanced Stability: A major focus of this version is refining the launch processes for new model types, ensuring that when you pull a fresh architecture, it runs smoothly without a hitch.

    Whether you’re building a private RAG pipeline or just experimenting with the latest open-source weights, this update keeps your local inference engine rock solid. ๐Ÿ› ๏ธ

    ๐Ÿ”— View Release

  • Ollama – v0.23.0-rc0

    Ollama – v0.23.0-rc0

    Ollama just dropped a fresh release candidate, v0.23.0-rc0, and itโ€™s looking like a major milestone for anyone running local LLMs! ๐Ÿš€

    If you aren’t using Ollama yet, it is the ultimate framework for getting models like Llama 3, DeepSeek-R1, and Mistral up and running on your own hardware without needing a massive cloud budget. It handles all the heavy lifting of downloading and configuring models so you can focus on building.

    Whatโ€™s new in this release:

    • Claude App Integration: This update includes significant work regarding the launch of Claude app support! The team is clearly focused on expanding how different model architectures and interfaces interact within the Ollama ecosystem. ๐Ÿค–
    • Release Candidate Status: Since this is an `rc0` build, itโ€™s the perfect playground for us tinkerers to test out the new plumbing and catch any bugs before the stable version hits the mainstream.

    This is a great time to pull the latest build and see how these architectural updates affect your local workflows! ๐Ÿ› ๏ธ

    ๐Ÿ”— View Release

  • Ollama – v0.22.1

    Ollama – v0.22.1

    Ollama just dropped v0.22.1, and itโ€™s a quick but tasty update for anyone running local LLMs! ๐Ÿฅ”

    If you haven’t tried Ollama yet, it is the ultimate toolkit for running powerful models like Llama 3, DeepSeek-R1, and Mistral directly on your own hardware without needing a cloud subscription. It handles all the heavy lifting of downloading and configuring models so you can focus on building.

    Here is whatโ€™s new in this release:

    • Gemma 4 Support: The star of this update is an updated renderer specifically optimized for Gemma 4. This ensures that when you’re pulling Google’s latest lightweight powerhouse, the architecture and output are handled perfectly by the Ollama backend.

    If youโ€™ve been waiting to experiment with the newest Gemma weights, now is the time to pull that update and get tinkering! ๐Ÿ› ๏ธ

    ๐Ÿ”— View Release

  • Ollama – v0.22.1-rc1

    Ollama – v0.22.1-rc1

    Ollama just dropped a fresh release candidate, v0.22.1-rc1, and itโ€™s all about keeping your local LLM experience up to date! ๐Ÿ› ๏ธ

    If you haven’t been playing with Ollama yet, it is the ultimate go-to tool for running large language models locally on your machine with zero friction. It makes pulling and running heavyweights like Llama 3, DeepSeek-R1, or Mistral as easy as a single command.

    Whatโ€™s new in this release:

    • Gemma 4 Support: The big highlight here is an update to the renderers to specifically support the brand-new Gemma 4 model family! ๐Ÿš€

    This is a perfect time to jump in and test out Google’s latest open weights model right on your own hardware. Happy tinkering! ๐Ÿค–

    ๐Ÿ”— View Release

  • Ollama – v0.22.1-rc0: New models (#15861)

    Ollama – v0.22.1-rc0: New models (#15861)

    Ollama just dropped a fresh release candidate (v0.22.1-rc0), and itโ€™s packed with some heavy-hitting model updates and precision improvements! If you’re running local LLMs, this one is definitely worth a look for better quantization and smarter logging. ๐Ÿ› ๏ธ

    Hereโ€™s the lowdown on whatโ€™s new:

    • New Model Support: The team has added support for the Laguna models (via both `mlx` and `ggml`) and implemented support for Nemotron 3 Nano Omni.
    • FP8 Precision Upgrades: A big win for efficiency! Ollama can now import FP8 safetensors. It intelligently handles decoding HF F8_E4M3 weights and uses source-precision metadata to decide the best quantization path (like defaulting FP8-sourced GGUFs to Q8_0). This means better quality when compressing models.
    • Improved Logprobs: The server now preserves `logprobs` during generation, even when using built-in parsers. Previously, logprob-only chunks could get dropped if the parser was buffering content; now, that data stays intact for much more accurate probability tracking. ๐Ÿ“ˆ
    • Poolside Integration: Added integration and updated documentation for Poolside, expanding your local ecosystem options.
    • Performance & Fixes: Includes various performance improvements for review comments, updates to the cache setup, and several bug fixes to keep things running smoothly.

    Time to pull that new image and test out those FP8 weights! ๐Ÿš€

    ๐Ÿ”— View Release

  • Ollama – v0.22.0

    Ollama – v0.22.0

    Ollama v0.22.0 is officially here! ๐Ÿ› ๏ธ

    If you’ve been looking for a way to run heavy-hitting LLMs like Llama 3, DeepSeek-R1, or Mistral directly on your own hardware without the cloud latency, Ollama remains the gold standard for local execution. It handles all the heavy lifting of model management and provides a slick REST API for your custom dev projects.

    Whatโ€™s new in this update:

    • Enhanced Model Support: The library continues to expand, making it even easier to pull and run the latest open-source weights with zero configuration.
    • Performance Optimizations: This release includes under-the-hood tweaks to the inference engine to ensure smoother token generation on both macOS and Linux.
    • Improved CLI Workflow: Smoother management for downloading and switching between different model versions via the command line.

    Whether you’re building a local RAG pipeline or just want a private chatbot that works offline, this update keeps your local ecosystem running at peak performance! ๐Ÿš€

    ๐Ÿ”— View Release

  • Ollama – v0.22.0

    Ollama – v0.22.0

    ๐Ÿš€ Ollama Update Alert! ๐Ÿš€

    If youโ€™re running your local LLMs on Apple Silicon, listen up! The latest release (v0.22.0-rc1) is officially here, and it’s bringing some massive performance optimizations via an MLX update. This is a huge deal for anyone trying to squeeze every bit of juice out of their Mac hardware.

    Hereโ€™s the breakdown of whatโ€™s new:

    • Batch Processing Power: The `mlxrunner` now supports batching the sampler across multiple sequences. If you’re working with large datasets or need to generate multiple outputs at once, this is a massive efficiency win! ๐Ÿ“ˆ
    • NVIDIA & MLX Bridge: In a super cool move for cross-platform workflows, MLX now supports importing models optimized via NVIDIA TensorRT. This makes it way easier to move your heavy-duty workflows between NVIDIA and Apple hardware without the headache.
    • Precision Tokenization: A bug fix for multi-regex BPE offset handling is included, ensuring your tokenization stays precise and error-free during complex text processing tasks.

    Time to pull that update and start benchmarking! ๐Ÿ› ๏ธ

    ๐Ÿ”— View Release

  • Lemonade – v10.3.0: Refine collection image reply behavior (#1726)

    Lemonade – v10.3.0: Refine collection image reply behavior (#1726)

    ๐Ÿ‹ Lemonade SDK v10.3.0 is officially live!

    If youโ€™ve been looking for a way to run high-performance LLMs locally without relying on the cloud, Lemonade is your new best friend. Itโ€™s a powerhouse toolkit designed to squeeze every bit of performance out of your hardware by leveraging NPUs (like AMD Ryzen AI) and GPUs via Vulkan support. Whether you’re using GGUF or ONNX models, it provides an OpenAI-compatible API endpoint so you can swap cloud services for local privacy in a snap.

    The latest update focuses on polishing the collection management experience, specifically making image handling within replies much more predictable and stable. ๐Ÿ› ๏ธ

    Whatโ€™s new in v10.3.0:

    • Refined Image Behavior: No more jumping layouts! Image replies are now properly anchored, and rendering is much more consistent across collections.
    • Smoother Navigation: The logic for scrolling and sizing collection images has been streamlined, making the UI feel much more fluid.
    • Single Source of Truth: To prevent “configuration drift,” the fixed image size (512×256) is now centralized in a single TypeScript constant (`collectionImageConfig.ts`). This keeps your CSS, runtime properties, and build-time tools perfectly synced.
    • Stability Boosts: Includes important bug fixes for the backend manager layout and general system stability improvements.

    Perfect for anyone building custom local AI interfaces or managing large model collections! ๐Ÿš€

    ๐Ÿ”— View Release

  • ComfyUI – v0.20.1

    ComfyUI – v0.20.1

    ComfyUI v0.20.1 ๐ŸŽจ

    If you live for node-based generative workflows, you know ComfyUI is the gold standard for granular control over Stable Diffusion and other models. It’s the ultimate playground for anyone who loves connecting nodes to build complex, custom AI image pipelines!

    This specific update is a quick patch! The developer pushed v0.20.1 primarily to fix a release hiccup caused by GitHub technical issues. ๐Ÿ› ๏ธ

    • Stability Fix: This tiny increment ensures your installation stays stable and correctly tagged following recent GitHub technical glitches.
    • Reliable Deployment: While there aren’t massive new features in this specific patch, itโ€™s a crucial “under the hood” fix to keep your workflows running smoothly without versioning headaches.

    ๐Ÿ”— View Release