Category: AI

AI Releases

  • Home Assistant Voice Pe – 25.12.0

    Home Assistant Voice Pe – 25.12.0

    Home Assistant Voice PE just dropped v25.12.0 ๐Ÿš€ โ€” and itโ€™s a game-changer for offline voice control!

    Now backed by the Open Home Foundation, this ESPHome-powered gem lets you command your smart home without the cloud โ€” perfect for privacy-first tinkerers.

    Biggest update? Music Assistantโ€™s Sendspin multi-room audio support (public preview)! Sync tunes across speakers in real-time โ€” think whole-home playlists, party mode activated by voice. ๐ŸŽถ

    Shoutout to @theHacker for their first contribution โ€” welcome to the crew! ๐Ÿ™Œ

    78 releases strong and growing.

    Full changelog: 25.11.0…25.12.0

    Your home just got smarterโ€ฆ and louder.

    ๐Ÿ”— View Release

  • ComfyUI – v0.5.0

    ComfyUI – v0.5.0

    ComfyUI v0.5.0 just landed โ€” and itโ€™s a game-changer for workflow builders ๐ŸŽจโœจ

    • LatentUpscale Node โ€” Upscale in latent space first for smoother, cleaner details with less noise.
    • Smarter Node Search โ€” Type “upscale” and it gets you. No more scrolling through 50 nodes.
    • Memory Management Upgrade โ€” Fewer crashes, better GPU predictability when running big models.
    • Custom Nodes Overhauled โ€” Install, update, and manage third-party nodes like a pro. No more “why isnโ€™t it showing up?!”
    • Dark Mode 2.0 โ€” Sleeker, higher contrast. Your eyes wonโ€™t scream after midnight prompt sessions.
    • Save as Template โ€” Turn your favorite chains into reusable templates. Perfect for teams or recurring styles.

    Plus: 40+ bugs squashed, including the dreaded “random node disconnects on reload.” ๐Ÿžโœ…

    If youโ€™re deep in Stable Diffusion workflows โ€” this update is your new best friend. Go tweak, test, and create! ๐Ÿš€

    https://www.comfy.org/

    ๐Ÿ”— View Release

  • MLX-LM – v0.29.0

    MLX-LM – v0.29.0

    ๐Ÿš€ MLX LM v0.29.0 is live โ€” and itโ€™s a beast!

    • Batch generation just got 2x faster thanks to `wired_limit` fixes โ€” your server will thank you.
    • RoPE & SuScaledRoPE fixed for `rnj-1` and others โ€” smoother attention, less drift.
    • Dequantize bug squashed โœ… Now using the right function โ€” cleaner outputs, better precision.
    • Repetition penalty defaults to 0.0 โ€” less annoying repetition from day one. ๐ŸŽฏ
    • DSV32 & Gemma3 โ€” bugs gone, stable and ready to deploy.
    • SSM batching fixed โ€” state-space models now behave on the server. ๐Ÿ’ก
    • Nemotron 3 added! ๐ŸŽ‰ Go ahead, test it.
    • Devstral-2 now works properly โ€” no more surprises. ๐Ÿ‘

    Big shoutout to first-time contributors: @otarkhan, @devnamrits, @DePasqualeOrg, and @inferencers โ€” welcome to the crew! ๐Ÿ™Œ

    Update now โ€” your LLMs are ready for a speed run. ๐Ÿ› ๏ธ

    Full changelog: [v0.28.4…v0.29.0](link)

    ๐Ÿ”— View Release

  • Ollama – v0.13.4

    Ollama – v0.13.4

    ๐Ÿš€ Ollama v0.13.4 just dropped โ€” tiny update, big impact for Nemotron users!

    Fix: The `think` token now actually listens to your custom prompts. No more ignored reasoning directives โ€” if you told the model to “think step by step,” it finally will. ๐Ÿค”โœจ

    Perfect for RAG builders, prompt engineers, and anyone relying on structured reasoning. No new features, just cleaner, more reliable internal reflection.

    Upgrade if youโ€™re using Nemotron models โ€” your prompts just got smarter. ๐Ÿ› ๏ธ

    ๐Ÿ”— View Release

  • Ollama – v0.13.4-rc2

    Ollama – v0.13.4-rc2

    ๐Ÿš€ Ollama v0.13.4-rc2 just droppedโ€”and itโ€™s all about speed and stability!

    If youโ€™ve ever sat there watching “Loading modelโ€ฆ” like itโ€™s a slow-loading game, this update is your cheat code.

    โœจ Whatโ€™s new:

    • โšก Faster model init โ€” Memory mapping + CUDA context tweaks slash startup time.
    • ๐Ÿค– Multi-GPU love โ€” Better resource splitting so your 2x or 4x GPUs actually work together, not fight.
    • ๐Ÿ› ๏ธ Smarter cache โ€” Fewer crashes from corrupted downloads or interrupted pulls.

    ๐Ÿ’ก Pro tip: Running Llama 3 or Mistral on a multi-GPU rig? You could see 20%+ faster load times.

    Still a release candidateโ€”but if youโ€™re ready to cut the wait time and boost reliability, upgrade now. Your GPUโ€™s doing backflips in gratitude. ๐Ÿค–๐Ÿ’ป

    ๐Ÿ”— View Release

  • Deep-Live-Cam – 2.4

    Deep-Live-Cam – 2.4

    Deep-Live-Cam 2.4 just droppedโ€”and itโ€™s a game-changer for real-time face swaps ๐ŸŽญ๐Ÿ’ป

    โœ… Dropdowns fixed โ€” No more menu stutters. Smooth sailing through all your deepfake edits.

    โšก Forced GPU mode on laptops โ€” Finally, your iGPU wonโ€™t hold you back. Full CUDA/CoreML/DirectML power unlocked.

    ๐ŸŽจ Poisson Blending upgraded โ€” Translucent artifacts? Vanished. Messy ear edges? Gone. Swaps now look planted, not pasted.

    ๐Ÿ‘„ Mouthmask optimized for Inswapper โ€” Lips sync perfectly with speech. No more robot-smiles. Realistic expressions, zero effort.

    Plus tiny but massive polish tweaksโ€”because your deepfake shouldnโ€™t have glitches.

    All updates live only on the official site. Grab it before your GPU starts sending you love letters ๐Ÿ’ฌ๐Ÿ”ฅ

    ๐Ÿ”— View Release

  • Text Generation Webui – v3.21

    Text Generation Webui – v3.21

    ๐Ÿš€ Text Generation WebUI v3.21 just dropped โ€” and itโ€™s lighter, faster, smarter!

    The portable builds are now leaner: no more bloated llama.cpp symlinks (Python .whl quirks, we see you ๐Ÿ˜…). They auto-recreate on first launch โ€” clean, efficient, zero hassle.

    ๐Ÿ”ฅ Backend upgrades galore:

    • llama.cpp โ†’ updated to latest ggml-org commit (5c8a717) โ€” smoother inference, fewer crashes
    • ExLlamaV3 v0.0.18 โ€” better quantization + smarter memory use
    • safetensors v0.7 โ€” faster load times, tighter security
    • triton-windows 3.5.1.post22 โ€” CUDA ops on Windows? Smoother than ever

    ๐Ÿ“ฆ Portable builds now come in 4 flavors:

    • ๐Ÿ–ฅ๏ธ `cuda12.4` (NVIDIA)
    • ๐Ÿ’ป `vulkan` (AMD/Intel GPUs)
    • ๐Ÿง  `cpu` (no GPU? no problem)
    • ๐Ÿ `macos-arm64` (Apple Silicon optimized)

    ๐Ÿ”„ Update? Just unzip โ†’ replace only your `user_data/` folder. All your models, settings, themes โ€” untouched. No reconfiguring. No stress.

    Perfect for tinkerers who want power without the install drama. Grab it, unzip, and start generating ๐Ÿš€

    ๐Ÿ”— View Release

  • Ollama – v0.13.4-rc1

    Ollama – v0.13.4-rc1

    ๐Ÿš€ Ollama v0.13.4-rc1 just dropped โ€” and Gemma 3 just got a serious upgrade!

    Gemma 3โ€™s RoPE scaling is now set to 1.0, meaning:

    โœ… Better long-context handling

    โœ… Fewer hallucinations

    โœ… More stable reasoning on laptops and low-resource rigs

    Perfect for RAG apps, chatbots, or code assistants that need to stay sharp on long prompts.

    Under the hood:

    ๐Ÿ”ง Smoother model loading

    ๐Ÿง  Minor memory optimizations

    โšก Improved compatibility with newer GPU drivers

    This is a release candidate โ€” stable, tested, and ready for early adopters who want to ride the wave before it hits mainline.

    Grab it, tweak your prompts, and start gemma-ing harder ๐Ÿค–โœจ

    ๐Ÿ”— View Release

  • Ollama – v0.13.4-rc0

    Ollama – v0.13.4-rc0

    ๐Ÿš€ Ollama v0.13.4-rc0 is live โ€” and itโ€™s a game-changer for Gemma 3 users!

    Fixed: Global RoPE scale values are now properly applied, meaning longer prompts get way more accurate reasoning and smoother context handling. No more weird token scaling glitches โ€” your Gemma 3 finally behaves like it should.

    โœ… Verified release (GitHub-signed for trust & security)

    โœ… Smoother, more consistent inference across macOS, Windows, and Linux

    If youโ€™re running Gemma 3 locally? Update now. This isnโ€™t just a patch โ€” itโ€™s your model finally unlocking its full potential. ๐Ÿค–โœจ

    Grab it before the final drops โ€” and keep those LLMs running clean!

    ๐Ÿ”— View Release

  • Mantella – v0.14 Preview 1

    Mantella – v0.14 Preview 1

    Hey AI tinkerers! ๐Ÿš€

    Mantella v0.14_preview_1 just dropped โ€” and itโ€™s turning Skyrim and Fallout 4 NPCs into living, breathing conversationalists.

    • ๐Ÿ—ฃ๏ธ Auto greetings in group chats โ€” NPCs now start convos when someone new walks in. No more awkward silences.
    • ๐Ÿ’ฌ First sentence always speaks โ€” even if you interrupt mid-sentence. Dialogue flows like real life.
    • ๐Ÿšซ Fast responses disabled in radiant dialogue โ€” less spam, more soul. Perfect for immersive storytelling.
    • ๐ŸŽฎ New advanced actions โ€” trigger NPCs based on proximity, state, or events. Build smarter worlds.
    • โš™๏ธ Standalone LLM framework for function calling โ€” cleaner, more powerful logic behind the scenes.
    • Italics now parse correctly โ€” emphasis and lore-heavy lines? Done right.
    • ๐Ÿ“… Day tracking in prompts โ€” NPCs remember how long youโ€™ve been gone. World feels alive.
    • ๐Ÿ”ง All tool parameters exposed โ€” no black boxes. Tinker to your heartโ€™s content.

    This isnโ€™t just an updateโ€”itโ€™s the future of in-game dialogue. Go make NPCs that feel real. ๐ŸŒŸ

    ๐Ÿ”— View Release