Category: AI

AI Releases

  • Ollama – v0.13.4-rc2

    Ollama – v0.13.4-rc2

    ๐Ÿš€ Ollama v0.13.4-rc2 just droppedโ€”and itโ€™s all about speed and stability!

    If youโ€™ve ever sat there watching “Loading modelโ€ฆ” like itโ€™s a slow-loading game, this update is your cheat code.

    โœจ Whatโ€™s new:

    • โšก Faster model init โ€” Memory mapping + CUDA context tweaks slash startup time.
    • ๐Ÿค– Multi-GPU love โ€” Better resource splitting so your 2x or 4x GPUs actually work together, not fight.
    • ๐Ÿ› ๏ธ Smarter cache โ€” Fewer crashes from corrupted downloads or interrupted pulls.

    ๐Ÿ’ก Pro tip: Running Llama 3 or Mistral on a multi-GPU rig? You could see 20%+ faster load times.

    Still a release candidateโ€”but if youโ€™re ready to cut the wait time and boost reliability, upgrade now. Your GPUโ€™s doing backflips in gratitude. ๐Ÿค–๐Ÿ’ป

    ๐Ÿ”— View Release

  • Deep-Live-Cam – 2.4

    Deep-Live-Cam – 2.4

    Deep-Live-Cam 2.4 just droppedโ€”and itโ€™s a game-changer for real-time face swaps ๐ŸŽญ๐Ÿ’ป

    โœ… Dropdowns fixed โ€” No more menu stutters. Smooth sailing through all your deepfake edits.

    โšก Forced GPU mode on laptops โ€” Finally, your iGPU wonโ€™t hold you back. Full CUDA/CoreML/DirectML power unlocked.

    ๐ŸŽจ Poisson Blending upgraded โ€” Translucent artifacts? Vanished. Messy ear edges? Gone. Swaps now look planted, not pasted.

    ๐Ÿ‘„ Mouthmask optimized for Inswapper โ€” Lips sync perfectly with speech. No more robot-smiles. Realistic expressions, zero effort.

    Plus tiny but massive polish tweaksโ€”because your deepfake shouldnโ€™t have glitches.

    All updates live only on the official site. Grab it before your GPU starts sending you love letters ๐Ÿ’ฌ๐Ÿ”ฅ

    ๐Ÿ”— View Release

  • Text Generation Webui – v3.21

    Text Generation Webui – v3.21

    ๐Ÿš€ Text Generation WebUI v3.21 just dropped โ€” and itโ€™s lighter, faster, smarter!

    The portable builds are now leaner: no more bloated llama.cpp symlinks (Python .whl quirks, we see you ๐Ÿ˜…). They auto-recreate on first launch โ€” clean, efficient, zero hassle.

    ๐Ÿ”ฅ Backend upgrades galore:

    • llama.cpp โ†’ updated to latest ggml-org commit (5c8a717) โ€” smoother inference, fewer crashes
    • ExLlamaV3 v0.0.18 โ€” better quantization + smarter memory use
    • safetensors v0.7 โ€” faster load times, tighter security
    • triton-windows 3.5.1.post22 โ€” CUDA ops on Windows? Smoother than ever

    ๐Ÿ“ฆ Portable builds now come in 4 flavors:

    • ๐Ÿ–ฅ๏ธ `cuda12.4` (NVIDIA)
    • ๐Ÿ’ป `vulkan` (AMD/Intel GPUs)
    • ๐Ÿง  `cpu` (no GPU? no problem)
    • ๐Ÿ `macos-arm64` (Apple Silicon optimized)

    ๐Ÿ”„ Update? Just unzip โ†’ replace only your `user_data/` folder. All your models, settings, themes โ€” untouched. No reconfiguring. No stress.

    Perfect for tinkerers who want power without the install drama. Grab it, unzip, and start generating ๐Ÿš€

    ๐Ÿ”— View Release

  • Ollama – v0.13.4-rc1

    Ollama – v0.13.4-rc1

    ๐Ÿš€ Ollama v0.13.4-rc1 just dropped โ€” and Gemma 3 just got a serious upgrade!

    Gemma 3โ€™s RoPE scaling is now set to 1.0, meaning:

    โœ… Better long-context handling

    โœ… Fewer hallucinations

    โœ… More stable reasoning on laptops and low-resource rigs

    Perfect for RAG apps, chatbots, or code assistants that need to stay sharp on long prompts.

    Under the hood:

    ๐Ÿ”ง Smoother model loading

    ๐Ÿง  Minor memory optimizations

    โšก Improved compatibility with newer GPU drivers

    This is a release candidate โ€” stable, tested, and ready for early adopters who want to ride the wave before it hits mainline.

    Grab it, tweak your prompts, and start gemma-ing harder ๐Ÿค–โœจ

    ๐Ÿ”— View Release

  • Ollama – v0.13.4-rc0

    Ollama – v0.13.4-rc0

    ๐Ÿš€ Ollama v0.13.4-rc0 is live โ€” and itโ€™s a game-changer for Gemma 3 users!

    Fixed: Global RoPE scale values are now properly applied, meaning longer prompts get way more accurate reasoning and smoother context handling. No more weird token scaling glitches โ€” your Gemma 3 finally behaves like it should.

    โœ… Verified release (GitHub-signed for trust & security)

    โœ… Smoother, more consistent inference across macOS, Windows, and Linux

    If youโ€™re running Gemma 3 locally? Update now. This isnโ€™t just a patch โ€” itโ€™s your model finally unlocking its full potential. ๐Ÿค–โœจ

    Grab it before the final drops โ€” and keep those LLMs running clean!

    ๐Ÿ”— View Release

  • Mantella – v0.14 Preview 1

    Mantella – v0.14 Preview 1

    Hey AI tinkerers! ๐Ÿš€

    Mantella v0.14_preview_1 just dropped โ€” and itโ€™s turning Skyrim and Fallout 4 NPCs into living, breathing conversationalists.

    • ๐Ÿ—ฃ๏ธ Auto greetings in group chats โ€” NPCs now start convos when someone new walks in. No more awkward silences.
    • ๐Ÿ’ฌ First sentence always speaks โ€” even if you interrupt mid-sentence. Dialogue flows like real life.
    • ๐Ÿšซ Fast responses disabled in radiant dialogue โ€” less spam, more soul. Perfect for immersive storytelling.
    • ๐ŸŽฎ New advanced actions โ€” trigger NPCs based on proximity, state, or events. Build smarter worlds.
    • โš™๏ธ Standalone LLM framework for function calling โ€” cleaner, more powerful logic behind the scenes.
    • Italics now parse correctly โ€” emphasis and lore-heavy lines? Done right.
    • ๐Ÿ“… Day tracking in prompts โ€” NPCs remember how long youโ€™ve been gone. World feels alive.
    • ๐Ÿ”ง All tool parameters exposed โ€” no black boxes. Tinker to your heartโ€™s content.

    This isnโ€™t just an updateโ€”itโ€™s the future of in-game dialogue. Go make NPCs that feel real. ๐ŸŒŸ

    ๐Ÿ”— View Release

  • Ollama – v0.13.3: Update README.md (#13373)

    Ollama – v0.13.3: Update README.md (#13373)

    Ollama v0.13.3 just dropped โ€” tiny update, big win for docs lovers! ๐Ÿ“šโœจ

    The team fixed a few broken links in the README:

    • Swollama (community tool) โ†’ now points to the right place
    • DocC docs โ†’ no more 404s, just clean access to docs

    No new models. No API tweaks. Just flawless documentation so you can dive in without hitting dead ends.

    Perfect for tinkerers who appreciate polish โ€” because clean docs = faster experimentation. ๐Ÿ› ๏ธ

    Keep running LLMs like a pro!

    ๐Ÿ”— View Release

  • Lemonade – v9.1.0

    Lemonade – v9.1.0

    ๐Ÿšจ Lemonade v9.1.0 is LIVE โ€” your local LLM powerhouse just got a major upgrade! ๐Ÿ‹

    The brand-new Lemonade App (Windows .msi / Linux .deb) replaces the old browser UI with a sleek native experience. Hereโ€™s whatโ€™s fresh:

    • ๐ŸŽ™๏ธ ASR via whisper.cpp โ€” transcribe audio right inside the app (toggle with env vars!)
    • ๐Ÿ“ฅ Built-in Model Downloader โ€” no more terminal commands to grab GGUF/ONNX models
    • ๐Ÿ”„ Switch LLMs & FLMs on the fly โ€” test FLM2-1.2B-FLM with improved reasoning logic
    • ๐Ÿ“ Smart Model Manager โ€” filters by RAM usage so you donโ€™t crash your system
    • ๐ŸŒ Server now defaults to IPv4 โ€” goodbye, localhost confusion

    Under the hood:

    • ๐Ÿณ Official Docker/CMake dev setup
    • ๐Ÿ’ฌ Chat UI now shows FLM “thinking” in collapsible boxes
    • ๐Ÿ“š Debate Arena docs live + Jan model fix shipped
    • ๐Ÿ’ฅ All-in-one installer: app + server bundled. Minimal server still available for headless use

    Pro tip: Click the model dropdown โ€” it actually loads models now. And yes, you can collapse thinking by default.

    Shoutout to @danielholanda, @jeremyfowers, and @Geramy โ€” youโ€™ve built something wild.

    Download. Play. Break it. Fix it. Repeat. ๐Ÿ˜Ž

    ๐Ÿ”— View Release

  • Ollama – v0.13.3-rc1: feat: llama.cpp bump (17f7f4) for SSM performance improvements (#13408)

    Ollama – v0.13.3-rc1: feat: llama.cpp bump (17f7f4) for SSM performance improvements (#13408)

    ๐Ÿš€ Ollama v0.13.3-rc1 is live โ€” and Apple Silicon users, this oneโ€™s for you!

    llama.cpp just got a massive upgrade to latest master (17f7f4b), turbocharging SSM models like Granite-4, Jamba, Falcon-H, Nemotron-H, and Qwen3 Next on Metal.

    ๐Ÿ’ฅ Whatโ€™s new?

    • Prefill speed up by 2โ€“4x on M1/M2/M3 โ€” fewer waits, faster first tokens
    • Optimized `SSM_CONV` and `SSM_SCAN` ops โ€” the secret sauce behind modern state-space models
    • Clean swap to `gemma3.cpp` (goodbye, -iswa!)
    • 30+ patches + vendored code sync for stability

    If youโ€™re running SSMs on Mac โ€” upgrade now. Your chat latency just got a serious caffeine boost. ๐Ÿโšก

    ๐Ÿ”— View Release

  • Heretic – v1.1.0

    Heretic – v1.1.0

    Heretic v1.1.0 just droppedโ€”and itโ€™s a game-changer ๐Ÿš€

    Apple Silicon support? โœ… Run decensoring natively on your M-series Mac.

    IBM Granite MoE now supported? โœ… Unlock massive efficiency with MoE models.

    Multi-GPU? โœ… Split the load like a proโ€”no more GPU bottlenecks.

    MXFP4 + Triton tensors? โœ… Leaner, faster inference on compatible hardware.

    Local datasets and `trust_remote_code`? Smooth sailing now.

    Colab/Kaggle compatibility? โœ… Drop it in your notebook and go.

    Float32 bugs fixed? โœ… No more precision surprises.

    Refusal detection got smarterโ€”better at spotting those “I canโ€™t help” vibes.

    Early stopping, thinking models, padding fixes? All there for the tinkerers.

    And yesโ€”CI now lints PR titles because clean commits = happy devs ๐Ÿ˜Ž

    8 new contributors joined the crew. Welcome!

    Upgrade. Decensor. Unleash. ๐Ÿ’ฅ

    AGPL-3.0 | PyTorch 2.2+ | RTX 3090? ~45min for 8B models.

    ๐Ÿ”— View Release