Category: AI

AI Releases

  • Tater – Tater v38

    Tater – Tater v38

    Tater v38 just droppedโ€”and itโ€™s got Siri on speed dial ๐Ÿฅ”๐ŸŽ™๏ธ

    Ask your HomePod “Whatโ€™s trending on Overseerr?” or “Launch MiSTer” and Tater replies like a chill buddyโ€”short, smart, Siri-optimized. No more robotic jargon. Just clean JSON tool calls and natural voice flow.

    ๐Ÿง  Smarter memory: Each device keeps its own chat history (with TTL caps)โ€”your iPhone wonโ€™t remember you asked for pineapple pizza at 3am.

    ๐Ÿ’ฌ Tighter prompts: Less “according to the model,” more “hey buddy, hereโ€™s whatโ€™s hot.”

    ๐Ÿ”ง New plugins:

    • ๐ŸŽฎ MiSTer Remote โ€” launch your retro games from the couch
    • ๐Ÿ” Web Search โ€” instant answers, zero tabs open
    • ๐ŸŽฅ Overseerr Request & Trending โ€” see whatโ€™s trending and add films with your voice

    Now your kitchen counter has a potato-powered AI that knows your movie queue and your SNES library. No laptop needed. Just ask.

    Check the README for install tipsโ€”Docker or Python, you pick!

    ๐Ÿ”— View Release

  • Ollama – v0.12.10

    Ollama – v0.12.10

    ๐Ÿš€ Ollama v0.12.10 just dropped โ€” quiet, but mighty!

    โœ… Code signing is back on macOS & Windows โ€” safer installs, trusted binaries, zero sketchy downloads.

    ๐Ÿ”ง No flashy new models or UI glitzโ€ฆ just clean, solid housekeeping for devs who run LLMs in production.

    If you care about security, integrity, and smooth local inference โ€” this patch is your unsung hero.

    Keep those Llama 3, DeepSeek, and Mistral models humming! ๐Ÿค–๐Ÿ’ป

    ๐Ÿ”— View Release

  • Text Generation Webui – v3.17

    Text Generation Webui – v3.17

    ๐Ÿ”ฅ Text Generation WebUI v3.17 just dropped โ€” and itโ€™s a beast!

    ๐Ÿ’ฅ Security upgrade: `torch.load()` now defaults to `weights_only=True` โ€” no more sketchy pickle loads in Training_PRO. Safe mode = activated.

    ๐Ÿงฉ Hugging Face fix: Pinning `huggingface-hub==0.36.0` to squash those pesky venv install bugs. No more “why wonโ€™t this install?!” moments.

    ๐Ÿ“Š Training cleanup: `evaluation_strategy` โ†’ `eval_strategy` (thanks, @inyourface34456 โ€” clean code wins again!).

    ๐Ÿš€ Backend power-ups:

    • ๐Ÿ–ผ๏ธ Qwen3-VL support landed in llama.cpp โ€” multimodal models now run natively. Text + images? Yes, please.
    • ๐Ÿš€ `exllamav3` updated to v0.0.12 โ€” faster, smoother, less stuttering. Your GPU will thank you.

    ๐Ÿ“ฆ PORTABLE BUILDS ARE LIVE!

    No install. No fuss. Just download, unzip, run:

    • ๐Ÿ–ฅ๏ธ NVIDIA: `cuda12.4`
    • ๐Ÿ–ฅ๏ธ AMD/Intel: `vulkan`
    • ๐Ÿ’ป CPU-only: `cpu`
    • ๐ŸŽ Apple Silicon: `macos-arm64`
    • ๐Ÿ Intel Mac: `macos-x86_64`

    ๐Ÿ”ง Upgrading?

    Drop your old `user_data/` folder into the new zip โ€” all models, settings, and custom themes preserved.

    Go run Qwen3-VL on your old laptop. Your CPU wonโ€™t know what hit it. ๐Ÿค–๐Ÿ’ฅ

    ๐Ÿ”— View Release

  • Lemonade – v8.2.1

    Lemonade – v8.2.1

    ๐Ÿ”ฅ Lemonade v8.2.1 just dropped โ€” and itโ€™s production-ready now.

    C++ stack got a full glow-up: tray icons, installers, CLI glitches, Ctrl+C crashes โ€” all DEAD. This isnโ€™t a patch; itโ€™s a rebirth.

    ๐Ÿš€ Linux users: `.deb` installer landed! `apt install lemonade` on Ubuntu/Debian โ€” no more compiling from source. Mac support polished too.

    ๐Ÿง  Qwen3 GGUF now has `enable_thinking` โ€” unlock deeper reasoning on complex prompts. Think smarter, not just faster.

    โšก `ryzenai-serve` โ†’ `ryzenai-server`. Now with halt & reasoning, cleaner logs, and smoother shutdowns. AMD Ryzen folks โ€” this oneโ€™s for you.

    ๐Ÿ“Š Real-time memory tracking! Health endpoint shows context size + VRAM usage. No more guessing how much your model is eating.

    ๐Ÿงฉ Model Manager rebuilt: faster, cleaner, smarter. FLM updates? No more stale models sneaking in.

    ๐Ÿ’ป Web UI & CLI got polish: cleaner commands, better streaming logs, no more redundant callbacks. Feels like a real product now.

    ๐Ÿ“ฆ Python 8.2.1 synced too โ€” lm-eval fixes, bigger default context/batch sizes for embeddings.

    ๐Ÿงช Test workflows cleaned up. Fewer false CI failures, more confidence in every release.

    This isnโ€™t just an update โ€” itโ€™s Lemonade stepping out of beta and into the big leagues.

    Grab it. Break it. Tell us how ๐Ÿ˜„

    ๐Ÿ”— Full changelog

    ๐Ÿ”— View Release

  • Ollama – v0.12.10-rc1

    Ollama – v0.12.10-rc1

    ๐Ÿš€ Ollama v0.12.10-rc1 just dropped โ€” and itโ€™s all about trust, not flash!

    ๐Ÿ”’ Code signing is back on macOS & Windows โ€” your downloads are now cryptographically verified, so you know itโ€™s really Ollama, not a sneaky imposter.

    Perfect for enterprises, devs in locked-down environments, or anyone who hates “this file is untrusted” warnings.

    No new models. No UI tweaks. Just clean, quiet hardening โ€” because security shouldnโ€™t be an afterthought.

    ๐Ÿ’ก Pro tip: If you script installs or deploy via CI/CD, your checksums and signature checks are back in action.

    Keep it local. Keep it safe. ๐Ÿ›ก๏ธ

    ๐Ÿ”— View Release

  • Wyoming Openai – Maintenance release (0.3.9)

    Wyoming Openai – Maintenance release (0.3.9)

    ๐Ÿš€ Wyoming OpenAI v0.3.9 is live โ€” lean, mean, and ARM-ready!

    ๐Ÿ”ง CLI & Env Fixes: Now accepts real option names (no more internal ID confusion), and `STT_BACKEND`/`TTS_BACKEND` are case-insensitive. Typos? Clearer errors now!

    ๐ŸŽ ARM64 Support: Docker images built for Raspberry Pi, M1/M2 Macs โ€” no more emulation. Just `docker pull` and go!

    ๐Ÿ“ฆ Deps Upgraded: OpenAI v2.7.1, Ruff, pytest โ€” faster, stabler, future-proofed.

    โš™๏ธ Build Pipeline Upgraded: GitHub Actions now use `build-push-action@v6` with QEMU + Buildx for seamless multi-arch builds.

    ๐Ÿ“– Docs updated to v0.3.9 โ€” no more outdated examples!

    7 files changed, +162/-17 lines. Zero fluff. Pure utility.

    Install via `pip install wyoming-openai` or pull the new Docker image โ€” your Pi (or M2 Mac) will thank you. ๐Ÿง๐ŸŽ

    ๐Ÿ”— View Release

  • ComfyUI – v0.3.68

    ComfyUI – v0.3.68

    ComfyUI v0.3.68 is live โ€” quiet updates, massive stability wins! ๐Ÿ› ๏ธโœจ

    • Fixed a nasty crash when custom nodes had missing deps โ€” no more sudden workflow deaths.
    • Node loading now more resilient: one misbehaving node wonโ€™t bring down your whole pipeline.
    • Smoother UI: tooltips landed in the right place, canvas renders faster even on massive setups.
    • Under-the-hood dependency updates to squash security alerts โ€” clean, quiet, and secure.

    If you run custom nodes or heavy workflows? Update now. The best releases are the ones that justโ€ฆ work. ๐Ÿ˜Œ

    ๐Ÿ”— View Release

  • Ollama – v0.12.9

    Ollama – v0.12.9

    ๐Ÿ’ฅ Ollama v0.12.9-rc0 just dropped โ€” and itโ€™s a GAME CHANGER for CPU-only users!

    No more sluggish LLM inference on your old laptop or cloud instances. This update slays the performance regression thatโ€™s been holding back CPU-based runs.

    โœ… Snappier responses

    โœ… Smoother local workflows

    โœ… Full GGUF + Llama 3, DeepSeek-R1, Phi-4, Mistral support intact

    Perfect for devs prototyping on bare metal or running lightweight models without a GPU. No flashy features โ€” just pure, quiet speed gains. ๐Ÿš€

    Check the changelog โ€” this oneโ€™s a hero update youโ€™ll feel in every token.

    ๐Ÿ”— View Release

  • Ollama – v0.12.9-rc0: ggml: Avoid cudaMemsetAsync during memory fitting

    Ollama – v0.12.9-rc0: ggml: Avoid cudaMemsetAsync during memory fitting

    ๐Ÿš€ Ollama v0.12.9-rc0 just dropped โ€” and itโ€™s a quiet hero for GPU warriors!

    The secret sauce? `ggml` now skips `cudaMemsetAsync` during memory fitting when it hits invalid pointers.

    ๐Ÿ’ก Why it rocks:

    • No more crashes when checking if your 70B model fits on a 24GB GPU
    • Smoother `op_offload` workflows โ€” no more CUDA tantrums during sizing checks
    • Faster, more stable memory estimation under pressure

    Think of it like silencing a false alarm before you pack your suitcase โ€” no noise, just better packing.

    Perfect for folks running Llama 3, DeepSeek-R1, or Mistral on edge GPUs. No reinstall needed โ€” just update and let Ollama handle the heavy lifting. ๐Ÿค–โšก

    ๐Ÿ”— View Release

  • Ollama – v0.12.8: win: avoid ID mixups on refresh (#12869)

    Ollama – v0.12.8: win: avoid ID mixups on refresh (#12869)

    Ollama v0.12.8 just droppedโ€”and Windows AMD GPU users, this is your win! ๐ŸŽฏ

    The fix? No more sneaky GPU ID mixups. On Windows, AMDโ€™s device IDs would shuffle during refreshes, causing Ollama to accidentally pick your integrated iGPU instead of your powerful RX/RTX card. ๐Ÿ˜ฑ

    Now? Ollama respects GPU filters, ignores unsupported iGPUs entirely, and gives your discrete GPU the spotlight it deserves.

    โœ… No more misattributed VRAM

    โœ… Clean, accurate GPU detection on AMD Windows rigs

    โœ… Inference finally runs where it shouldโ€”on your real GPU

    Update now and stop letting your laptopโ€™s integrated graphics do all the work. ๐Ÿš€

    ๐Ÿ”— View Release