• Lemonade – 9.0.2: C++: General Availability (#549)

    Lemonade – 9.0.2: C++: General Availability (#549)

    🚨 Lemonade v9.0.2 just dropped — and C++ is officially GA! 🎉

    No more beta labels. The C++ server is now production-ready, faster, leaner, and built for real-world LLM serving. Here’s the breakdown:

    • C++ is now the future — Python NSIS installer and dev tools are gone. Focus fully on C++ for peak performance.
    • 🚫 Python server is deprecated — start migrating your workflows now.
    • 🔧 Fixed `make_http_request()` bugs + default host = localhost (no more weird network issues).
    • 📚 Docs overhauled — cleaner setup, less confusion.
    • 💥 Removed all `lemonade-server-dev` clutter — clean, production-ready codebase.
    • 📈 All version numbers bumped for clarity.

    If you’re running LLMs locally on Ryzen AI or Radeon GPUs — this is your moment. Drop the Python server, go C++, and unlock low-latency AI at scale. 🚀

    Time to sip something cold… and run LLMs faster than ever.

    🔗 View Release

  • Ollama – v0.12.11-rc1

    Ollama – v0.12.11-rc1

    Ollama v0.12.11-rc1 is here — and Windows GPU fans, rejoice! 🎉

    Vulkan support is officially back on track. No more “Vulkan not found” errors — your RTX, RX, or Arc cards can now accelerate LLM inference without a hitch.

    This is a quiet patch with huge impact: if you’ve been stuck on CPU-only runs, it’s time to fire up your GPU again.

    ⚠️ Still a release candidate — but if you’re on Windows and craving faster generations, this is the one to try.

    Pro tip: Update your Vulkan drivers first! Ollama’s fixed its end — now let your hardware do the heavy lifting. 🚀

    🔗 View Release

  • Lemonade – v8.2.2

    Lemonade – v8.2.2

    Lemonade v8.2.2 just dropped—and it’s a game-changer for local LLM tinkerers! 🚀

    • Vision-Language Models are live 🖼️🧠: Run LLaMA-based VLMs locally—image + text reasoning, no cloud needed.
    • Precise device control: `–device` flag now actually works—tune GPU/CPU with zero guesswork.
    • Linux stability fixed: No more crashes or phantom DLL deps. CLI’s solid now.
    • HF_HUB_CACHE supported: Smarter offline caching for Hugging Face models—perfect if your internet’s spotty.
    • Web UI glow-up: Cleaner layout + new enable_thinking toggle to make models pause & reason before replying.
    • Real-time stats endpoint: Monitor `prompt_tokens` live—ideal for optimizing prompts and performance.
    • FLM Chat Completions patched: No more broken mid-convo responses.

    All wrapped in faster inference and cleaner C++ code. If you’re running LLMs on Ryzen AI or Radeon GPUs—this is your must-update beta. 💪

    🔗 View Release

  • Deep-Live-Cam – Version 2.3c is out now!

    Deep-Live-Cam – Version 2.3c is out now!

    Deep-Live-Cam v2.3c is live 🎭✨

    Fixed that annoying dropdown bug in model & camera selection—no more weird options or blank menus. Now it just works.

    Only available via QuickStart for now (Windows/Mac Silicon users, this is your cue!).

    CUDA, CoreML, DirectML, OpenVINO—your GPU still gets the spotlight.

    Reload, swap faces in real-time, and keep breaking the internet (responsibly). 🚀

    🔗 View Release

  • Tater – Tater v38

    Tater – Tater v38

    Tater v38 just dropped—and it’s got Siri on speed dial 🥔🎙️

    Ask your HomePod “What’s trending on Overseerr?” or “Launch MiSTer” and Tater replies like a chill buddy—short, smart, Siri-optimized. No more robotic jargon. Just clean JSON tool calls and natural voice flow.

    🧠 Smarter memory: Each device keeps its own chat history (with TTL caps)—your iPhone won’t remember you asked for pineapple pizza at 3am.

    💬 Tighter prompts: Less “according to the model,” more “hey buddy, here’s what’s hot.”

    🔧 New plugins:

    • 🎮 MiSTer Remote — launch your retro games from the couch
    • 🔍 Web Search — instant answers, zero tabs open
    • 🎥 Overseerr Request & Trending — see what’s trending and add films with your voice

    Now your kitchen counter has a potato-powered AI that knows your movie queue and your SNES library. No laptop needed. Just ask.

    Check the README for install tips—Docker or Python, you pick!

    🔗 View Release

  • Ollama – v0.12.10

    Ollama – v0.12.10

    🚀 Ollama v0.12.10 just dropped — quiet, but mighty!

    ✅ Code signing is back on macOS & Windows — safer installs, trusted binaries, zero sketchy downloads.

    🔧 No flashy new models or UI glitz… just clean, solid housekeeping for devs who run LLMs in production.

    If you care about security, integrity, and smooth local inference — this patch is your unsung hero.

    Keep those Llama 3, DeepSeek, and Mistral models humming! 🤖💻

    🔗 View Release

  • Text Generation Webui – v3.17

    Text Generation Webui – v3.17

    🔥 Text Generation WebUI v3.17 just dropped — and it’s a beast!

    💥 Security upgrade: `torch.load()` now defaults to `weights_only=True` — no more sketchy pickle loads in Training_PRO. Safe mode = activated.

    🧩 Hugging Face fix: Pinning `huggingface-hub==0.36.0` to squash those pesky venv install bugs. No more “why won’t this install?!” moments.

    📊 Training cleanup: `evaluation_strategy` → `eval_strategy` (thanks, @inyourface34456 — clean code wins again!).

    🚀 Backend power-ups:

    • 🖼️ Qwen3-VL support landed in llama.cpp — multimodal models now run natively. Text + images? Yes, please.
    • 🚀 `exllamav3` updated to v0.0.12 — faster, smoother, less stuttering. Your GPU will thank you.

    📦 PORTABLE BUILDS ARE LIVE!

    No install. No fuss. Just download, unzip, run:

    • 🖥️ NVIDIA: `cuda12.4`
    • 🖥️ AMD/Intel: `vulkan`
    • 💻 CPU-only: `cpu`
    • 🍎 Apple Silicon: `macos-arm64`
    • 🍏 Intel Mac: `macos-x86_64`

    🔧 Upgrading?

    Drop your old `user_data/` folder into the new zip — all models, settings, and custom themes preserved.

    Go run Qwen3-VL on your old laptop. Your CPU won’t know what hit it. 🤖💥

    🔗 View Release

  • Lemonade – v8.2.1

    Lemonade – v8.2.1

    🔥 Lemonade v8.2.1 just dropped — and it’s production-ready now.

    C++ stack got a full glow-up: tray icons, installers, CLI glitches, Ctrl+C crashes — all DEAD. This isn’t a patch; it’s a rebirth.

    🚀 Linux users: `.deb` installer landed! `apt install lemonade` on Ubuntu/Debian — no more compiling from source. Mac support polished too.

    🧠 Qwen3 GGUF now has `enable_thinking` — unlock deeper reasoning on complex prompts. Think smarter, not just faster.

    ⚡ `ryzenai-serve` → `ryzenai-server`. Now with halt & reasoning, cleaner logs, and smoother shutdowns. AMD Ryzen folks — this one’s for you.

    📊 Real-time memory tracking! Health endpoint shows context size + VRAM usage. No more guessing how much your model is eating.

    🧩 Model Manager rebuilt: faster, cleaner, smarter. FLM updates? No more stale models sneaking in.

    💻 Web UI & CLI got polish: cleaner commands, better streaming logs, no more redundant callbacks. Feels like a real product now.

    📦 Python 8.2.1 synced too — lm-eval fixes, bigger default context/batch sizes for embeddings.

    🧪 Test workflows cleaned up. Fewer false CI failures, more confidence in every release.

    This isn’t just an update — it’s Lemonade stepping out of beta and into the big leagues.

    Grab it. Break it. Tell us how 😄

    🔗 Full changelog

    🔗 View Release

  • Ollama – v0.12.10-rc1

    Ollama – v0.12.10-rc1

    🚀 Ollama v0.12.10-rc1 just dropped — and it’s all about trust, not flash!

    🔒 Code signing is back on macOS & Windows — your downloads are now cryptographically verified, so you know it’s really Ollama, not a sneaky imposter.

    Perfect for enterprises, devs in locked-down environments, or anyone who hates “this file is untrusted” warnings.

    No new models. No UI tweaks. Just clean, quiet hardening — because security shouldn’t be an afterthought.

    💡 Pro tip: If you script installs or deploy via CI/CD, your checksums and signature checks are back in action.

    Keep it local. Keep it safe. 🛡️

    🔗 View Release

  • Wyoming Openai – Maintenance release (0.3.9)

    Wyoming Openai – Maintenance release (0.3.9)

    🚀 Wyoming OpenAI v0.3.9 is live — lean, mean, and ARM-ready!

    🔧 CLI & Env Fixes: Now accepts real option names (no more internal ID confusion), and `STT_BACKEND`/`TTS_BACKEND` are case-insensitive. Typos? Clearer errors now!

    🍎 ARM64 Support: Docker images built for Raspberry Pi, M1/M2 Macs — no more emulation. Just `docker pull` and go!

    📦 Deps Upgraded: OpenAI v2.7.1, Ruff, pytest — faster, stabler, future-proofed.

    ⚙️ Build Pipeline Upgraded: GitHub Actions now use `build-push-action@v6` with QEMU + Buildx for seamless multi-arch builds.

    📖 Docs updated to v0.3.9 — no more outdated examples!

    7 files changed, +162/-17 lines. Zero fluff. Pure utility.

    Install via `pip install wyoming-openai` or pull the new Docker image — your Pi (or M2 Mac) will thank you. 🐧🍎

    🔗 View Release