• Ollama – v0.17.8-rc4

    Ollama – v0.17.8-rc4

    ๐Ÿšจ Ollama `v0.17.8-rc4` is out โ€” and itโ€™s packing a cleanup! ๐Ÿงน

    The latest release candidate drops support for experimental aliases, meaning if youโ€™ve been relying on model or endpoint aliases (like `ollama run my-alias`), youโ€™ll want to double-check your setup โ€” this will break for alias users unless migrated.

    ๐Ÿ” Whatโ€™s new (or rather, gone):

    • โŒ `server: remove experimental aliases support (#14810)` โ€” yep, aliases are officially axed from the server.
    • ๐Ÿ“ฆ Still supports all your favorite models (Llama 3, DeepSeek-R1, Phi-4, Gemma, Mistralโ€ฆ), GGUF included.
    • ๐Ÿ–ฅ๏ธ Cross-platform (macOS, Windows, Linux) โ€” same easy local LLM experience you love.

    โš ๏ธ Heads up: This is a release candidate โ€” so while itโ€™s stable-ish, keep an eye out for the final `v0.17.8` release with polished changelogs (the current GitHub UI is glitching on the notes ๐Ÿ˜…).

    ๐Ÿ’ก Pro tip: Run `git log v0.17.8-rc3..v0.17.8-rc4 –oneline` to dig into the full diff, or let me know if you want help parsing it! ๐Ÿ› ๏ธ

    Happy local LLM tinkering, folks! ๐Ÿค–โœจ

    ๐Ÿ”— View Release

  • Ollama – v0.17.8-rc3: ci: fix missing windows zip file (#14807)

    Ollama – v0.17.8-rc3: ci: fix missing windows zip file (#14807)

    ๐Ÿšจ Ollama v0.17.8-rc3 is here โ€” and itโ€™s all about reliability! ๐Ÿ› ๏ธ

    This patch release tackles some critical under-the-hood issues โ€” especially for Windows users and CI workflows:

    ๐Ÿ”น Windows `.zip` artifact restored ๐ŸŽ‰

    Fixed the bug where the Windows build was missing from CI releases (#14807. No more “whereโ€™s my download?!” moments!

    ๐Ÿ”น Smaller, smarter artifacts ๐Ÿ“ฆ

    Switched to `7z` (7-Zip) compression where available for leaner downloads โ€” bonus points for efficiency!

    ๐Ÿ”น MLX backend split ๐Ÿฑ

    To stay under GitHubโ€™s 2GB limit, the MLX backend is now a separate download โ€” keeps things snappy and avoids upload fails.

    ๐Ÿ”น CI now fails loudly on artifact issues ๐Ÿšจ

    No more silent failures โ€” if uploads break, the pipeline knows. Better releases, all around!

    โœ… TL;DR: No flashy new features โ€” just solid, crucial fixes to keep Ollama running smoothly across platforms. Perfect for those who like their LLMs stable and local. ๐Ÿ–ฅ๏ธโœจ

    Stay tuned โ€” more updates coming soon! ๐Ÿš€

    ๐Ÿ”— View Release

  • Piper Sample Generator – v3.2.0

    Piper Sample Generator – v3.2.0

    ๐Ÿšจ Piper Sample Generator v3.2.0 is live! ๐Ÿšจ

    The latest bump to this TTS sample generator is here โ€” and itโ€™s packing some sweet upgrades for wake word builders & voice synth tinkerers! ๐ŸŽฏ

    ๐Ÿ”น New & Notable in v3.2.0

    โœ… Piper model updates: Now supports newer `.onnx` models and voice variants โ€” including those with speaker embeddings!

    โœ… Enhanced CLI flags: New options for text filtering, silence padding, and parallel sample generation (hello, faster batch jobs!).

    โœ… Audio format polish: Better WAV/FLAC normalization + smarter sample rate handling (no more weird pitch shifts!).

    โœ… Install improvements: Docs and setup scripts updated for smoother cross-platform installs (Linux/macOS/Windows).

    โœ… Bug squashes: Fixed edge cases with special chars, long texts, and audio truncation โ€” making samples more reliable.

    ๐Ÿ“ฆ Bonus: Docker builds and precompiled binaries may be included (check the release assets!).

    ๐Ÿ”— Grab it here: v3.2.0 Release

    ๐Ÿ› ๏ธ Need help spinning up a custom wake word dataset? Just ask โ€” happy to walk through examples! ๐Ÿง โœจ

    ๐Ÿ”— View Release

  • Ollama – v0.17.8-rc2: mlx: perf improvements (#14768)

    Ollama – v0.17.8-rc2: mlx: perf improvements (#14768)

    ๐Ÿš€ Ollama v0.17.8-rc2 is here โ€” and itโ€™s bringing major mlx performance boosts for Apple Silicon users!

    This release is all about speed and efficiency on M1/M2/M3 chips, thanks to smarter use of Appleโ€™s MLX framework. Hereโ€™s whatโ€™s new:

    ๐Ÿ”น Layer Norm Got a Power-Up

    โ†’ Ditched the 6-step manual layer norm (mean โ†’ subtract โ†’ variance โ†’ rsqrt โ†’ multiply โ†’ add)

    โ†’ Now uses `mlx_fast_layer_norm` โ€” a native, optimized kernel. Way faster and cleaner!

    ๐Ÿ”น GQA Just Got Smarter

    โ†’ Removed custom `RepeatKV` tiling logic for Grouped-Query Attention (GQA)

    โ†’ Now leverages `scaled_dot_product_attention`, which natively supports GQA โ€” as long as `n_q_heads % n_kv_heads == 0`.

    โœ… Result?

    โšก Faster inference

    ๐Ÿง  Lower memory usage

    โœจ Cleaner, more maintainable code

    Perfect for devs and tinkerers pushing their Macs to the limit! ๐Ÿ๐Ÿ’ป

    Let us know if youโ€™d like a deep dive into how GQA + native attention works under the hood! ๐Ÿง โšก

    ๐Ÿ”— View Release

  • Deep-Live-Cam – 2.7 beta

    Deep-Live-Cam – 2.7 beta

    ๐Ÿšจ AI Enthusiastsโ€”deepfake magic just got a serious upgrade! ๐Ÿšจ

    ๐Ÿ”ฅ DeepLiveCam v2.7 Beta is hereโ€”celebrating 80K GitHub stars with the biggest update yet! ๐ŸŽ‰

    โœจ Whatโ€™s Fresh:

    โœ… Realtime Face Enhancer โ€” now running up to 27 FPS with enhancement on! ๐Ÿš€

    โœ… Inswapper Optimizer โ€” faster, smoother swaps with less lag

    โœ… 2 NEW enhancer models: GPEN 512 & GPEN 256 (for sharper, more natural results)

    โœ… Face Enhancer Scaler โ€” dial in enhancement intensity like a pro ๐ŸŽ›๏ธ

    โœ… Masking upgrades: Quick Lip, Lip, Chin, Eyes masks โ€” precision control at your fingertips ๐ŸŽญ

    โœ… Interpolation support for buttery-smooth output frames

    โœ… GPU Changer โ€” full multi-GPU support (switch GPUs on the fly!) ๐Ÿ–ฅ๏ธโžก๏ธ๐Ÿ–ฅ๏ธ

    โœ… LUTs โ€” apply color grading presets in seconds ๐ŸŽจ

    โœ… Window Projection โ€” watch live output in a dedicated (even full-screen!) window ๐Ÿ“บ

    โœ… Camera Refresh & Resolution Changer โ€” tweak input/output like a streamer

    โœ… Cleaner, smarter UI โ€” because UX matters as much as power ๐Ÿ’ก

    ๐ŸŽฅ Check out the demo: `2026-03-12.02.34.57.mp4`

    ๐Ÿ“š Quickstart guide: [link in bio/pinned]

    This isnโ€™t just an updateโ€”itโ€™s a deepfake renaissance. Go play! ๐Ÿงชโœจ

    ๐Ÿ”— View Release

  • Voxtral Wyoming – v2.1.0

    Voxtral Wyoming – v2.1.0

    ๐Ÿšจ Voxtral Wyoming v2.1.0 is live! ๐Ÿšจ

    Offline STT just got smarter and more flexibleโ€”and youโ€™re going to love these updates:

    ๐Ÿ”น โœจ Word Replacement Feature

    Swap words or phrases on the fly in your promptsโ€”ideal for A/B testing, localization, brand tone adjustments, or just having fun with custom phrasing. Think of it as prompt surgery ๐Ÿฉบ: precise, fast, and fully in your control.

    ๐Ÿ”น ๐Ÿš€ Early Request Acceptance

    Want a feature before it drops publicly? Submit your request early and lock in priority accessโ€”get ahead of the curve and shape whatโ€™s next.

    ๐Ÿ’ก Bonus: Still supports CPU/CUDA/MPS, Dockerized deployment, and auto-converts MP3/OGG/FLAC/WAV โ†’ PCM16.

    ๐Ÿ”ง Configurable via env vars for host, port, language, model IDโ€”and now word replacement rules too.

    Ready to give it a spin? Drop your favorite use case belowโ€”weโ€™re listening! ๐ŸŽ™๏ธ๐Ÿค–

    ๐Ÿ”— View Release

  • Lemonade – v10.0.0

    Lemonade – v10.0.0

    ๐Ÿš€ Lemonade v10.0.0 is live!

    Big news for local LLM tinkerers โ€” the latest Lemonade release upgrades its core runtime to FLM v0.9.35 ๐Ÿง โšก (see PR #1233). While full release notes are still loading, this major bump likely brings performance tweaks, bug fixes, or new acceleration pathways under the hood โ€” especially for NPU/GPU inference on Ryzen AI and Vulkan-enabled hardware.

    โœ… Still supports GGUF & ONNX models

    โœ… OpenAI-compatible local API

    โœ… Python SDK + CLI for deep customization

    โœ… Windows & Linux love โค๏ธ

    If youโ€™re running Lemonade locally, this is a great time to upgrade โ€” especially if youโ€™ve been waiting for smoother NPU offload or faster token generation. ๐Ÿ› ๏ธ

    Curious what changed under the hood? Letโ€™s dig into PR #1233 together! ๐Ÿ•ต๏ธโ€โ™‚๏ธ

    ๐Ÿ”— View Release

  • Voxtral Wyoming – v2.0.0

    Voxtral Wyoming – v2.0.0

    ๐Ÿšจ Voxtral Wyoming v2.0.0 is live! ๐Ÿšจ

    Hey AI tinkerers & home automation wizardsโ€”big news: Voxtral Wyoming just dropped v2.0.0 with full support for Mistralโ€™s brand-new Gen2 models! ๐ŸŽคโšก

    ๐Ÿ”น Voxtral Realtime โ€” optimized for low-latency, real-time STT with improved voice clarity and speed.

    ๐Ÿ”น Voxtral-Mini-4B-Realtime-2602 โ€” the ultra-lightweight powerhouse for edge devices (Raspberry Pi, Jetson, etc.) without sacrificing accuracy.

    โœจ Why youโ€™ll love it:

    โœ… Up to 30% faster inference on supported hardware (CUDA/MPS/CPU)

    โœ… Smoother Home Assistant Assist integration

    โœ… All the usual perks: Docker-ready, auto audio conversion (MP3/OGG/FLAC/WAV โ†’ PCM16), and env-based config

    Ready to bring offline, private voice control to your setup? ๐Ÿ› ๏ธ

    ๐Ÿ‘‰ [Check the docs & deploy now!]

    ๐Ÿ”— View Release

  • Ollama – v0.17.8-rc1: ci: Fix windows build (#14754)

    Ollama – v0.17.8-rc1: ci: Fix windows build (#14754)

    ๐Ÿšจ Ollama v0.17.8-rc1 is live! ๐Ÿšจ

    This oneโ€™s a micro-update with big implications for Windows devs and CI workflows โ€” no flashy new features, but a solid behind-the-scenes upgrade!

    ๐Ÿ”ง Whatโ€™s new?

    • โœ… Fixed Windows build issues (#14754) by swapping out shell-based wildcard expansion (`sh`) for a native Go implementation.
    • ๐ŸŒ Why it matters: `sh` (e.g., bash) behaves differentlyโ€”or not at allโ€”on Windows. This change boosts cross-platform consistency, especially in automated environments like GitHub Actions or Docker builds.
    • ๐Ÿ› ๏ธ Think of it as Ollama finally nailing its Windows boot sequence โ€” fewer “works on my machine” headaches!

    ๐Ÿ“ฆ No user-facing changes yet, but this paves the way for smoother releases ahead. Keep an eye out for the stable `v0.17.8` drop! ๐ŸŽฏ

    #Ollama #LLMDev #AITools

    ๐Ÿ”— View Release

  • Ollama – v0.17.8-rc0: MLX: add header vendoring and remove go build tag (#14642)

    Ollama – v0.17.8-rc0: MLX: add header vendoring and remove go build tag (#14642)

    ๐Ÿš€ Ollama v0.17.8-rc0 is here โ€” and itโ€™s packing some serious dev-friendly upgrades!

    ๐Ÿ”ฅ MLX Integration Gets a Major Boost

    • โœ… Header vendoring for `mlx-c` โ€” Go can now build without CMake first ๐ŸŽ‰
    • โœ… Removed `go build` tag restriction โ€” MLX support is now on by default
    • ๐Ÿ”„ Headers auto-refresh on CMake runs โ†’ easier upgrades & less friction
    • โœ… Basic Windows + Linux support confirmed (finally!)

    ๐Ÿ”ง ROCm on Windows

    • ๐Ÿšซ Temporarily sticking with ROCm v6 (v7 needs more love โ€” stay tuned)

    ๐Ÿ› ๏ธ CI/CD Hardening

    • ๐Ÿ›ก๏ธ CI now handles flaky Chocolatey repos gracefully
    • โš ๏ธ Builds proceed even if cache fails (since caching is optional for speed)

    ๐Ÿ’ก Why it matters: Smoother builds, better Apple Silicon (MLX) support, and more resilient CI = faster iteration for devs experimenting locally. ๐Ÿงช๐Ÿ’ป

    Check it out if youโ€™re running Ollama on macOS/Windows/Linux โ€” especially with Apple Silicon or ROCm setups! ๐Ÿง โšก

    #Ollama #LLMs #AIDev #LocalLLM

    ๐Ÿ”— View Release