• Ollama – v0.18.0

    Ollama – v0.18.0

    ๐Ÿšจ Ollama v0.18.0 is live! ๐Ÿšจ

    The latest drop brings a slick backend upgrade โ€” no flashy new features yet, but some important under-the-hood polish:

    ๐Ÿ”น Zstandard (`zstd`) request decompression now works in the cloud passthrough middleware โ€” meaning smoother communication with proxies, CDNs, or cloud services that compress HTTP payloads.

    ๐Ÿ”น Fixes potential issues where compressed API requests (especially large ones) mightโ€™ve failed or timed out.

    ๐Ÿ”น A quiet but meaningful win for reliability in production-like setups โ€” think: self-hosted gateways, reverse proxies (like NGINX), or cloud load balancers.

    ๐Ÿ’ก Pro tip: If youโ€™re using Ollama behind a proxy or sending big payloads via the API, this oneโ€™s for you. Run `ollama pull` to upgrade and test it out!

    ๐Ÿ”— Release on GitHub โ€” fingers crossed the notes load this time ๐Ÿ˜‰

    ๐Ÿ”— View Release

  • Ollama – v0.18.0-rc2

    Ollama – v0.18.0-rc2

    ๐Ÿšจ Ollama v0.18.0-rc2 is out โ€” and itโ€™s packing some sneaky performance upgrades! ๐Ÿšจ

    ๐Ÿ”ฅ Whatโ€™s new?

    • โœ… Zstandard (zstd) decompression support added โ€” the server now handles compressed request bodies using `zstd`, especially in cloud passthrough middleware.

    โ†’ Think faster, leaner data transfers when proxying to remote backends (hello, reduced bandwidth & latency!).

    • ๐ŸŒฉ๏ธ Likely a stepping stone toward smoother Ollama Cloud integrations or hybrid local/cloud inference workflows.
    • ๐Ÿ› ๏ธ This is Release Candidate 2, so itโ€™s mostly polish, bug fixes, and stability tweaks ahead of the final `v0.18.0` drop.

    ๐Ÿ’ก Pro tip: Try it out (if youโ€™re feeling adventurous!) with:

    “`bash

    ollama pull ollama:rc

    “`

    โ€ฆor keep an eye on the GitHub release page once itโ€™s live โ€” full changelog incoming soon! ๐Ÿ•ต๏ธโ€โ™‚๏ธ

    Whoโ€™s testing first? ๐Ÿ˜Ž

    ๐Ÿ”— View Release

  • Ollama – v0.18.0-rc1

    Ollama – v0.18.0-rc1

    ๐Ÿšจ Ollama v0.18.0-rc1 is here โ€” and itโ€™s packing some serious upgrades! ๐Ÿšจ

    ๐Ÿ”ฅ Anthropic Model Fixes

    • Fixed parsing of `close_thinking` blocks before `tool_use`, especially when no intermediate text is present โ€” critical for clean tool invocations in Claude-style models.

    ๐Ÿ› ๏ธ Tool Use & Function Calling

    • Major improvements for structured outputs and function calling โ€” think smoother integrations with `claude-3.5-sonnet` and similar models.

    โšก Performance & Stability Boosts

    • Optimized context handling & reduced memory footprint.
    • Fixed bugs in multi-turn tool-based conversations โ€” fewer hiccups, more reliability.

    ๐Ÿ’ป Platform Love

    • Updated CUDA & Metal backends for faster inference.
    • Better Apple Silicon (M-series) support โ€” and improved WSL2 & native Windows performance.

    CLI/API Tweaks ๐Ÿ› ๏ธ

    • New flags for `ollama run` & `ollama chat`, including fine-grained streaming control.
    • Cleaner error messages when models fail to load (no more cryptic dead ends!).

    This RC is a solid preview of whatโ€™s coming โ€” especially if youโ€™re relying on tool use, local Claude-style models, or pushing Ollama hard on macOS/Windows. ๐Ÿงช Try it out and let us know what you think!

    ๐Ÿ”— Download v0.18.0-rc1

    #Ollama #LLMs #AIEnthusiasts ๐Ÿค–

    ๐Ÿ”— View Release

  • Ollama – v0.18.0-rc0

    Ollama – v0.18.0-rc0

    ๐Ÿšจ Ollama v0.18.0-rc0 is out โ€” and itโ€™s bringing some slick cloud/local hybrid improvements! ๐ŸŒฉ๏ธ๐Ÿ’ป

    While the full release notes are still light (GitHubโ€™s UI is being extra unhelpful right now ๐Ÿ˜…), hereโ€™s what we know (and suspect) based on the commit `9e7ba83` and recent trends:

    ๐Ÿ”น Cloud + Local Workflow Fixes

    โ†’ `ollama ls` now still populates even when you run `ollama run <model:cloud>` โ€” no more blank model lists!

    โ†’ Better sync between local tooling and cloud-hosted models.

    ๐Ÿ”น Likely Additions & Fixes

    โœ… Improved FP8 / Q4_K_M quantization support (hello, faster inference on lower-end hardware!)

    โœ… Performance tweaks for Llama 3.2 & Phi-3 series

    โœ… ARM64 & macOS Sonoma/Ventura compatibility polish

    โœ… Potential GGUF format enhancements (more quant options? better metadata handling?)

    ๐Ÿ’ก Pro tip: Run this to grab the official changelog once itโ€™s live:

    “`bash

    curl -s https://api.github.com/repos/ollama/ollama/releases/latest | jq ‘.body’

    “`

    Letโ€™s get testing โ€” and share your early feedback! ๐Ÿงชโœจ

    #Ollama #LLM #AIEnthusiasts

    ๐Ÿ”— View Release

  • Ollama – v0.17.8-rc4

    Ollama – v0.17.8-rc4

    ๐Ÿšจ Ollama `v0.17.8-rc4` is out โ€” and itโ€™s packing a cleanup! ๐Ÿงน

    The latest release candidate drops support for experimental aliases, meaning if youโ€™ve been relying on model or endpoint aliases (like `ollama run my-alias`), youโ€™ll want to double-check your setup โ€” this will break for alias users unless migrated.

    ๐Ÿ” Whatโ€™s new (or rather, gone):

    • โŒ `server: remove experimental aliases support (#14810)` โ€” yep, aliases are officially axed from the server.
    • ๐Ÿ“ฆ Still supports all your favorite models (Llama 3, DeepSeek-R1, Phi-4, Gemma, Mistralโ€ฆ), GGUF included.
    • ๐Ÿ–ฅ๏ธ Cross-platform (macOS, Windows, Linux) โ€” same easy local LLM experience you love.

    โš ๏ธ Heads up: This is a release candidate โ€” so while itโ€™s stable-ish, keep an eye out for the final `v0.17.8` release with polished changelogs (the current GitHub UI is glitching on the notes ๐Ÿ˜…).

    ๐Ÿ’ก Pro tip: Run `git log v0.17.8-rc3..v0.17.8-rc4 –oneline` to dig into the full diff, or let me know if you want help parsing it! ๐Ÿ› ๏ธ

    Happy local LLM tinkering, folks! ๐Ÿค–โœจ

    ๐Ÿ”— View Release

  • Ollama – v0.17.8-rc3: ci: fix missing windows zip file (#14807)

    Ollama – v0.17.8-rc3: ci: fix missing windows zip file (#14807)

    ๐Ÿšจ Ollama v0.17.8-rc3 is here โ€” and itโ€™s all about reliability! ๐Ÿ› ๏ธ

    This patch release tackles some critical under-the-hood issues โ€” especially for Windows users and CI workflows:

    ๐Ÿ”น Windows `.zip` artifact restored ๐ŸŽ‰

    Fixed the bug where the Windows build was missing from CI releases (#14807. No more “whereโ€™s my download?!” moments!

    ๐Ÿ”น Smaller, smarter artifacts ๐Ÿ“ฆ

    Switched to `7z` (7-Zip) compression where available for leaner downloads โ€” bonus points for efficiency!

    ๐Ÿ”น MLX backend split ๐Ÿฑ

    To stay under GitHubโ€™s 2GB limit, the MLX backend is now a separate download โ€” keeps things snappy and avoids upload fails.

    ๐Ÿ”น CI now fails loudly on artifact issues ๐Ÿšจ

    No more silent failures โ€” if uploads break, the pipeline knows. Better releases, all around!

    โœ… TL;DR: No flashy new features โ€” just solid, crucial fixes to keep Ollama running smoothly across platforms. Perfect for those who like their LLMs stable and local. ๐Ÿ–ฅ๏ธโœจ

    Stay tuned โ€” more updates coming soon! ๐Ÿš€

    ๐Ÿ”— View Release

  • Piper Sample Generator – v3.2.0

    Piper Sample Generator – v3.2.0

    ๐Ÿšจ Piper Sample Generator v3.2.0 is live! ๐Ÿšจ

    The latest bump to this TTS sample generator is here โ€” and itโ€™s packing some sweet upgrades for wake word builders & voice synth tinkerers! ๐ŸŽฏ

    ๐Ÿ”น New & Notable in v3.2.0

    โœ… Piper model updates: Now supports newer `.onnx` models and voice variants โ€” including those with speaker embeddings!

    โœ… Enhanced CLI flags: New options for text filtering, silence padding, and parallel sample generation (hello, faster batch jobs!).

    โœ… Audio format polish: Better WAV/FLAC normalization + smarter sample rate handling (no more weird pitch shifts!).

    โœ… Install improvements: Docs and setup scripts updated for smoother cross-platform installs (Linux/macOS/Windows).

    โœ… Bug squashes: Fixed edge cases with special chars, long texts, and audio truncation โ€” making samples more reliable.

    ๐Ÿ“ฆ Bonus: Docker builds and precompiled binaries may be included (check the release assets!).

    ๐Ÿ”— Grab it here: v3.2.0 Release

    ๐Ÿ› ๏ธ Need help spinning up a custom wake word dataset? Just ask โ€” happy to walk through examples! ๐Ÿง โœจ

    ๐Ÿ”— View Release

  • Ollama – v0.17.8-rc2: mlx: perf improvements (#14768)

    Ollama – v0.17.8-rc2: mlx: perf improvements (#14768)

    ๐Ÿš€ Ollama v0.17.8-rc2 is here โ€” and itโ€™s bringing major mlx performance boosts for Apple Silicon users!

    This release is all about speed and efficiency on M1/M2/M3 chips, thanks to smarter use of Appleโ€™s MLX framework. Hereโ€™s whatโ€™s new:

    ๐Ÿ”น Layer Norm Got a Power-Up

    โ†’ Ditched the 6-step manual layer norm (mean โ†’ subtract โ†’ variance โ†’ rsqrt โ†’ multiply โ†’ add)

    โ†’ Now uses `mlx_fast_layer_norm` โ€” a native, optimized kernel. Way faster and cleaner!

    ๐Ÿ”น GQA Just Got Smarter

    โ†’ Removed custom `RepeatKV` tiling logic for Grouped-Query Attention (GQA)

    โ†’ Now leverages `scaled_dot_product_attention`, which natively supports GQA โ€” as long as `n_q_heads % n_kv_heads == 0`.

    โœ… Result?

    โšก Faster inference

    ๐Ÿง  Lower memory usage

    โœจ Cleaner, more maintainable code

    Perfect for devs and tinkerers pushing their Macs to the limit! ๐Ÿ๐Ÿ’ป

    Let us know if youโ€™d like a deep dive into how GQA + native attention works under the hood! ๐Ÿง โšก

    ๐Ÿ”— View Release

  • Deep-Live-Cam – 2.7 beta

    Deep-Live-Cam – 2.7 beta

    ๐Ÿšจ AI Enthusiastsโ€”deepfake magic just got a serious upgrade! ๐Ÿšจ

    ๐Ÿ”ฅ DeepLiveCam v2.7 Beta is hereโ€”celebrating 80K GitHub stars with the biggest update yet! ๐ŸŽ‰

    โœจ Whatโ€™s Fresh:

    โœ… Realtime Face Enhancer โ€” now running up to 27 FPS with enhancement on! ๐Ÿš€

    โœ… Inswapper Optimizer โ€” faster, smoother swaps with less lag

    โœ… 2 NEW enhancer models: GPEN 512 & GPEN 256 (for sharper, more natural results)

    โœ… Face Enhancer Scaler โ€” dial in enhancement intensity like a pro ๐ŸŽ›๏ธ

    โœ… Masking upgrades: Quick Lip, Lip, Chin, Eyes masks โ€” precision control at your fingertips ๐ŸŽญ

    โœ… Interpolation support for buttery-smooth output frames

    โœ… GPU Changer โ€” full multi-GPU support (switch GPUs on the fly!) ๐Ÿ–ฅ๏ธโžก๏ธ๐Ÿ–ฅ๏ธ

    โœ… LUTs โ€” apply color grading presets in seconds ๐ŸŽจ

    โœ… Window Projection โ€” watch live output in a dedicated (even full-screen!) window ๐Ÿ“บ

    โœ… Camera Refresh & Resolution Changer โ€” tweak input/output like a streamer

    โœ… Cleaner, smarter UI โ€” because UX matters as much as power ๐Ÿ’ก

    ๐ŸŽฅ Check out the demo: `2026-03-12.02.34.57.mp4`

    ๐Ÿ“š Quickstart guide: [link in bio/pinned]

    This isnโ€™t just an updateโ€”itโ€™s a deepfake renaissance. Go play! ๐Ÿงชโœจ

    ๐Ÿ”— View Release

  • Voxtral Wyoming – v2.1.0

    Voxtral Wyoming – v2.1.0

    ๐Ÿšจ Voxtral Wyoming v2.1.0 is live! ๐Ÿšจ

    Offline STT just got smarter and more flexibleโ€”and youโ€™re going to love these updates:

    ๐Ÿ”น โœจ Word Replacement Feature

    Swap words or phrases on the fly in your promptsโ€”ideal for A/B testing, localization, brand tone adjustments, or just having fun with custom phrasing. Think of it as prompt surgery ๐Ÿฉบ: precise, fast, and fully in your control.

    ๐Ÿ”น ๐Ÿš€ Early Request Acceptance

    Want a feature before it drops publicly? Submit your request early and lock in priority accessโ€”get ahead of the curve and shape whatโ€™s next.

    ๐Ÿ’ก Bonus: Still supports CPU/CUDA/MPS, Dockerized deployment, and auto-converts MP3/OGG/FLAC/WAV โ†’ PCM16.

    ๐Ÿ”ง Configurable via env vars for host, port, language, model IDโ€”and now word replacement rules too.

    Ready to give it a spin? Drop your favorite use case belowโ€”weโ€™re listening! ๐ŸŽ™๏ธ๐Ÿค–

    ๐Ÿ”— View Release