• Ollama – v0.18.3-rc1

    Ollama – v0.18.3-rc1

    ๐Ÿšจ Ollama v0.18.3-rc1 is out! ๐Ÿšจ

    A quick heads-up for our fellow AI tinkerers โ€” the latest release candidate for Ollama is here, and while itโ€™s a small RC, it packs a useful fix:

    ๐Ÿ”ง Whatโ€™s new?

    โœ… Windows CGO compiler error fixed (PR #15046) โ€” this resolves a CI/CD hiccup that was causing build issues on Windows. If youโ€™ve been hitting weird compilation errors or CI failures on Windows, this oneโ€™s for you!

    ๐Ÿ” Note: Due to a temporary GitHub UI glitch, the full release notes didnโ€™t load cleanly โ€” so this RC currently only has one confirmed change. If youโ€™re feeling adventurous and want to test the latest fixes (especially on Windows), go grab `v0.18.3-rc1` from the GitHub Releases page.

    Let us know if you run into anything โ€” happy prompting! ๐Ÿง โœจ

    ๐Ÿ”— View Release

  • Ollama – v0.18.3-rc0: mlx: add mxfp4/mxfp8/nvfp4 importing (#15015)

    Ollama – v0.18.3-rc0: mlx: add mxfp4/mxfp8/nvfp4 importing (#15015)

    ๐Ÿšจ Ollama v0.18.3-rc0 is here โ€” and itโ€™s a quantization powerhouse! ๐Ÿšจ

    The latest release adds supercharged import support for MLX (Appleโ€™s Metal framework) and NVIDIAโ€™s new FP8 formats โ€” meaning you can now run even more efficient, low-bit models locally. Hereโ€™s the breakdown:

    ๐Ÿ”น New Quantization Imports

    โœ… Import BF16 models โ†’ convert on-the-fly to:

    • `mxfp4` (Metaโ€™s 4-bit mixed-precision)
    • `mxfp8` (Metaโ€™s 8-bit mixed-precision)
    • `nvfp4` (NVIDIAโ€™s 4-bit floating-point format)

    โœ… Import FP8 models โ†’ convert directly to `mxfp8`

    ๐ŸŽฏ Why this rocks:

    • ๐Ÿ Apple Silicon users (M1/M2/M3): Run ultra-efficient MLX-native models with minimal memory footprint.
    • ๐ŸŽฎ NVIDIA fans: Get early access to NVFP4 โ€” a promising new format for faster, smaller inference.
    • โšก Smaller models + less VRAM = more models on your laptop, fewer cloud trips.

    This is a big leap toward truly portable, hardware-agnostic LLM inference โ€” all from your desktop. ๐Ÿง ๐Ÿ’ป

    Curious how `mxfp4` stacks up against `nvfp4`? Let us know โ€” happy to deep dive! ๐Ÿงต

    ๐Ÿ”— View Release

  • Home Assistant Voice Pe – 26.3.0

    Home Assistant Voice Pe – 26.3.0

    ๐Ÿšจ Home Assistant Voice PE 26.3.0 is live! ๐Ÿšจ

    Big updates in this releaseโ€”letโ€™s break it down:

    ๐ŸŽง Media Playback Stability

    No more audio/video hiccups! Expect smoother, rock-solid playback across devicesโ€”perfect for voice-triggered media or smart speakers.

    ๐ŸŒ Multiple Sendspin Servers

    Now supports multiple Sendspin backends simultaneously. More redundancy, better failover, and improved scalability for larger setups.

    โฑ๏ธ TTS Timeout Fix

    Say goodbye to cut-off voice responses! Text-to-speech now waits properlyโ€”your AI replies play fully before moving on.

    ๐ŸŒŸ New Contributor Alert!

    Shoutout to @akloeckner for their first PR (#558)โ€”welcome to the crew! ๐Ÿ™Œ

    โœจ Bonus: The project is now officially sponsored by the Open Home Foundationโ€”a huge vote of confidence in open, private-by-design voice control!

    ๐Ÿ“ฆ Full changelog: [25.12.4 โ†’ 26.3.0](#)

    Letโ€™s make our homes smarterโ€”privately and offline-capable. ๐Ÿ ๐Ÿ—ฃ๏ธ

    ๐Ÿ”— View Release

  • Ollama – v0.18.2: launch: fix openclaw not picking up newly selected model (#14943)

    Ollama – v0.18.2: launch: fix openclaw not picking up newly selected model (#14943)

    ๐Ÿšจ Ollama v0.18.2 is live! ๐Ÿšจ

    This patch fixes a sneaky bug in OpenCLaW (Ollamaโ€™s model-launching engine) where switching models mid-session would not actually switch the active model โ€” yikes! ๐Ÿ˜…

    ๐Ÿ”ง Whatโ€™s fixed?

    • Previously: Changing the primary model in GUI/CLI wouldnโ€™t update active sessions โ€” youโ€™d keep running the old model, even if it looked like youโ€™d switched.
    • Now: Sessions properly refresh when the model changes โ€” no more stale model confusion! โœ…

    ๐ŸŽฏ Why it matters:

    • Perfect for devs testing multiple models in one session (e.g., comparing Llama 3 vs. DeepSeek).
    • Critical for demos or workflows where model switching is part of the flow โ€” reliability restored!

    ๐Ÿ”— PR #14943

    ๐Ÿ“ฆ Verified by BruceMacD on Mar 18, 20:20 UTC

    Ready to upgrade? `ollama update` && happy model-hopping! ๐Ÿง โœจ

    ๐Ÿ”— View Release

  • Ollama – v0.18.2-rc1: launch: fix openclaw not picking up newly selected model (#14943)

    Ollama – v0.18.2-rc1: launch: fix openclaw not picking up newly selected model (#14943)

    ๐Ÿšจ Ollama v0.18.2-rc1 is live! ๐Ÿšจ

    Quick, focused fix in this release candidate โ€” perfect for keeping your local LLM workflows smooth and reliable.

    ๐Ÿ” Whatโ€™s New?

    โœ… Fixed `openclaw` model-switching bug โ€” previously, if you changed the active model mid-session without restarting, `openclaw` would keep using the old (stale) model. Now it correctly picks up the new selection instantly!

    ๐Ÿ’ก Why Youโ€™ll Care:

    • ๐Ÿ”„ Makes dynamic model switching (e.g., testing Llama 3 vs. DeepSeek-R1) way more reliable in dev/test loops
    • ๐Ÿงช Critical for tooling and integrations that rely on runtime model changes
    • โšก Clean, minimal change โ€” no side effects or breaking changes

    ๐Ÿ‘‰ Grab the RC and give it a spin before final `v0.18.2` drops!

    Let the team know if you spot anything weird ๐Ÿžโžก๏ธโœ…

    #Ollama #LLM #LocalAI #DevTools

    ๐Ÿ”— View Release

  • Ollama – v0.18.2-rc0

    Ollama – v0.18.2-rc0

    ๐Ÿš€ Ollama v0.18.2-rc0 is out โ€” and itโ€™s bringing live web search!

    This release candidate (not quite stable yet, but very promising) adds support for web search integration via the `openclaw` tool โ€” likely enabling models to pull in real-time info from the web during inference. ๐ŸŒ๐Ÿ”

    ๐Ÿ”น New Tool Registered: `openclaw` is now available as a callable tool โ€” think of it like giving your local LLM a live browser.

    ๐Ÿ”น Tool-Calling Boost: Aligns with the growing trend of models (like Llama 3.1) supporting function/tool calling โ€” making Ollama even more dynamic and up-to-date.

    ๐Ÿ”น Use Cases? Think: real-time Q&A, research assistants, news-aware chatbots โ€” all running locally (or on your own server).

    โš ๏ธ Note: This is an `rc0`, so things may shift before the final `v0.18.2` lands โ€” but itโ€™s a huge step toward truly intelligent, always-current local AI.

    Curious about how `openclaw` works? Or want to test it out? Letโ€™s dive in! ๐Ÿงช๐Ÿ’ก

    ๐Ÿ”— View Release

  • Ollama – v0.18.1

    Ollama – v0.18.1

    ๐Ÿšจ Ollama v0.18.1 is live โ€” and itโ€™s bringing seriously useful upgrades! ๐Ÿšจ

    ๐Ÿ”ฅ Web Search is BACK & fixed โ€” no more broken links or silent failures.

    ๐ŸŒ New Web Fetch feature โ€” Ollama can now pull in live web content (think docs, articles, news) to enrich its responses.

    โœ… Both features are ON by default โ€” zero config needed! Just `ollama run` and go.

    ๐Ÿ’ป Works locally too โ€” even when running on your machine (as long as youโ€™ve got internet access).

    ๐Ÿ’ก Why youโ€™ll love this:

    • Real-time context for RAG pipelines
    • Fact-checking on the fly
    • Fresh info without retraining or API keys

    Perfect for devs building local, context-aware apps โ€” or just curious tinkerers who want smarter, up-to-date answers. ๐Ÿ› ๏ธโœจ

    Check it out: ollama.com

    Need a quick demo or config tweaks? Drop a ๐Ÿง  below!

    ๐Ÿ”— View Release

  • Ollama – v0.18.1-rc1

    Ollama – v0.18.1-rc1

    ๐Ÿšจ Ollama v0.18.1-rc1 is here โ€” and itโ€™s bringing web smarts to your local LLMs! ๐ŸŒโœจ

    ๐Ÿ”ฅ Whatโ€™s new?

    โœ… Web search is back & fixed! No more silent failures โ€” Ollama can now reliably pull live results.

    ๐ŸŒ New `web fetch` tool โ€” ask Ollama to retrieve up-to-the-minute web content and use it directly in responses.

    โšก Both features are ON by default โ€” just update, and youโ€™re ready to go (no config needed!).

    ๐Ÿ’ก Why care?

    This means your local models can now answer questions with real-time context โ€” think news, docs, or live data โ€” all without hitting external APIs. Perfect for building smarter, self-contained AI apps ๐Ÿง ๐Ÿ’ป

    ๐Ÿ“ฆ Release type: Release Candidate (`v0.18.1-rc1`) โ€” ideal for testing and feedback before the stable drop!

    Ready to test it out? ๐Ÿ› ๏ธ Let us know how your web-aware Ollama experiments go! ๐Ÿš€

    ๐Ÿ”— View Release

  • Ollama – v0.18.1-rc0: cmd/launch: skip –install-daemon when systemd is unavailable (#14883)

    Ollama – v0.18.1-rc0: cmd/launch: skip –install-daemon when systemd is unavailable (#14883)

    ๐Ÿš€ Ollama v0.18.1-rc0 is here โ€” and itโ€™s fixing a pesky containerization hiccup!

    ๐Ÿ”ฅ Whatโ€™s new?

    • ๐Ÿ› ๏ธ `ollama launch openclaw –install-daemon` no longer fails in environments without systemd (like Docker, WSL, Alpine, or CI runners).
    • โœ… Smart detection: Ollama now checks if systemd is actually available before trying to install the daemon.
    • ๐Ÿงช Falls back gracefully to foreground mode when systemd isnโ€™t present โ€” meaning your local LLMs keep running, no matter the setup!

    ๐Ÿ“ฆ Why it matters:

    Perfect for devs testing in containers or minimal Linux setups โ€” less friction, more inference! ๐Ÿณ๐Ÿ’ป

    ๐Ÿ‘‰ Try the RC and let us know how it plays in your envs!

    #Ollama #LLM #AIEnthusiasts

    ๐Ÿ”— View Release

  • Ollama – v0.18.0

    Ollama – v0.18.0

    ๐Ÿšจ Ollama v0.18.0 is live! ๐Ÿšจ

    The latest drop brings a slick backend upgrade โ€” no flashy new features yet, but some important under-the-hood polish:

    ๐Ÿ”น Zstandard (`zstd`) request decompression now works in the cloud passthrough middleware โ€” meaning smoother communication with proxies, CDNs, or cloud services that compress HTTP payloads.

    ๐Ÿ”น Fixes potential issues where compressed API requests (especially large ones) mightโ€™ve failed or timed out.

    ๐Ÿ”น A quiet but meaningful win for reliability in production-like setups โ€” think: self-hosted gateways, reverse proxies (like NGINX), or cloud load balancers.

    ๐Ÿ’ก Pro tip: If youโ€™re using Ollama behind a proxy or sending big payloads via the API, this oneโ€™s for you. Run `ollama pull` to upgrade and test it out!

    ๐Ÿ”— Release on GitHub โ€” fingers crossed the notes load this time ๐Ÿ˜‰

    ๐Ÿ”— View Release