• Home Assistant Voice Pe – 26.3.0

    Home Assistant Voice Pe – 26.3.0

    ๐Ÿšจ Home Assistant Voice PE 26.3.0 is live! ๐Ÿšจ

    Big updates in this releaseโ€”letโ€™s break it down:

    ๐ŸŽง Media Playback Stability

    No more audio/video hiccups! Expect smoother, rock-solid playback across devicesโ€”perfect for voice-triggered media or smart speakers.

    ๐ŸŒ Multiple Sendspin Servers

    Now supports multiple Sendspin backends simultaneously. More redundancy, better failover, and improved scalability for larger setups.

    โฑ๏ธ TTS Timeout Fix

    Say goodbye to cut-off voice responses! Text-to-speech now waits properlyโ€”your AI replies play fully before moving on.

    ๐ŸŒŸ New Contributor Alert!

    Shoutout to @akloeckner for their first PR (#558)โ€”welcome to the crew! ๐Ÿ™Œ

    โœจ Bonus: The project is now officially sponsored by the Open Home Foundationโ€”a huge vote of confidence in open, private-by-design voice control!

    ๐Ÿ“ฆ Full changelog: [25.12.4 โ†’ 26.3.0](#)

    Letโ€™s make our homes smarterโ€”privately and offline-capable. ๐Ÿ ๐Ÿ—ฃ๏ธ

    ๐Ÿ”— View Release

  • Ollama – v0.18.2: launch: fix openclaw not picking up newly selected model (#14943)

    Ollama – v0.18.2: launch: fix openclaw not picking up newly selected model (#14943)

    ๐Ÿšจ Ollama v0.18.2 is live! ๐Ÿšจ

    This patch fixes a sneaky bug in OpenCLaW (Ollamaโ€™s model-launching engine) where switching models mid-session would not actually switch the active model โ€” yikes! ๐Ÿ˜…

    ๐Ÿ”ง Whatโ€™s fixed?

    • Previously: Changing the primary model in GUI/CLI wouldnโ€™t update active sessions โ€” youโ€™d keep running the old model, even if it looked like youโ€™d switched.
    • Now: Sessions properly refresh when the model changes โ€” no more stale model confusion! โœ…

    ๐ŸŽฏ Why it matters:

    • Perfect for devs testing multiple models in one session (e.g., comparing Llama 3 vs. DeepSeek).
    • Critical for demos or workflows where model switching is part of the flow โ€” reliability restored!

    ๐Ÿ”— PR #14943

    ๐Ÿ“ฆ Verified by BruceMacD on Mar 18, 20:20 UTC

    Ready to upgrade? `ollama update` && happy model-hopping! ๐Ÿง โœจ

    ๐Ÿ”— View Release

  • Ollama – v0.18.2-rc1: launch: fix openclaw not picking up newly selected model (#14943)

    Ollama – v0.18.2-rc1: launch: fix openclaw not picking up newly selected model (#14943)

    ๐Ÿšจ Ollama v0.18.2-rc1 is live! ๐Ÿšจ

    Quick, focused fix in this release candidate โ€” perfect for keeping your local LLM workflows smooth and reliable.

    ๐Ÿ” Whatโ€™s New?

    โœ… Fixed `openclaw` model-switching bug โ€” previously, if you changed the active model mid-session without restarting, `openclaw` would keep using the old (stale) model. Now it correctly picks up the new selection instantly!

    ๐Ÿ’ก Why Youโ€™ll Care:

    • ๐Ÿ”„ Makes dynamic model switching (e.g., testing Llama 3 vs. DeepSeek-R1) way more reliable in dev/test loops
    • ๐Ÿงช Critical for tooling and integrations that rely on runtime model changes
    • โšก Clean, minimal change โ€” no side effects or breaking changes

    ๐Ÿ‘‰ Grab the RC and give it a spin before final `v0.18.2` drops!

    Let the team know if you spot anything weird ๐Ÿžโžก๏ธโœ…

    #Ollama #LLM #LocalAI #DevTools

    ๐Ÿ”— View Release

  • Ollama – v0.18.2-rc0

    Ollama – v0.18.2-rc0

    ๐Ÿš€ Ollama v0.18.2-rc0 is out โ€” and itโ€™s bringing live web search!

    This release candidate (not quite stable yet, but very promising) adds support for web search integration via the `openclaw` tool โ€” likely enabling models to pull in real-time info from the web during inference. ๐ŸŒ๐Ÿ”

    ๐Ÿ”น New Tool Registered: `openclaw` is now available as a callable tool โ€” think of it like giving your local LLM a live browser.

    ๐Ÿ”น Tool-Calling Boost: Aligns with the growing trend of models (like Llama 3.1) supporting function/tool calling โ€” making Ollama even more dynamic and up-to-date.

    ๐Ÿ”น Use Cases? Think: real-time Q&A, research assistants, news-aware chatbots โ€” all running locally (or on your own server).

    โš ๏ธ Note: This is an `rc0`, so things may shift before the final `v0.18.2` lands โ€” but itโ€™s a huge step toward truly intelligent, always-current local AI.

    Curious about how `openclaw` works? Or want to test it out? Letโ€™s dive in! ๐Ÿงช๐Ÿ’ก

    ๐Ÿ”— View Release

  • Ollama – v0.18.1

    Ollama – v0.18.1

    ๐Ÿšจ Ollama v0.18.1 is live โ€” and itโ€™s bringing seriously useful upgrades! ๐Ÿšจ

    ๐Ÿ”ฅ Web Search is BACK & fixed โ€” no more broken links or silent failures.

    ๐ŸŒ New Web Fetch feature โ€” Ollama can now pull in live web content (think docs, articles, news) to enrich its responses.

    โœ… Both features are ON by default โ€” zero config needed! Just `ollama run` and go.

    ๐Ÿ’ป Works locally too โ€” even when running on your machine (as long as youโ€™ve got internet access).

    ๐Ÿ’ก Why youโ€™ll love this:

    • Real-time context for RAG pipelines
    • Fact-checking on the fly
    • Fresh info without retraining or API keys

    Perfect for devs building local, context-aware apps โ€” or just curious tinkerers who want smarter, up-to-date answers. ๐Ÿ› ๏ธโœจ

    Check it out: ollama.com

    Need a quick demo or config tweaks? Drop a ๐Ÿง  below!

    ๐Ÿ”— View Release

  • Ollama – v0.18.1-rc1

    Ollama – v0.18.1-rc1

    ๐Ÿšจ Ollama v0.18.1-rc1 is here โ€” and itโ€™s bringing web smarts to your local LLMs! ๐ŸŒโœจ

    ๐Ÿ”ฅ Whatโ€™s new?

    โœ… Web search is back & fixed! No more silent failures โ€” Ollama can now reliably pull live results.

    ๐ŸŒ New `web fetch` tool โ€” ask Ollama to retrieve up-to-the-minute web content and use it directly in responses.

    โšก Both features are ON by default โ€” just update, and youโ€™re ready to go (no config needed!).

    ๐Ÿ’ก Why care?

    This means your local models can now answer questions with real-time context โ€” think news, docs, or live data โ€” all without hitting external APIs. Perfect for building smarter, self-contained AI apps ๐Ÿง ๐Ÿ’ป

    ๐Ÿ“ฆ Release type: Release Candidate (`v0.18.1-rc1`) โ€” ideal for testing and feedback before the stable drop!

    Ready to test it out? ๐Ÿ› ๏ธ Let us know how your web-aware Ollama experiments go! ๐Ÿš€

    ๐Ÿ”— View Release

  • Ollama – v0.18.1-rc0: cmd/launch: skip –install-daemon when systemd is unavailable (#14883)

    Ollama – v0.18.1-rc0: cmd/launch: skip –install-daemon when systemd is unavailable (#14883)

    ๐Ÿš€ Ollama v0.18.1-rc0 is here โ€” and itโ€™s fixing a pesky containerization hiccup!

    ๐Ÿ”ฅ Whatโ€™s new?

    • ๐Ÿ› ๏ธ `ollama launch openclaw –install-daemon` no longer fails in environments without systemd (like Docker, WSL, Alpine, or CI runners).
    • โœ… Smart detection: Ollama now checks if systemd is actually available before trying to install the daemon.
    • ๐Ÿงช Falls back gracefully to foreground mode when systemd isnโ€™t present โ€” meaning your local LLMs keep running, no matter the setup!

    ๐Ÿ“ฆ Why it matters:

    Perfect for devs testing in containers or minimal Linux setups โ€” less friction, more inference! ๐Ÿณ๐Ÿ’ป

    ๐Ÿ‘‰ Try the RC and let us know how it plays in your envs!

    #Ollama #LLM #AIEnthusiasts

    ๐Ÿ”— View Release

  • Ollama – v0.18.0

    Ollama – v0.18.0

    ๐Ÿšจ Ollama v0.18.0 is live! ๐Ÿšจ

    The latest drop brings a slick backend upgrade โ€” no flashy new features yet, but some important under-the-hood polish:

    ๐Ÿ”น Zstandard (`zstd`) request decompression now works in the cloud passthrough middleware โ€” meaning smoother communication with proxies, CDNs, or cloud services that compress HTTP payloads.

    ๐Ÿ”น Fixes potential issues where compressed API requests (especially large ones) mightโ€™ve failed or timed out.

    ๐Ÿ”น A quiet but meaningful win for reliability in production-like setups โ€” think: self-hosted gateways, reverse proxies (like NGINX), or cloud load balancers.

    ๐Ÿ’ก Pro tip: If youโ€™re using Ollama behind a proxy or sending big payloads via the API, this oneโ€™s for you. Run `ollama pull` to upgrade and test it out!

    ๐Ÿ”— Release on GitHub โ€” fingers crossed the notes load this time ๐Ÿ˜‰

    ๐Ÿ”— View Release

  • Ollama – v0.18.0-rc2

    Ollama – v0.18.0-rc2

    ๐Ÿšจ Ollama v0.18.0-rc2 is out โ€” and itโ€™s packing some sneaky performance upgrades! ๐Ÿšจ

    ๐Ÿ”ฅ Whatโ€™s new?

    • โœ… Zstandard (zstd) decompression support added โ€” the server now handles compressed request bodies using `zstd`, especially in cloud passthrough middleware.

    โ†’ Think faster, leaner data transfers when proxying to remote backends (hello, reduced bandwidth & latency!).

    • ๐ŸŒฉ๏ธ Likely a stepping stone toward smoother Ollama Cloud integrations or hybrid local/cloud inference workflows.
    • ๐Ÿ› ๏ธ This is Release Candidate 2, so itโ€™s mostly polish, bug fixes, and stability tweaks ahead of the final `v0.18.0` drop.

    ๐Ÿ’ก Pro tip: Try it out (if youโ€™re feeling adventurous!) with:

    “`bash

    ollama pull ollama:rc

    “`

    โ€ฆor keep an eye on the GitHub release page once itโ€™s live โ€” full changelog incoming soon! ๐Ÿ•ต๏ธโ€โ™‚๏ธ

    Whoโ€™s testing first? ๐Ÿ˜Ž

    ๐Ÿ”— View Release

  • Ollama – v0.18.0-rc1

    Ollama – v0.18.0-rc1

    ๐Ÿšจ Ollama v0.18.0-rc1 is here โ€” and itโ€™s packing some serious upgrades! ๐Ÿšจ

    ๐Ÿ”ฅ Anthropic Model Fixes

    • Fixed parsing of `close_thinking` blocks before `tool_use`, especially when no intermediate text is present โ€” critical for clean tool invocations in Claude-style models.

    ๐Ÿ› ๏ธ Tool Use & Function Calling

    • Major improvements for structured outputs and function calling โ€” think smoother integrations with `claude-3.5-sonnet` and similar models.

    โšก Performance & Stability Boosts

    • Optimized context handling & reduced memory footprint.
    • Fixed bugs in multi-turn tool-based conversations โ€” fewer hiccups, more reliability.

    ๐Ÿ’ป Platform Love

    • Updated CUDA & Metal backends for faster inference.
    • Better Apple Silicon (M-series) support โ€” and improved WSL2 & native Windows performance.

    CLI/API Tweaks ๐Ÿ› ๏ธ

    • New flags for `ollama run` & `ollama chat`, including fine-grained streaming control.
    • Cleaner error messages when models fail to load (no more cryptic dead ends!).

    This RC is a solid preview of whatโ€™s coming โ€” especially if youโ€™re relying on tool use, local Claude-style models, or pushing Ollama hard on macOS/Windows. ๐Ÿงช Try it out and let us know what you think!

    ๐Ÿ”— Download v0.18.0-rc1

    #Ollama #LLMs #AIEnthusiasts ๐Ÿค–

    ๐Ÿ”— View Release