• Ollama – v0.18.3: api/show: overwrite basename for copilot chat (#15062)

    Ollama – v0.18.3: api/show: overwrite basename for copilot chat (#15062)

    ๐Ÿš€ Ollama v0.18.3 is live!

    This patch fixes a subtle but important quirk in how Ollama interacts with GitHub Copilot Chat โ€” specifically around model naming.

    ๐Ÿ”น Whatโ€™s fixed?

    The `/api/show` endpoint now returns `req.Model` (the actual model name from `/api/tags`) instead of the generic `general.basename`.

    โœ… Why youโ€™ll love this:

    • No more confusing name collisions when multiple models share the same basename.
    • Copilot Chat now displays exactly the model name you expect โ€” making selection cleaner and more intuitive.
    • Better alignment with how Ollama itself labels models elsewhere.

    Itโ€™s a small but mighty polish update โ€” perfect for Copilot users who want seamless, accurate model switching. ๐Ÿ› ๏ธโœจ

    ๐Ÿ‘‰ Grab the update and keep local LLM-ing!

    `ollama pull ollama` ๐ŸŽฏ

    ๐Ÿ”— View Release

  • Ollama – v0.18.3-rc2: api/show: overwrite basename for copilot chat (#15062)

    Ollama – v0.18.3-rc2: api/show: overwrite basename for copilot chat (#15062)

    ๐Ÿšจ Ollama v0.18.3-rc2 is out! ๐Ÿšจ

    This update fixes a pesky bug in GitHub Copilot integration โ€” specifically, how model names are displayed when using `/api/show`.

    ๐Ÿ”น Whatโ€™s fixed?

    • Previously, Copilot would see internal `general.basename` values (e.g., `deepseek-r1:7b-v3-fp16`) instead of clean, user-friendly model names like `deepseek-r1`.
    • Now, `/api/show` returns the actual requested model name (`req.Model`) โ€” matching what you see in `ollama list` or `/api/tags`.

    โœ… Why itโ€™s a win:

    • Cleaner, more intuitive model names in Copilot.
    • No more duplicate or confusing basenames when switching between similar models (e.g., `llama3:8b` vs. `llama3:70b`).

    ๐Ÿง  Bonus: This small tweak makes Copilot + Ollama feel way more polished โ€” especially for those of us juggling multiple local models.

    ๐Ÿ“ฆ RC2 dropped on Mar 25, 2024 โ€” grab it and give Copilot a spin! ๐Ÿ› ๏ธ

    ๐Ÿ”— GitHub PR #15062

    Let us know how it feels! ๐Ÿ™Œ

    ๐Ÿ”— View Release

  • Ollama – v0.18.3-rc1

    Ollama – v0.18.3-rc1

    ๐Ÿšจ Ollama v0.18.3-rc1 is out! ๐Ÿšจ

    A quick heads-up for our fellow AI tinkerers โ€” the latest release candidate for Ollama is here, and while itโ€™s a small RC, it packs a useful fix:

    ๐Ÿ”ง Whatโ€™s new?

    โœ… Windows CGO compiler error fixed (PR #15046) โ€” this resolves a CI/CD hiccup that was causing build issues on Windows. If youโ€™ve been hitting weird compilation errors or CI failures on Windows, this oneโ€™s for you!

    ๐Ÿ” Note: Due to a temporary GitHub UI glitch, the full release notes didnโ€™t load cleanly โ€” so this RC currently only has one confirmed change. If youโ€™re feeling adventurous and want to test the latest fixes (especially on Windows), go grab `v0.18.3-rc1` from the GitHub Releases page.

    Let us know if you run into anything โ€” happy prompting! ๐Ÿง โœจ

    ๐Ÿ”— View Release

  • Ollama – v0.18.3-rc0: mlx: add mxfp4/mxfp8/nvfp4 importing (#15015)

    Ollama – v0.18.3-rc0: mlx: add mxfp4/mxfp8/nvfp4 importing (#15015)

    ๐Ÿšจ Ollama v0.18.3-rc0 is here โ€” and itโ€™s a quantization powerhouse! ๐Ÿšจ

    The latest release adds supercharged import support for MLX (Appleโ€™s Metal framework) and NVIDIAโ€™s new FP8 formats โ€” meaning you can now run even more efficient, low-bit models locally. Hereโ€™s the breakdown:

    ๐Ÿ”น New Quantization Imports

    โœ… Import BF16 models โ†’ convert on-the-fly to:

    • `mxfp4` (Metaโ€™s 4-bit mixed-precision)
    • `mxfp8` (Metaโ€™s 8-bit mixed-precision)
    • `nvfp4` (NVIDIAโ€™s 4-bit floating-point format)

    โœ… Import FP8 models โ†’ convert directly to `mxfp8`

    ๐ŸŽฏ Why this rocks:

    • ๐Ÿ Apple Silicon users (M1/M2/M3): Run ultra-efficient MLX-native models with minimal memory footprint.
    • ๐ŸŽฎ NVIDIA fans: Get early access to NVFP4 โ€” a promising new format for faster, smaller inference.
    • โšก Smaller models + less VRAM = more models on your laptop, fewer cloud trips.

    This is a big leap toward truly portable, hardware-agnostic LLM inference โ€” all from your desktop. ๐Ÿง ๐Ÿ’ป

    Curious how `mxfp4` stacks up against `nvfp4`? Let us know โ€” happy to deep dive! ๐Ÿงต

    ๐Ÿ”— View Release

  • Home Assistant Voice Pe – 26.3.0

    Home Assistant Voice Pe – 26.3.0

    ๐Ÿšจ Home Assistant Voice PE 26.3.0 is live! ๐Ÿšจ

    Big updates in this releaseโ€”letโ€™s break it down:

    ๐ŸŽง Media Playback Stability

    No more audio/video hiccups! Expect smoother, rock-solid playback across devicesโ€”perfect for voice-triggered media or smart speakers.

    ๐ŸŒ Multiple Sendspin Servers

    Now supports multiple Sendspin backends simultaneously. More redundancy, better failover, and improved scalability for larger setups.

    โฑ๏ธ TTS Timeout Fix

    Say goodbye to cut-off voice responses! Text-to-speech now waits properlyโ€”your AI replies play fully before moving on.

    ๐ŸŒŸ New Contributor Alert!

    Shoutout to @akloeckner for their first PR (#558)โ€”welcome to the crew! ๐Ÿ™Œ

    โœจ Bonus: The project is now officially sponsored by the Open Home Foundationโ€”a huge vote of confidence in open, private-by-design voice control!

    ๐Ÿ“ฆ Full changelog: [25.12.4 โ†’ 26.3.0](#)

    Letโ€™s make our homes smarterโ€”privately and offline-capable. ๐Ÿ ๐Ÿ—ฃ๏ธ

    ๐Ÿ”— View Release

  • Ollama – v0.18.2: launch: fix openclaw not picking up newly selected model (#14943)

    Ollama – v0.18.2: launch: fix openclaw not picking up newly selected model (#14943)

    ๐Ÿšจ Ollama v0.18.2 is live! ๐Ÿšจ

    This patch fixes a sneaky bug in OpenCLaW (Ollamaโ€™s model-launching engine) where switching models mid-session would not actually switch the active model โ€” yikes! ๐Ÿ˜…

    ๐Ÿ”ง Whatโ€™s fixed?

    • Previously: Changing the primary model in GUI/CLI wouldnโ€™t update active sessions โ€” youโ€™d keep running the old model, even if it looked like youโ€™d switched.
    • Now: Sessions properly refresh when the model changes โ€” no more stale model confusion! โœ…

    ๐ŸŽฏ Why it matters:

    • Perfect for devs testing multiple models in one session (e.g., comparing Llama 3 vs. DeepSeek).
    • Critical for demos or workflows where model switching is part of the flow โ€” reliability restored!

    ๐Ÿ”— PR #14943

    ๐Ÿ“ฆ Verified by BruceMacD on Mar 18, 20:20 UTC

    Ready to upgrade? `ollama update` && happy model-hopping! ๐Ÿง โœจ

    ๐Ÿ”— View Release

  • Ollama – v0.18.2-rc1: launch: fix openclaw not picking up newly selected model (#14943)

    Ollama – v0.18.2-rc1: launch: fix openclaw not picking up newly selected model (#14943)

    ๐Ÿšจ Ollama v0.18.2-rc1 is live! ๐Ÿšจ

    Quick, focused fix in this release candidate โ€” perfect for keeping your local LLM workflows smooth and reliable.

    ๐Ÿ” Whatโ€™s New?

    โœ… Fixed `openclaw` model-switching bug โ€” previously, if you changed the active model mid-session without restarting, `openclaw` would keep using the old (stale) model. Now it correctly picks up the new selection instantly!

    ๐Ÿ’ก Why Youโ€™ll Care:

    • ๐Ÿ”„ Makes dynamic model switching (e.g., testing Llama 3 vs. DeepSeek-R1) way more reliable in dev/test loops
    • ๐Ÿงช Critical for tooling and integrations that rely on runtime model changes
    • โšก Clean, minimal change โ€” no side effects or breaking changes

    ๐Ÿ‘‰ Grab the RC and give it a spin before final `v0.18.2` drops!

    Let the team know if you spot anything weird ๐Ÿžโžก๏ธโœ…

    #Ollama #LLM #LocalAI #DevTools

    ๐Ÿ”— View Release

  • Ollama – v0.18.2-rc0

    Ollama – v0.18.2-rc0

    ๐Ÿš€ Ollama v0.18.2-rc0 is out โ€” and itโ€™s bringing live web search!

    This release candidate (not quite stable yet, but very promising) adds support for web search integration via the `openclaw` tool โ€” likely enabling models to pull in real-time info from the web during inference. ๐ŸŒ๐Ÿ”

    ๐Ÿ”น New Tool Registered: `openclaw` is now available as a callable tool โ€” think of it like giving your local LLM a live browser.

    ๐Ÿ”น Tool-Calling Boost: Aligns with the growing trend of models (like Llama 3.1) supporting function/tool calling โ€” making Ollama even more dynamic and up-to-date.

    ๐Ÿ”น Use Cases? Think: real-time Q&A, research assistants, news-aware chatbots โ€” all running locally (or on your own server).

    โš ๏ธ Note: This is an `rc0`, so things may shift before the final `v0.18.2` lands โ€” but itโ€™s a huge step toward truly intelligent, always-current local AI.

    Curious about how `openclaw` works? Or want to test it out? Letโ€™s dive in! ๐Ÿงช๐Ÿ’ก

    ๐Ÿ”— View Release

  • Ollama – v0.18.1

    Ollama – v0.18.1

    ๐Ÿšจ Ollama v0.18.1 is live โ€” and itโ€™s bringing seriously useful upgrades! ๐Ÿšจ

    ๐Ÿ”ฅ Web Search is BACK & fixed โ€” no more broken links or silent failures.

    ๐ŸŒ New Web Fetch feature โ€” Ollama can now pull in live web content (think docs, articles, news) to enrich its responses.

    โœ… Both features are ON by default โ€” zero config needed! Just `ollama run` and go.

    ๐Ÿ’ป Works locally too โ€” even when running on your machine (as long as youโ€™ve got internet access).

    ๐Ÿ’ก Why youโ€™ll love this:

    • Real-time context for RAG pipelines
    • Fact-checking on the fly
    • Fresh info without retraining or API keys

    Perfect for devs building local, context-aware apps โ€” or just curious tinkerers who want smarter, up-to-date answers. ๐Ÿ› ๏ธโœจ

    Check it out: ollama.com

    Need a quick demo or config tweaks? Drop a ๐Ÿง  below!

    ๐Ÿ”— View Release

  • Ollama – v0.18.1-rc1

    Ollama – v0.18.1-rc1

    ๐Ÿšจ Ollama v0.18.1-rc1 is here โ€” and itโ€™s bringing web smarts to your local LLMs! ๐ŸŒโœจ

    ๐Ÿ”ฅ Whatโ€™s new?

    โœ… Web search is back & fixed! No more silent failures โ€” Ollama can now reliably pull live results.

    ๐ŸŒ New `web fetch` tool โ€” ask Ollama to retrieve up-to-the-minute web content and use it directly in responses.

    โšก Both features are ON by default โ€” just update, and youโ€™re ready to go (no config needed!).

    ๐Ÿ’ก Why care?

    This means your local models can now answer questions with real-time context โ€” think news, docs, or live data โ€” all without hitting external APIs. Perfect for building smarter, self-contained AI apps ๐Ÿง ๐Ÿ’ป

    ๐Ÿ“ฆ Release type: Release Candidate (`v0.18.1-rc1`) โ€” ideal for testing and feedback before the stable drop!

    Ready to test it out? ๐Ÿ› ๏ธ Let us know how your web-aware Ollama experiments go! ๐Ÿš€

    ๐Ÿ”— View Release