Ollama – v0.18.0-rc1

Ollama – v0.18.0-rc1

🚨 Ollama v0.18.0-rc1 is here β€” and it’s packing some serious upgrades! 🚨

πŸ”₯ Anthropic Model Fixes

  • Fixed parsing of `close_thinking` blocks before `tool_use`, especially when no intermediate text is present β€” critical for clean tool invocations in Claude-style models.

πŸ› οΈ Tool Use & Function Calling

  • Major improvements for structured outputs and function calling β€” think smoother integrations with `claude-3.5-sonnet` and similar models.

⚑ Performance & Stability Boosts

  • Optimized context handling & reduced memory footprint.
  • Fixed bugs in multi-turn tool-based conversations β€” fewer hiccups, more reliability.

πŸ’» Platform Love

  • Updated CUDA & Metal backends for faster inference.
  • Better Apple Silicon (M-series) support β€” and improved WSL2 & native Windows performance.

CLI/API Tweaks πŸ› οΈ

  • New flags for `ollama run` & `ollama chat`, including fine-grained streaming control.
  • Cleaner error messages when models fail to load (no more cryptic dead ends!).

This RC is a solid preview of what’s coming β€” especially if you’re relying on tool use, local Claude-style models, or pushing Ollama hard on macOS/Windows. πŸ§ͺ Try it out and let us know what you think!

πŸ”— Download v0.18.0-rc1

#Ollama #LLMs #AIEnthusiasts πŸ€–

πŸ”— View Release