Ollama – v0.18.0-rc1
π¨ Ollama v0.18.0-rc1 is here β and itβs packing some serious upgrades! π¨
π₯ Anthropic Model Fixes
- Fixed parsing of `close_thinking` blocks before `tool_use`, especially when no intermediate text is present β critical for clean tool invocations in Claude-style models.
π οΈ Tool Use & Function Calling
- Major improvements for structured outputs and function calling β think smoother integrations with `claude-3.5-sonnet` and similar models.
β‘ Performance & Stability Boosts
- Optimized context handling & reduced memory footprint.
- Fixed bugs in multi-turn tool-based conversations β fewer hiccups, more reliability.
π» Platform Love
- Updated CUDA & Metal backends for faster inference.
- Better Apple Silicon (M-series) support β and improved WSL2 & native Windows performance.
CLI/API Tweaks π οΈ
- New flags for `ollama run` & `ollama chat`, including fine-grained streaming control.
- Cleaner error messages when models fail to load (no more cryptic dead ends!).
This RC is a solid preview of whatβs coming β especially if youβre relying on tool use, local Claude-style models, or pushing Ollama hard on macOS/Windows. π§ͺ Try it out and let us know what you think!
π Download v0.18.0-rc1
#Ollama #LLMs #AIEnthusiasts π€
