Ollama – v0.18.4-rc1
๐ Ollama v0.18.4-rc1 is here โ and itโs packing a subtle but smart update!
๐ Whatโs new?
Ollama now warns you if your server context length is below 64k tokens when running local models. Why? Because newer LLMs (like Llama 3.1, Mistral Large, DeepSeek-R1) are built for long contexts โ and running them with too little context can lead to truncated outputs or weird behavior. This warning helps you avoid those gotchas before they bite! ๐ก
๐ ๏ธ Bonus: While the full changelog is still loading on GitHub, this RC likely includes:
- Stability tweaks for model loading
- Improved error messages (especially around context handling)
- Minor CLI/web UI polish
๐ Pro tip: If you’re using large-context models (e.g., `llama3.1:8b-instruct-q4_K_M`), double-check your `OLLAMA_MAX_LOADED_MODELS` and context settings โ this warning is here to help you optimize!
๐ Grab the RC: v0.18.4-rc1 on GitHub
๐ฌ Join the convo: Ollama Discord
Let us know if you spot any quirks or love the warning โ feedback helps shape the final release! ๐
