Ollama – v0.13.0
๐ Ollama v0.13.0 is live โ and itโs a game-changer for local LLM folks!
Meet DeepSeek-V3.1 (aka Deepseek2) โ now officially supported with 128K context, razor-sharp reasoning, and killer coding skills. But hereโs the kicker: itโs running on Ollamaโs brand-new engine with MLA (Multi-Layer Attention) โ meaning faster token generation, lower latency, and no more sluggish long-context hangs.
โจ Whatโs new?
- โ DeepSeek-V3.1 support โ perfect for complex prompts, multilingual tasks & code generation
- ๐ MLA engine = smoother, faster inference on both CPU and GPU (NVIDIA/AMD)
- ๐ก Optimized streaming โ ideal for chat apps, agents, and real-time LLM workflows
Just run `ollama pull deepseek2` and feel the difference. No more waiting. Just pure, local LLM power. ๐ค๐ป
