Ollama – v0.13.0

Ollama – v0.13.0

๐Ÿš€ Ollama v0.13.0 is live โ€” and itโ€™s a game-changer for local LLM folks!

Meet DeepSeek-V3.1 (aka Deepseek2) โ€” now officially supported with 128K context, razor-sharp reasoning, and killer coding skills. But hereโ€™s the kicker: itโ€™s running on Ollamaโ€™s brand-new engine with MLA (Multi-Layer Attention) โ€” meaning faster token generation, lower latency, and no more sluggish long-context hangs.

โœจ Whatโ€™s new?

  • โœ… DeepSeek-V3.1 support โ€” perfect for complex prompts, multilingual tasks & code generation
  • ๐Ÿš€ MLA engine = smoother, faster inference on both CPU and GPU (NVIDIA/AMD)
  • ๐Ÿ’ก Optimized streaming โ€” ideal for chat apps, agents, and real-time LLM workflows

Just run `ollama pull deepseek2` and feel the difference. No more waiting. Just pure, local LLM power. ๐Ÿค–๐Ÿ’ป

๐Ÿ”— View Release