Ollama – v0.15.0-rc6

Ollama – v0.15.0-rc6

πŸš€ Ollama v0.15.0-rc6 just dropped β€” and it’s a quiet hero for GPU users!

If you’ve been hitting CUDA MMA errors when running quantized Llama models on your RTX card, breathe easy. This patch slays those sneaky crashes during inference.

βœ… Fixed: CUDA MMA bugs in release builds

🚫 No more mysterious GPU crashes β€” stable, fast, local LLMs back on track

Perfect for devs pushing limits on NVIDIA hardware. GGUF? Still supported. API? Still sweet. Just… smoother.

Run it hard. Run it local. πŸ–₯️πŸ”₯

πŸ”— View Release