Ollama – v0.30.0-rc3

Ollama – v0.30.0-rc3

Ollama just dropped v0.30.0-rc3, and it looks like the team is hard at work smoothing out the edges for Windows users! 🛠️

If you haven’t tried Ollama yet, it’s the ultimate framework for running powerful LLMs like Llama 3, DeepSeek-R1, and Mistral locally on your own machine. It’s a total game-changer for privacy-focused devs and anyone wanting to experiment with AI without worrying about API costs or limits.

What’s new in this release candidate:

  • Windows ROCm Fix: The big highlight here is a specific fix for the Windows ROCm build. This is huge news for anyone trying to leverage AMD GPUs on Windows to accelerate their local model inference! 🚀
  • CI Improvements: The update includes much-needed continuous integration (CI) fixes to ensure more stable, reliable builds moving forward.

This is a targeted release focused on stability and hardware compatibility, making sure your local AI setup stays buttery smooth!

🔗 View Release