• Ollama – v0.30.0-rc11

    Ollama – v0.30.0-rc11

    Ollama v0.30.0-rc11 is here! 🛠️

    If you’re running your own local LLM playground with models like Llama 3 or DeepSeek-R1, listen up! Ollama is the ultimate framework for getting powerful open-source models up and running on your machine without the cloud headache. This latest release candidate is all about polishing the experience and smoothing out those pesky edge cases.

    What’s new in this update:

    • Windows Optimization: Big win for my Windows tinkerer friends! 🪟 The team implemented a fix to prevent issues caused by spaces in the compiler path, making builds and installations much more stable on Windows environments.
    • Stability Focus: Since this is an `rc` (release candidate) update, the primary goal is bug squashing. It’s designed to refine the stability of the current version before the full stable rollout hits your machine.

    Keep testing these RC builds—they are the secret sauce to catching bugs before they hit your production workflows! 🚀

    đź”— View Release

  • Ollama – v0.30.0-rc9

    Ollama – v0.30.0-rc9

    Ollama just dropped v0.30.0-rc9, and it looks like the team is focusing on polishing up the Windows experience! 🛠️

    If you’re looking to run heavy hitters like Llama 3, DeepSeek-R1, or Mistral locally without relying on expensive cloud APIs, Ollama is your best friend. It handles all the heavy lifting of downloading and setting up models right on your machine.

    What’s new in this release candidate:

    • Windows Build Fix: The big news here is a fix for the Windows build assets. This update aims to ensure much smoother installations and more reliable builds for our fellow Windows power users. 🪟

    It might seem like a small tweak, but stability is everything when you’re managing local inference! Keep an eye on these release candidates as they move toward the final stable version. 🚀

    đź”— View Release

  • Ollama – v0.30.0-rc8: Merge remote-tracking branch ‘upstream/main’ into llama-runner-phase-0

    Ollama – v0.30.0-rc8: Merge remote-tracking branch ‘upstream/main’ into llama-runner-phase-0

    Ollama v0.30.0-rc8 🦙

    If you love running heavy-hitting models like Llama 3, DeepSeek-R1, or Mistral right on your own hardware, keep your eyes peeled! Ollama is making some serious moves under the hood to prep for the next generation of local LLM performance.

    This release candidate is all about synchronization and stability as the team merges the latest upstream updates into the `llama-runner-phase-0` branch. Here’s the breakdown:

    • Core Engine Sync: The development branch is being synced with the main upstream codebase. This is a huge step toward ensuring stability for upcoming runner features!
    • Configuration Refinement: Updates to `envconfig/config.go` mean smoother handling of environment variables and runtime configurations. ⚙️
    • Hardened Architectures: New updates to integration tests (specifically for model architectures and context windows) mean the engine is getting much more robust at handling complex, diverse models.

    It’s a “behind-the-scenes” kind of update, but it’s exactly what we need to see for smoother, more reliable local workflows! 🛠️

    đź”— View Release

  • Lemonade – v10.4.0: Fix ollama tool calling (#1780)

    Lemonade – v10.4.0: Fix ollama tool calling (#1780)

    🍋 Lemonade SDK v10.4.0 is officially here!

    If you’ve been looking for a way to squeeze every bit of performance out of your local hardware—especially leveraging NPUs and GPUs—Lemonade is the toolkit you need. It brings high-performance, private LLM serving right to your Windows or Linux machine with OpenAI API compatibility.

    This latest update is a massive win for anyone building AI apps that rely on Ollama for local model execution. We’re seeing much better stability and much clearer communication between the SDK and your local engines.

    What’s new in v10.4.0:

    • Enhanced Model Visibility: You can now pull specific model capabilities and context window sizes directly via the Ollama API. No more “context window guesswork” when building your prompts! đź§ 
    • Fixed Tool Calling Logic: We’ve squashed a major bug where tool calls were being sent in streaming mode, which was causing malformed responses on the client side. Your agents should behave much better now.
    • Protocol Alignment: Fixed a mismatch between the Ollama protocol and the OpenAI-compatible protocol used by `llama-server`. This ensures requests forwarded through Lemonade are no longer rejected as malformed.

    Pro Tip: This patch was specifically tested with Zed v1.0.0. If you use Zed’s agent chat with built-in Ollama support, your coding workflow just got a whole lot more reliable! 🛠️

    đź”— View Release

  • Ollama – v0.30.0-rc7

    Ollama – v0.30.0-rc7

    Ollama just dropped v0.30.0-rc7, and it looks like the team is fine-tuning things for an even smoother local LLM experience! 🛠️

    If you’re looking to run heavyweights like Llama 3, DeepSeek-R1, or Mistral directly on your hardware without a massive cloud budget, this is the tool you need in your stack. This latest release candidate is all about stability and polishing the engine before the official stable rollout.

    Here’s what’s new in this update:

    • OpenMP Optimization: The team has implemented a change to disable OpenMP. This is a big deal if you’ve been running into threading conflicts or stability issues during model execution—it helps prevent clashes with other parallel processing libraries on your machine.
    • Final Testing Phase: As an `rc7` build, this version is in the home stretch of bug-squashing. It’s the perfect time to test it out and see if these tweaks resolve any crashes you’ve been seeing during long inference sessions. 🚀

    đź”— View Release

  • Ollama – v0.30.0-rc6

    Ollama – v0.30.0-rc6

    Ollama v0.30.0-rc6 🛠️

    If you’re running local LLMs, you know Ollama is the go-to for getting models like Llama 3 and DeepSeek-R1 up and running with zero friction. This latest release candidate is a targeted update focused on stability for our Windows users!

    What’s new:

    • Windows Fix: The primary update in this release addresses an issue with dependency gathering on Windows systems. 🪟

    This is a great little patch to keep your local inference engine running smoothly without those pesky missing dependency errors during setup or updates. Keep tinkering! 🚀

    đź”— View Release

  • Ollama – v0.30.0-rc5

    Ollama – v0.30.0-rc5

    Ollama just dropped v0.30.0-rc5, and it’s looking like a vital stability patch for our local LLM setups! 🛠️

    If you haven’t tried Ollama yet, it is the ultimate toolkit for running powerful models like Llama 3, DeepSeek-R1, and Mistral directly on your own hardware. It handles all the heavy lifting of downloading and configuring models so you can focus on building and experimenting.

    What’s new in this release:

    • Windows Stability Fix: This update specifically targets Windows dependency issues. If you’ve been hitting roadblocks or runtime errors while trying to get things running on a PC, this patch is designed to smooth out those installation hiccups. 🪟

    It’s a release candidate, so it’s a focused effort to ensure the Windows experience is as seamless and friction-free as possible!

    đź”— View Release

  • Ollama – v0.30.0-rc4: ci: windows mlx tuning

    Ollama – v0.30.0-rc4: ci: windows mlx tuning

    Ollama v0.30.0-rc4 🦙

    If you’re looking to run large language models locally with ease, Ollama remains the gold standard for managing and interacting with LLMs on your own machine. This latest release candidate brings some much-needed optimization and fixes to the build process!

    What’s new in this release:

    • Windows MLX Tuning: Significant updates to the CI pipeline specifically focused on tuning MLX performance for Windows environments. If you’re running locally on Windows, expect smoother execution! 🛠️
    • Build Optimization: The team has shortened the “long-tail” during the build process, making the development and deployment cycle much snappier.
    • Installer Fix: A crucial fix to bring `OllamaSetup.exe` back under the 2GB limit. This ensures smoother downloads and prevents those pesky installation hurdles caused by file size limits.

    Great news for the Windows crowd looking to squeeze more performance out of their local setups! 🚀

    đź”— View Release

  • Ollama – v0.30.0-rc3

    Ollama – v0.30.0-rc3

    Ollama just dropped v0.30.0-rc3, and it looks like the team is hard at work smoothing out the edges for Windows users! 🛠️

    If you haven’t tried Ollama yet, it’s the ultimate framework for running powerful LLMs like Llama 3, DeepSeek-R1, and Mistral locally on your own machine. It’s a total game-changer for privacy-focused devs and anyone wanting to experiment with AI without worrying about API costs or limits.

    What’s new in this release candidate:

    • Windows ROCm Fix: The big highlight here is a specific fix for the Windows ROCm build. This is huge news for anyone trying to leverage AMD GPUs on Windows to accelerate their local model inference! 🚀
    • CI Improvements: The update includes much-needed continuous integration (CI) fixes to ensure more stable, reliable builds moving forward.

    This is a targeted release focused on stability and hardware compatibility, making sure your local AI setup stays buttery smooth!

    đź”— View Release

  • Ollama – v0.30.0-rc1

    Ollama – v0.30.0-rc1

    Ollama v0.30.0-rc1 🦬

    If you haven’t jumped on the Ollama train yet, now is the time! It’s the ultimate go-to tool for running powerful large language models like Llama 3, DeepSeek-R1, and Mistral locally on your machine with zero friction. It handles all the heavy lifting of model management and serving so you can focus on building cool stuff.

    This latest release candidate (rc1) is a focused stability update:

    • Windows MLX Build Fix: The team has pushed a fix specifically for the Windows MLX build process. If you’ve been experimenting with MLX-related workflows on Windows, this should smooth out those compilation hiccups! 🛠️

    It looks like a targeted patch to keep your local LLM engine running buttery smooth across different environments. Keep an eye out for more feature-heavy lifting in the upcoming full release!

    đź”— View Release