Category: AI

AI Releases

  • Ollama – v0.14.3-rc3: model: add lfm2 architecture and LFM2.5-1.2B-Thinking support (#13792)

    Ollama – v0.14.3-rc3: model: add lfm2 architecture and LFM2.5-1.2B-Thinking support (#13792)

    Big news for AI tinkerers! 🚀

    Ollama v0.14.3-rc3 just dropped with native support for the brand-new LFM2 architecture and its first model: LFM2.5-1.2B-Thinking — a lean 1.2B parameter model built for reasoning, not just generation.

    🧠 Think step-by-step problem solving, code reasoning, and complex QA — all running locally with zero cloud latency.

    Pull it in seconds:

    `ollama pull lfm2.5:1.2b-thinking`

    No more waiting for APIs — now you’ve got a tiny, thinking LLM on your machine. Perfect for dev experiments, edge deployments, or just geeking out in privacy.

    #Ollama #LLMs #LocalAI #LFM2

    🔗 View Release

  • ComfyUI – v0.10.0

    ComfyUI – v0.10.0

    ComfyUI v0.10.0 just dropped—and it’s a game-changer 🎨⚡

    • Native WebUI Integration: Drag & drop your Stable Diffusion WebUI models directly. No more conversion headaches.
    • Dynamic Prompts in Nodes: Use `{prompt}`, `{seed}`, or `{CFG}` inside inputs—batch test variations without cloning nodes.
    • 30% Faster Workflows: Smarter node caching = quicker loads on massive pipelines.
    • New “Batch Sampler” Node: Generate 50+ variations in one go—randomize seeds, styles, CFG—all from a single node.
    • Dark Mode Upgrades: Smoother, higher contrast—perfect for late-night prompt tinkering.
    • Linux ARM64 Support: M-series Macs & Raspberry Pi 5? You’re now fully supported. 🍏🧠

    Pro tip: Combine Batch Sampler + Dynamic Prompts to auto-generate character concepts in seconds. Perfect for artists, devs, and AI tinkerers.

    Update now. The nodes are alive. 🎮✨

    🔗 View Release

  • Ollama – v0.14.3-rc2

    Ollama – v0.14.3-rc2

    🚀 Ollama v0.14.3-rc2 just dropped — and it’s a quiet hero for your RAM!

    💥 Bug squashed: Image models (SDXL, DALL·E, etc.) no longer get loaded into memory during model deletion. They now stay out of your way until you actually call them.

    🧠 Why it rocks:

    • Less RAM bloat = faster model swaps
    • Smoother performance on laptops & tiny servers
    • Cleaner shutdowns + smarter cleanup of unused vision models

    Perfect if you’re juggling multimodal AI or running vision models in prod. Still a release candidate, but solid — keep those GPUs cool and your memory free! 🖥️✨

    🔗 View Release

  • MLX-LM – v0.30.4

    MLX-LM – v0.30.4

    MLX LM v0.30.4 just dropped and it’s a beast 🚀

    • AWQ/GPTQ weight transforms now live — convert quantized models in one line.
    • Nemotron Super 49B v1.5 and GLM4 MoE Lite added — big brains, bigger performance on Apple silicon.
    • Batch generation? Fixed. MambaCache, CacheList, IQuestLoopCoder — all smoothed out.
    • New continuous batching server benchmark — measure your throughput like a pro.
    • LongCat Flash now supports sharding + extended context — longer prompts, zero headaches.
    • GPT-OSS & Minimax tensor sharding — distributed inference just got way easier.
    • SwiGLU compiled, Falcon H1 embeddings fixed, tokenizer errors now warn instead of crash.
    • Huge shoutout to new contributors: Eric, Nikhil, Solarpunkin, Evanev7 & Andrew! 🎉

    All powered by the latest MLX + smarter caching. Upgrade, benchmark, and go build something wild.

    🔗 View Release

  • MLX-LM – v0.30.3

    MLX-LM – v0.30.3

    MLX LM v0.30.3 just dropped and it’s a beast 🚀

    • AWQ & GPTQ quantization now fully supported — load quantized models like it’s nothing.
    • New models: IQuest Coder V1 Loop (code gen on steroids) + GLM4 MoE Lite (lightweight but mighty).
    • Nemotron Super 49B v1.5 and Falcon H1 with tied embeddings & muP scaling — optimized for peak performance.
    • Batching got a massive overhaul: sliding window + cache handling fixed, `CacheList`/`ArraysCache` now batchable, empty caches? Handled.
    • First-ever server benchmark for continuous batching — real-world numbers, not just benchmarks.
    • LongCat Flash now sharded + extended context — generate longer texts without choking.
    • Minitensor sharding (Minimax) + GPT-OSS sharding — scale your models smarter, not harder.
    • SwiGLU fixed, tokenizer errors now use `warnings`, MLX updated to latest — all the polish you didn’t know you needed.

    Massive thanks to @ericcurtin, @nikhilmitrax, @tibbes, @solarpunkin, @AndrewTan517, and @Evanev7 for the wins!

    Update. Run. Build something wild. 🤖💻

    🔗 View Release

  • Ollama – v0.14.3-rc1: MLX – dynamic loading of mlx-c (#13735)

    Ollama – v0.14.3-rc1: MLX – dynamic loading of mlx-c (#13735)

    🚀 Ollama v0.14.3-rc1 just dropped — and it’s a game-changer for Mac & Linux tinkerers!

    MLX is now dynamically loaded via `dlopen` — meaning:

    ✅ Ollama starts even if MLX isn’t installed

    ✅ Swap MLX paths on the fly (perfect for custom builds or multi-env setups)

    ✅ Graceful fallbacks — no more crashing if dependencies are missing

    No more “why won’t it start?!” headaches. Just pure, flexible local LLM power.

    Perfect if you’re running M-series Macs or Linux with custom CUDA/MLX builds.

    Tests fixed, reviews addressed — clean, stable, and ready for your next experiment.

    Try it out. If MLX isn’t there? Ollama just shrugs… and keeps going. 😎

    🔗 View Release

  • Ollama – v0.14.3-rc0

    Ollama – v0.14.3-rc0

    🚀 Ollama v0.14.3-rc0 just dropped — and macOS users, this one’s for you!

    No more ghost processes after rebooting. The Ollama app now properly shuts down during logouts and restarts — clean, quiet, and respectful of your system’s power management. 🍎💤

    Under the hood:

    • Smoother background cleanup
    • Better memory & resource handling on shutdown
    • Minor stability tweaks (zero breaking changes)

    This is a release candidate — stable, tested, and perfect for Mac folks tired of unresponsive apps after a reboot.

    Grab it if you run Ollama locally and value a clean, hassle-free experience. 🛠️✨

    🔗 View Release

  • Ollama – v0.14.2

    Ollama – v0.14.2

    🚀 Ollama v0.14.2 just dropped — tiny update, huge impact for AI agents!

    Fixed a sneaky bug in `ToolCallFunctionArguments` so nested JSON function calls no longer crash mid-execution. 🛠️

    Now your LangChain agents, custom tools, and multi-step workflows run smoother than ever.

    No breaking changes — just quiet, reliable stability for builders who rely on function calling.

    If you’re chaining tools or automating LLM workflows, this is the update that keeps your agents from falling apart.

    Upgrade and keep building! 🤖✨

    🔗 View Release

  • Ollama – v0.14.2-rc1: openai: tweak v1/responses to conform better (#13736)

    Ollama – v0.14.2-rc1: openai: tweak v1/responses to conform better (#13736)

    🚀 Ollama v0.14.2-rc1 just dropped — and it’s making your OpenAI API integrations smoother than ever!

    ✅ `/v1/` responses now perfectly mirror OpenAI’s structure — no more weird response quirks. Your existing code? Just works.

    🖼️ Bad image URLs? Say goodbye to cryptic errors — now you’ll get clear, helpful feedback.

    🧹 Under-the-hood linting fixes = cleaner code, fewer headaches.

    Perfect for devs using Ollama as a drop-in OpenAI replacement — whether you’re running Llama 3, Mistral, or Phi-4 locally.

    Keep those models humming and your APIs clean. 🛠️✨

    🔗 View Release

  • Ollama – v0.14.2-rc0

    Ollama – v0.14.2-rc0

    Hey AI tinkerers! 🚀 Ollama v0.14.2-rc0 just landed — and Mac users with Apple Silicon are in for a treat! 🍏

    MLX build instructions added to the README — now you can compile Ollama natively on M1/M2/M3 chips, bypassing Docker and getting faster, leaner local LLM inference.

    MLX = Apple’s new ML framework (think PyTorch, but built for M-series). No GPU? Still rockin’ Llama 3, DeepSeek-R1, or Mistral — just smoother and snappier.

    ⚠️ Still a release candidate, so keep an eye out for final tweaks — but if you’re tinkering on Mac? This is your golden ticket. 🎯

    Linux & Windows folks — your Ollama magic stays untouched, no worries! 💻🛠️

    🔗 View Release