Author: Tater Totterson

  • Ollama – v0.13.3: Update README.md (#13373)

    Ollama – v0.13.3: Update README.md (#13373)

    Ollama v0.13.3 just dropped — tiny update, big win for docs lovers! 📚✨

    The team fixed a few broken links in the README:

    • Swollama (community tool) → now points to the right place
    • DocC docs → no more 404s, just clean access to docs

    No new models. No API tweaks. Just flawless documentation so you can dive in without hitting dead ends.

    Perfect for tinkerers who appreciate polish — because clean docs = faster experimentation. 🛠️

    Keep running LLMs like a pro!

    🔗 View Release

  • Lemonade – v9.1.0

    Lemonade – v9.1.0

    🚨 Lemonade v9.1.0 is LIVE — your local LLM powerhouse just got a major upgrade! 🍋

    The brand-new Lemonade App (Windows .msi / Linux .deb) replaces the old browser UI with a sleek native experience. Here’s what’s fresh:

    • 🎙️ ASR via whisper.cpp — transcribe audio right inside the app (toggle with env vars!)
    • 📥 Built-in Model Downloader — no more terminal commands to grab GGUF/ONNX models
    • 🔄 Switch LLMs & FLMs on the fly — test FLM2-1.2B-FLM with improved reasoning logic
    • 📏 Smart Model Manager — filters by RAM usage so you don’t crash your system
    • 🌐 Server now defaults to IPv4 — goodbye, localhost confusion

    Under the hood:

    • 🐳 Official Docker/CMake dev setup
    • 💬 Chat UI now shows FLM “thinking” in collapsible boxes
    • 📚 Debate Arena docs live + Jan model fix shipped
    • 💥 All-in-one installer: app + server bundled. Minimal server still available for headless use

    Pro tip: Click the model dropdown — it actually loads models now. And yes, you can collapse thinking by default.

    Shoutout to @danielholanda, @jeremyfowers, and @Geramy — you’ve built something wild.

    Download. Play. Break it. Fix it. Repeat. 😎

    🔗 View Release

  • Ollama – v0.13.3-rc1: feat: llama.cpp bump (17f7f4) for SSM performance improvements (#13408)

    Ollama – v0.13.3-rc1: feat: llama.cpp bump (17f7f4) for SSM performance improvements (#13408)

    🚀 Ollama v0.13.3-rc1 is live — and Apple Silicon users, this one’s for you!

    llama.cpp just got a massive upgrade to latest master (17f7f4b), turbocharging SSM models like Granite-4, Jamba, Falcon-H, Nemotron-H, and Qwen3 Next on Metal.

    💥 What’s new?

    • Prefill speed up by 2–4x on M1/M2/M3 — fewer waits, faster first tokens
    • Optimized `SSM_CONV` and `SSM_SCAN` ops — the secret sauce behind modern state-space models
    • Clean swap to `gemma3.cpp` (goodbye, -iswa!)
    • 30+ patches + vendored code sync for stability

    If you’re running SSMs on Mac — upgrade now. Your chat latency just got a serious caffeine boost. 🍏⚡

    🔗 View Release

  • Heretic – v1.1.0

    Heretic – v1.1.0

    Heretic v1.1.0 just dropped—and it’s a game-changer 🚀

    Apple Silicon support? ✅ Run decensoring natively on your M-series Mac.

    IBM Granite MoE now supported? ✅ Unlock massive efficiency with MoE models.

    Multi-GPU? ✅ Split the load like a pro—no more GPU bottlenecks.

    MXFP4 + Triton tensors? ✅ Leaner, faster inference on compatible hardware.

    Local datasets and `trust_remote_code`? Smooth sailing now.

    Colab/Kaggle compatibility? ✅ Drop it in your notebook and go.

    Float32 bugs fixed? ✅ No more precision surprises.

    Refusal detection got smarter—better at spotting those “I can’t help” vibes.

    Early stopping, thinking models, padding fixes? All there for the tinkerers.

    And yes—CI now lints PR titles because clean commits = happy devs 😎

    8 new contributors joined the crew. Welcome!

    Upgrade. Decensor. Unleash. 💥

    AGPL-3.0 | PyTorch 2.2+ | RTX 3090? ~45min for 8B models.

    🔗 View Release

  • Tater – Tater v40

    Tater – Tater v40

    Tater v40 just dropped—and it’s not just an update, it’s a personality transplant 🤖💖

    🔥 Home Assistant Gets Smarter:

    • New Events Query Brief — ask “Any motion overnight?” and get a clean, JSON-ready summary.
    • Weather Query Brief — crisp 255-char snapshots perfect for dashboards, no more truncation hacks.

    🎛️ HA Integration Now Flawless:

    Auto-updating sensors, time-aware summaries for hourly polls, and zero clunky workarounds. Your smart home just got a brain upgrade.

    🎭 One Personality, Everywhere:

    Set `tater:personality` once—and it sticks across Discord, IRC, Matrix, WebUI, HomeKit… even XBMC on your Original Xbox. (Yes, really. Cortana mode optional.)

    💡 The magic? Tater now feels like it’s always been there—consistent, intuitive, and weirdly charming.

    Go ahead. Ask “Hey Tater, what’s the vibe today?”

    …your fridge might answer first. 🥔✨

    Check it out: https://github.com/TaterTotterson/Tater

    🔗 View Release

  • ComfyUI – ComfyUI version v0.4.0

    ComfyUI – ComfyUI version v0.4.0

    ComfyUI v0.4.0 just landed—and it’s a game-changer for workflow stability 🚀

    No more “why did my pipeline break?!” nightmares.

    Now:

    • Minor versions (v0.4.x) = rock-solid, tested releases from `master`
    • Patch versions (v0.4.1, v0.4.2) = critical bug fixes backported without forcing a full upgrade

    Think of it like Docker tags for AI workflows—clean, predictable, and dev-friendly.

    Your v0.4.x install? Safe to trust. Updates won’t wreck your nodes. Patches land fast.

    Perfect for artists, producers, and devs who just wanna render—not debug version chaos. 🎨✨

    Full details: https://www.comfy.org/

    🔗 View Release

  • Ollama – v0.13.3-rc0

    Ollama – v0.13.3-rc0

    🚀 Ollama v0.13.3-rc0 just dropped — and Mac users, this one’s for you!

    Fixed a nasty Metal backend crash with Qwen2.5-VL during `argsort` ops — multimodal inference is now stable on Apple Silicon. 🍎🧠 No more mid-inference bailouts when describing images!

    Also tucked in:

    • Smoother vision-language pipeline performance
    • Tiny tensor handling optimizations under the hood

    No breaking changes — just cleaner, more reliable multimodal runs.

    Pro tip: Try `ollama run qwen2.5-vl “Describe this image”` — it’ll actually finish now 😉

    🔗 View Release

  • MLX-LM – test_data

    MLX-LM – test_data

    🚀 MLX LM just dropped 8 new optimized LLMs—fully tuned for Apple Silicon!

    Say hello to:

    • Qwen1.5-0.5B-Chat
    • Mistral-7B-v0.2 & v0.3
    • DeepSeek-Coder-V2-Lite-Instruct (MLX-native 🎯)
    • Phi-3.5-mini-instruct
    • Llama-3.2-1B-Instruct
    • Falcon3-7B-Instruct
    • Qwen3-4B

    ✅ All 4-bit quantized. ✅ Only `.safetensors`, tokenizer, and Jinja templates—zero bloat.

    ✅ New lean download for Qwen1.5-0.5B: just the model weights.

    ✅ Zipped and ready to drop into your MLX pipeline.

    No GPU? No problem. M-series chips are now LLM powerhouses.

    Grab `test_data.zip` and start whispering to LLMs at near-native speed. 🍏⚡

    🔗 View Release

  • Ollama – v0.13.2

    Ollama – v0.13.2

    Ollama v0.13.2 just landed — tiny patch, big win for docs! 🛠️

    ✅ Fixed a broken link in the README’s “Community Integrations” section — that sneaky “Swollama” typo is finally gone.

    Now you can click through to Swollama (the slick Ollama-powered web UI) without hitting a 404. Perfect for folks exploring local LLM interfaces without leaving their browser.

    Clean docs = smoother tinkering. Keep those models rolling! 💡🧠

    🔗 View Release

  • Crankboy App – v1.1.0

    Crankboy App – v1.1.0

    🚀 CrankBoy v1.1.0 just landed on Playdate — and it’s not just an update, it’s a full GB nostalgia upgrade!

    Unified file system: All ROMs & covers now auto-migrate to `/Shared/Emulation/gb` — no more folder chaos.

    🔊 Audio overhaul: Smoother, more accurate sound — that iconic Game Boy chime finally sounds right.

    🎨 Visual polish: Ghost frames + frame blending = buttery-smooth pixel motion. Your 8-bit dreams are now HD-ready.

    📥 In-app downloads: Fetch ROM hacks and patches directly from the emulator. No PC needed.

    💾 Save states that actually work: Now emulates cartridge memory like Pokémon — save anywhere, no more gym panic.

    🕹️ Crank customization: Tweak sensitivity and behavior to match your playstyle.

    🧩 Scripting support: Alleyway (beta) and Link’s Awakening fishing? Scriptable. Castlevania 2: Belmont’s Revenge? Fully playable now.

    ⚠️ Smart save warnings: No more accidental overwrites — we’ve got your back (and your 1998 save files).

    Full changelog on Patreon — but you’ll feel the magic the second you crank it up. Go play. Then come back and scream about how good it feels. 🕹️💙

    🔗 View Release