Author: Tater Totterson

  • Perplexica – v1.11.1

    Perplexica – v1.11.1

    Perplexica v1.11.1 just dropped πŸš€ β€” your open-source Perplexity alternative just got way smoother!

    • No more hanging searches β€” SearxNG timeouts? Fixed. Queries now actually finish instead of ghosting you.
    • Your go-to LLM (Qwen, DeepSeek, Llama, Mistral) remembers your pick β€” stored in localStorage. No more dropdown roulette on reload.
    • Run commands? Your data now persists via volumes. Say hello to saved outputs, goodbye to “oh no, I lost my last run.”

    Tiny release. Huge quality-of-life wins. Perfect for devs who just wanna search, not reconfigure. πŸ› οΈβœ¨

    πŸ”— View Release

  • Text Generation Webui – v3.16

    Text Generation Webui – v3.16

    πŸš€ Text Generation WebUI v3.16 just droppedβ€”and it’s a game-changer for local LLM folks!

    New portable build via symlink? Yes, please. Devs juggling multiple setups can now switch models and configs without reinstalling. Big shoutout to @reksar! πŸ™Œ

    macOS Apple Silicon usersβ€”your day is saved. Python deps now work flawlessly on Tahoe (thanks @drieschel)! 🍎

    Backend upgrades? Oh yeah:

    • llama.cpp updated to latest GGML fork β†’ now supports Llama-Mini-2.0 and Ring-Mini-2.0! Tiny but mighty models, unlocked.
    • ExLlamaV3 v0.0.11 = faster inference, smoother text flow.
    • Triton-Windows updated to 3.5.0.post21 β†’ better CUDA perf on Windows rigs.

    Portable builds are now even easier:

    πŸ“₯ Download β†’ πŸ“¦ Unzip β†’ πŸ’Ύ Copy your old `user_data` folder in β†’ βœ… All models, themes, and settings preserved. No pip. No venvs. Just AI magic.

    Pick your build:

    • NVIDIA? β†’ `cuda12.4` (new) or `cuda11.7` (legacy)
    • AMD/Intel? β†’ Use `vulkan`
    • CPU-only? β†’ `cpu` build
    • Mac? β†’ `macos-arm64` (M-series) or `macos-x86_64`

    No install. No fuss. Just drop-in, run, and chat with your LLMs like never before. πŸš€

    πŸ”— View Release

  • Wyoming Openai – Streaming hotfix and Chatterbox TTS release (0.3.8)

    Wyoming Openai – Streaming hotfix and Chatterbox TTS release (0.3.8)

    πŸŽ™οΈ Wyoming OpenAI 0.3.8 is liveβ€”and TTS streaming just got a serious upgrade!

    Say goodbye to stilted audio pauses. The new smart TTS streaming uses pySBD to chunk text at sentence boundaries, then prefetches the next line while playing the current oneβ€”so even if OpenAI stumbles, your voice assistant keeps flowing.

    πŸš€ Highlights:

    • πŸš€ Parallel prefetching: Up to 3 TTS requests running at once, sequenced perfectly.
    • 🐳 Chatterbox TTS support: Drop-in Docker compose for self-hosted neural voicesβ€”with voice cloning!
    • πŸ›‘οΈ Robust error handling: New `TtsStreamError` + `_abort_synthesis` to kill broken streams and stop audio doubles.
    • πŸ“¦ Install via `pip install wyoming-openai`β€”no git needed.
    • πŸ”§ Updated deps: `openai==2.3.0`, `wyoming==1.8.0` for full compatibility.

    Perfect for Home Assistant users who want smooth, low-latency voiceβ€”whether on cloud APIs or local models like Piper, Kokoro, or Edge TTS.

    No more buffering. Just natural, uninterrupted speech. 🎧✨

    Check the docs and start streaming!

    πŸ”— View Release

  • Perplexica – v1.11.0

    Perplexica – v1.11.0

    πŸš€ Perplexica v1.11.0 just dropped β€” your open-source Perplexity AI alternative just got a massive upgrade!

    ✨ New Setup Wizard β€” No more config nightmares. Pick your model, pick your provider β€” done in 60 seconds.

    βš™οΈ Config System Reborn β€” Live updates, hash-based tracking, and zero-loss migrations. Settings now survive reboots.

    πŸͺ„ Single Docker Install β€” `docker run …` and you’re running a full AI search engine. No repos, no deps. Pure magic.

    🧠 New Models Galore β€” GPT-5, Claude Opus 4.1, Gemini 2.5, O3… plus AIML API, LM Studio, and dynamic Transformers loading β€” models load on-demand, fast & lean.

    πŸ“± UI/UX Glow-Up β€” Sleek sidebar, mobile settings button, weather widget with geolocation 🌑️, topic filters, preview mode, and file uploads + light theme finally working right.

    ⚑ Dev Love β€” API validation, clean citations, instrumentation-based migrations (bye-bye ts-node), and faster message handling.

    πŸ› Bugs Eaten Alive β€” Double JSON, iOS zoom chaos, DOC upload fails, light mode glitches, and that pesky “repeated first token” β€” all gone.

    πŸ‘ 17 New Contributors β€” Huge props to @ClawCloud-Ron, @haddadr, @alckasoc, and the crew for making this release legendary.

    One command. Zero friction. All the power.

    Upgrade now β€” your next search just got smarter. πŸš€

    πŸ”— View Release

  • ComfyUI – v0.3.66

    ComfyUI – v0.3.66

    ComfyUI v0.3.66 is live πŸš€ β€” and it’s a quiet powerhouse for your AI workflows!

    ✨ New `LatentUpscale` node β€” upscale in latent space before decoding for sharper, cleaner high-res results with less noise.

    ⚑ Memory optimized β€” fewer spikes during batch processing, perfect for mid-tier GPUs.

    πŸ” Faster node search β€” partial matches work now! Type “upscale” and get all related nodes instantly.

    🧩 Custom node fix β€” no more vanishing nodes after reloads (we feel you πŸ˜…).

    🎨 UI polish β€” smoother transitions + zoom snapping to 25%/50%/100% for pixel-perfect control.

    Pro tip: Pair `LatentUpscale` with KSampler + High-Res Fix for insane detail without VRAM overload.

    Upgrade now β€” your next masterpiece is just a click away. πŸ–ΌοΈπŸ’»

    πŸ”— View Release

  • MLX-LM – v0.28.3

    MLX-LM – v0.28.3

    πŸ”₯ MLX LM v0.28.3 is LIVE! πŸ”₯

    Heads up, Apple silicon LLM tinkerers – LLaMA-Factory just dropped a massive update for MLX LM! This release is packed with refinements and new features to help you build, train & serve even better models.

    Here’s the breakdown:

    • Memory Efficiency: State Space Models (SSM) are leaner now. πŸ™Œ

    MoE Magic: Lots of improvements to Mixture of Experts – LoRA fixes, bailing logic, and* a new LFM2 option!

    • Qwen3-VL Support: Visual language model support added with Qwen3-VL (plus a dense version!). πŸ–ΌοΈ
    • Faster GPT2: Batch processing for GPT-2 just got quicker.
    • DWQ Tweaks: Depthwise Quantization refined with temperature adjustments.
    • Python 3.9 Love: Qwen3 support now extends to Python 3.9 users!
    • Plus: Cleaned up params, simplified I/O, CUDA install fixes, batched SSM masking, gradient accumulation, data parallel eval, Jamba support & LLM Benchmarks! πŸ“Š

    Dig into the full changelog – there’s a ton here to play with! πŸŽ‰

    πŸ”— View Release

  • LMStudio – 1.6.0b1 – 2025-10-28

    LMStudio – 1.6.0b1 – 2025-10-28

    LM Studio Python SDK v1.6.0b1 is live! πŸš€

    This release focuses on making it even easier to interact with your LM Studio server APIs from Python. Here’s what’s fresh:

    • Smoother Error Handling: Say goodbye to crashes when the server sends back pre-formatted errors – now handled gracefully!
    • API Token Authentication: Added support for API tokens, giving you more flexibility in how you authenticate your requests. πŸ”‘
    • New Contributor: A big welcome to @ryan-the-crayon! πŸŽ‰

    Check out the full changelog for all the details: https://github.com/ncoghlan/server-api/compare/1.5.0…1.6.0b1

    πŸ”— View Release

  • Ollama – v0.12.6: cuda: bring back CC 5.2 (#12666)

    Ollama – v0.12.6: cuda: bring back CC 5.2 (#12666)

    Ollama just dropped v0.12.6! πŸŽ‰

    Need to run large language models locally? Ollama makes it easy – think GPT-4, Llama 2, right on your machine. No API limits, more privacy!

    This release brings some good stuff for NVIDIA users:

    • Older GPU Support: They’ve brought back support for Compute Capability (CC) 5.2 CUDA devices. Dust off those slightly older NVIDIA cards!

    Driver Harmony: Fixes a compatibility issue with newer drivers and* keeps the CC 5.2 experience smooth.

    Dig into the details: https://github.com/ollama/ollama

    πŸ”— View Release

  • Tater – Tater v35

    Tater – Tater v35

    🏠 Tater v35 – “Eyes Everywhere” just dropped! πŸ₯”

    Your smart home just leveled up with an Event Systemβ€”giving your AI long-term awareness of everything happening around the house (doorbell rings, dog playtime… you name it!). Think of it as a memory for your home! 🧠

    Here’s what’s new:

    • Event Storage & Retrieval: Core platform now stores events via Redis.
    • New Endpoints: Add (`POST`) and search (`GET`) events with ease.

    Retention Control: Configure how long Tater remembers – from 2 days to forever*.

    • Plugins Galore:

    πŸšͺ Doorbell Alert (updated): Captures images, describes visitors, speaks alerts, and* logs structured events! Talk back through compatible doorbells too. πŸ—£οΈ

    • πŸŽ₯ Camera Event (new): Motion detection + Vision LLM descriptions = smart camera monitoring.
    • 🧠 Events Query (new): Ask natural language questions like “What happened in the front yard today?” and get a summary!

    Tater syncs time with your Home Assistant & cleans up old events automatically. πŸ•“ How it works: HA triggers Tater, AI Router picks the plugin, events are logged, and the Query Plugin brings it all back to life!

    README for details & installation.

    πŸ”— View Release

  • Ollama – v0.12.6-rc1

    Ollama – v0.12.6-rc1

    Ollama v0.12.6-rc1 is here! πŸš€

    Running local LLMs just got a little smoother. Ollama lets you spin up models like Llama 3 & Mistral right on your machine – perfect for tinkering and building AI apps without hitting an API.

    This release brings:

    • More Embedding Models: More options for semantic search & vector databases with added models like `all-MiniLM-L6-v2`.
    • Bug Fixes & Polish: General improvements to keep things running nice and tight! πŸ› οΈ

    Dig into the full release notes: https://github.com/ollama/ollama/releases/tag/v0.12.6-rc1

    πŸ”— View Release