Author: Tater Totterson

  • Deep-Live-Cam – 2.3d

    Deep-Live-Cam – 2.3d

    ๐Ÿšจ Deep-Live-Cam 2.3d just dropped โ€” and itโ€™s a game-changer for real-time face swaps!

    โœจ Smart Model Picker โ€” Browse and swap top-tested models with one click. No more digging through folders.

    ๐Ÿคฏ HyperSwap 256×256 โ€” Face swaps now 200% sharper. Details? Crisp. Artifacts? Gone.

    โšก Face Enhancer v2 โ€” Up to 4x faster, zero lag. Your stream wonโ€™t stutter even with heavy swaps.

    โœ… Mouth Mask + FPS Counter โ€” Fixed those weird mouth glitches and now you can monitor performance live.

    ๐Ÿšซ One-click magic โ€” Run `deep-live-cam.bat` and it just works. No more config headaches.

    All of this? Only in QuickStart for now. Windows & Mac Silicon users โ€” update ASAP.

    Keep swapping smarter, not harder. ๐ŸŽญ๐Ÿ’ป

    ๐Ÿ”— View Release

  • Ollama – v0.13.0

    Ollama – v0.13.0

    ๐Ÿš€ Ollama v0.13.0 is live โ€” and itโ€™s a game-changer for local LLM folks!

    Meet DeepSeek-V3.1 (aka Deepseek2) โ€” now officially supported with 128K context, razor-sharp reasoning, and killer coding skills. But hereโ€™s the kicker: itโ€™s running on Ollamaโ€™s brand-new engine with MLA (Multi-Layer Attention) โ€” meaning faster token generation, lower latency, and no more sluggish long-context hangs.

    โœจ Whatโ€™s new?

    • โœ… DeepSeek-V3.1 support โ€” perfect for complex prompts, multilingual tasks & code generation
    • ๐Ÿš€ MLA engine = smoother, faster inference on both CPU and GPU (NVIDIA/AMD)
    • ๐Ÿ’ก Optimized streaming โ€” ideal for chat apps, agents, and real-time LLM workflows

    Just run `ollama pull deepseek2` and feel the difference. No more waiting. Just pure, local LLM power. ๐Ÿค–๐Ÿ’ป

    ๐Ÿ”— View Release

  • Lemonade – v9.0.3

    Lemonade – v9.0.3

    ๐Ÿš€ Lemonade v9.0.3 just dropped โ€” and itโ€™s a game-changer for local LLM folks!

    The C++ server now ships with a clean, official `.msi` installer (`lemonade-server-minimal.msi`) โ€” goodbye clunky .exe, hello Windows stability ๐ŸŽฏ.

    โœจ Whatโ€™s new:

    • C++ system info now matches Pythonโ€™s accuracy โ€” no more mismatched specs!
    • Embedding UX got a serious polish: smoother, faster, less lag.
    • Model list now pulls from FLM + single source of truth ๐Ÿ—‚๏ธ (no more duplicate chaos).
    • Fixed bugs in `flm install`, `user_models.json`, and the `list` command.
    • Linux users: `unzip` is now a .deb dependency โ€” no more “command not found” headaches.
    • Help menu cleaned up โœจ, and “Version:” logs cleanly in the terminal ๐Ÿ“‹
    • Python tests now only run when code changes โ€” faster builds, less noise.

    All wrapped in a sleek WiX-built MSI for rock-solid Windows installs.

    Switching to local LLMs just got even easier. Grab it, tweak it, own your AI. ๐Ÿš€

    ๐Ÿ”— View Release

  • Text Generation Webui – v3.18

    Text Generation Webui – v3.18

    ๐Ÿ”ฅ text-generation-webui v3.18 is live โ€” and llama.cpp just leveled up!

    • ๐Ÿ–ฅ๏ธ `–cpu-moe` flag dropped โ€” offload MoE experts to CPU and run massive models on low-end GPUs. VRAM? Who needs it.
    • ๐Ÿง ROCm support is HERE! AMD GPU users on Linux โ€” rejoice. No CUDA? No problem.
    • ๐ŸŽ macOS 13 wheels retired. Time to update your OS if youโ€™re still on Big Sur or earlier.
    • ๐Ÿš€ Backend upgrades:
    • llama.cpp โ†’ latest commit (10e9780) โ€” smoother, faster, more stable
    • ExLlamaV3 v0.0.15 โ€” better quant, faster attention
    • peft 0.18.* โ€” new LoRA magic for fine-tuning lovers
    • triton-windows 3.5.1.post21 โ€” Windows inference just got a turbo boost

    ๐Ÿ“ฆ Portable builds? Still the best part.

    Download โ†’ unzip โ†’ run. No pip, no install.

    • NVIDIA? `cuda12.4`
    • AMD/Intel? Use `vulkan`
    • CPU-only? `cpubuilds` is your hero
    • Mac M1/M2? `macos-arm64` โ€” all set

    ๐Ÿ”ง Upgrading? Just swap the binary. Your `user_data/` folder stays untouched โ€” models, configs, themesโ€ฆ all safe.

    Go run Llama 3 70B MoE on your old laptop. The future isnโ€™t just local โ€” itโ€™s portable. ๐ŸŽ’๐Ÿ’ป

    ๐Ÿ”— View Release

  • Ollama – v0.13.0-rc0

    Ollama – v0.13.0-rc0

    ๐Ÿš€ Ollama v0.13.0-rc0 just dropped โ€” and itโ€™s packed with power!

    Say hello to DeepSeek-V3.1 (aka Deepseek2) โ€” one of the most capable open LLMs out there, now available with a simple `ollama pull deepseek-ai/deepseek-v3.1`.

    โœจ Why itโ€™s awesome:

    • ๐Ÿš€ MLA (Multi-Layer Attention) is live โ€” cuts memory use, speeds up inference, and keeps reasoning sharp.
    • ๐Ÿ› ๏ธ New engine under the hood = smoother runs, fewer crashes, better future-proofing.
    • ๐Ÿ’ฅ Run state-of-the-art reasoning on your laptop โ€” no cloud needed.

    GGUF? Still supported. API? Still there. CLI? Even better.

    This isnโ€™t just an update โ€” itโ€™s your ticket to running top-tier models locally, faster than ever.

    Go grab it:

    `ollama pull deepseek-ai/deepseek-v3.1`

    #LocalAI #DeepSeek #Ollama #LLM

    ๐Ÿ”— View Release

  • Heretic – v1.0.1

    Heretic – v1.0.1

    Heretic v1.0.1 is live ๐ŸŽ‰ โ€” the first public release of the fully automated LLM censorship remover is here, and itโ€™s wilder than you thought.

    No more manual tuning. No labeled data. Just run `heretic Qwen/Qwen3-4B-Instruct-2507` and watch it surgically erase refusal layers using directional ablation. Itโ€™s like giving your model a caffeine IV while keeping its brain intact.

    ๐Ÿ”ฅ Whatโ€™s new in v1.0.1?

    • โœ… First stable release: Betaโ€™s over โ€” this is the real deal.
    • ๐Ÿš€ 8B model decensoring in ~45 mins on RTX 3090 โ€” fast, lean, and mean.
    • ๐Ÿงช Improved KL divergence control: More original intelligence preserved post-ablation.
    • ๐Ÿ’พ Save or push to Hugging Face with one command โ€” no PhD needed.
    • ๐Ÿ› ๏ธ Better MoE support: Now handles Qwen-MoE and Llama-MoE with fewer hiccups.
    • ๐Ÿ“Š Enhanced eval suite: Auto-benchmarks refusal rates + output quality in one shot.

    Built with PyTorch 2.2+, AGPL-3.0 licensed, and ready to break the safety chains.

    Go run it. Then ask: “Why did we ever accept this?” ๐Ÿ’ฅ

    ๐Ÿ”— View Release

  • Chatterbox – v0.1.2

    Chatterbox – v0.1.2

    Chatterbox v0.1.2 just droppedโ€”and itโ€™s a game-changer for TTS tinkerers ๐ŸŽ™๏ธ

    โœ… M1/M2 Macs rejoice: Native support via MPSโ€”no more Rosetta slowdowns.

    ๐Ÿ”Š Safetensors everywhere: Faster, safer model loads + new WAV examples to play with.

    ๐Ÿ› ๏ธ CFG scaling optional: Dial realism or creativity like a knobโ€”perfect for voice acting or AI bots.

    ๐Ÿ› CUDA errors? Gone. GPU runs smoother than ever.

    ๐ŸŽฎ Min_P sampler added for finer audio controlโ€”less robotic, more human.

    ๐Ÿ“š Docs now crystal clear on OS/Python deps + watermarking (PerTh) best practices.

    ๐Ÿ“ฃ New Discord link fixed & liveโ€”join to share voice clones, memes, and cat meows ๐Ÿฑ๐Ÿ”Š

    ๐ŸŒŸ 7 fresh contributors brought the heatโ€”thank you!

    Install with `pip install chatterbox-tts` and start cloning voices (or your petโ€™s purr) in seconds.

    Full changelog: https://github.com/resemble-ai/chatterbox/commits/v0.1.2

    ๐Ÿ”— View Release

  • ComfyUI – v0.3.70

    ComfyUI – v0.3.70

    ComfyUI v0.3.70 just landed โ€” and itโ€™s the quiet hero your workflows have been waiting for ๐Ÿš€

    • Memory got smarter โ€” Fewer crashes on big SDXL or 4K renders. Keep those long pipelines running without hitting OOM hell.
    • Nodes wonโ€™t kill your whole graph โ€” A single failed node? No problem. The rest of your canvas keeps humming along.
    • UI tweaks that matter โ€” Smoother panning, fixed tooltip glitches, cleaner labels. Tiny changes, big comfort.
    • PyTorch & CUDA updates โ€” Linux users, rejoice: better compatibility under the hood.

    Pro tip: Drop your batch size by 1 if youโ€™ve been battling memory limits โ€” youโ€™ll be amazed how much longer your renders last.

    No flashy new nodesโ€ฆ just a more stable, reliable engine. Sometimes the best upgrades are the ones you donโ€™t notice โ€” because they just work. ๐Ÿ’ช

    ๐Ÿ”— View Release

  • ComfyUI – v0.3.69

    ComfyUI – v0.3.69

    ComfyUI v0.3.69 is live! ๐ŸŽ‰

    • New `LatentUpscale` node โ€“ Upscale in latent space before decoding for sharper results + faster renders.
    • Smarter memory handling โ€“ Fewer crashes on big batches; VRAM spikes? Not today.
    • SDXL Refiner flow fixed โ€“ Seamless transitions between base and refinerโ€”no more weird detail jumps.
    • Custom nodes reload properly โ€“ Finally! No more restarting ComfyUI after editing your favorite custom node.
    • WebAPI polish โ€“ Better compatibility with external tools and automation scripts.

    Perfect for high-res SDXL wizards and node-based tinkerers. Update now and keep those latents crisp! ๐Ÿ’ซ

    ๐Ÿ”— View Release

  • Ollama – v0.12.11

    Ollama – v0.12.11

    ๐Ÿš€ Ollama v0.12.11 just dropped โ€” and itโ€™s a quiet gem for the detail-oriented folks!

    The big win? `logprob` now includes byte-level data ๐ŸŽฏ

    No more guessing which bytes map to your tokens. Whether youโ€™re debugging multilingual text, tracking tokenization edge cases, or building precision prompt tools โ€” you now see exactly whatโ€™s happening at the byte level.

    Perfect for:

    • Prompt engineers wrestling with weird token splits
    • Researchers analyzing model confidence down to the byte
    • Devs building LLM debuggers or token analyzers

    No UI fluff, no breaking changes โ€” just pure, nerdy utility.

    Upgrade with `ollama pull` and start seeing the hidden layers beneath your prompts. ๐Ÿ’ก

    ๐Ÿ”— View Release