• Deep-Live-Cam – 2.7 beta

    Deep-Live-Cam – 2.7 beta

    ๐Ÿšจ AI Enthusiastsโ€”deepfake magic just got a serious upgrade! ๐Ÿšจ

    ๐Ÿ”ฅ DeepLiveCam v2.7 Beta is hereโ€”celebrating 80K GitHub stars with the biggest update yet! ๐ŸŽ‰

    โœจ Whatโ€™s Fresh:

    โœ… Realtime Face Enhancer โ€” now running up to 27 FPS with enhancement on! ๐Ÿš€

    โœ… Inswapper Optimizer โ€” faster, smoother swaps with less lag

    โœ… 2 NEW enhancer models: GPEN 512 & GPEN 256 (for sharper, more natural results)

    โœ… Face Enhancer Scaler โ€” dial in enhancement intensity like a pro ๐ŸŽ›๏ธ

    โœ… Masking upgrades: Quick Lip, Lip, Chin, Eyes masks โ€” precision control at your fingertips ๐ŸŽญ

    โœ… Interpolation support for buttery-smooth output frames

    โœ… GPU Changer โ€” full multi-GPU support (switch GPUs on the fly!) ๐Ÿ–ฅ๏ธโžก๏ธ๐Ÿ–ฅ๏ธ

    โœ… LUTs โ€” apply color grading presets in seconds ๐ŸŽจ

    โœ… Window Projection โ€” watch live output in a dedicated (even full-screen!) window ๐Ÿ“บ

    โœ… Camera Refresh & Resolution Changer โ€” tweak input/output like a streamer

    โœ… Cleaner, smarter UI โ€” because UX matters as much as power ๐Ÿ’ก

    ๐ŸŽฅ Check out the demo: `2026-03-12.02.34.57.mp4`

    ๐Ÿ“š Quickstart guide: [link in bio/pinned]

    This isnโ€™t just an updateโ€”itโ€™s a deepfake renaissance. Go play! ๐Ÿงชโœจ

    ๐Ÿ”— View Release

  • Voxtral Wyoming – v2.1.0

    Voxtral Wyoming – v2.1.0

    ๐Ÿšจ Voxtral Wyoming v2.1.0 is live! ๐Ÿšจ

    Offline STT just got smarter and more flexibleโ€”and youโ€™re going to love these updates:

    ๐Ÿ”น โœจ Word Replacement Feature

    Swap words or phrases on the fly in your promptsโ€”ideal for A/B testing, localization, brand tone adjustments, or just having fun with custom phrasing. Think of it as prompt surgery ๐Ÿฉบ: precise, fast, and fully in your control.

    ๐Ÿ”น ๐Ÿš€ Early Request Acceptance

    Want a feature before it drops publicly? Submit your request early and lock in priority accessโ€”get ahead of the curve and shape whatโ€™s next.

    ๐Ÿ’ก Bonus: Still supports CPU/CUDA/MPS, Dockerized deployment, and auto-converts MP3/OGG/FLAC/WAV โ†’ PCM16.

    ๐Ÿ”ง Configurable via env vars for host, port, language, model IDโ€”and now word replacement rules too.

    Ready to give it a spin? Drop your favorite use case belowโ€”weโ€™re listening! ๐ŸŽ™๏ธ๐Ÿค–

    ๐Ÿ”— View Release

  • Lemonade – v10.0.0

    Lemonade – v10.0.0

    ๐Ÿš€ Lemonade v10.0.0 is live!

    Big news for local LLM tinkerers โ€” the latest Lemonade release upgrades its core runtime to FLM v0.9.35 ๐Ÿง โšก (see PR #1233). While full release notes are still loading, this major bump likely brings performance tweaks, bug fixes, or new acceleration pathways under the hood โ€” especially for NPU/GPU inference on Ryzen AI and Vulkan-enabled hardware.

    โœ… Still supports GGUF & ONNX models

    โœ… OpenAI-compatible local API

    โœ… Python SDK + CLI for deep customization

    โœ… Windows & Linux love โค๏ธ

    If youโ€™re running Lemonade locally, this is a great time to upgrade โ€” especially if youโ€™ve been waiting for smoother NPU offload or faster token generation. ๐Ÿ› ๏ธ

    Curious what changed under the hood? Letโ€™s dig into PR #1233 together! ๐Ÿ•ต๏ธโ€โ™‚๏ธ

    ๐Ÿ”— View Release

  • Voxtral Wyoming – v2.0.0

    Voxtral Wyoming – v2.0.0

    ๐Ÿšจ Voxtral Wyoming v2.0.0 is live! ๐Ÿšจ

    Hey AI tinkerers & home automation wizardsโ€”big news: Voxtral Wyoming just dropped v2.0.0 with full support for Mistralโ€™s brand-new Gen2 models! ๐ŸŽคโšก

    ๐Ÿ”น Voxtral Realtime โ€” optimized for low-latency, real-time STT with improved voice clarity and speed.

    ๐Ÿ”น Voxtral-Mini-4B-Realtime-2602 โ€” the ultra-lightweight powerhouse for edge devices (Raspberry Pi, Jetson, etc.) without sacrificing accuracy.

    โœจ Why youโ€™ll love it:

    โœ… Up to 30% faster inference on supported hardware (CUDA/MPS/CPU)

    โœ… Smoother Home Assistant Assist integration

    โœ… All the usual perks: Docker-ready, auto audio conversion (MP3/OGG/FLAC/WAV โ†’ PCM16), and env-based config

    Ready to bring offline, private voice control to your setup? ๐Ÿ› ๏ธ

    ๐Ÿ‘‰ [Check the docs & deploy now!]

    ๐Ÿ”— View Release

  • Ollama – v0.17.8-rc1: ci: Fix windows build (#14754)

    Ollama – v0.17.8-rc1: ci: Fix windows build (#14754)

    ๐Ÿšจ Ollama v0.17.8-rc1 is live! ๐Ÿšจ

    This oneโ€™s a micro-update with big implications for Windows devs and CI workflows โ€” no flashy new features, but a solid behind-the-scenes upgrade!

    ๐Ÿ”ง Whatโ€™s new?

    • โœ… Fixed Windows build issues (#14754) by swapping out shell-based wildcard expansion (`sh`) for a native Go implementation.
    • ๐ŸŒ Why it matters: `sh` (e.g., bash) behaves differentlyโ€”or not at allโ€”on Windows. This change boosts cross-platform consistency, especially in automated environments like GitHub Actions or Docker builds.
    • ๐Ÿ› ๏ธ Think of it as Ollama finally nailing its Windows boot sequence โ€” fewer “works on my machine” headaches!

    ๐Ÿ“ฆ No user-facing changes yet, but this paves the way for smoother releases ahead. Keep an eye out for the stable `v0.17.8` drop! ๐ŸŽฏ

    #Ollama #LLMDev #AITools

    ๐Ÿ”— View Release

  • Ollama – v0.17.8-rc0: MLX: add header vendoring and remove go build tag (#14642)

    Ollama – v0.17.8-rc0: MLX: add header vendoring and remove go build tag (#14642)

    ๐Ÿš€ Ollama v0.17.8-rc0 is here โ€” and itโ€™s packing some serious dev-friendly upgrades!

    ๐Ÿ”ฅ MLX Integration Gets a Major Boost

    • โœ… Header vendoring for `mlx-c` โ€” Go can now build without CMake first ๐ŸŽ‰
    • โœ… Removed `go build` tag restriction โ€” MLX support is now on by default
    • ๐Ÿ”„ Headers auto-refresh on CMake runs โ†’ easier upgrades & less friction
    • โœ… Basic Windows + Linux support confirmed (finally!)

    ๐Ÿ”ง ROCm on Windows

    • ๐Ÿšซ Temporarily sticking with ROCm v6 (v7 needs more love โ€” stay tuned)

    ๐Ÿ› ๏ธ CI/CD Hardening

    • ๐Ÿ›ก๏ธ CI now handles flaky Chocolatey repos gracefully
    • โš ๏ธ Builds proceed even if cache fails (since caching is optional for speed)

    ๐Ÿ’ก Why it matters: Smoother builds, better Apple Silicon (MLX) support, and more resilient CI = faster iteration for devs experimenting locally. ๐Ÿงช๐Ÿ’ป

    Check it out if youโ€™re running Ollama on macOS/Windows/Linux โ€” especially with Apple Silicon or ROCm setups! ๐Ÿง โšก

    #Ollama #LLMs #AIDev #LocalLLM

    ๐Ÿ”— View Release

  • Text Generation Webui – v4.0

    Text Generation Webui – v4.0

    ๐Ÿš€ Text-Generation-WebUI v4.0 is here! ๐Ÿš€

    The big v4.0 update brings major improvements across the board โ€” think of it as the “big bang” moment for the UIโ€™s modernization! Hereโ€™s whatโ€™s new:

    ๐Ÿ”น Brand-new Gradio-based UI

    A complete redesign with a sleek, responsive interface โ€” faster, cleaner, and way more intuitive. Think Stable Diffusion WebUI vibes, but for LLMs ๐ŸŽจโœจ

    ๐Ÿ”น Backend Agnostic Architecture

    Now even easier to switch between backends (llama.cpp, Transformers, ExLlamaV3/V2, TensorRT-LLM) โ€” with better hot-swapping and config management. ๐Ÿ”„

    ๐Ÿ”น Built-in Document Loader (PDF, DOCX, TXT)

    Upload and chat with your files directly โ€” no more copy-paste or external tools! ๐Ÿ“„โžก๏ธ๐Ÿง 

    ๐Ÿ”น Web Search Integration (Optional)

    Add real-time context via DuckDuckGo or SERP API โ€” perfect for RAG without the hassle. ๐Ÿ”๐ŸŒ

    ๐Ÿ”น New Extensions API

    Smoother, more powerful extension support โ€” TTS, translation, custom UI hooks, and more. ๐Ÿงฉ

    ๐Ÿ”น OpenAI-Compatible API Improvements

    Better parity with the real deal โ€” including streaming, chat completions, and tool calling. ๐Ÿ“ก

    ๐Ÿ”น One-Click Portable Builds & Installer

    Still as easy as ever to get up and running โ€” Windows, Linux, macOS. ๐Ÿงชโšก

    ๐Ÿ”ฅ Bonus: Better memory management, improved prompt templates, and a much more stable chat mode.

    Grab the update โ€” your local LLM playground just got a serious upgrade! ๐Ÿ› ๏ธ๐Ÿ’ป

    ๐Ÿ‘‰ github.com/oobabooga/text-generation-webui

    ๐Ÿ”— View Release

  • Ollama – v0.17.7

    Ollama – v0.17.7

    ๐Ÿšจ Ollama v0.17.7 is out! ๐Ÿšจ

    This patch brings a subtle but important fix under the hood:

    ๐Ÿ”น Stale context window entries now get properly overridden โ€” meaning outdated prompt/chat history data wonโ€™t linger and mess with your inference accuracy. ๐Ÿ’ก

    ๐Ÿง  Why youโ€™ll care:

    • Cleaner, more reliable multi-turn conversations
    • Better token efficiency (no hidden bloat from old context!)
    • Smoother long-context handling โ€” especially helpful if youโ€™re pushing model limits

    ๐Ÿ“ฆ No flashy new models or API changes this time, but itโ€™s a solid reliability bump for everyday use.

    ๐Ÿ”— Full details: v0.17.7 Release

    Happy local LLM tinkering! ๐Ÿ› ๏ธ๐Ÿค–

    ๐Ÿ”— View Release

  • Ollama – v0.17.7-rc2

    Ollama – v0.17.7-rc2

    ๐Ÿš€ Ollama v0.17.7-rc2 is out!

    This release candidate brings a handy fix for context window management โ€” specifically, overriding stale entries in the context tracking logic. ๐Ÿง โœจ

    ๐Ÿ”น Whatโ€™s fixed?

    • Stale context data (e.g., outdated conversation history) no longer lingers and messes with model responses.
    • Improves reliability in multi-turn chats, especially for longer sessions or when switching between conversations.

    ๐Ÿ’ก Why it matters: Cleaner context = more accurate, consistent responses โ€” and fewer “wait, why did it say that?!” moments. ๐Ÿ˜…

    Since this is an rc2, itโ€™s a pre-release focused on polish and stability ahead of the final `v0.17.7`. No flashy new features yet, but solid under-the-hood improvements!

    ๐Ÿ‘‰ Grab it and test: v0.17.7-rc2 on GitHub

    Let us know how it behaves in the wild! ๐Ÿ› ๏ธ

    ๐Ÿ”— View Release

  • Ollama – v0.17.7-rc1

    Ollama – v0.17.7-rc1

    ๐Ÿšจ Ollama v0.17.7-rc1 is out! ๐Ÿšจ

    This release is a tiny but tidy patch candidate โ€” only one commit landed:

    ๐Ÿ”ง `cmd/config: fix cloud model limit lookups in integrations (#14650)`

    โœ… Whatโ€™s fixed:

    • Resolves a bug where Ollama was misfetching or misapplying model usage limits when integrated with cloud services (e.g., Ollama Cloud or third-party APIs).
    • Ensures smoother, more accurate rate-limit handling in hybrid/localโ€“cloud workflows.

    ๐Ÿ“Œ Why it matters:

    • If youโ€™re using Ollama with cloud backends or integrations (like LangChain, LlamaIndex, or custom tooling), this fix helps avoid unexpected throttling or config mismatches.
    • No new features, no breaking changes โ€” just more reliability ๐Ÿ› ๏ธ

    ๐Ÿ“… Tagged: Mar 5, 2024

    ๐Ÿ”— Release on GitHub

    โš ๏ธ RC = Release Candidate โ€” test it out, but maybe wait for the stable drop before pushing to prod.

    Let me know if you want a deep dive into PR #14650 or how this affects your integrations! ๐Ÿค–โœจ

    ๐Ÿ”— View Release