Discussion of Free and Open Source Text-to-Image/Video ModelsPrev: >>107693072https://rentry.org/ldg-lazy-getting-started-guide>UIComfyUI: https://github.com/comfyanonymous/ComfyUISwarmUI: https://github.com/mcmonkeyprojects/SwarmUIre/Forge/Classic/Neo: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassicneoSD.Next: https://github.com/vladmandic/sdnextWan2GP: https://github.com/deepbeepmeep/Wan2GP>Checkpoints, LoRAs, Upscalers, & Workflowshttps://civitai.comhttps://civitaiarchive.com/https://openmodeldb.infohttps://openart.ai/workflows>Tuninghttps://github.com/spacepxl/demystifying-sd-finetuninghttps://github.com/ostris/ai-toolkithttps://github.com/Nerogar/OneTrainerhttps://github.com/kohya-ss/musubi-tunerhttps://github.com/kohya-ss/sd-scriptshttps://github.com/tdrussell/diffusion-pipe>Z Image Turbohttps://huggingface.co/Tongyi-MAI/Z-Image-Turbo>WanXhttps://github.com/Wan-Video/Wan2.2https://comfyanonymous.github.io/ComfyUI_examples/wan22/>NetaYumehttps://civitai.com/models/1790792?modelVersionId=2485296https://nieta-art.feishu.cn/wiki/RY3GwpT59icIQlkWXEfcCqIMnQd>Chromahttps://huggingface.co/lodestones/Chroma1-Basehttps://rentry.org/mvu52t46>Illustrioushttps://rentry.org/comfyui_guide_1girlhttps://tagexplorer.github.io/>MiscLocal Model Meta: https://rentry.org/localmodelsmetaShare Metadata: https://catbox.moe|https://litterbox.catbox.moe/GPU Benchmarks: https://chimolog.co/bto-gpu-stable-diffusion-specs/Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-oneTxt2Img Plugin: https://github.com/Acly/krita-ai-diffusionArchive: https://rentry.org/sdg-linkBakery: https://rentry.org/ldgcollage>Neighbors>>>/aco/csdg>>>/b/degen>>>/r/realistic+parody>>>/gif/vdg>>>/d/ddg>>>/e/edg>>>/h/hdg>>>/trash/slop>>>/vt/vtai>>>/u/udg>Local Text>>>/g/lmg
blessed thread of frenship
ramtorch when
is there like a beginner friendly guide to get started with your own model rig like the kind that Pewdie uses?
I need to seamlessly blend someone into a group of people. Gemini Banana Pro can't do it apparently.. :(
guys how do I upscale, please discuss local diffusion with me I thought you were my friends
how do I get rid of this stupid piece of shit?
it's over...
>>107700541Chinese cultureād
>>107700522What, you don't like clicking three times to cancel the queue and back out or have several things piled on top of each other? That's $17mil worth of design right there, Anon!
>>107700576I thought you had to go, fuck off already
>>107700582You must be mistaken, I'm not anybody in these threads.
>>107700604I would love to discuss the pros and cons of all software pertaining to local image diffusion, but exactly half of that discusion will cop me a ban, so I'd rather not discuss any of it, hence this empty thread
>>107700568What do you mean? Is it actually out?
>>107700612anon you are free to discuss anything on-topic and constructive if you keep personalities and schizo drama out of it
>>107700629you're not allowed to impersonate janitors or moderation staff
>>107700522>>107700576>>107700582>>107700604>>107700612>>107700629...I just wanted to know if I could disable it
>>107700637that's not what i'm doing. just stating the obvious. nobody ever got banned for discussion local diffusion in a local diffusion general, you're spreading disinfo
Where rentry
>>107700643You can actually roll the frontend back a bit, I'm not sure how far though.
>>107700656https://www.youtube.com/watch?v=X1osnpVqY_k
Finally got Latentsync 1.6 working, spent all fucking morning on it. It doesn't replace random shit like InfiniteTalk does. But infinitetalk is better at the lipsync.I can't find jack shit on how to mask a face, or even mouth, with infinitetalk. Does anyone know?
>>107700676So, there's no way to turn it off with the current frontend?
so what are the current top dog local modelswhen last I checked it was I think either flux or chroma for realistic stuffand noobai/illustrious for weebI saw that pony 7 came out but I have no idea if that had any impact or not
>>107700409based
>>107700763No, unfortunately.
>addicted to watching avr_loss go up and down for hours>bouncing between its over and we are back
Why was this thread created 2 hours before the previous thread hit bump limit?
FUCKING WHY
Why is Anistudio is such an useless piece of shit software?
>>107700879things made out of spite often are, it's not a good motivator for quality
>>107700883ThisMost of his actions are out of spite plus he never delivers any of the software he promises. Still no high res fix and most UI had that in weeks not months.
>>107700883I feel genuinely sorry for whoever uses it in their workflow
>>107700866>2025-02-27I can think of why perhaps
>>107700866>>107700927AAAAAAAAAAAWHY ISN'T REQUIREMENTS INSTALLED AUTOMATICALLY WITH COMFY ANY MORE REEEEEEEEEEEEEEEEEEEEEEEEEEEEEEEEFUCK OPEN SOURCE
nooooticing
>>107700957noticing? noticing what? the time? the weather? that's so ambiguous it could mean anything, most of which are on topic
>>107700951Yeah we really need an alternative UI that preferably isn't Python. Do any exist?
>>107700966There's always stable-diffusion.cpp :)Not many other good alternatives, though
I can't get custom WAN checkpoints to work for some reason. They completely ruin the generation quality in all my tests compared to default checkpoint. WTF is this, am I missing something?
>>107701015are you using light2x? some wan shitmixes already have that embeded in them, so using light2x on top of it is like doubling the strength. only thing i can think of is try running it without it
>>107700966>muh pythonkill yourself retard, language has nothing to do with how software works, all non python alternatives are currently WORSE.
Any model to edit pixel art sprites to do different actions like, walking and so on?Maybe qwen edit? Is there even a lora for it?
Given the insane amount of hype around base its obviously going to flood the thread when it drops. I think we should preemptively come up with a temporary containment thread for it like how /wait/ was made to stop deepseek taking over /lmg/. Thoughts on /zog/ - z omni general?
>>107701123python brings in the jeets to ruin the project. see: comfyui
Ok legit what the fuck is the proper way to generate videos locally?I just tried generating text to video 720p with 20 steps in comfyUI with hunyuanvideo1.5 model and this shit came out:https://streamable.com/nag39nI have no idea why those black squares are on the video.This took like 1h30m to make. If i wanted to use like 50 steps it would probably take like 3-4h. Also there doesn't see to be any proper multi-gpu support either.Is image-to-video any better in this regard?
>>107701135Comfy is still faster than the failure you're trying to shill
>>107701123>>107701135why isn't yours in the UI section of OP then julien?
>>107701135never gonna use your trash uicomfyui live forever, ani seethe forever