Yume Edition Discussion of Free and Open Source Text-to-Image/Video Models and UIPrev: >>106688541https://rentry.org/ldg-lazy-getting-started-guide>UIComfyUI: https://github.com/comfyanonymous/ComfyUISwarmUI: https://github.com/mcmonkeyprojects/SwarmUIre/Forge/Classic/Neo: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassicneoSD.Next: https://github.com/vladmandic/sdnextWan2GP: https://github.com/deepbeepmeep/Wan2GP>Checkpoints, LoRAs, Upscalers, & Workflowshttps://civitai.comhttps://civitaiarchive.com/https://openmodeldb.infohttps://openart.ai/workflows>Tuninghttps://github.com/spacepxl/demystifying-sd-finetuninghttps://github.com/Nerogar/OneTrainerhttps://github.com/kohya-ss/sd-scripts/tree/sd3https://github.com/derrian-distro/LoRA_Easy_Training_Scriptshttps://github.com/tdrussell/diffusion-pipe>WanXhttps://comfyanonymous.github.io/ComfyUI_examples/wan22/https://github.com/Wan-Video>Chromahttps://huggingface.co/lodestones/Chroma1-BaseTraining: https://rentry.org/mvu52t46>Neta Luminahttps://huggingface.co/neta-art/Neta-Luminahttps://civitai.com/models/1790792?modelVersionId=2203741https://neta-lumina-style.tz03.xyz/>Illustrious1girl and Beyond: https://rentry.org/comfyui_guide_1girlTag Explorer: https://tagexplorer.github.io/>MiscLocal Model Meta: https://rentry.org/localmodelsmetaShare Metadata: https://catbox.moe | https://litterbox.catbox.moe/GPU Benchmarks: https://chimolog.co/bto-gpu-stable-diffusion-specs/Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-oneTxt2Img Plugin: https://github.com/Acly/krita-ai-diffusionArchive: https://rentry.org/sdg-linkBakery: https://rentry.org/ldgcollage>Neighbours>>>/aco/csdg>>>/b/degen>>>/b/realistic+parody>>>/gif/vdg>>>/d/ddg>>>/e/edg>>>/h/hdg>>>/trash/slop>>>/vt/vtai>>>/u/udg>Local Text>>>/g/lmg>Maintain Thread Qualityhttps://rentry.org/debo
Anyone try out the new Wan 2.5 yet? Made possible by ComfyUI
>>106691543kek
Bro what is this pr
I know people don't unironically talk about the API nodes here, but does anyone actually know how they work? Like could I pass in all the same conditioning I pass into VACE/wan-animate into 2.5? Or is it just very specific types of inputs?
>>106691532>"Gigantabooba!"https://files.catbox.moe/724riu.webmThis isn't mine, just an AI video I saw on JoyReactor the other day.
>>106691562Tranime retards are called that for a reason
>>106691594That's gud.
>>106691594Prompt or artist? Badass.
>>106691294i'm not using stable diffusion to make the videos retarded nogen>>106691519>What do you use for negative prompt in wan2.2 negative prompts?tattoo, tattooschink pasta makes the quality worse imoit doesn't help very much even with NAG
>nunchaku r128Are the leftover pixels caused by the non-standard aspect ratio or the original image being clear png? Should I remove the alpha first before editing?
>>106691602without the leftover pixels the leggings would lose the 3d look
>>106691602>nunchaku r128Not sure what it is but your image looks like a failure to premultiply alpha, what exactly are you trying to do?
>>106691610The jagged pixels are along the entire edge of the pic tho.
Can I use a 'high' lora on both the high and low lora slots? I'm worried that this breast bounce lora will only apply to the high denoiser and then when details are added in low, the breasts will lose all the jiggle detail.
>>106691623Just testing replacing clothes. The orig had clear background so wondering if I should just turn it white first.
>>106691600nomura tetsuya was the artist.
>>106691543Quite ironic but that's Purple Witch if you know what I mean.
>>106691669Nobody cares about your ban evading tranny debo, he's such a thin skin little bitch he seethes all night when EU makes fun of him.
>raise num_blocks_gpu to 10>still no increase in vram use>raise to 12>oomThis nunchaku shit is fucking retarded. 60% of my card is doing nothing.
Why even bother? Local will forever be stuck where it is now till the end of time. What's the point of doing anything knowing it will never get better.
Let's say you are making a finetune checkpoint.And the checkpoint isn't something that can be trivially a lora instead. (I dunno say beating photorealism into an anime model, feel free to give a better example)How many images would you need for this task? How many total steps with Prodigy cosine would you expect to be necessary?Does anyone have a ballpark number?
Windows 10 is in its death throes. I saved so much slop onto my Desktop and Downloads folder I think I completely broke Explorer because it can't show thumbnails anymore and if I try to do anything like sort by date it hangs up until I have to restart explorer in Task Manager. I will never, ever make a Microsoft account
>>106691839wangblowz breaks the os when its (((update))) timeyou can clear/replace all thumbnails under disk cleanup but it will only marginally help
can I post anime in here?
>>106691860Of course
>>106691860As long as it's original and inspiring.
>>106691848tried that it didnt work. I heard Fedora KDE is more lightweight. Win10 isn't built to handle massive outputs of local AI slop
Trying out new Qwen Image Edit, I change my mind from my initial testing on the demo online, it's much better than old QIE overall, but still needs to be pixelspace and hopefully MoE at some point too
>>106691821>(I dunno say beating photorealism into an anime model, feel free to give a better example)This is easier than you think
>>106691877either your drive is failing or you have an obscene amount of files in a single folder
>>106691866>>106691869Ok, thank you
>>106691901what is that, nyte tyde? you cranked the shit out of it
>>106691895>you have an obscene amount of files in a single folderson you have no idea
comfyui is a mass slop production factory
>>106691913huaishen
>>106691916>son you have no ideai do, file explorer starts slowing down in the 10-50 thousand range just segment them
>>106691893Well yeah but that was just an example, not a great one admittedly.I was actually thinking of something like unfucking chroma's broken anatomy and other faults. (Pic not related)I made a few loras in the past but never finetuned a model.If it was easy someone probably would have done by now so I feel like I am missing something important.
>>106691827
>>106691932>fluxtch
>Look in /h/ thread>Ask simple question>Also has a dedicated schizos>They all seem to talk about SaaS and waste the thread with stupid one liners This dude is going cross country on everyone that ever slighted him
>>106691839>retards acting as if your computer magically stops working
crazy how the official QIE workflows still dont autocrop all dimensions so they are divisible by 112 to fix a lot of the quality loss lmao
>>106691821i think noob tuned illustrious on something like 10m images so...
>>106691969Has illustrious caught up yet? The dog fucker booru really pushed it above and beyond
>>106691971>Has illustrious caught up yet?There was new version recently with updated dataset.
>>106691971Nope
>>106691969>noob tuned illustrious on something like 10m images so...holy fuck, what are the best noobs for anime and realism then?
>>106691945i like how it did the eyes
https://github.com/lodestone-rock/RamTorchCan I vibecode this into a comfy node or does it need to be installed on a python level?
>>106691984You might as well train your own lora for realism. For anime, base has always been king.
>>106691886Wtf QIE 2509 is better than nano banana with my limited testing single image editing prompts I previously tried with nano banana, this shit is very good, are there any prompts people tried and failed at that I can try with new QIE?I think people who complain are probably the usual ones that use fp8 scaled or even worse Q2-6 quants instead of Q8 and fp16 8 step v2 lightning lora
>>106691959It's actual mental illness. Like I don't like the SaaS stuff, but going around shitting up every local image diffusion thread with it is not the right play. He's basically turned the site into his toilet.
>>106691969Hopefully that many images wouldn't be necessary
>>106692011Yeah It's been pretty good in my testing. Just don't expect to throw all the parts into a cauldron and tell it what to spit out. You still need to take things in logical baby steps.
https://x.com/bdsqlsz/status/1971022216675590380
>>106692032I'm not even gonna open it. That's the liar and fraud that told us wan 2.5 would be open source.
Double exposure being used as a test now thanks to based Pixart being the only model that could do it during its time.
why does chroma take so fucking long to gen a single image when illustrious takes like 5 seconds. i dont understand
>>106691839just use CachyOS like someone who knows how to go on the computer
>>106692050one is SDXL the other is dogshit
>>106692058i think chroma's nice, but it takes too fucking long>inb4 buy a 5090 and 128gbs of ram lolololol
>>106691994>Your question is nonsensical, you need to install both the pip library and import the module into code>gguf nodes already do something like thisI am actually lowkey curious how well this performs though, maybe it can be useful for LORA training on large models for us VRAMlets
>>106692050why does sdxl take so fucking long to gen a single image when sd 1.4 takes like 0.5 seconds. i dont understand
>>106692068Flux and Qwen are faster, and better, I can totally use the with my good old 3080
>>106692076illu takes like 5 seconds, chroma takes like 1-2 minutes. its ridiculous
>>106691821What's your goal? If it's something pretty specific like a single artist's style then you don't need that much. I made a noobai finetune with about 800 images of a specific artist's style and it did a very good job.
>>106692032>>106692044there's a chance it'll be open source
>>106692050>why does the model that is 3.5 times the size of the other work slower than the other?
>>106692072The upside is suppossed to be less slowdowns thanks to heavy offloads to ram unlike current solutions.
>>106692099Sorry I don't believe anyone after the horrific backstabbing we all just received this week.
>>106692092illust doesnt take 5 seconds if you actually do a proper second pass and dont use some cope quants/faster loraschroma doesnt need a second pass and it still can create more details than illust which can only do tranimeif you want realism, then you need chroma, otherwise stay on ilust/noob simple as
>>106692097Better anatomy and composition >>106691941I am curious if it can be done without spending a fortune.
>>106692090Flux is faster since it's distilled, Qwen is NOT faster
>>106692113Hmm can't comment on that, seems like it would take a lot more data and you would need a pristine dataset.
>>106692112>chroma doesnt need a second passIt absolutely does. It does wonders.
>>106692126post workflow
What workflow are you guys using for Wan 2.2? I've been shopping around for one that uses lightning and supports an input lora, that will safely run on 24gb VRAM (when I've tried adding a lora to the one that otherwise worked, I get an allocation error)
i shan't post mine workfloweth
>>106692145concession accepted
>>106692128https://files.catbox.moe/jgqobf.pngNormally I'd set the denoise ~0.65. If you go higher it starts changing/improving/worsening features.
>>106692155accept this*unzips lora of my penis*
>>106692138You can replace the boards 4chan org with desuarchive in your URL bar, and then go to the previous thread and search for "json" to find my workflow. It's t2v but can easily be turned into i2v by switching the model and loras to the i2v versions and changing the empty hunyuan video latent to an image encode node if you want With 24gb of vram you can probably do 720p in under 5 minutes per video since it'll all be inside the gpu at Q8_0
>>106692138When I was a VRAMlet I used the UnetLoaderGGUFMultiGPU node to load the model, which allows you to specify some amount of the model to offload to system memory (at the cost of speed), if you're not using that node already then you just need to switch to that and then play with the number until you no longer oom. Unfortunately I'm no longer a VRAMlet and I just load the models in full so I can't just send you my workflow.
>>106692097>800 imagesTips on scraping? Seems like most sites make it near impossible.
>>106692223Well in my case it was part of the game's files but if you're scraping danbooru just usehttps://github.com/Bionus/imgbrd-grabberIt has support for other sites but I think a lot of them will require an account or you'll get heavily throttled.
>>106692109What happened?
>>106692239Wan 2.5 API.
>>106692097
>>106692274Interesting style. I would ask for a catbox but it looks like it was made with a meme ui
>>106692286Why would you need a catbox to replicate a prompt?
>>106692274Looks like absolute shitWhat does Ani even offer anymore?He used to larp saying his animation work was saving the space and we see how that failed, now he larped saying that he's important for his vibe code UI and we see how that went, his gens are shit and he's a drunken sperg that advertises every fucking day
>>106692231Nice. Thanks anon.
>>106692286>>106692325for multiple reasons one can deduce that image as not being created with anistudio. you were duped.
>>106692336I simply asked a question about our persistent shameless drunk slob shill that will never touch the power he so desperately craves because everyone and their mother can tell he would abuse it and burn everything to the ground.
for qwen edit v2, how do you reference the second or third image? image2? I know you can describe it but can you reference the node? (ie: what if you have two girls in both nodes)
>>106692366k, image2 did worka computer with image2 on the a CRT monitor is in front of the man with the black pistol.
>>106692366I use the numbering on the image hooks.
>>106692379*on the
>>106692099they are liars. they will never come back for free. hailuo is now 100% paid. they literally removed all free daily points. only new accounts still have limited points. chinks never change.
the man in sunglasses is holding a white CRT monitor with image2 on the screen. keep his pose and expression the same.this model is so fun and versatile. this + wan, and noob/illustrious for anime gens, can do basically anything.
>>106692403This. Never underestimate how ruthless they are once they believe they have the other hand. It's like switch is flipped. I'm not commenting on the morality of that behavior, but pretending it doesn't exist is foolish.
does comfyui glow when local?
>>106692336What do you mean?
Change the text "Deus Ex" to "LDG General". The man in sunglasses is sitting at a computer typing, a white CRT monitor is on his desk. He is wearing a black trenchcoat. (qwen edit v2)
Qwen edit V1 was pretty good. V2 is even better. it's not quite nano bana at home for a multitude of reasons, but it can do what Nanobana does with some creativity.
>>106691959I'm offended you asked /h/ instead of here
Ahem, this was the last thread, any objections /ldg/?
yes, /ldg/ is a saas shillware thread, we know this. now remove comfyui from the OP already
Can someone explain to me what Comfy's "output" vs "denoised output" means on the custom sampler nodes? Because "denoised output" seems to behave identically to the regular ksampler's "return with leftover noise: enabled" option, and I'm struggling to understand what kind of concept could be described as both "denoised" and "returned with leftover noise"
>>106691959Seems like you are obsessed.
>15-40 minutes between gens thanks to glorious new i2v paradigm2022 was a special year and the thread will never be that exciting again.
>106692646>As he cries because his sole function in life is to be a human gnat on multiple boards and threads
>>106692670It's because everyone is genning themselves fondling their favourite hot woman with i2v which can't be posted
the japanese girl in image1 is wearing the outfit of the girl in image2.covered her boobs but it's a very good sailor outfit swap.
there has never been a clearer demonstration of what "sovl" vs "soulless" is than the Chroma vs LoRA pics in this op collage >>106691532
>>106692681a blouse/skirt image, also worked fine:
>>106692692and of course, it works with anime too. this can do a lot more than the first version.the japanese girl in image1 is wearing the outfit of the anime girl in image2.of course, image2 is miku.
>106692686Swing and a miss
>>106692709lora is very obviously slopped tf up
Begging once again for someone to run the wan 2.2 template workflow and post the outpust. No matter what I do I cannot get correct outputs from fp8 models but quants work fine. I've used a fresh comfy install, I've upgraded pytorch, downgraded pytorch, nothing works and I am being driven insane
asuka looks a bit different...
>>106692754>and I am being driven insanea new schizo is about to be born
>>106692801I've been trying to figure out why this is happening for DAYS nowAnother example using some chink lora from civitExample videofp8_scaledQ8
>>106692754>fp8 modelsKijai's and other fp8 variants aren't interchangeable between workflows, at least in my experience (things might have changed since I experimented though).
>>106692754>cannot get correct outputs from fp8 models but quants work finefp8 is a quant, and a cope one at that, Q8 or go home
>ranfaggot
>>106692839Using the template workflow with the correct models just produces garbage and I don't know why>>106692840I do use Q8, but I need to test my shit on fp8 for the copers before uploading them and and I cannot get fp8_scaled models to work
and one more test, seems to work very well: just reference the images as image1 or image2.ie: replace the outfit of the woman in image1 with the outfit of the anime girl in image2. keep her expression the same.
>>106692864>I do use Q8, but I need to test my shit on fp8 for the copers before uploading them and and I cannot get fp8_scaled models to workThink you already posted something like this before, as said, I think it might be some lora training setting that ultimately fucks with the lower than the full precision model inference.There was a case with horizontal artifacting happening even on Q8 with some prompts on Chroma without a lora compared to bf16, so it's probably a deeper training issue that can't easily be fixed.If it really matters to you and you can get similar lora performance with it, try using another trainer to train the loras. Try the lora on different fp8 scale types, e4 e5, try all those fp8 quants on both kj and comfy nodes, and if nothing works then just say that your lora verison there only works with Q8 for now.
>>106692892
hi anonis there flux redux equivalent for qwen yet?
>>106692899That was me, It's not my lora that's the problem that was just some civitai retard with an unrelated issue, in testing that though I found that using fp8 is just completely fucked for me, this >>106692754 is the completely standard comfyui lightx2v template workflow with the listed models. If someone could run it and post their result I'd really appreciate it
>>106692904
>>106692922Post the exact workflow you used and links to the exact versions of the models and loras you are using and ill do it
>>106692932it's pretty cool how qwen edit/kontext can do all this which would take a lot of effort with inpainting or controlnets and other stuff, plus edits with this model treat elements like their own layer. so you can change/remove elements without altering the composition otherwise.
>>106692940The girl on the right is asian so she still gets some /ldg/ points.
>i'm a nigbophile
>>106692948like, how would you change a plugsuit into a cammy outfit, high denoise + prompt, right? but then how would you get it in perspective, or done properly, even with openpose? and you'd need a mask to get it pixel perfect.it's a pretty neat tool, inpainting is still useful but this is another option to make stuff with and is very versatile.
i miss schizo anon
>>106692849nice
>>106692940I'm honestly getting an urge to train a lora on SD3 'girl lying in grass' output, if nothing else you could upload it as Cronenberg style
>>106692948what is the container size?
>>106692960replace the outfit of the red hair anime girl in image1 with the outfit of the girl in image2. keep her expression the same, and her red hair the same.and poof, asuka but dressed as cammy.
>>106692935You should already have the workflow, it's in templates but here anywayWhat it should behttps://files.catbox.moe/8y5xjs.mp4fp8https://files.catbox.moe/tmwx3y.mp4Q8https://files.catbox.moe/k4nahp.mp4fp8 models/textencoder are linked in the workflow. Or find them here https://docs.comfy.org/tutorials/video/wan/wan2_2Q8 models here https://huggingface.co/QuantStack/Wan2.2-I2V-A14B-GGUF/tree/main
*sip*
>finally can train loras and craft all the goofy shit i want>only limitations is vram for loading wan framesWish it was like animatediff where you can load 1000 frames without issue. If any kind anon with a 32gb card or higher can confirm how many frames at, lets say 512 x 512 with wan context can you load for i2v and does it slop out or hold context? I can only get up to like 190
I'll just say it. I think the collages lately have been lacking in the care and effort department.
Serious question. How did comfyui get to the point that I’m getting fucking Nvidia emails advertising it? Anyone else remember when the comfy dev himself was in these generals back during the very first threads shilling comfy so much that it almost got banned from the OP? From shilling comfy on 4chan to partnering with Nvidia is insane. How did he do it?
>>106693163By never engaging the schizos and going for the throat of big businesses.
>>106693080>'It's hip to be /ldg/'
>>106693163start emailing Nvidia and tell them comfyui is shit and anistudio has faster inference
So how many users does Ani studio have and what models does it support? I might jump ship if it's worth it.And don't try and pawn another gradio shit interface on me. Not interested.
>>106693197pros:model loading is incredibly fastinference is noticably fasterit's a pure C/C++ applicationcons:memory management isn't ready so you have to reload the checkpoint every genno qwen support for nowdoesn't support all the diffusers Lora formats eitherupscaling only supports two esrgan modelsalso, ani hasn't said anything about it being ready to switch over and he is in Japan to drum up funding and support. he's playing too many roles to have things done in a timely manner but who knows what will happen when he is finished in tokyo
>>106693227if ani brings Japan back into the AI race to beat the chinks I would be forever grateful but he's still a fag rn. at least he is a hard working fag
>>106693239I've worked in Japan for quite a while and Japanese people some of the most clueless motherfuckers when it comes to AI, it's genuinely astounding. Like the ones who know their shit are genuinely good, but for the most part companies still haven't gotten over CV.
>>106693227>he is in Japanlucky nigger>to drum up funding and supportI have a bad feeling this could end up being a comfy situation but at least he knows he can't do it alone>playing too many roles to have things done in a timely mannercomfy was incapable of running his own company let alone a fucking frontend and just let the grift chink sell everything out. maybe I should just trust ani because he doesn't seem like a greedy sellout like the alternatives>who knows what will happen when he is finished in tokyoplease be good end
>>106693138That is because the best prompters are posting API gens.
>>106693138just skip niggerjak bakes
>>106693288>comfy was incapable of running his own company let alone a fucking frontend and just let the grift chink sell everything out.lol so true!
>>106693227honestly, just sounds like it needs more time to mature but it's on the right track. I find it wild ggml beat out pytorch despite being some random balkan basement experiment.
>>106693227LMAO. Everything about this sounds like it is an absolute meme that will get abandoned in two months.Can you guys try working on it and maybe get it to a respectable state instead of relentlessly shilling it here?
>>106693343>abandoned in two months.The real joke is that this has taken way more than two months to get to this point.
>>106693343>guysit's just one guy compared to an army of chink sloppers comfy has
>>106693343He's working since 3 years to get to this state kek
>>106693354to be fair, it takes 2 years on average for homebrew game engines to get to a stable state
>almost solved the color shifting for loops>introduced a large stutter in motion instead
niggerjak woke up and immediately started seething I see
>>106693334if it has all the ggml options, multigpu is finally solved
nvidia sissies...>China's latest GPU arrives with claims of CUDA compatibility and RT support — Fenghua No.3 also boasts 112GB+ of HBM memory for AIhttps://archive.is/jQVZo
>>106693402>China>claims
>>106693365I don't know how you specifically always seems to have these issues.
>ran took everything from me
recommended comfyui img2prompt model?
>>106693413Try doing it on more complex images. I'm doing seamless loops of my older gens which are far from simple.
>>106693428joycaption
>>106693428If it's single images just use joycaption on HF or ask Gemini
>>106693428https://www.reddit.com/r/LocalLLaMA/comments/1not4up/qwen3vl235ba22bthinking_and/
>>106693440>GeminiI know it's SaaS shit by my god Gemini understands images so well.
>>106692754>>106693010I've finally fucking narrowed down what the issue is, it's how native comfy nodes merge the lora weights to the modelhttps://huggingface.co/Kijai/WanVideo_comfy/discussions/52#689196f3665bcc325ec1dbac>When using GGUF in native Comfy or the unmerged LoRA mode in my wrapper, the LoRA weights are not merged and instead handled like I explained.Picrel is fp8 on KJ nodes, working as intended
lets remove comfyui from the OP first and figure out the technicals after
>>106693463So is the error something that can be fixed in the official comfy workflow or
>>106693463Does this affect all models or just wan?
>>106693434thanks>>106693440>ask GeminiI'm sure Gemini will gladly img2txt my nsfw images>>106693450I'm a vramlet so no qwen for me unfortunately
>>106693480use taggui with local joycaption modelhttps://github.com/jhc13/taggui
>>106693473I don't know, it could be a broken node, or something wrong with my setup. I have a 40xx series card so I should have issues with fp8 and I've tried every pytorch+cu combination under the sun. At this point I'm just going to swap to KJ nodes>>106693475Also don't know I pretty much exclusive use comfyui for Wan
>>106693498>so I should have NO* issues with fp8
>>106693163By being an agressive sperg who hires bots to damage anything remotely resembling competition. Remember his feud with lllyas right here, and then he tried the same with invoke, even though nobody literally cares about invoke. Except comfy, because they're in his space. Don't cross comfy's pass, he'll go right for the throat.
>>106693480Pretty sure Gemini stops giving a shit once you put it in silly tavern.
>>106693521>agressive sperg who hires bots to damage anything remotely resembling competition>Every accusation is an admission
ComfyUI is SaaS adware and should be removed from the OP. If you are fine with ComfyUI being in the OP, you are fine with SaaS gens and API shilling.
>the most memory optimized official comfyui workflow (QIE2)
what if you worked on your trash wrapper instead of endlessly sperging here?
>>106693577>everyone I don't like is aniwhy don't you contribute to something instead of shitting your schizo diaper every 5 mins itt?
>>106693580>>106693586>>106693589>>106693594>>106693600Why are you posting sora gens with the assets_task in the filename replaced with comfyui_2Ftask anon?Did you want (You)s, here are some.Please fuck off now.
>>106693589pissmaxxedgrainmaxxedslopmaxxedredditmaxxedcuckmaxxed
They appear to be locally generated though? They have comfyui in the filename
>>106693639>>106693569
>>106693589It would take over 5 hours to gen something like this in chroma btw
>>106693589People pay for this shit, lmao?
>>106693649Why would anyone want to gen this piss ?
>>106693486interestingthanks
Which one is better, Flux Kontext Dev, Pro, or Max?
>>106693911leave before you become a schizo too.
>>106693911Qwen Image Edit
>>106693911Max of course.
>>106693943Thanks. And between Qwen vs Seedream 4?
>>106693911wan2.5
>>106693954Seedream 4 easily. It's the best model available currently.
>>106693963*Diffused locally via API*
>>106693963For editing I suppose? I have tried it for image gen and it didn't seem that good, at least for anime-ish stuff, though I'm a promptlet so it might be a skill issue
>>106693975Neither are good for anime, you will want illust/noob or novelai for that.
>ups died while I was genning cunnyfucking APC>it's still in warrantyI'm fucking fuming the cunny was COMING so prime I could feel it, the combo of samplers and tags I was using made it GLISTENING, but no, the APSHIT had to fail.
>>106693498>At this point I'm just going to swap to KJ nodesnot q8 gguf? i think you indicated earlier the "quants work fine", i took hat to mean the gguf quants
Qwen Image Edit 2509v2 lightning loras seem the best versions compared to v1 and v1.14 step lora quality is actually very close to 8 step, due to RNG its not uncommon to be better for a set seedBut both are not too great for very large changes, where you need cfg 4, thus 50 steps.Solid jump in improvement of the model overall, especially for cartoony styles, very similar to Wan 2.1 to 2.2 jump, it seems like Qwen had a lot more cartoony data to add to their models.Single image editing is almost nano banana level overall, and for a large amount of prompt types it's better, although with the same somewhat opinionated changes that the model makes. Not unexpected as they basically certainly trained on their outputs.Image height and width needing to be divisible by 112px so it crops them less and it still doing it a little is still annoying.VAE quality loss is still there, this is a huge problem for image edit models, we obviously need pixel space models asap, like Chroma radiance.Outside of this, the biggest thing to improve when it comes to the model itself is for it to be able to copy the likeness of the images even better, it still slops them up too much into non-candid plastic when doing bigger changes but this is just something that will need everything about the model training process and datasets to get better over time to continue to go away slowly.
>>106691532why hasn't the SAAS troonware been removed from the OP yet?
>>106694081>VAE quality loss is still thereit royally destroys complex backgrounds, smudges them all over from my experience. very sad.
>>106694094If you keep it divisible by 112, and probably around 1 megapixel, rng it a little, it can keep most of the details.But it's probably better to try some more advanced masking workflow which I didn't yethttps://civitai.com/models/1983350/ultimate-qwen-image-edit-plus-2509https://civitai.com/models/1986315?modelVersionId=2248464
>>106694094>it royally destroys complex backgroundsI only had that happen once. Other times it stayed the same. I don't know what the cause was.
>>106694108>>106694130it's when it fucking zooms in, which can be mitigated when you set multiplier to 112 like the other anon said (you also need to not apply the vae to the text encode node and instead use the classic reference latent conditioning)But even then, it CAN still happen, and when it does it's a wasted gen.To mitigate this I'm personally just using the lighting 4step so at least I only waste 8 seconds of my life if the zoom in happens.
If I could get any superpower I wanted, it'd be deleting people from existence through civitai.
>>106694040what a shamesurely you still have the workflow so you can genn it later
>>106694074Basically for me>Default Comfy nodesQ8 works finefp8 broken>KJ nodesQ8 works finefp8 works fineFrom what I can gather this is due to how comfy merges loras into fp8 before inferenceYou can read KJ explain it better here https://huggingface.co/Kijai/WanVideo_comfy/discussions/52#689196f3665bcc325ec1dbac
>>106694180Surely there's a comfy API node that can fix this?
Is there a general purpose inpainting model that can just "pick up" style from the surrounding image?I just need to unfuck a small hand.
>>106694193no, if you want a classic masking and/or SEGS/detailer, your best bet is using either the edit models (qie/kontext) or dedicated inpaint models (flux inpaint). Usually they 'pick up' the original style but sometimes you either gotta lora it up or switch to a model that most resembles the style youre aiming for
>>106694134What am I looking at, did you hide people that got rid of their account?
>>106694216Blocked users. All the fuckers uploading furry, gay etc.
>>106694220holy fucking based
the fucking chink promised the qie+ lightning models today, WHERE THE FUCK ARE THEY YOPU FUCKING CHINMKOID
>>106694211I don't want Kontext/QIE since I don't want it to touch the rest of the image (Plus they may not be too happy about the booba.).I guess I can give a shot to flux fill, it's mask based as I want if I understand correctly.I hope it doesn't stand out too much or produce those weird inpaint smear artifacts.
>>106694180thanks, nice findings
>>106694220Oh I see, wouldn't it be better to just hide the tags? (If that's possible)
>>106694242you can use the edit models for INPAINT workflows, that's what I mean, not in pure edit mode.
>try qwen edit>it adheres so well to the dress design that it removes the cleavage>change dressSuccess. This is extremely useful.
>>106694247It's right above the blocked users section and there's even a toggle to get rid of furry content. They're either retarded or autistic
>>106694249Oh, so I can give Kontext a mask and it will only touch that mask?Let me try this.
>>106694277the mask literally forces the models to only touch the masked part, this is valid for ALL models
>>106694274>>106694247It would be so lovely, but these fuckers doesn't tag their shit, at all. So it's useless.
>>106694255Kek, it even works on less coherent images. I even managed to prompt for it to hide the entire body.
>>106694284Pretty sure civitai auto tags images with no way to edit them yourself
I haven't used these types of editing models before, so I'm pretty blown away like a boomer.(this was grayscale)
>mix the physical features of the two women together into oneFinally works with QIE now with 2509, big
>>106694367make him fall onto a walmart parking lot
"turn the two men on horses into office workers wearing black suits and fashionable pants. remove their hats and replace with short stylish black hair."The overall quality gets destroyed, but you can just do further edits in photoshop, masking them out etc.
>>106694081 >>106694255 >>106694376got an interesting workflow for this or is it the example one?
>>106694389can't you just add a grain/gauss node or something?
>>106694393Example with Q8 gguf loader, you can also try >>106694108
>>106694380
>>106694376Actually, it doesn't really work
Holy shit I lost myself in this for an hour already.>>106694393Yeah the example one with q8 as well.>>106694408I'd rather do it manually in photoshop along with the rest of the editing.
There used to be an eye symbol in ComfyUI that quickly hid all connections/noodles. It seems to be gone now for some retarded reason, and that's after reinstalling after not using AI for a couple months. Anyone know how to get it back? I don't want to have to go to the options menu and set render mode to none every fucking time I want to hide them.
>>106694549
How compatible are base flux loras to kontext and fill?
>>106693660Second pass with 0.95 denoise turned it kino
>>106694568on a scale of are to not; not
>>106694574>Second pass with 0.95 denoiseIndistinguishable from just doing a first pass.Also which model is this, do be aware that many won't like you rawdogging 1440x1440 resolution too much.
>>106694559Ugh, I see, that quick bar only shows up in focus mode. Lame.Thanks though.
>>106694597View > Bottom Panel, then drag the logs window down so you don't see it. That way, you keep the quickbar
>>106694594Chroma. When you start pumping the denoise 0.80+ on the second sampler it starts doing big changes and 0.90+ it just uses the original pic as a i2i reference
what happened to the voice diffusion from microsoft? need to catfish people
>>106694611That did it, thanks!
>>106694617Yes you are just doing a whole new image.Just use empty latent, first pass becomes a pointless waste of time above 0.7.
Is this comfy ui alert anything to be concerned about?"Invalid workflow against zod schema:Validation error: Invalid format. Must be 'github-user/repo-name' at "nodes[43].properties.aux_id""
>>106694621They pulled it, but it's an open license so they can't take down anyone who hosts it.>https://github.com/diodiogod/TTS-Audio-SuiteUse that guys. It auto-downloads it. Get 7B and don't quantize to 4bit if you're not a poorfag with a sub 24GB.
>>106694588I am just disappointed that there is no lora to help NSFW stuff with Kontext.
>>1066946341st pass + 2nd with upscale gets better gens than just trying to smash it raw and hoping the model keeps coherence on high Mpx count
what do you reckon is the size of wan2.5? it has a bunch of extra shit this time, with the audio model and such. wan2.2 was 16ish gb on q8/fp8.i wonder if the audio/speech is done during video generation or afterwards.man i want to know the tech details so badly.
>qwen edit refuses nudityShieeet.>>106694621Shouldn't you be out shitting on the streets of canada?
>>106694662audio model?are we going to be able to gen audio for i2v and t2v?
>>106694666blame trudeau
>>106694621>need toNo you don't, Pranjesh
>>1066946591. The resolution of both images you posted were same so no, it's not an upscale.2. No it helps jack shit at such high denoising. The model sees next to nothing from the original image. It will have the exact level of coherency as drawing from scratch.
>>106694694The first was upscale on lower denoise
>>106694673it can generate speech/audio for the video. the quality is.. shit to be honest but you can always just dub over it as the lip movement will be there.
>>106694662FUCK OFF SAAS SHILL
>>106694393>example onewhich one is that?
>>106694814are you retarded? if i were a shill i'd say how amazing it is. but it isn't. the audio is shit and it's slow as fuck. man kys, i just want to know the model details. i do not give a fuck if it stays api or not.
>>106694876ok you can ask about the model details in the NON LOCAL diffusion general. you're welcome
>>106695094Where's that?
localjeets never recovered from dall-e 3
>>106695149yes, yes, local is toy, look upon saas and despair, yadayada
we need someone to leak the seedream model
>>106694648Cracks me up how they purged their github of the models, crying over people "abusing" the model. The fuck did they expect people to use a zero shot voice cloning model for?
>>106695167>create robot intended to cook for you, but it can also jerk you off real good>majority of users use said robot to jerk them offsurprised pikachu face
>Replace the outfit of the man in picture 1 with the outfit in picture 3. >Change nothing else about picture 2. Only the outfit.
heh
neta yume is pretty good ngl, I've been gooning for the past two hours genning nekopara bitches
>>106695359meh, it didnt humanize the red bitch
>>106694862All the "example" workflows for any model I think of so far were in comfyui (browse the workflows) OR -especially before this feature- on github/huggingface where the model was published... also I just trust the anons to know if there's something remarkably special in their workflow or if it's just ~the thing anyone would have>>106694410ty for that too
>>106695362>the red bitchdon't insult Akari like that!
You are reporting bad things, right, anon?After you downloaded it of course.
>>106692099lol, it's over
>>106695392yuru yuri (not real yuri btw) is a garbage sol moeshit anime, sorry.
>>106695415>not real yuri btwit is real yuri, and a funny one>pircelyour anime suck, it's "omg I'm a straight girl but a lesbian declared her love for me what should I do?" only subhumans would enjoy this shit
kek
>>106695427>it is real yuriit is as yuri as nichijou is, you dumb faggot>rest of the postholy shit taste
>>106695436real usecase for wan, removing faggots from movies.BASED CHINA
>>106695446>it is as yuri as nichijou is(You)
>>106695362Why would I want it to do something I didn't ask it to?
>>106695451will we get a nexus mods site but for movies this time? top kek
>>106695415It's not real yuri, it's yuru yuri and the word yuru is doing a lot of heavy lifting.
>>106695460fucked up eyes, ultra slopped.qie the SPRO treatment
>>106695396catb0x it, or at least tell what it was
https://xcancel.com/bdsqlsz/status/1971055141001605307#m>China just released a new GPU>CUDA-compatible>112GB of HBM memory,HOLD UP LET THEM COOK
>genning porn>the girl randomly turns her head and looks at the camera with a confused and slightly disgusted expressionS-should I stop..?
>>106695488it's a propaganda post most likely, we'll need to wait until the card is actually out, also imagine the pricing on it.
>>106695499it's a sign the model is recognizing you're raping it, so continue
>>106695468yuri doesn't mean "the characters end up being in couple" it means there's lesbians in there and it's true, they're almost all are (except Akari I guess)
>>106695382thanks, civitai "examples" are always like "150 nodes and 20 new custom ones to install" and it's a bit annoying when I just want to test the model itself in a simple way
>>106695504>ACKSHUALLYlmao keep coping bro
>>106695500>Chinese company shows their new GPU>PROPAGANDA >:O>American company shows their new GPU>Advertising :)
>>106695488>claimsit doesn't exist, anon
>>106695528lurk more, there's also movies where the characters don't end up in relationship, and we still call that "straight romance"
How many more years until models can create / replicate facial expressions? This is my biggest frustration with image generation, it always sloppifies any facial expression / head positioning and angle to that 'staring at viewer with a blank expression' or at most smiling.
>>106695470Unfortunately, I don't think it's really possible to unslop once the sloppening has begun.
>>106695562better than before but yeah. I think it'd be better if he walked with miku and teto desu, akari is a slut
>>106695562why does she have dirt on her knees? did she.. you know..
anons is there any big difference btw fp8 and q8 on wan2.2?
>>106695597remind me of the teto meme where she's tired as fuck and miku ends up getting the employee of the month or something kek
>>106695597it's a pantyhose>>106695602q8 is always better than fp8, so go for it.
I hate the vocaloids
>>106695602Not a huge difference from my tests, but q8 will look better, at the price of slower inference for 5000 cards since they are optimized for fp8.
>>106695652still better than Vutubers desu
>>1066956564000 cards are also optimized for fp8, wasnt the 5000 meme about fp4?
>>106695664absolutely
>>106695665I only have a 5090 and 3090, so I dunno, but I tested q8/fp8 scaled, and while for the 3090 there was no difference, the 5090 was faster for fp8.
>>106695700strange, for inference everything gets casted to fp16 from what I remember
is fp16 fp8_e5m2 dtype going to look better than q8 default dtype?>>106695656
>>106695713no, no one talk about e5 because it's worse than the e4 one
>>106695713it goes like this:fp16bf16Q8fp8_e4m3 (for 4000s/5000s)fp8_e5m2 (for 3000s)If you go any lower than this you'll get shit quality
>>106695708I got 450s vs 600s in wan when I tested, from memory.>>106695713All I know is:fp8_e5m2 (scaled or not) -> recommended for 3000 cards and belowfp8_e4m3fn (scaled or not) -> 4000 and up
>>106695499>the girl randomly turns her head and looks at the camera with a confused and slightly disgusted expressiondid it also say your full name and social security number? i hate it when that happens
>>106695488>>106695500Please, based Chinaman, liberate us all from the nvidia menace>>106695499Kek, this the twerking loras, skinwalker bitch looks directly at the camera
>>106695499>confused and slightly disgusted expressionis wan even able to understand that?
>>106695740so if i use fp16 model with fp8_e5m2 dtype, it's the same as using an fp8 model?
>>106695800bro what the fuck are you doing/babbling about?assuming you're talking about kijai's wan nodes, the dtype has to match your actual model type. I doubt you downloaded the full fp16, no? literally go ask chatgpt about this, you seem to be lacking a fundamental understanding of how this shit works
>>106695809>the dtype has to match your actual model type.he literally said he downloaded a fp16 model, so he's allowed to run it on fp8 e4 or e5
>>106695800>it's the same as using an fp8 model?if the type of quanting to fp8 of that fp8 model is e5m2, then yes
>>106695451I'm waiting for Ryan Goslings "Roots"
>>106695800yes, >fp16 + you run on fp8 e5 mode = fp8 e5 >fp16 + you run on fp8 e4 mode = fp8 e4
>>106694231where's the lightning nunchaku model? is the chink sleeping?
> nunchaku> can't run lorawew lad.also, what is the logic of this? is the 8step lora higher quality than the 4step one?
>>106694231I noticed that using the lightning loras on qwen image (and edit) made the images more plastic and slopped, since QIE is already a highly plastic shit, we're going to the radioactive territory with that one
>>106695898>is the 8step lora higher quality than the 4step one?obviously
>>106695898it can run loras for flux (the other supported model) but not for qwen (wip, very close to release).
>>106695980>fluxdead model lmaoi'll try nunchaku once it has lora support. i can run the native model so the only interesting thing here is the apparent speed increase.
>>106695991>dead model lmaofar from it, there's a reason Chroma was made from Flux Schnell and SPRO from flux dev, Qwen Image hasn't replaced flux at all (and it's humiliating to know that since QI is a 20b model, almost twice as big as Flux)
>>106695991qie is way more slopped than flux, I'm hoping it gets the SPRO treatment
>>106696009>I'm hoping it gets the SPRO treatmentno one will do that on QIE since the Alibaba fags will release an "improvement" every month
>>106696016I know they called it 'monthly' release, but will they really?
>>106696016I'm ok with that
SPRO just looked like low noise injection but i know far too little about it to really REE about it.>>106696003i get what you are saying but am having far more fun with qwen image compared to flux. are there any flux finetunes you could recommend that enhance it's general understanding and nsfw capabilities? because needing a lora for every single little thing is tedious.
>>106696042it also destroyed details, but maybe that was because they applied it to a distill.
>>106696052hello biuteful, show bobs
>>106696042nta, i havent found a nfsw flux that isnt dogshit. having to get loras for everything is gay
>>106696070>having to get loras for everything is gaytrue that, that's why I'm rooting for edit models, at least you don't need characters lora anymore if they work well (so far only nano banana works well)
bread?
>>106696076Never used nano banana, I assumed is unlewdable?
>>106696076I didn't test the new qie but my main issue was that it was kind of dog shit at even understanding what underwear types or a lot of clothes types (the more daring ones) were.
>>106696093it's google anon, it's so safe, you will have a safety orgasm
>>106696093it is hardcore censored. people have to use all kinds of retarded language to get anything wearing a bikini to not cause google to dronstrike your house.
Neta yume bros, why are there basically no loras for this model? Is it hard to train?
>>106696109They're losing to SDXL because their base model isn't trained well enough.
>>106696128some sad ass there
>>106696128Why go to such extent to just generate the video it was trained on?Same thing baffles me with all the porn people are doing. Just replicating the trained data.
>>106696042Flux Dev SRPO at guidance 3.5 is equivalent to like guidance 2 on normal Flux Dev. The apparent advantages of SRPO are way more apparent on seed-to-seed comparisons if you take it up to around guidance 4.5. That said I far prefer Flux Krea regardless, it just has quite noticeably better adherence than normal Dev or SRPO and a better understanding of a lot of style-related stuff, along with the improved realism.
>>106696130That's definitely not the reason, lack of awareness of Lumina 2 -> Neta Lumina -> NetaYume Lumina even existing plus the added hardware requirements are likely more the cause. A lot of people really can't handle anything that's even a bit more demanding than SDXL, even now.
>>106696130it's fucking sad really, if you check the lora page on civitai it's 21 loras, most of which fucking SUCK. I just wanted to gen some bitches from the currently airing anime, but NO LORAS.
>>106696190everything is for FUCKING illustrious
>check youtube for 2.5>check plebbit for 2.5Kek, when 2.1 and 2.2 dropped it was everywhere, now almost radio silence
>>106696178Up to around Wan 2.2, we know people will use it if the output quality is sufficient, so reasons like long computation time have now been ruled out. It simply came down to poor model performance.
>>106696190AlphaVLLM had their own lora training scripts for the underlying Lumina 2.0 but there was not support in e.g. Kohya until very recently. Kohya does have it now in the "SD3" branch, not sure about other trainers.
>>106696212The NetaYume guy has been steadily improving the quality of Neta Lumina 1.0 with his finetune though, and Neta Lumina 1.0 was always quite objectively better than stock Illustrious 0.1 (which people continue to train off of)
>>106696247yeah I mean all the gens I posted were done with neta yume v3, I couldnt believe my eyes with the QUALITY I'm seeing for anime at least. It can also generate up to 2048x2048Also being able to use both tags and boomer prompting at the same time. I don't think I can go back.>>106696225I guess I'll need to train my own loras, will check kohya's scripts
>>106696274>>106696274>>106696274>>106696274
>>106695740Sorry but isn't fp8 supposed to be better than Q8 since it uses floating point or does it entirely depend on the model and application?