Never Converge EditionDiscussion of Free and Open Source Text-to-Image/Video Models and UIPrev: >>106540158https://rentry.org/ldg-lazy-getting-started-guide>UIComfyUI: https://github.com/comfyanonymous/ComfyUISwarmUI: https://github.com/mcmonkeyprojects/SwarmUIre/Forge/Classic/Neo: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassicneoSD.Next: https://github.com/vladmandic/sdnextWan2GP: https://github.com/deepbeepmeep/Wan2GP>Checkpoints, LoRAs, Upscalers, & Workflowshttps://civitai.comhttps://civitaiarchive.com/https://openmodeldb.infohttps://openart.ai/workflows>Tuninghttps://github.com/spacepxl/demystifying-sd-finetuninghttps://github.com/Nerogar/OneTrainerhttps://github.com/kohya-ss/sd-scripts/tree/sd3https://github.com/derrian-distro/LoRA_Easy_Training_Scriptshttps://github.com/tdrussell/diffusion-pipe>WanXhttps://rentry.org/wan22ldgguidehttps://github.com/Wan-Videohttps://alidocs.dingtalk.com/i/nodes/EpGBa2Lm8aZxe5myC99MelA2WgN7R35y>Chromahttps://huggingface.co/lodestones/Chroma1-BaseTraining: https://rentry.org/mvu52t46>Neta Luminahttps://huggingface.co/neta-art/Neta-Luminahttps://civitai.com/models/1790792?modelVersionId=2122326https://neta-lumina-style.tz03.xyz/>Illustrious1girl and Beyond: https://rentry.org/comfyui_guide_1girlTag Explorer: https://tagexplorer.github.io/>MiscLocal Model Meta: https://rentry.org/localmodelsmetaShare Metadata: https://catbox.moe | https://litterbox.catbox.moe/GPU Benchmarks: https://chimolog.co/bto-gpu-stable-diffusion-specs/Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-oneTxt2Img Plugin: https://github.com/Acly/krita-ai-diffusionArchive: https://rentry.org/sdg-linkBakery: https://rentry.org/ldgcollage>Neighbours>>>/aco/csdg>>>/b/degen>>>/b/realistic+parody>>>/gif/vdg>>>/d/ddg>>>/e/edg>>>/h/hdg>>>/trash/slop>>>/vt/vtai>>>/u/udg>Local Text>>>/g/lmg>Maintain Thread Qualityhttps://rentry.org/debo
Made a gif of this >>106540916 for anyone that is interested. Just looping in celebration.
Cursed thread of foeboat
>>106542161Thanks for making that, the cup part was silly.
I did a few more tests with hunyuan image. Each one with the "reprompt" enabled and then disabled. The initial prompts I used are captions generated from original illustrations Qwen 7B VL.spoiler: sloppy slop but maybe the higher resolution and vae will make it good for upscaling output from other models.
>>106542182That's the reason I edited it. The best bit was at the start.
>>106542161I'll always celebrate Futaba and 4chan.
blessed thread of anime
>>106542163Ahaha based ty anon.
>>106542184>3dpd background with anime faceThis model is so over.
The reprompt didn't change much here
>>106542182catbox?
>>106542184>real ass shoes
>>106542182You're welcome. I downloaded it from Civit, I have no idea who she is, I don't like anime, kek.Remember to post anime here! It's important for /ldg/ unity!
>>106542192i just love me pixar girls so much. i know everyone got it out of their system back with sd1.4, i'm a late bloomer.
The reference for this prompt was a knights of sidonia manga cover
>>106542206y-you don't know yotsuba, the mascot of 4chan?
>>106542196>>106542087"The girl wiggles her breasts before she smiles and blinks at the camera."
No reprompt. The reprompter deleted information about the character / franchise name, very helpful.
>>106542097that works i guess but the 4step loras make it shit the bed
Midjourney-flavored slop?
What's the best model for horror?
Why did the quality of this thread suddenly drop so low?
No reprompt, followed the original prompt better.
>>106542184They're made to look like slop on purpose. Being open source, they intentionally make it look unrealistic
>>106542239640x640
>>106542184Does it know artists?
Last one, reference was a technical diagram for a Soviet rocket. Wondering if it can do fine detail lines/text with the high res vae.
>>106542251I tried a few so far and got nothing. Actually couldn't get it to deviate from its base generic moe slop style at all for anything anime-like.
>>106542223What language version is this because Vol4 is this.
>chroma-unlocked-v50-annealedwhat the fuck is annealed in this context?
>>106542272some schizo bake experiment. avoid
>>106542272dont even bother. stick to wan and qwen
>>106542195
>>106542272Why are you using v50 at all? Stick to Base or HD.
>>106542217No, no idea, never watched it.But that doesn't matter, keep posting it!
>>106542288>>106542290uh, i was already downloading it kek. damn, that bad? im gonna try it
>>106542272Use either Chroma base or Chroma HD. The numbered versions 49, 50 and 50A are abortions. You can also get 48 as a benchmark.
>never watched it
>>106542241
>>106542293>>106542299I like testing models to see the differences on fixed seeds. That's why. I downloaded base and HD too so i'll keep those.Gonna try Qwen and the others later.
>>106542135Now that Nvidia will be finally offering 24 gb at a somewhat not completely insane price, what is the state of the current open source video models? Can they finally support start and end frames? Or is it still just prompt based?
>try comfyui native t2v>blurry and noisy gonna have to bite the bullet on that wrapper huh
>>106542248stay mad vramlet
>>106542266
>>106542184>>106542195Crazy how one is SEXO! SEXO! SEXO! while the other is Ehhh, Yeah. I guess I would.
>>106542294
>>106542314i'm doing start and end frames right now but it's for looping pron
>>106542314Why buy a 5080 super when you can just buy a 5090?
>>106542310I don't blame you since lode explains literally anything ever. If I remember right, annealed was a merge that one random autist requested on his discord server and has next to no use case.
>>106542314>not completely insane priceYou do realize it's going to be vastly more expensive than the 4090 still. You wont get it for MSRP.
Videogen peaked with HunVid.
>>106542347
>>106542357
>>106542340I can't imagine spending tens of thousands and not even documenting anything. Like if he wants this to be a base model to be trained, how about some basic training settings for a full finetune?
Am I tripping or does Qwen not look much different than Noob/Illustrious semi-realistic generations?
>>106542383qwen looks like coherent sd 1.5
>>106542383get your eyes checked
>>106542366prompt for this one?
>>106542383it was probably trained on slopped aibooru outputs so...
>>106542192
>>106542368Lode is a super autist so I don't really think much of him not being straightforward. What does surprise me is how *nobody* in his circle has gone out of their way to help out the community and get some basic guides out for the model.
>>106542334>>106542345The 5070 Ti Super is rumored to be sub 800 bucks. Better buck for money than a 5090 or 4090.
After training a few loras and using it pretty often... chroma is just too hit or miss for me to want to sit here and gen constantly. The variance of quality when it comes to seed is actually insane. Super frustrating to deal with.
>>106542444what part of You wont get it for MSRP do you not understand
>>106542424I don't know. It's from anon https://desuarchive.org/g/thread/103457674/#q103458870
>>106542444>bang for your buckIf you're pinching pennies you're in the wrong hobby
>>106542237depends on what you mean by horror but any good model should play nice with "horror \(theme\)"
>>106542444How fucking poor are you that you can't just get a 5090? You could probably save up in a few months if you're really want it. I agree with the other guy, wrong hobby to be a fucking miser.
How do I run pseudo hiresfix/upscale/detailer or whatever you wanna call it, on chroma? The external hiresfix script doesn't on flux architecture and the flux upscale controlnet doesn't seem to do anything on Chroma. If it can be done with just UltimateSD upscale, then what settings? All I currently get are just enhanced image artifacts and not polishing.
>>106542192>>106542211
>>106542482holy shit, I was just about to ask that.
>>106542482
>>106542444Not sure why you're getting so much negative attention. It's the best price/performance ratio. The 5000 series have been gathering dust for over a year at this point at MSRP so past the initial rush I imagine the same will be true of these.
>>106542522see>>106542449
>>106542445Chroma really needs a long, detailed prompt to get consistent results with. The less detail you give it, the wackier the output is going to be.
>>106542482vae decode. upscale image with upscaler node. vae encode, sampler at low denoise
>>106542528I already mentioned that in my post. 5000 series hasn't exactly been flying off the shelves and I don't expect the revised models to either considering we're in a recession.
>>106542496Reminds me of those old esurance commercials. Also>she won’t stop yappingKek, it’s almost charming at this point
>>106542445it made me go back to slopmixes on sdxl.the prompt adherence can be superior but if i get another random art style change even though it has been described in detail in the prompt i'm just going to delete chroma and shitpost here about
>>106542535Not latent upscale?
>>106542347>Videogen peaked with HunVid.It didn't, but I am excited to go back to legacy video models with a GPU from 2030 that can make a 5 second video in 15 seconds in fp16 at 720p and see what could have been if only we had the compute at the consumer level>>1065425445090 still sells out like hotcakes. I expect the super to as well because it is above 16gb of vram
>>106542469post your 4x 6000 pro sis, you're not a retarded poorfag, right? lool
>>106542583I guess that depends on your country. Here it's never gone out of stock, hence my perspective.
>>106542583>Videogen peaked with HunVid.NTA, but I think it would be more correct to say that Hunyuan peaked with Hunvid.
>>106542586There's a big difference between a 5070, a 5090 and a 6000 Pro. But yeah bro have fun genning on SDXL or whatever you will do on that shitty card. But I get it, you're poor, you don't have expendable income, so even $800 is a bad choice.
how the fuck do i load gguf's with kijai's wrapper
>>106542606>me, NTA, but an sdxl genner on a 5070ti just catching random strays for no reasonDamn anon you don’t have to do us like that
>>106542586You got told to buy a 5090 and you skip to demanding a x4 6000 Blackwell setup, you're retarded. A 5090 is the best bang for your buck. No need to settle for a 5070 LMAO.
>>106542611You just load the model in the model loader node.
fucking around with infinite talker and it's cool but still has janky lipsynnc issues because it's on outdated wan 2.1. Got it to work with a janky 2.2 allinone porn model and it was obviously better with lipsyncing even through all the noise and jank.I'm assuming all the comfyui nerds are making 2.1 workflows for a reason though. What is it waiting on?
>>106542637No he wants a card that's already obsolete and can't even run modern models of 2025 let alone models in the next 2 years.
>>106542535>sampler at low denoisehow low? i keep getting these black scanlines on the upscale.
>>106542637>>106542606>poorfag copeLol. Imagine not having like 5-7k$ after like month of work, poorfag children.
>>106542706broski is larping
>>106542706I don't know why you seethe like this
More hunyuan slop!
Hunyuan seems to really like this inset photo frame composition. Not sure what triggers it
Was aiming for a flatter shading style here, but it least went with the limited palette.
>>106542733I want to know how to seemingly unrelated companies managed to produced the exact same slopped anime style.
>>106542760They really may be training on eachother's outputs + the same prepared datasets. I think that at the least they're doing their deranged "aesthetic tuning" pass with very similar targets.
>>106542760A general model is going to be reduced to the average of the dataset, ultimately what the consensus of what "anime" means.
>>106542710>>106542709It's ok you're a poorfag who larps as having money for spending 2k$, you can just find another hobby lol
>>106542767Aka synthetic data, well known secret the chinks just use ai to train the ai
This one is interesting because the prompt actually avoided including the name "asuka" or anything that specific, while still trying to describe her.
>>106542774>no proof>no terminal of your 4x6000PRO LOL
>>106542780He's a tourist.
>>106542774Do you actually think it's difficult to have more than $10,000 in a bank account? Get a job anon. Even a normal incel can work 40 hours a week and make $1000/week and keep all that money living at home with mommy.
reposting this in case anyone missed:>Wansisters, we've OFFICIALLY and natively escaped 5 second hell. I got up to 15 secs and can't notice the seem, here's the info:>How?- the new Wan Context Windows (manual) node>Workflow- https://files.catbox.moe/aw54aq.jsonHere video example, hopefully some of you anons will get better results (yyyyeeeeess, I know, lightx2v fried the first frame, kek)
Was supposed to be Sailor Mars. No idea how it became... this.
>>106542790Can you believe it, the problem isn't you *can't* do 15 seconds, the problem is no one does variations of the same movements for 15 seconds. You can tell it's just doing the same prompt on repeat.
Still not much luck deviating from the unistyle
I think that despite the higher resolution vae/output, the actual level of detail is lower than what we already get from the ~1MP models.
>>106542790Clown girl ai gf when?
>>106542812That's a general digital illustration, you should browse ArtStation to see what real artists do. Most artists don't do 1000 hour hyper-detail projects.
what sampler/schedulers do you all use for wan
>>106542825I'm in the media industry and quite familiar. In the case of that test, the goal of the prompt was a high level of detail.
>>106542847Well "high level of detail" isn't what any model is captioned with, but you would know that as an expert.
>>106542780>well known secret the chinks just use ai to train the aikek, nothing is more slopped than western models, with Flux dev being the most slopped of all due to being entirely trained on synthetic data, as in nothing but output from Flux ProOnly unslopped base model since SD1.5 is Chroma, even SDXL was doing a lot of synthetic training data
>>106542790What does this new node even do?
>>106542790slow motion
>>106542801True, its a hit or miss but yes, a lot of the gens do tend to repeat. The plus side is, its a step in the right direction away from the hacky workflows. Plus vramlets like me can load 240+ frames without OOMing. Reminds me of animatediff days.>>106542856>These nodes let you sample in sliding context windows instead of all at once, opening up new workflows for long sequences. Currently, only manual control is supported, and some WAN models still need tuning, but this lays the groundwork for more advanced scheduling and custom nodes.Source: https://blog.comfy.org/p/comfyui-now-supports-qwen-image-controlnet?open=false#%C2%A7context-window-support
>>106542790Okay but, can you make her do a long sequence of things within that timeframe or is it going to loop like that? It's also in slow motion.
>>106542895The right direction would be allowing first frame + X frames which can be the last frames from another video with a new prompt. The problem never was "the videos aren't long enough" or rather that's a naive understanding of the problem, the problem is we need to be able to control the video over long sequences and a single prompt doesn't cut it, so you need a sliding window and you need to be able to interpolate the prompt across the sliding window.
>>106542897Fuck knows. Try it out yourself, workflows there or load the new context nodes after updating comfy if you havn't. I only mentioned it because didn't see anyone talking about it.
>>106542935Message the devs, brother.
>>106542095>>106542163>>106542434>>106542496Stop touching the mix with those latex gloves, you're ruining it dumb bitch (Also still fuckign around with VibeVoice)https://vocaroo.com/1cbgLlwQ15FK
>>106542934>256 color VGAThe good old days
>>106542948KEK
holy shit ani is actually onto something with this sdcpp stuff
Hunyuan did this without even asking
>>106542999(the nip, not the censoring)
whats the verdict on the new chink model? total slop? better or worse than qwen?
>>106542999this the image model? wonder if it's fully uncensored like their video model
Name a bigger scam than fp16 fast. You can't.
>>106543022Yes, the new Hunyuan Image.>>106543010Many examples in this thread. Slop level is similar to Qwen.
>>106542999looks like a gay man with long hair
>>106542790i bet if you sped it up so its not longer slow motion it would be 5 seconds
>>106542806This image belongs to Anime Diffusion Thread.
The stupid post above me is going to get like three (You)s
>>106541117show some respect to your king...I think there are like 4 people here with 6000s I'm working on a C210 build with 4 in it :3
>one more seed >i know the next seed will be better
anyone have that one 3 sampler workflow for wan?
>>106543065CFG skimmers
>>106543170I still think fp16 fast is worse. It absolutely butchers quality.
>update ComfyUI>suddenly basic image generation takes up 100% VRAMcan anybody think of a fix or do I need to reinstall from scratch?
>>106543317
don't prompt large eyes on noob
>>106543334download more ram
>>106543334Just change the git head to the last commit that didn't do that?
>>106543334do a basic, separate clean install and test it. shouldnt take that long.or use a backup of you comfy when it worked, you DO keep a backup, right?
>>106543340I used to love playing Beach Babe Killer when I was a kid
>>106542434
>>106543367>you DO keep a backup, right>what is git reflog
>>106543400i know what it is but i just like to start shit while i wait for my prompt to finish
>>106543414I also like to start shit in this thread while my gens are genning.
Are there any small local LLMs for prompt generation? Either booru or boomer prompts
Maybe i'm fucking retarded but, what words do I have to give Chroma for it to make a room look like this. I tried>a dark room with barely any light>dimly lit room>dark ambience>horror theme>television is the only source of light in the roomAnd they're still bright or fully lit.
>just use the 2.1 4step lora on 2.2 bro! >get this
>>106543455yes, multiple. one is even called promptgen IIRC, then dantaggen/tipo and so oni am not very convinced of their usefulness as most of them don't have a particularly good grasp on what tags the models know or the creativity isn't necessarily very goodwildcard lists seem to perform better to me
>>106543490Looks like you're using the wrong vae to me.
>>106543501Makes sense. Maybe editing models are more suitable for my purposes actually.
>>106543505this is a screenshot of the sampler preview
>>106543455Qwen and probably Hunyuan use Qwen 7B VL, so it would probably be the best. Hunyuan actually rewrites your prompt by default using it, but it enslopifies it.
>>106543490works on my machine
there must be more to life than this
>>106543533you're here forever, chud
>>106543512Looks like you're using the wrong vae to me.
>>106543533>than thisthan what?
>>106542314
>>106542314I don't think I could convince myself to buy another 24gb card knowing a 32gb card exists. Like my brain just won't let me.
What u guys think? Is the slight lost of detail is worth the smoother fps?Default 16 fps: https://files.catbox.moe/6dwlns.mp4Interpolate 4x (64fps) https://files.catbox.moe/gsca58.mp4
Found something interesting about Hunyuan and artist names. It has an interesting behavior.If it doesn't recognize the name at all:Generate character with the assumed ethnicity of the name. "masamune shirow" or "yoji shinkawa" produce asians and "greg rutkowski" and "ilya kuvshinov" produce euros (sameface within that group).If it recognizes the name, apply generic style associated with the name. For example, "yoshitaka amano" produces the generic moe anime slop style shown previously. Pic related.Classical painters like Michelangelo and Caravaggio produce the same generic renaissance-like style. Make of this what you will.
>>106543566proooooooompting
>>106543675I make nothing of it because as far as I can tell you are the one human being on earth using this model.
>>106543668Definitely prefer default, looks much more real
>>106543683>is there le MORE to this in lifeyou could say that about any hobby or anything really. if you're depressed, that's your problem but i quite like prooooooooompting and hearing about the advancements in the tech
>>106543685Comfy I think just added support so most were probably waiting for that.
So I've come into the need to use *gags a little* SDXL for some img2img stuff. Is there anything on comfy UI that would let me inpaint faces at a higher resolution?
>>106543724For you? No.
>>106543722i dont know why but for a while i thought krea was a closed source model
>>106543728Fine I'll just use the crop and stitch node.