Discussion of Free and Open Source Text-to-Image/Video Models and UIPrev: >>106609272https://rentry.org/ldg-lazy-getting-started-guide>UIComfyUI: https://github.com/comfyanonymous/ComfyUISwarmUI: https://github.com/mcmonkeyprojects/SwarmUIre/Forge/Classic/Neo: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassicneoSD.Next: https://github.com/vladmandic/sdnextWan2GP: https://github.com/deepbeepmeep/Wan2GP>Checkpoints, LoRAs, Upscalers, & Workflowshttps://civitai.comhttps://civitaiarchive.com/https://openmodeldb.infohttps://openart.ai/workflows>Tuninghttps://github.com/spacepxl/demystifying-sd-finetuninghttps://github.com/Nerogar/OneTrainerhttps://github.com/kohya-ss/sd-scripts/tree/sd3https://github.com/derrian-distro/LoRA_Easy_Training_Scriptshttps://github.com/tdrussell/diffusion-pipe>WanXhttps://comfyanonymous.github.io/ComfyUI_examples/wan22/https://github.com/Wan-Video>Chromahttps://huggingface.co/lodestones/Chroma1-BaseTraining: https://rentry.org/mvu52t46>Neta Luminahttps://huggingface.co/neta-art/Neta-Luminahttps://civitai.com/models/1790792?modelVersionId=2122326https://neta-lumina-style.tz03.xyz/>Illustrious1girl and Beyond: https://rentry.org/comfyui_guide_1girlTag Explorer: https://tagexplorer.github.io/>MiscLocal Model Meta: https://rentry.org/localmodelsmetaShare Metadata: https://catbox.moe | https://litterbox.catbox.moe/GPU Benchmarks: https://chimolog.co/bto-gpu-stable-diffusion-specs/Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-oneTxt2Img Plugin: https://github.com/Acly/krita-ai-diffusionArchive: https://rentry.org/sdg-linkBakery: https://rentry.org/ldgcollage>Neighbours>>>/aco/csdg>>>/b/degen>>>/b/realistic+parody>>>/gif/vdg>>>/d/ddg>>>/e/edg>>>/h/hdg>>>/trash/slop>>>/vt/vtai>>>/u/udg>Local Text>>>/g/lmg>Maintain Thread Qualityhttps://rentry.org/debo
>>106613605If it weren't for that crying girl the entire bottom row would be mine. I'd have a collage bingo.
https://files.catbox.moe/qphnpf.jpgRepeated reminder to not use Chroma HD/Flash HD. Base/2K + flash lora is a good speedy starting point. Base is also the most suited for second pass/upscale.
>>106613615not bad, anon
>>106613629Kek that glitch skirt.
nunchaku team, wtf are you doing, where is the promised wan support??
whats better, scaled fp8 or q8??? BROS??
>>106613641Sorry some literal who just released a model nobody will use so we've diverted all our resources to making that work.
>>106613648same quality, scaled is pretty good
>>106613647nice SD1.4 image anon, I too love nostalgia
>>106613641Bro Wan3 is dropping soon. Give up.>>106613648Q8>>106613655No
>>106613663>Wan3 is dropping soonsource??
>>106613648q8 might be better but it's not guaranteed
>>106613668The blue dragon probably. Some say he is wisest in all of China.
>>106613648I prefer scaled, but both are fine.
>>106613663>>106613668let's hope they got rid of the dual model meme, with the lightvx lora, it's taking more time to unload/reload the second model than doing the inference part
Has anyone experimented with the wanvideo context options node? Supposedly allowing you to gen a bit longer stuff?>>106613648Scaled fp8/16 seems to give me better results for more static videos for loops, while q8 can do a lot of motion. This is for a first frame-last frame loop workflow.
>>106613702>Has anyone experimented with the wanvideo context options node? Supposedly allowing you to gen a bit longer stuff?that doesn't seem to be a usable thing with the way wan 2.2 works
>>106613661perhaps so, it'd however be very inflexibletrain 1.4 model to support more fetish boots with ballet outfit, probably only get these after that
so is small/flat chests impossible on wan2.2? I want to gen some porn of fit track runners.
>>106613729do it with qwen
>>106613709The bane of open source I guess. New things come out and the previous thing doesn't work.
>>106613729in i2v as far as I can tell it's almost only that huge breasts shrink, not that small ones grow (specific lora excluded)
>>106613758>It's a testament to the perils of the sunk cost falacy. He's burnt so much money and obviously hasn't released v7 just because the results were so shockingly bad that it it would instantly make ponysisters rope. This can't end well.I want him to release v7 though, it would be so funny
https://www.reddit.com/r/comfyui/comments/1niddkv/the_comfy_oath_carved_in_stone_free_forever/holy cringe.
>>10661380810 years old me would have been very impressed.
is this snakeoil?
>>106613808Its reddit so they need to pander to their brand of retardation a bit
>>106613850Nag isn't. But delete torch compile.>>106613853Not even reddit is buying it lol.
>>106613850NAG works, but radial attention is piss
>>106613850nag works really well on kontext, dunno for wan though
>>106613850No its WanVideo
>>106613808Cringe yes, but at least he kept his word
>>106613872>he kept his word... yet
the man carrying boxes on his back runs to his left into an amazon warehouse, where a large amazon logo is above the door.amazon stranding is real.disabled high 2.2 lightx2v, low enabled, 6 steps. works like a charm, high enabled kills the motion.
>>106613872to appease the peasants while they laugh, sure.
>>106613882Ehh, ok
>>106613883top kek, if this game wins the GOTY it won't be funny at all though
>>106613808the wording is definitely a bit... but the core of it is greati suppose one day we can generate eminence in the shadows: comfy edition
>>106613883that turned out great.
>>106613648Forget the 4/5xxx series copers, Q8 is basically fp16 while fp8_scaled is quite different every time.
>>106613989yep, nothing can beat Q8, I wished the nunchaku guys focused on making Q8 fast instead of coping with some fp4 shit
>>106613989>b-b-but fp8_scaled CAN look OK!!!Yeah, you can RNG your way into something that looks OK since images have a high capacity of containing error but in places where it doesn't matter. But none of that is relevant when you going away from base fp16 model is objectively gonna be worse in general and especially for details.
>>106614009damn its basically a different model
go amazon man go!
>>106613688you dont have enough ram, it should only take half a sec
>>106614023and it gets worse the further you go
>>106614005nunchaku is even better. The parts of it that would degrade are fp16
before the issue was slow motion. now it can be sanic fast with the high 2.2 lora disabled.
>>106614059I think three steps is too low desu. 4 was around what the paper outlines.
>>106614056>nunchaku is even better.it's not better than Q8 you're delusional
>>106614080it legit is closer to fp16 than Q8
>>106614056>nunchaku is even better.You should be in an insane asylum. Nunchaku is good, but it's not better than q8
>>106614094Q8 has more detail degradation, nunchunu only looks different style wise per seed
>>106614092prove it, show a comparison image between bf16, Q8 and nunchaku
>>106614005They did in the paper, their 8 bit method is basically perfect and also supports SDXL
>>106614112>in the papernigga
>>106614112they compare that to INT8, this shit is worse than fp8 (and even worse than Q8), it's not a good comparison
>>106614112Sir, I'm from the asylum. Please come with us, you need help.
>>106614112our 0.7b LLM model beats the <latest top trillion param model> on this benchmark we specifically finetuned it for its basically better than that model now!!!!!! tier retardation
All the AI papers are fucking useless. Only thing that holds any value is same seed comparison between models.
>>106614112don't cite papers here, they can tell the truth, only trust your gut and tell stupid shit with confidence
>>106614159>only trust your gut*eyes >>106614050
>>106614159>qwen has high aesthetic quality! the paper said so!!
>>106614140are you retarded? it's their own int8 method not naive int8their int4 and nvfp4 are better than q4
>>106614110youll have to wait till I get home but they have this
>>106614165there is neither scaled nor nunchaku stuff there, so you're right, be even more confident!
>>106614159>they can tell the truth30% of the time yeshttps://en.wikipedia.org/wiki/Replication_crisis>A 2016 survey by Nature on 1,576 researchers who took a brief online questionnaire on reproducibility found that more than 70% of researchers have tried and failed to reproduce another scientist's experiment results
>>106614175>be even more confident!>>106614056>nunchaku is even better.yep, that's confidence, always trust a random anon, if he says so, that's true
I trust myself.I made videos with fp8 scaled, and ones with q8, no difference in output, but the fp8 scaled was faster.
Assuming I've got a shitrig of an old server from the 2010s runnin nextcloud and lyrion, How viable would putting a modern gpu there for SD be? Im wondering how much of a bottleneck old chipset/cpu/ram would be?
Have you ever made a claim so retarded the entire general fell into chaos?
>>106614173the problem with this comparison was always that its too basic with a huge room for error in the image, you can fuck it up during inference a lot and as long as its vaguely a book shop of books with correct words on it, its goodgen a realistic crowd of different people of different clothes/races all holding different objects engaged in battle for example or other similar complex prompts, it will shit itself
>>106614222aesthetic af
>>106614200Surely you tested it on multiple seeds on complex motion and action prompts, right... right? Oh...fp8 scasled blurrs the motion
>>106614200>no difference in outputif you only asked for "1girl, walking" then yeah you don't need a solid quant to do this, it depends on each case
>>106614241>aestheticLucky gen. had to (badly) airbrush the little man out of it.
>>106614228not hard, when the entire general already has below average intelligence.
>>1066142252010s is a bit vague. Probably most important is that it's at least PCIE 4.0, and you want your models to be on a fast nvme ssd. If you offload to the CPU (you most likely will for video gen unless you get a 5090 at minimum) then the System RAM speed matters a lot and then the CPU speed.
>>106614273>when the entire general already has below average intelligence.It's your fault, your score is so low that it brought the average down to a ridiculous level.
just unfucked my lora thanks /g/
Radiance is strange because it loves to slap super fine threads throughout the image.
So is it a better idea to train a character lora and a pose lora, or trian the character and the pose in one lora?
My friend is an architect and he wants to use AI to enhance his images. I haven't image genned since the Dreambooth days (I primarily just video gen now), how should I go about this? SD with some realism loras + control net with depth map?
There's not gonna be a real VACE 2.2 is there?
>>106614364What model? For poses you can use controlnet.
not sure if this is the right place to ask this, but can image to video gens be profitable? or there’s a good chance that the original owner can sue your ass into oblivion?
>>106614409You'd be the first.
>>106614335>furshit
>>106614409Do you mean like having a porn patreon focused on i2v content? In that case I think you'd want to gen your own images.
nunchaku wan WHEN WHEN WHEN WEHN