Pruned Edition Discussion of Free and Open Source Text-to-Image/Video Models and UIPrev: >>106642301https://rentry.org/ldg-lazy-getting-started-guide>UIComfyUI: https://github.com/comfyanonymous/ComfyUISwarmUI: https://github.com/mcmonkeyprojects/SwarmUIre/Forge/Classic/Neo: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassicneoSD.Next: https://github.com/vladmandic/sdnextWan2GP: https://github.com/deepbeepmeep/Wan2GP>Checkpoints, LoRAs, Upscalers, & Workflowshttps://civitai.comhttps://civitaiarchive.com/https://openmodeldb.infohttps://openart.ai/workflows>Tuninghttps://github.com/spacepxl/demystifying-sd-finetuninghttps://github.com/Nerogar/OneTrainerhttps://github.com/kohya-ss/sd-scripts/tree/sd3https://github.com/derrian-distro/LoRA_Easy_Training_Scriptshttps://github.com/tdrussell/diffusion-pipe>WanXhttps://comfyanonymous.github.io/ComfyUI_examples/wan22/https://github.com/Wan-Video>Chromahttps://huggingface.co/lodestones/Chroma1-BaseTraining: https://rentry.org/mvu52t46>Neta Luminahttps://huggingface.co/neta-art/Neta-Luminahttps://civitai.com/models/1790792?modelVersionId=2122326https://neta-lumina-style.tz03.xyz/>Illustrious1girl and Beyond: https://rentry.org/comfyui_guide_1girlTag Explorer: https://tagexplorer.github.io/>MiscLocal Model Meta: https://rentry.org/localmodelsmetaShare Metadata: https://catbox.moe | https://litterbox.catbox.moe/GPU Benchmarks: https://chimolog.co/bto-gpu-stable-diffusion-specs/Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-oneTxt2Img Plugin: https://github.com/Acly/krita-ai-diffusionArchive: https://rentry.org/sdg-linkBakery: https://rentry.org/ldgcollage>Neighbours>>>/aco/csdg>>>/b/degen>>>/b/realistic+parody>>>/gif/vdg>>>/d/ddg>>>/e/edg>>>/h/hdg>>>/trash/slop>>>/vt/vtai>>>/u/udg>Local Text>>>/g/lmg>Maintain Thread Qualityhttps://rentry.org/debo
cope thread of vramletniggers and mental cases
all the rage against the KJ machine aside, i am getting utterly blown the fuck out by sharing this benchmarking page given they directly recommend his nodes. so. lol.
>>106647244has anyone done a good comparison using official vs kijai? how much movement loss are we talking about?
What the fuck is this new movement copying tech, what did I miss, where are the workflow jsons?You guys need a News section in the OP like /lmg/
>>106647244are they retarded? so far kijai has always been the far heavier workflow.
Blessed thread of frenship
>>106647244>Native version (the so-called official workflow)Kek, Comfyui is such a shitshow
>>106647273I wonder what model they use to parse previous threads for news, migus, etc
>>106647244> the kijai version is better because it uses kijai nodesholy fucking esl. whoever wrote this is a room temp mango.
will comfyorg pat the h1b premium to keep comfy, the chinks and the jeets in the US?
>normies waking up to the idea that AI can generate children Oh no...
>>106647305lmao, trump killed the org
>>106647300the original is in Japanese, that's an automatic translation
shitposting aside, is flux kontext actually better than qwen edit at anything? i'm cleaning out old models and i genuinely can't think of a reason to keep it around. loras for kontext are slim compared to qwen too.
>>106647325Better non-destructive smaller changes, when it works
>>106647325I keep it for very niche use cases>flux has better controlnets than Qwen right now>processing hundreds of images quickly with nunchaku + lightning lora>generating chibi manlet versions of characters without having to prompt it
>>106647305They are Canadians anyway.
>>106647347>>106647344i see. wasn't aware kontext had controlnets (nor do i know what i would use them for i suppose)
>>106647356Asians that live in Canada are rich, no reason to live in slum like San Francisco
>https://github.com/FizzleDorf/AniStudio/releasesNew releases, ya'll.
>>106647436The only release anyone cares about is the police release statement of your ACKing, sis
>>106647436>4 days ago Not new, tourist.
>>106647436I thought Comfy was insufferable but that dev takes the cake
>>106647436congrats
>>106647305they will send comfy to bangladesh by mistake. it will be so funny
>>106647533holy shit this is so hot
>>106647305comfy is an immigrant?
is there a good sdxl checkpoint to go for if you're looking to gen stuff that's NOT 1girl nsfw?
>>106647570he's from quebecistan
>>106647294>modelhe does it manually
>>106647347how do you generate chibi manlet version of character?
>>106647593Explains why he is so insufferable
>>106647628lurk more
>>106647582"NOT 1girl" is a very broad category. Care to elaborate?
Anyone done grid comparisons between identical loras trained on differrent resolutions?
>>106647627By using kontext. Will do that naturally.
>motion is significantly better in wan 480p than 720pwut duh, so the loras DEFINITELY need to be trained on the specific checkpoint to actually use them? dammit.>480p did my prompt perfectly but gave her two nipples on each tit>720p has near nonexistant physics and kinda stiff motion
>>106647436Pedo
>>106647689is the prompt just "generate chibi manlet verion of character"?
>>106647680I'm just looking for something to play around with for a wide range of stuff and all these 1girl checkpoints feel a bit limitedI just need something different
>>106647759SDXL is a shit model if you want to stray beyond 1girl.
>>106647724No, just have outpaint or even change poses. Chances are they will end up a manlet. It's a fault of kontext not really a feature lol
Just use Seedream. Most versatile model, no censorship, and insane base res
>>106647582unless a checkpoint is literal ass it should be able to do non 1girl fine
>>106647796>no censorshippost futa oneesans pegging a shota
>>106647796Diffusing... locally... ComfyUI... API... nodes...
>>106647796Post vagina or gtfo
>last new thing didn't pan out>anon back to replying to bottom of the barrell b8Next new thing when
>>106647567yes. either go and save endangered regions of asia from depopulation or make more with wan
>>106647796>no censorshipcome on bruh, no need to lie like that, the model is a great SFW model, that's it
>coomberboomers seething over seedreamChina won, the west lost.
>>106647852what are you using for img gen?
kek it masked both after the cut
>>106647863Those shoulders nigga lmao
apparently seedream can't actually do 4k and it just uses esrgan upscaling. they just hide that behind the api
>>106647875duh
>>106647875Anyone could have told you that with how blurry the damn model is
>>106647863Wtf I was literally just in the tdkr /wsg/ thread looking for a vid because I was about to do something like this
apparently seedream can do native 4k in under 40 seconds, openkeks can't figure out how because they cleverly hide it within comfyui's API
>>106647863>troonkuWhat would happen if I took the rope around your neck off?
apparently my cock is in your ass right now
>>106647928it would be extremely painfulalso i'm getting the hang of this but this shit is gonna be absolute gold for memes, like VACE on steroids.
>>106647935Oh that's what that was...barely felt it
>>106647947also, at the same time you could do a qwen edit swap, or shoop, then i2v with wan. but this has openpose and so on so it can be a direct swap.
>>106647947You're a big guy.
>>106647952yeah because your hole is so loose
>>106647201I so want a Mayli LoRa, I would generate her for years if I got one
the anime girl is talking and holding a silver briefcase.gonna try with a full body miku. but, it works.
>>106647962FOR YOU
>>106647967just take any photo and use qwen edit + clothes remover lora. viable alternative if no lora.
>>106647615Would be cool if it were true
>>106647973yep, full body source is the way.
>>106647211>vramletniggers>>106637352>mental cases>>106643240
>>106647852please don't stop
>>106648000now we're talking
>>106647989>just take any photo and use qwen edit + clothes remover lora. viable alternative if no lora.I'll have to try that. She's so gorgeous. I'm glad she did porn once I just wish it wasn't facialabuse. Then again, maybe abuse and humilation is something she deserves.
Anyone else was able to use qwen image nunchaku?I wanted to test it but it just crashes comfy silently for me.It's just the standard qwen image wf with nunchaku model loader.
>>106648063but what if my penis can't handle it?
hi new learner here, i want to add some loras to a premade image to video workflow that i found following a guide. the tldr is i want to use the breast bounce lora. how do i fit another lora into a premade workflow? is it as simple as adding another lora and chaining it or is there something else I have to do?
>>106648094Yes chain them in series or just use a multiple lora loader node which does the same thing.
>>106648081only one way to find out
>>106648105thank you. does it matter which order the loras are in, or is that more of a trial and error thing?
the anime girl is wearing a black mask like bane from batman the dark knight rises, the background is white. remove the text and other elements aside from the anime girl.qwen edit made a neat persona miku
>>106648114Order doesn't matter, only weight does.
>>106648128No, only height matters.#ItsOver2025 #ItNeverBegan2025
>>106647861wan 2.2 t2v, no image reference. adult women with glossy clothes in a disco. the "fuck me" text also is promptedrecommending trying stuff that wan can do physics with 'cause it's fucking cool
>>106648155post workflow? how are you going above 5 seconds without issues?
the problem with animate wan is it's a lot of work and you could just use wan 2.2 after an edit with qwen edit/kontext then prompt.
>>106648201interpolate and then run it a little slower
change the position of the anime girl to a side profile.
>>106648210aw well thats not interesting thats just wasting my time with no new information. thanks anyways
>>106647244so native is for professionals and kijai is for vramlets, simple as
>>106648255what a weird response, would you rather just get no answer?
>>106648261are you mentally ill? why did you interpret my response as hostile? i was just upset it wasn't any actual new frames since i don't need longer gens, i need more information because only 5 seconds of mommy worship is usually not enough
>>106648155>lightsticks not in the training data, yetwhich group would you want to see first?
>>106648298>why did you interpret my response as hostile?>thats just wasting my time with no new informationgee I wonder
https://files.catbox.moe/fhdlss.mp4https://files.catbox.moe/k1hor8.mp4https://files.catbox.moe/d2dbuy.mp4https://files.catbox.moe/6bc02e.mp4https://files.catbox.moe/3a1ktg.mp4https://files.catbox.moe/aq276p.mp4https://files.catbox.moe/m3gn41.mp4https://files.catbox.moe/u9tffg.mp4https://civitai.com/models/1818841/wan-22-workflow-t2v-i2v-t2i-kijai-wrapperhttps://civitai.com/models/1952945/wan-22-massage-tits-by-mq-labhttps://civitai.com/models/1874153/oral-insertion-wan-22https://civitai.com/models/1923528/sex-fov-slider-wan-22
>>106648078OK I found out why, the wheel downloader node was downloading the wrong torch version for some reason (2.9 instead of 2.10).
>>106648310ok now it's time to be actually rude, because you're a fucking retard with no reading comprehension did i say YOU'RE wasting my time? no fucking retarded monkey, I said that interpolating slower wastes my time because it takes longer to watch the video and there's no actual new information. holy shit kill yourself you fucking mongoloid
>>106648317>https://civitai.com/models/1952945/wan-22-massage-tits-by-mq-labcrazy this wasn't blasted by some retard yet
>>106648336anon, absolutely no one would read it as thatyou're austistic
>>106648351you have proven that you're retarded so why would anyone listen to your opinion on "absolutely no one"? kill yourself subhuman. even brown people know to apologize and shut the fuck up when they're wrong, so you must be a mutt with 1-3% nigger DNA floating around inside of you
>>106648240change the position of the anime girl to a rear profile.
/lmg/ anon here. Haven't used imagegen since flux came out and was wondering what the meta is today. Looking to mimic the AI figure stuff /v/ has got going on using gemini. Anyone have recommendations for frontends and models?
>>106648375you're clearly a very smart and level headed person, anon, I bow down to you
>>106646564So adding on to discussion from the previous thread.I downloaded Q8 and Int4 nunchaku versions of Krea and experimented a bit.Some images turn out fine. Noticeable changes but coherent without major defects. Some though indeed get noticeably degraded.However, to be fair to it, it's running 4 to 6 times faster than the Q8. Can you not make the argument that it is better to gen more images in the same time frame and pick the best seed? The quality of this is much higher than an LCM distill while providing comparable speed up.Or maybe I am indeed a coping VRAMlet I dunno.They seem to have only released 128 rank versions for Qwen, which I also want to test but it would to take a while to experiment on my system.
>>106648383noob/illustrious for anime (wai v15 is good)wan 2.2 for videoflux/qwen for realismyou could use a lora for anime figures to get something like that.
>>106648387I'm testing right now with the prompt from the example comfy wf.bf16 vs q8 vs svdqfp4.
>>106648386dude actually kill yourself, i literally said "that's not interesting" do you think I was referring to YOU being not interesting you egotistical fucking retard? of course not, i was referring to the method. and then the second "that's" obviously was referring to the same subject as the first "that's" because that's how it works in the english language you retarded fucking ESL. holy shit kill yourself I refuse to let you accept that you are in the right here
>>106648383Flux Kontext and Qwen Image Edit if you want to give a reference picture of a character and turn it into a figure. (No idea how well they work for this though)If you don't intend to use a reference image and just prompt the character of your choice with a figure/toy lora. Someone out there is bound to have trained one for a model.
>(wai v15 is good)rope
>>106648417For Qwen (128)? Oh yeah post the results later then.
>>106648317Impressive.
>>106648419of course anon, you're right as always and you're generous in your interactions with othersagain I bow down to you, this time a little further down
>>106648433Yeah qwen image, with the following prompt : "A vibrant, warm neon-lit street scene in Hong Kong at the afternoon, with a mix of colorful Chinese and English signs glowing brightly. The atmosphere is lively, cinematic, and rain-washed with reflections on the pavement. The colors are vivid, full of pink, blue, red, and green hues. Crowded buildings with overlapping neon signs. 1980s Hong Kong style. Signs include:"龍鳳冰室" "金華燒臘" "HAPPY HAIR" "鴻運茶餐廳" "EASY BAR" "永發魚蛋粉" "添記粥麵" "SUNSHINE MOTEL" "美都餐室" "富記糖水" "太平館" "雅芳髮型屋" "STAR KTV" "銀河娛樂城" "百樂門舞廳" "BUBBLE CAFE" "萬豪麻雀館" "CITY LIGHTS BAR" "瑞祥香燭莊" "文記文具" "GOLDEN JADE HOTEL" "LOVELY BEAUTY" "合興百貨" "興旺電器" And the background is warm yellow street and with all stores' lights on.
>>106648404for example I used this:https://civitai.com/models/656994prompt: masterpiece, best quality, amazing quality, hatsune miku, waving, <lora:Figure:1> Figure, Figma, pedestalif the eyes aren't perfect just do an adetailer pass.
>>106648317that is a really long penis
>>106648387it doesn't mean a thing if there's not a fp16 comparison too, we don't know which one is closer to the real deal
>>106648454whats with the incredibly cursed starting image
wtf is wrong with miku normalniggers?
>>106648459Aren't Q8s very close to fp16 overwhelming majority of the time? I don't recall seeing any counter examples to this ever.But I can make a quick fp16 vs int4 test with the SDXL version they have recently released if you insist.
>>106648305lightsticks just seemed appropriate to try>groupno preference. ideally checkpoints would have a range of clothing like freckledvixon/ruanyi/[...] loras.
>>106648452that's the average yellow fever enjoyer's penis
>>106648404Thanks! Will try out wai first.>>106648422I assumed it would be as simple as image to image using forge or something. I used to have an older lora for SDXL that did figures pretty well but its been a long time so I'm trying to get back into it.
>>106648094>bounceVthe fuck is this?
>>106648305>>106648490hey ESLs, the term is "glowstick" not lightstick lol. have you never played Terraria?
>>106648497>image to imageThat won't work, it would also destroy the character, you can't simply transfer style with just de-noising. The model needs to also know the character and the concept of figures. (Either by itself or by using loras)
>>106648502Seed Bouncer Vendradium or SBV. it bounces your seed through a Vendradium Entropic Variation algorithm. Essentially, it prevents predictable image duplication.
>>106648551Thanks. I assume something like Flux Kontext and Qwen Image Edit work just fine in forge, right?
make the image in an avant garde style. keep the same pose.any art anons? I just tried a random style I vaguely remember (qwen edit)
>>106648586https://en.wikipedia.org/wiki/Black-figure_pottery
>>106648572I don't know.I abandoned Forge months ago due to lack of support for major models (video diffusion).Flux Kontext possibly does because base Flux works on it. But no idea about Qwen.The auto111/forge ecosystem kinda died while you were away.
>>106648636Well shit. Are we all using comfy now?
>>106648636>>106648640neo forge is looking promising because it actually has an active maintainer but the memory management that illya added years ago is outdated for newer models and is just causing worse memory issues than comfy has.
>>106648640Well I moved on to the spaghetti yes.There has been some drama around it recently, maybe eventually the community moves on to something else but for now it is the most prominent all rounder.
>>106648586same prompt
>>106648692i genuinely lmao'd. might be a decent way to get them to stop stealing shit from here
can you solve the puzzle?
>>106648450My tests show something weird : it seems like svdquant fp4 r128 (the one for blackwell), is non deterministic.I get different outputs when I retry using the same parameters.wtf.I'll retry with int4 instead, too bad because fp4 is twice as fast as q4 on a 5090.
https://www.liblib.art/modelinfo/99d2d7a0bf0e41bd9275bdbc9a84995d?from=feed&versionUuid=5a5b4e055ed4485db884d26a440eb018&rankExpId=RVIyX0wyI0VHMTEjRTM3X0wzI0VHMjUjRTM4china wins again
API nodes bros, another source to use to use See Dream 4 for free? I'm not feeling very powerful...
>>106648786>completely impossible to download without an accountyay
>>106648757Come on, this is ridiculous.Thankfully I have a 3090, but it's so dumb.
What do you actually want to run for generating images with 24GB of VRAM? Is there a workflow that makes use of that or is a faster GPU only better for speed? The rentry post doesn't seem very thorough, so I've stuck to asspull loras and checkpoints and random workflows from civitai.
>>106648786does it do actual nudity?
>>106648807Any workflow is fine. More vram just allows you to offload less and run unmolested model quants/full sizes. Also faster lora training
>remove the yellow censorship bar and restore the nipples on the anime characters breasts.like magic. some minor touching up needed but it was a big yellow bar, before. great example of what the edit model can do (plus a lora).https://files.catbox.moe/lcdu3y.png>>106648829sure can!
>>106648486Well I can't do it today because they haven't released nodes for SDXL lol.>>106648757I doubt this. Are you certain you are using a converging sampler?
>>106648155oh dang I haven't tried t2v but maybe I should >>106648530I think they were going for kpop lightsticks which are these things the fans in the audience sometimes have
>>106648851and the original image (ad from here)
>>106648851no, do it with real people, full body
>>106648786can you share the lora?
>>106648869yes, that works. sample from a gravure shot or w/e:https://files.catbox.moe/78jqz9.pngbasically removes the sfw filter from the model, more or less.>>106648871sec, upload is slow
>>106648861>I doubt this. Are you certain you are using a converging sampler?No, I didn't use an ancestral sampler or anything of the sort.Try it if you have a blackwell card, I catboxed the result :https://files.catbox.moe/pbg4t8.pngGen one time, then close comfy, open then gen again.
>>106648885what about genitals?
>>106648900yeah that works too, havent tested it a lot but it can do that apparently.
the cartoon frog is sitting at a computer wearing a blue shirt, and red shorts, and sandals. keep their expression the same. a white CRT monitor is on the computer desk.
huggingface/civitai aren't allowed to host it (where I originally got it), because...it can lewd.so here's a mirror of the same lora: https://limewire.com/d/sBUPu#GclImNhwoG
>>106648925same reason you have to change a line of code on reactor/face swap stuff to remove nsfw restrictions, rules are stupid>you can download this photo software to lewd>but this extension is bad!
>>106648925>limewireblast from the past
>>106648925Thanks anon.
>>106648966I googled fast file upload and got https://www.file.io/ as the first linkoddly enough, it's a part of limewire.
anybody manage to go past 5 seconds with animate at 1280x720? if i try to go longer than 5s shit just ooms.https://github.com/comfyanonymous/ComfyUI/issues/9937about to say fuck it and try this.
god DAMN wan is good.once we get a model that does this shit in 10 seconds instead of 40+... i don't know what will happen to me
>>106648692Catbox PLEASE
>>106648897OK found out why I think : https://nunchaku.tech/docs/nunchaku/faq/usage.html#why-do-the-same-seeds-produce-slightly-different-images-with-nunchaku>This behavior is due to minor precision noise introduced by the GPU’s accumulation order. Because modern GPUs execute operations out of order for better performance, small variations in output can occur, even with the same seed. Enforcing strict accumulation order would reduce this variability but significantly hurt performance, so we do not plan to change this behavior.The difference in picrel is less than what I see, but I think it's because I generate bigger images (2048x2048) with a more complex prompt, so the variations are bigger.
>>106648998nah fuck thatwe need a model that can actually do 10 second videos instead of 5.
>>106648897>>106649046I see.Would keep this in mind.
>>106648990damn that workflows a mess now that i look at it
>>106649143yeah the comfy wiki guy barely understands the program so im not surprised
>>106648433>>106648450Here it is : https://imgsli.com/NDE2NzEzI used : euler/simple/40steps 2048x2048Gen times : BF16: 260sQ8: 230sQ4KM: 216sSVDQUANT FP4: 102s!Keep in mind the size and complex prompt is probably a worst case scenario.Quality wise, Q8 is fine, nunchaku is slightly worse but honestly too different to compare, and Q4 is really bad.Speed wise everything was more or less the same with the 5090, but man fp4 is fast, insane speed compared to the rest.My conclusion: I don't really need svdquant for imagegen for qwen (or flux) on a 5090, but I'd trade the weird non deterministic behaviour of svdquant to have faster speed, and gen way more things using wan. (Image gen using nunchaku is probably more worth it for people on 40xx and 30xx cards.)The best use case being wan is very annoying, because they keep making everything except it.
>>106649034>MP4Reminded me of this.
>>106649159Thanks anon.It seems to just double speed for Qwen.I will probably keep not bothering with it on my budget setup, but it's good to know.I would probably use bf16 if I were you with a 5090.
>>106649153yeah kijais is fucked for me at 720p can't go past 5s even with everything set to offload.https://github.com/kijai/ComfyUI-WanVideoWrapper/issues/1267glad i'm not the only one having issues tho.
>>106649277>24fpsfeels like 8
various ice cream containers in an ice cream shop. (qwen image)looks pretty good desu
>>106649306Can you use an add grain node so it doesn't look so plasticky?
>>106649210For fun I did a speed test on my other 3090, and the speed : svdq int4 : 434sQ4: 635s(I had to disable sage attention on the 3090 because it gave me black screens, so it didn't help.)>I would probably use bf16 if I were you with a 5090.Yeah, I only care for svdquant for their future (hopefully) wan model. And only if they also release support for loras.
>>106649306a man puts the vanilla ice cream into a bowl with an ice cream scoop. (wan 2.2)should have said cone, but it works
>>106648990>seedreanHow did you do it?It's free?
>>106649367the bloody bastard is mixing the flavors
>>106649367yeah that scooping animation made me salivate a bit
>>106649367the ice cream in the ice cream shop melts into a liquid.wan is such a neat model.
>>106647593
>>106649393they're almost a week old back when it was on lmarena. don't really have anything to post since /g/ doesn't allow vids with audio and the file size limit is ass anyways.
>>106647201I wanna do extracurricular studies with Hatsune Miku and her baloonbies…
anime girls made out of ice cream, come out of the ice cream bowls on the table.again, wan is pretty cool.
>>106648807this i would actually believe is a real picture if someone used it on a dating app or something
>>106648807model?
https://github.com/kijai/ComfyUI-WanVideoWrapper/issues/1262#issuecomment-3314926799curious to see what this is all going to look like once he's finished. it's not that bad already if you don't use the speed lora.
>>106649455Miku Hatsune grabs the vanilla ice cream bowl and walks out the door of the ice cream shop.need more time but grab success
>>106649548
>>106648240Which model is this?
used qwen edit to make the sign edit, wan to move it
Why do my chroma loras always end up being the same size no matter the dataset size or training settings?
>>106649567qwen edit (q8)
>>106649592It depends on your --network_dim and --network_alpha settings
>>106649587
>>106649592shut up ran
>>106649630the possibilities are endless for the burger king.
>>106649628I don't see it anywhere in Onetrainer. Would giving it more space produce a higher quality result or is that not how it works?
>>1066484524 her
>>106648383>sabamenExtremely based
the anime figure of miku hatsune rotates 360 degrees.
>>106649630>>106649633what was his BK order though? tendies? :3
>>106649703GOT DAYUM
you know qwen edit is a pretty capable model not just for slight changes, it can take an image and do an entire scene.
>>106649662It's in the lora settings tabThe higher the rank, more of the underlying model will be affected by the lora, but contrary to what you may think this does not automatically mean better quality, the best results are had by scaling rank according to the number of images and their resolution.For example, lets say you have 30 images of a person and train a lora on those, going over rank 16 won't get you a better result, most likely a worse one, you could probably even go down to rank 8 and have as good a result.If you have 500 images or you are training several people / concepts at once, you should look to increase rank to 32-64.Alpha is this rather stupid extra parameter which affects the strenght of the LR, it just adds pointless complexity, my suggestion is keeping it the same as your rank, so rank 16, alpha 16, which essentially nullifies its effect.Also when training a Chroma lora, I would suggest using the 'blocks' preset in the lora tab, for me it gives the best results.
>>106649570Catbox pls.>>106649662Higher dim can also turn it into overfitted turd.Optimal Dim number depends on what the lora is about and training dataset quality.>>106649704Generally yeah but you can experiment with lower values for some loras.
>>106649724>my suggestion is keeping it the same as your rank, so rank 16, alpha 16, which essentially nullifies its effectIn OT the default value is 1.0 which I guess is a ratio to the rank?
>>106649767NTA and I think you should set it as half or lower or experiment different values yourself but no 1 is the minimum, the exact opposite of setting it as the same...
>>106649499https://civitai.com/models/1950841/intorealism-ultra +https://civitai.com/models/573152?modelVersionId=2155386 setup as a refinerI have no idea what I'm doing tho, just working off of someone else's existing node setup switching things around
>>106649710>>106649703>"she wiggles her bunny tail at the camera, she remains facing away, she is smirking, playfully wiggling tail back and forth"
>>106649767Yes, you can leave it at that or use the same as your rank to nullify its effect, as in rank 16 / alpha 16, rank 32 / alpha 32.Overall it doesn't really matter as long as you continue using whatever you decide, since if you keep switching this option you will have a harder time figuring out which LR to use, since it affects LR.Kohya and Diffusion-Pipe uses the same rank alpha by default, OneTrainer uses alpha 1.0 by default, I don't know the defaults of other trainers.I use OneTrainer, but I use the same rank / alpha because I want to remove that extra variable from my training.
>>106649721the anime girl is on an album cover playing a teal colored rock guitar.
>>106649455they look so confused lol
Queuing up a batch of low-CFG gens and letting the sampler just go completely hog wild is a real trip.Mostly you end up with a bunch of insane tentacle hands and Megaman 1 box art tier proportions and faces but every now and again it drops a surprisingly decent and creative gen.
i've come to the realization that my prompt doesn't do anythingit keeps generating the same thingcan you tell me what's going wrong from this screenshot, or do you need more info?
>>106649858authentic indian video
>>106649859fixed seed on second sampler?
>>106647201y does she laugh like tht
>>106649886Only the first stage generates the starting noise, the second stage denoises the output of the first.
Can any of these new Chinese image models generate seamless repeating/tiling/texture images? It's crazy to me that last I checked, A1111 + SD1.5 was still the gold standard.
>>106649891im not sure why a pepe appeared on the screen, I just said "the text LDG appears on the CRT monitor"
>>106649859share your workflow
>>106649893baker is spiteful\insane
So how do I get rid of the retarded horizontal scanlines in every Chroma gen I make?
>>106649914teh baker made tht? woaw
I need someone with the specs to make lewd videos of Anna P on the beach.
the green cartoon frog walks to a nearby water cooler and fills his cup with water.uhh, not quite...
>>106649954better:
Not what I prompted at all but I like it.
>>106649973when will you stop pulling the trigger early on the shit gens and just post the better one?
>>106649920These are typically the result of:Training on bad images, as in noisy with artifacts, such as linesUsing low quant models (same as with Flux which caused a grid pattern)Too long promptsUsing a lot of different loras at the same time (3 or more)Also depends on the base model you are using, Chroma1-HD seem to be largely immune
>>106649954The funny kind of AI slop nonsense.
cute!
>>106650035stop yapping
>>106649978Catbox?
>>106650038even with talking\singing in the neg-field she still does it often ;_;
>>106649991Looking at it again it might actually be that my monitor has something burnt into lmao. Thanks for the help though.
>>106650035>silent>waifu appears smiling>happy conversationNah wan got it right that time
>>106648419Anon with respect, this is classic autism. You had no insight into how your message would read to someone other than you (lack of theory of mind) and are now sperging out because it was interpreted other than how you intended it. They are entirely correct that it is pure autism on your part.Getting mad about it after being annoyed that THEY got a bit mad about it is just double down 'tism.You can be as mad about THIS as you like, it's still true.
>>106649982I liked the first one better, I had never seen a monitor do that, the second was boring, prompt adherence is overrated.
>>106648419least hostile response
>>106650047I am not Brazilian or Portuguese btw, just copied the prompt from Sora.https://litter.catbox.moe/t5cdeilmgwt4jeb4.png
>>106649913it's the "Wan2.2 14B I2V Image-to-Video Workflow Example" from here https://docs.comfy.org/tutorials/video/wan/wan2_2#wan2-2-14b-i2v-image-to-video-workflow-example
>>106650119And no translating didn't help, how proficient t5 is at Portuguese doesn't seem to be the problem.Still amusing image though.
>>106650068SEXO!
I wonder how many languages qwen and wan understand outside of English, Chinese.
>>106648671kek
>>106648671>>106650182:c
>>106650175>wan>UMT5 is pretrained on the an updated version of mC4 corpus, covering 107 languages: Afrikaans, Albanian, Amharic, Arabic, Armenian, Azerbaijani, Basque, Belarusian, Bengali, Bulgarian, Burmese, Catalan, Cebuano, Chichewa, Chinese, Corsican, Czech, Danish, Dutch, English, Esperanto, Estonian, Filipino, Finnish, French, Galician, Georgian, German, Greek, Gujarati, Haitian Creole, Hausa, Hawaiian, Hebrew, Hindi, Hmong, Hungarian, Icelandic, Igbo, Indonesian, Irish, Italian, Japanese, Javanese, Kannada, Kazakh, Khmer, Korean, Kurdish, Kyrgyz, Lao, Latin, Latvian, Lithuanian, Luxembourgish, Macedonian, Malagasy, Malay, Malayalam, Maltese, Maori, Marathi, Mongolian, Nepali, Norwegian, Pashto, Persian, Polish, Portuguese, Punjabi, Romanian, Russian, Samoan, Scottish Gaelic, Serbian, Shona, Sindhi, Sinhala, Slovak, Slovenian, Somali, Sotho, Spanish, Sundanese, Swahili, Swedish, Tajik, Tamil, Telugu, Thai, Turkish, Ukrainian, Urdu, Uzbek, Vietnamese, Welsh, West Frisian, Xhosa, Yiddish, Yoruba, Zulu.The quality might vary but it should at least somewhat follow along what you are prompting in these languages.>qwen>The dataset is primarily composed of Chinese and English data, with supplementary multilingual entries to support broader linguistic diversity.I am guessing it only excels at English and Chinese but curious how it would fare against a German or French or any other major language prompt.I am too much of a VRAMlet but maybe someone else can test.
>>106650239Interesting, thanks anon.
>>106649982There should be a rule that you can't post gens that are less than six hours old. Well, not a rule since it would be unenforceable, more of a gentlemen's understanding. It would cut down on the number of people who post a dozen minor variations on the same prompt.Delayed gratification, people. It's what separates men from animals.
>generated wan video (nsfw) with fp8 scaled 5/7 steps 97 frames 800x1200->618s>generated wan video (nsfw) with q8 5/7 steps 97 frames 800x1200->852sI don't know for generations prior to 5000, but for the 5000 fp8 (e4m3fn scaled) is way faster than q8. I also didn't see any visible difference in quality, so it's not worth it for me, especially with such a speed penalty.
>>106650119Danke
>>106650333There is definitely a quality degradation with respect to Q8 and fp16 if you test it enough.But yes, on Blackwell fp8 is noticeably faster due to hardware acceleration (which previous generations lack).One can argue whether it is worth it or not.
>>106650359I like the interface effects
>>106650333q8 technically is more quality but its slower than fp8 for newer generation cards.
>>106650333>I also didn't see any visible difference in qualitygen it with the same settings and seed
>>106650371>>106650392Yeah 30% faster is worth it for me if I don't see any horrible degradation.Looking at it again the only thing I noticed was slightly different lighting, but nothing even bad, just a tad bit different.>>106650403That's what I did of course.
>>106650433>That's what I did of course.i guess you didnt gen anything that needed anatomic consistency or movement clarity
>>106650448It was a nude woman walking forward, and it was fine, she didn't get another limb, her face wasn't weird, and her movements were ok.I would expect bad anatomy from lower quants, but not from what I tested.
>>106650460try anything with hand movement
Where are the AI NSFW videos?
>>106650464OK will try later, once my big queue has cleared.
>>106649858nice cope anon, but that's the GNU/Linux car! We don't use macbook cars when we have that!
>>106650486Have you looked at /gif/?
>>106649548nice
>>106647325kontext is better at detecting and editing text
>>106648201>girl dancingyou can easily just increase the length for that
man, can we get real time video gen mixture of experts models already?
>>106650633Accurate
the anime girl throws the green leek vegetables in her hands towards the camera.qwen edit to make the leeks, wan to animate
Being a face guy is suffering. It is so hard to get facial variety and even harder to actually control facial features. I just don't want everybody to look the same. Qwen can't even handle different body types, nevermind facial features.
neat, actually workedthe anime girl transforms the green leek vegetables in her hands into a green rock guitar.
>>106650840imo qwen edit is better for transforming people despite being an edit model primarily, take a base figure or image then prompt them with diff traits.
>>106650840Chroma has quite a few different faces. I suppose you could want all facial features tagged and that isn't the case, but the diversity is there.
>>106650840thats why you dont use the model with the least seed variety that qwen is or some giga slopped model like base fluxuse chroma hd/2k for realism and noob/illustrious for tranime
>>106649703cute
that was easy bait for the chroma shills
>>106650916
>>106650840That's the price you pay for almost always perfect anatomy, they get this by massively overtraining on a small pool of faces and human posesThere's really no workaround at the moment, you quickly notice when you train loras on models like qwen where you use other people and poses, then the almost always perfect anatomy starts degrading
>>106650914>npc gets triggered as he hears his activation phrase amongst multiple recommendationstextbook rent free
too easy
>>106650869>>106650871Chroma is way better in this regard, but it is hard to go back when it can't do fine details and Qwen can. These patterns aren't perfect, but they are way better than what Chroma can do. Chroma is a very viable choice and I won't shit-talk it, but there's a lot I like about Qwen.>>106650856I'll have to give this a shot.
>>106650840Another face guy here. Use chroma. It randomly creates unique features with very little to no prompting. I was in the exact same position a few months back where anon suggested chroma because I wanted to make wonky unique features. Haven't used any other models since.
The token bleed in chroma is kind of fucking insane, it's pretty fucking bad and it can even force you to have to make loras stronger
>>106651004beautiful
>>106650885Is it seedream girl?
>>106651024Train a fuckhuge 'face types' Lora to serve as basis vectors for face guys like him: >>106650840
>>106648990You are meant to chain the wananimatetovideo nodes using the frame offset connection. tie a ksampler to each one, decode the latents and feed into the continue motion connection of the next ksampler. and also batch the images of each before sending to combine video. The idea here is that you can use a different prompt for each chunk and probably chnage character mid way through the video and use different shift value or lora's.>>106649241KJ nodes oom? no surprise there...
>>106649241That's a real photo
>>106651083Flux Krea girl actually.>>106651024Examples?
By the way if you want to detect AI you can use tellif.aiIts free and it works really well
>>106652478wow it's not plastic, it's AI!
>>106652478Seems to work, but why only 78%?
>>106652478kek
>>106652478>80%. Likely created by human. Natural imperfections detected.
>>106652679 idk, for me generated
>>106652738Guess it's just retarded then
>>106652679Dalle-2 had soul
>>106651485>Flux Krea girl actually.Total local victory
>>106651350I wrote this and read it and i see that i am very shite at explaining things because autism and it really does irritate me that people want everyone on a silver plate. They complain there is no basic native workflow, well there is but its shite because it does not fully use the wananimate model to its fullest. I don't like writing walls of text and do usually share workflows but I can answer specific questions about how to use the node properly. for starters (and important) you need to resize and pad the reference image to the exact size of the control video frames (the original video) and it can be any person in any background. if you do get this right the control net poses won't work properly and you will end up with body horrors and other glitches as the model tries but fails to align properly. You DON'T resize the frames of the control video, you just pass them directly into the dwpose node and then the controlnet images go into the control video connection on the wanAnimateVideo node, then connect the resized reference image to the reference image connection on wanAnimateVideo and it must be the same resolution as the original video or it won't work! I use ResizeAndPadImage node to do this and link from video info node the source video width and height and I use interpolation lanczos for best quality. pic related to this post >>106651350shows roughly the layout of the nodes, i will probably share the workflow later once I tidy everything up. but the image should be good enough for reference to how to set it up in native, also the speed lora's i'm using. BTW this model is fucking insane when you get it working right, you only need to set it up like i've tried to explain and it just works, it takes care of everything. So fuck all those whiny ass bitches wanting everything on a silver plate, they as stupid as they are fucking lazy. FUCK EM!I wouldn't change much of a thing with the native nodes or the model it works very well as is.
>>106653158So i'll continue to help explain. I pad the reference image with white borders, though any image so long as its resized works, its background will heavily change the video so be aware of that, be creative etc. Control the frame rate on the video combine node using the source FPS so that the audio is in sync.Oh and the shift value I'm using is 1.00 and seems to give nice results. So i use math to calculate the total frames from the video load node. chunks of 77 * 3 samplers = 233 frames but i use math nodes to do all that. You could indeed extend enough ksamplers to remix an entire video. I use 5 continue motion frames and i drop those frames because they are always garbage which makes sense because when you drop them you don't get jump cuts and only a smooth video. I use math to drop the frames 77 - 5 = 72 frames and select them from batch at index 4 for the length of 72