Discussion of Free and Open Source Text-to-Image/Video ModelsPrev: >>106975747https://rentry.org/ldg-lazy-getting-started-guide>UIComfyUI: https://github.com/comfyanonymous/ComfyUISwarmUI: https://github.com/mcmonkeyprojects/SwarmUIre/Forge/Classic/Neo: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassicneoSD.Next: https://github.com/vladmandic/sdnextWan2GP: https://github.com/deepbeepmeep/Wan2GP>Checkpoints, LoRAs, Upscalers, & Workflowshttps://civitai.comhttps://civitaiarchive.com/https://openmodeldb.infohttps://openart.ai/workflows>Tuninghttps://github.com/spacepxl/demystifying-sd-finetuninghttps://github.com/Nerogar/OneTrainerhttps://github.com/kohya-ss/sd-scripts/tree/sd3https://github.com/derrian-distro/LoRA_Easy_Training_Scriptshttps://github.com/tdrussell/diffusion-pipe>WanXhttps://comfyanonymous.github.io/ComfyUI_examples/wan22/https://github.com/Wan-Video>Chromahttps://huggingface.co/lodestones/Chroma1-BaseTraining: https://rentry.org/mvu52t46>Neta Luminahttps://civitai.com/models/1790792?modelVersionId=2298660https://neta-lumina-style.tz03.xyz/https://huggingface.co/neta-art/Neta-Lumina>Illustrious1girl and Beyond: https://rentry.org/comfyui_guide_1girlTag Explorer: https://tagexplorer.github.io/>MiscLocal Model Meta: https://rentry.org/localmodelsmetaShare Metadata: https://catbox.moe | https://litterbox.catbox.moe/GPU Benchmarks: https://chimolog.co/bto-gpu-stable-diffusion-specs/Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-oneTxt2Img Plugin: https://github.com/Acly/krita-ai-diffusionArchive: https://rentry.org/sdg-linkBakery: https://rentry.org/ldgcollage>Neighbors>>>/aco/csdg>>>/b/degen>>>/b/realistic+parody>>>/gif/vdg>>>/d/ddg>>>/e/edg>>>/h/hdg>>>/trash/slop>>>/vt/vtai>>>/u/udg>Local Text>>>/g/lmg>Maintain Thread Qualityhttps://rentry.org/debo
>>106978583I hope it keeps motion with multiple frames otherwise I don't see the point
is it even worth putting anything in the negative prompt with chroma hd flash
Where do i get the new lightx2v?
anyone has nsfw vibevoice examples?
>>106978624https://huggingface.co/lightx2v/Wan2.2-Distill-Loras/tree/main
>>106978608don't you use a cfg of 1 with that?then no
>>106978608>>106978662you can use negative prompts with cfg 1 if you go for NAG, dunno if there's a Chroma NAG though
>>106978641thanks
>>106978629you can make it moan and do (very gacha) kisses, though nothing reliably. you basically need to give it porn as a reference
>>106978693ok thank you anon, I will check then
Blessed thread of frenship
>>106978722bye bye microsoft safety researcher
Which one you use?Also, does it matter?
>>106978652>>106978781Didn't see we are already into a new thread. I have some thoughts on how it could work using native nodes apart from some helper nodes. In any case comfyanon is probably going to add nodes for it within days.
>>106978775Do we even use clip vision? It was breaking my gens on the first day I got 2.2 and ive left it off
when will API keks be able to do this, again?https://files.catbox.moe/0bv1ex.mp4
Wansisters, we long video now (maybe), settings and findings>wan 2.1 i2v 14b Q6 gguf>10 seconds (165 frames for 16 frame_rate)>euler a, beta, 1 cfg>lightx2v_I2V_480p_cfg_step_distill_rank256 and svi-shot (https://huggingface.co/vita-video-gen/svi-model/tree/main/version-1.0) loras>no context nodes, no last frame, no tricks>prompt: portrait video of a cute asian woman waving at the camera then she pulls out her camera phone to mockingly take a photo>it did repeat the waving at the end, will try with longer prompt
>>106978843This cured my porn addiction.
>>106978589> SVI will iteratively generate a video clip for each prompt within the prompt stream, which uses the last frames of the previous generation as the conditions.> SVI-Film supports end-to-end long filming controlled with a storyline-based prompt stream. We use five motion frames and replace the padding frame with zero for the image latent.but i've though i2v wan can only work with 1 input imageneed to read i2v paper too
>>106979047>but i've though i2v wan can only work with 1 input imageThis is why I'm also scratching my head, and wouldn't it then only just be treated as a new video using lastframe? The only thing I can think of is either taking last 5 latents and combining somehow for the next sampler, or using wanAnimate nodes in a slightly unintended way. promising but as always shit presentation and almost zero instruction on how to actually get it working correctly within comfyui.
>>106978926workflow where
>>106979086the thing is wanimagetovideo node output latent has no initial frame, it's conditionings using it according to the code
>>106979151Just use any of your wan 2.1 workflows, download svi-shot from https://huggingface.co/vita-video-gen/svi-model/tree/main/version-1 into your loras folder, set your frames, that's literally it
>>106978926so you didn't use a new prompt, it's one prompt for the full gen
>>106979193What? I'm still in the middle of testing.
>>106979177pic relatedwill try
>>106979181Yes but it will 100% always miss the nipple with the mouth.
>>106979216>will tryNo you wont
>>106979177but it will oom after so many frames... they say it requires frames from previous? Anyway I'm currently building a test workflow using imagetovideo for first sampler then wanAnimate node and i will select from range the last 5 frames and feed into continue motion. But again its not wan doesn't treat each as a new video, it has no context.
From what I understand, this SVI lora thing doesn't make long videos possible magically, it just stops the last frame you reuse before the next video part from drifting too much from the original one, including brightness, colors etc.So your original "long video" workflows will work the same, with less errors from each extracted last frame, that's it.If it's just that it's great but doesn't solve motion issues from reusing last frame.Ah and apparently this is only for wan 2.1, they're working on 2.2... 5B. Why?"PS: Wan 2.2-5B-SVI is coming."
>>106979255lol so it's literally garbage
>>106979255And the lora to do that multiple stitched videos is : "SVI-Film": this has been trained for multiscene generation in mind, so it can be used for the last frame feeding a new video use case.
>>106979275No, it's pretty cool to not have color and brightness issues, it's a major pain in the ass.But it'll be unusable for most people unless they go back to wan2.1 or use a shitty 5B model.At least until the team makes a wan 2.2 14B version lora, which they didn't promise.I hope they don't do a nunchaku and just disappear or do random models after the 5B one.
>>106979229Depends on your hardware I guess? I can load up 245 frames on 16gb vram. I was having ram issues before with regular unet loader, you could try the multi gpu nodes>picrel>>106979255Yeah, it seems to be repeating the motion but its only slightly better than the context nodes.
>>106979306>I hope they don't do a nunchaku and just disappear or do random models after the 5B one.then it'll be another promising tech dead in the water because they only used it on irrelevant modelsmany such cases
>>106979317>>106979255oh well then just use the context window durr, it won't oom with that, and it should work already on wan2.2 if its a lora as most every wan2.1 lora just works with 2.2
>>106979340It's a lora especially finetuned to solve errors for wan2.1, so not sure it would work with 2.2 at all.I didn't check sliding windows yet, but as far as I know you cannot prompt every n frames with it, so it's kind of useless for long gens if you want to have multiple things happening over time.
>>106979255and where did they mention that? Or are you just talking out your ass? I don't see them providing a workflow...
So this is like the 6th attempt. While I'm impressed that it holds quality, consistency and movement, it didn't follow the end part of the prompt. I'm sure there's something we're missing but over all, I like it. I'm out, think there's another anon doing test somewhere too.>portrait video of a cute asian woman waving at the camera, she then pulls out her phone to mockingly take a photo, next the phone then flashes, she then puts away the phone and finally she walks out of shot at the end
>>106979367https://github.com/vita-epfl/Stable-Video-Infinity
>>106979357>but as far as I know you cannot prompt every n frames with it, so it's kind of useless for long gens if you want to have multiple things happening over time.but they literally claim the opposite, which why anons are left wondering if something needs to be changed for it to work as they intended/advertised.
>>106979384>5b
>>106979394I was talking about the context window node.
>>106979381how do you set that up? couldn't figure it out for the life of me.
is there a word token limit in noobai pos prompt? wasnt there some limit back in the day or am i remembering it wrong? seems like there isnt one now
>>106979384>5bkek
>>106979411See >>106979177
god chroma sucks so bad at doing anything POV beyond muh sex
>>106979322Based edgekinoposter, catbox?
>>106979427>god chroma sucks so bad at doing anythingyou could've stopped there lol
>>106979425I'm not convinced it would be that easy. do you understand what unlimited video with a 20 minute test with no drift means in terms of wan? You're not gonna be doing 20 minutes of video on your 16GB vram card. Unless it truly is just using the last frame as input for each 5 seconds of video, but that don't make sense when pic related. Why would they go out of their way to make so many versions?
>>106979456>>106979425like wan has no concept of a previous gen, so i don't understand what they fucking mean. fucking why u no explain properly god damn it.
>>106979478>>106979255
>>106979432https://files.catbox.moe/p44x6s.png>You are an assistant designed to generate anime images based on textual prompts. <Prompt Start>>@Ash Thorp, @Amu Aoi, @82 PIGEON, @Chris Bachalo, @Carpet Crawler Ccrawler Art, traditional media, painting \(medium\), canvas \(medium\), scan artifacts, magazine scan, scan, artbook, production art, novel illustration, horror \(theme\), non-web source, original, commission, dark, >abstract, abstract background, >outdoors, landscape, scenery, castle, battlement, palace, fortress, sky, cloud, >You are an assistant designed to generate anime images based on textual prompts. <Prompt Start> >@Avogado6, @Katsuya Terada, ai-generated, stable diffusion, midjourney, ai-assisted, 3d, render, blender \(render\), sepia, obese, fat, overweight, digital media, toon, cartoonized, western comics, >bad quality, worst quality, worst detail, sketch, censor, transparent background, 1girl, solo, solo focus, english text, magazine cover, photograph \(medium\), 1other,
svi does not work in comfy natively plug and play retards, it samples 5 frames in sections, I2V just gives it 1 frame, someone will have to implement it for comfy
>>106979456kek I never said anything about being able to do unlimited/20 minutes, either way, I dont care, feel free to try it out your self
here is SVI
>>106979381nothing for 2.1 t2v?
>>106979525incredible
>>106979525Wow, it's shit.
>>106979512>SVI will iteratively generate a video clip for each prompt within the prompt stream, which uses the last frames of the previous generation as the conditions.image + video 1 + svi lora -> last image(last image = first image) + video 2 + svi lora -> last image 2etc
>>106979525>time to rape
>>106979525i genned a few with his film one to try it out but it's p wonky. not enough time to try without the self forcing lora.
>>106979571>>106979552this is 2 + 2 steps with light, cba to wait for testing
>>106979581oh, yea, it somewhat works with 2.2 btw, same with other 2.1 loras, use at a higher weight on high noise, dont use on low noise
why does chroma so easily drift between normal realism and SD 1.5 2.5D hyperslop "realism"
>>106979398
>>106979585so about 4 - 5 strength in high then? Thanks for testing it on 2.2, did you compare outputs with and with out the lora? It really needs to be tested with longer and more complex videos.
>>106979585tried 1.0 in high, got endless loop of same motion ignoring prompt. will try higher now
>>106979604because its fucking shit, stop talking about it. qwen is arguable better if you know what you're doing. only probably is moral fags reporting nsfw loras on civitai, or just a lack of them. Chroma is the same but only because people lost interest.
>>106979525this is so fucking bad lmao
>>106979628>>106979581And have you tried stringing together gens before? this is night and day better, and this is not implemented right
>>106979623>qwen is arguable better if you know what you're doingseed variety is utter horseshit
>>106979636this, qwen needs loras or every image looks exactly the same, and then you gota train a fucking lora for any little thing, so qwen is useless cept for super specific shit
>>106979529where did it go
>aitoolkit FINALLY got offloading for wan 2.2about time
>>106979636>seed variety is utter horseshitProbably because qwen has much stronger prompt adherence? As in it gives you what you prompt no matter the seed?
>>106979664qwen is overcooked and so is super locked across seeds with no variety
>>106979664Nope. You can give it a minimal prompt, and despite there being many ways to depict it, it will rigidly do the same exact thing.
>>106979664I think that's because of the flow architecture, flux also has that issue
>>106979664alibaba's models are just overtrained with synthetic shit, this company is overrated as fuck, the llm community is making fun of them because they're too obvious they're gaming the benchmarks
>>106979672flux was also a bit overcooked, that was a large part of what chroma fixed, it destroyed the lack of variety, of course it also destroyed the aesthetic training but I would rather have the more flexible model
>>106979676>>106979669fair enough
>>106979671not 100% on this but this feels like it may be a consequence of doing aesthetic reinforcement with a smaller, limited dataset, and overcooking it that way
>>106979652https://desuarchive.org/g/thread/106978567/#106979529you should install https://github.com/TuxedoTako/4chan-xt
>>106979678>it also destroyed the aesthetic trainingif only it only destroyed that, but unfortunately it also destroyed good anatomy and good details
>>106979676The benchmarks are worthless and everyone is gaming them anyway, but the chinese companies are the main ones actually releasing new top models for end users and especially trying new things
>>106979691blame BFL for intentionally doing super destructive distillation with the intent of making it very hard to finetune
>>106979691its really not that bad there considering the crazy nsfw shit it can do, and again, unlike flux / qwen it was not aesthetic trained to hell and back, even illustrious trained the fuck out of their model on that
>>106979690i have x, he deleted the image too quickly
the main thing is people need to stop using shitty ass T5, and use gemma instead. T5 is the real thing holding models back, its too complex and models can not fit well to the entire English lauange like that, there is a reason why even fuck huge LLMs are all retarded
>>106979633Exactly, these retards continuously shit on everything that releases. While not perfect, the demos are pretty impressive, especially compared to all of the previous tricks we had to do before it. With these new loras and context nodes, I honestly don't miss daisy chaining frankenstien workflows for sake of mild color change.>>106979623Chroma is fine...if you know what you're doing.
>>106979712>While not perfect, the demos are pretty impressivewe live in the sora 2 era, I can't pretend that's impressive when the API fags are eating so good
here is what im mean btwhttps://civitai.com/models/1790792/netayume-lumina-neta-luminalumina-image-20this will be the proper illustrious 2, trained on actual gemma 2 instead of shitty T5. Its super undertrained though
>>106979721fuck off. have you actually used sora? it's shit
>>106979721even sora 2 does not have unlimited length video gen (if you had enough vram) though
>>106979721>API fags are eating so goodDeboonked already sweaty: >>106978843
>>106979652i didnt think it was very good after posting
>>106979658Does this mean we can train on potato pcs yet?
>>106979729>have you actually used sora?I did lol >>>/wsg/6007532
>>106979737>2boy, talkingwhooooooa buddy
>>106979721Ma'am, this is a local ai thread.
>>106979664That’s not a good thing
>>106979743>the 1girl community is making fun of 2charactersKEEEK
Ive been out of the loop for about 2 monthsI just use wan 2.2 with the lightx2v lora, which I just updated. Did any other secret sauce come out? When does 2.5 hit?
>>106979754new light loras are better, use with https://github.com/VraethrDalkr/ComfyUI-TripleKSampler for best results, wan2.1 got a unlimited length gen lora that kind of sorta works on 2.2 high, there are like 3 character replacement models I cba to list that came out this / last week...
>>106979754>Did any other secret sauce come outno>When does 2.5 hitit's out and not shared
>>1069797542.5 is already available through the ComfyUI API, which is included as part of the download linked in the OP
>>106979765>character replacement modelswhat is that?
lol
>>106979772replacing a character in a video with another character you give images of
>>106979721>we live in the sora 2 era, I can't pretend that's impressive when the API fags are eating so good>Prompt meal prep video>guardrail kicks in>knifes are dangerous>anon bad>account bannedfuck off
>>106979774ah ok
like thishttps://files.catbox.moe/5exhaa.mp4
https://files.catbox.moe/ffjiwq.mp4
https://files.catbox.moe/hnmys7.mp4
>>106979669but the entire point of qwen was to be able to edit shitty SDXL nsfw gens, such as fixing anatomy or change pose/position without changing the scene or character. Unfortunately nsfw lora's for edit models = bad and the dickheads report them as fast as they are produced. They should just ban all nsfw content then and be fucking done with it, because anyone with a clue can take any image and edit it the manual way using any decent SDXL based model. but they won't because their website will just die.
>>106979830not all nsfw loras are bad. the new qwen edit remove clothes one is very effective desuhttps://limewire.com/d/AvpLO#Gd7AyXiz1r
ai pornography has fried my brain
>>106979845utter kino
>>106978567this collage fucking blows, kys
>>106979766>>106979767Do we have a name for the cycle where new AI firm releases a local model to get their name out, then they pull the rug? Is Hunyun going to be on the cloud too? Will the next upgrade be some new model I've never heard of then the cycle repeats?
>>106979729>>106979731>>106979733>>106979744>>106979777y cum u fall for it
>>106979845catbox for that one too?
replace the blonde girl in the pink shirt on the right, with the anime girl in image2.
oh, and for the netayume model this is a early aesthetic tune for it https://civitai.com/models/1974130/reakaakas-enhancer-lumina-2
>>106979830>20b bloated base models so bad they can only serve as a cleanup for superior 3b sdxlgrim state of local, clean it up chinakeks
>>106979926kek, this
>no lumina support in onetrainer :(
>>106979926NetaYume Lumina is the real next gen model, its already close to illustrious with a fraction of the training and its prompt understanding is as good as qwen
>>106979935note, it is trained on 2 sets of captions, tags and natural lauange, switch between them each epoch
>>106979937>its prompt understanding is as good as qwenit's not quite that good at understanding prompts, no. you'll notice when you prompt multiple characters or just a lot of stuff.
>>106979912https://files.catbox.moe/tap3oo.png>>106979937>its already close to illustriousit surpasses it desu i cant think of anything ilu does better other than the fact that the community has had more time with it
Qwen is the best local base model by far, but it’s too expensive to train. But on the other hand look at the absolute disaster that is Chroma, where if he just spent the $200k on 5 epochs of Qwen it would be way better than the fluxenstein abomination we got
>>106979976no, if he spent that on lumina 2 we would have had a better model, I said from day one any model trained on T5 was a disaster. The only way to make a model good with T5 is to overcook the fuck out of it
>>106979976>Qwen is the best local base model by farand that's sad, this is a plastic factory
>>106979981and I am praying next wan if it ever releases switched to qwen or gemma or some new form of clip instead of using T5
>>106979937>its prompt understanding is as good as qwenabsolutely delusional
>>106979963>https://files.catbox.moe/tap3oo.pngbased
>>106979981Lumina 2 is nowhere near as good as qwen, sorry
>>106979994enjoy your one image per prompt / style of character per prompt unless you train a lora for any little thing you want
>>106979994and if you mean aesthetics wise its not aesthetic trained like qwen is (overly) Here is a lora though that is half way there: https://civitai.com/models/1974130?modelVersionId=2309365
replace the pink hair anime girl waving hello with the anime girl in image2. replace the text on the left saying "I need to make 1girls" with "I need to make more 2B's".edit of a previous text edit (was testing)I love how you can emulate fonts/styles too, you can dupe a font that you couldn't find a .ttf for without an issue.
>>106979963>that many artistsjesus christ
Is there honestly a API that does better genning than local?
>>106980024Midjourney
>>106980024NovelAI V4.5
>>106980028novelai is better and does nsfw if you wana pay pig it
Holy bait, batman
>>106980024Every single one. Midjourney, Seedream, Sora, NovelAI. Local is behind in all fields
>>106979086>>106979047yes, seems like regular i2v accept multiple frames, but still denoise masked and masks aply not per frame, but per 4 frames chunks or somethingthe video is 16(12 masked) first and last frames as copies of one input image
>>106980036it's a harsh truth but it's true yeah
>>1069800424 first and last frames are copies
>>106980020you only ever use one artist?
APIs can't do degenerate porn so they automatically lose to Chroma
>>106980064>muh coomthe only cope of localkeks
>>106980064Please understand anon. ALL FIELDS. ALL. FIELDS.
>>106980067>muh cat at mcdonalds
>>106980064NovelAI can.
>>106980075how about hitler ads >>>/wsg/6006269
>>106980061only a handful, not this many though. trying it currently and it does seem netayume benefits from throwing a million artists at the board
>>106980019ssr teto has arrived
>>106979963>prompt weights>plasma noise snakeoil>scheduler randomizer>clip in loras when the TE is fucking gemmathe only interesting thing really is the low CFG (but I guess it goes with the euler pp cfg sampler)all in all, a bad workflowjust curious, which impact wildcard file are you using? or are you rocking your own?
any process to deslop qwen edit skin slopping
>>106980078It can't do realism for shit though thereby it's irredeemably shit compared to Chroma.
>>106980086how about coom hitler ads?concession accepted.
>>106980146>coom hitler ads?can wan 2.2 do that? prove it
>>106980113their whole thing is anime which it is the best at
>>106980091okay, now it's better.
>>106980113Chroma can’t do realism either, only blurry meltyslop
>>106980155You'll have to pay me a sub fee.https://www.youtube.com/watch?v=2d6A_l8c_x8
>>106980166>Chroma can’t do realismskill issue
>>1069800538 first only
>>106980176Correct, the baker has a skill issue where the model somehow got worse over time. Perhaps try contacting him about it
>>106980166>Chroma can’t do realism either, only blurry meltyslopfacts, and the skin texture ain't what it used to
>>106980020>>106980087i used to do the same if not more with ilu >>106980099wf schizo but outputs kino. its the same wf i use with noobvpred so much is superfluous and i dont care to clean it up. plasma isnt snakeoil desu. but same wf, picrel is what i see anyway im using my own wildcards
it looks like illustrious is also trying to train a lumina 2 model https://www.illustrious-xl.ai/model/19 fuck their stardust system though
>>106980205it's beautiful, and some people have the nerve to say that we don't need artist tags, we definitely do
>>106980205>stable diffusion and midjourney in negativeshuh, guess I should do that
>>106980205>hiding the spaghettiFOR SHAME
>>106980205>he transformed spagghetiUi into a regular UiHOW???
>>106980231>being this newdo you really just look at the ugly spaghetti when gening?
>>106980234yes? :(
>>106980231https://github.com/chrisgoringe/cg-controller
>>106980249based, thanks anon
https://civitai.com/models/1134895/2000s-analog-coreQwen Image is saved!
>>106980266Welcome back, kino boringreality LoRA.
>>106979916I wonder how long it took him and with what hardware.
>>106980292its on the page>10 million images. Training was conducted over a period of 3 weeks on 8× NVIDIA B200 GPUs.
>>106980295so about $30k rounding up a bit for storage
>>106980295only 3 weeks? impressive, took that furry fag 6 months to finish chroma's finetune with 5 millions images
>>106980295>>106980309What? The description says "Trained with total ~7k images." and no indication of hardware or time. He does mention some settings though "rank 16 alpha 16, adawm, constant lr 0.0001"
>>106979916this lora is fucking garbage tho
>>106980313chroma is 50% bigger (2x as much compute) and again, T5 is fucking trash, also flux was distilled which is another level of fuckery
>>106980315? ah, I was talking about the actual model, not the lora, I didn't notice
>>106980316well its merely an aetheric tune so
>>106980316no its not?
>>106980340it is, crushes details and slops the output. if you don't think so you need to get your fucking eyes checked
>>106980343show me comparisons
>>106979086>>106979047yes, i'm a retardmultiple images input was here all the time81 frames from the first gen + 76 (-5 starting) from the second genleft is both vids with svi_film lora at 1.0i see no difference
>>106980266>correct number of strings and pegsThis is impossible for chromakeks btw
>>106980442if only Qwen Image wasn't so slopped it would be an incredible model
>>106980447now this is the good stuff
>>106980447>broken strings>sovl style>>106980266>good strings details>slopped base modelwhen will we reach the both at the same time?
>>106980470the problem is to get perfect details like that requires cooking a model until it becomes slopped, that is how it works. Its all about balancing it instead of a light aesthetic tune
>>106980487>instead ofinstead withqwen gets those small details right so well because they threw a shit ton of alibaba's compute at it until it pretty much fit to images to the point where a prompt just gives that single image
how long til qwen edit but it handles nsfw concepts well
Wansisters, do you think this could work on wan? Wonder if we can use this to set it to a higher cfg to get the movement but still keep sampler cfg to 1. I cant test, not at home https://github.com/Extraltodeus/Skimmed_CFG>A powerful anti-burn allowing much higher CFG scales for latent diffusion models (for ComfyUI)
>>106980515I already saw WFs with that before, it wasn't very good compared to light lora and did not work well with it
>>106980515Skimming, thresholding, etc. is aesthetic snakeoil.
>>106980515I'm using NAG to get better prompt understanding on Wan personally (that way I still keep cfg 1 and get that 2x speed increase over cfg > 1)https://github.com/ChenDarYen/ComfyUI-NAG
>>106980515>literal snake oileven worse than ''''plasma''' latents
>>106980399you can do 1girl, shaking ass 161 frames natively
>>106980529> ComfyUI-NAGyet another snake oil
>>106980549wrong, nag actually is great
>>106980529Yea nag is pretty good, thats in all of my 2.1 workflows
>>106980549>another snake oil*Extremely Loud Incorrect Buzzer*
>>106980546yes, but 561 frames?what if need 81 frames of promptA, then 41 frames of promptB, then 201 frames of promptC?>>106980557[citation needed]
>>106980559works great with 2.2 as well, just plug a different nag in for each model, wan2.2's default negatives make a big diffrence, speaking, talking, moving mouth and the same in chinese is 100% required or characters will always fucking talk
>>106980560nag for gwen when?
Reminder that comfyui logs your prompts
>>106980588[citation needed]
>>106980567so you just paste the default chinese negatives in there? why do you need a seperate nag for each model, and do you have any other tips? it always seemed like voodoo to me.
>>106980578>nag for gwen when?I'm also surprised it hasn't been implemented at all
>>106980594he's probably using the lightning loras so he's at cfg 1 (and therefore can't use negative prompts unless he activates NAG)
>>106980562>[citation needed]Nta but see >>106978926 I have nag in it, notice theres no slow mo>>106980567Interesting, I'll have to give it a test later when Im home again
>>106980589>he doesn't know
>>106980597last I read up on it he was working on making it work for flux nunchaku
>>106980610>he was working on making it work for flux nunchaku
>>106980559>>106980560>>106980606i had nag too but then turned it off and have seen that my gens are better without
>>106980606yea, always double up your negatives in chinese, I found that works way better, having it just in chinese or english only works half the time I found
>>106980515>>106980529might need to tweak depending on what lightx version and strength used. And yes it does follow prompt much better obviously.I read about these settings here.https://civitai.com/models/1889070/camera-tilt-down-undershotdon't use NAG its fucking excrement.
>>106980623the only negative is very slightly slower gens, it only increases quality when you have a negative prompt with stuff like blurry, low res and stuff in it, not using NAG if using CFG 1 is just retarded
>>106980635>not using NAG if using CFG 1 is just retardedthis
didn't know ComfyUi has a channelhttps://www.youtube.com/watch?v=JIBba5zZ38k
>>106980647>literal head of onions in the bottom right kek
>don't use NAG its fucking excrement>not using NAG if using CFG 1 is just retarded
>>106980661its easily testable, nag only improves quality if using CFG 1
>>106980661for kontext dev it was really effective, I'm not sure for Wan though, I didn't see much difference
>>106980661>CFG 1 is just retardedusing that is also fucking retarded because without cfg it won't follow complex prompts. but keep being a fucking idiot like most people here.
>Totally organic anti-Chroma posting and Qwen shilling
>>106980678have you ever heard of light lora? that is the entire point of it, you must be new as fuck
>>106980678NAG is relevant nowdays because we all use lightning loras (which put the cfg back at 1)
>>106980656hes done some wild animatediff vids back in the day, even had a motion lora training tutorial for https://github.com/kijai/ComfyUI-ADMotionDirector could think of it like training a wan video lora but instead its 1 video and its for animatediff lol>>106980656kek
wan2.1 with stable video infinityso it can belightx2vvaesome work under the hood requiredyet another snake oil from china
>>106980686yeah i have
>>106980706huh? besides the low quality it seems to be doing what is advertised
>>106980707what value do you put in the skimmed cfg node?
>>106980707>5 strength, 4 cfgjesus, your shit must be burnt as fuck, also the old 2.1 light lora sucks now compared to the newest one, especially for 2d animationsUse the latest one at 1.2 weight, 2 + 2 or 3 + 3 steps, also use the triple k sampler so you actually use correct time steps for 2.2
>>106980722>skimmed cfg nodei don't use that for wan.
>>106980723https://github.com/VraethrDalkr/ComfyUI-TripleKSampler
>>106980711no it doesn'tcompare frames from the middle of both videosyou can't make long vids with such quality drop
>>106980706>motion snapthis is worse than just using the last frame to gen a new video and stitching
>>106980688What are the settings? 11, 0.25, 2.5?
>>106980744works in vace with 8-15 frames tho
Wow! What an amazing upscaler! I totally must check this out!
Oh I'm gonna get real sloppy later with>https://github.com/stduhpf/ComfyUI-WanMoeKSampler>https://github.com/ChenDarYen/ComfyUI-NAG>https://huggingface.co/Kijai/WanVideo_comfy/tree/main/Lightx2v>https://github.com/Extraltodeus/Skimmed_CFG>https://huggingface.co/vita-video-gen/svi-model/tree/main/version-1.0
>>106980723>>106980627i don't care or give a fuck, you can see examples of videos using those settings by the person who created this lora. Also he new light lora I've read mixed opinions and i'm not going to waste my fucking time until someone else figures out the best settings. I tweaked the settings based on what that guy said and i was more than happy because it follows prompt. Otherwise i'd not even bother using the light lora's because again 1 CFG is fucking shit and i don't care what you or anyone has to say about it.
>>106980758>no smoothmix
>>106980758>SlopMaxxxingbased!
$17m must be enough for hiring one or two shills from india
>>106980747for wan I went for those values
>>106980765Thats what https://huggingface.co/Phr00t/WAN2.2-14B-Rapid-AllInOne is for when I wanna get reeeeally sloppy, oh yeah
>>106980758dont use skimmed cfg, its useless
>>106980758also it looks like kijai does not have the latest light lora, use this https://civitai.com/models/1585622
>>106980777for wan it is yeah, but for when i really wanna force shit and then process in controlnet with another sampler.
>>106980765>no gguf
>>106978641>https://huggingface.co/lightx2v/Wan2.2-Distill-Loras/tree/main>lora key not loaded:what's their fucking problem? why can't they make their shit compatible with ComfyUi? it's always the same shit with them
>>106980828just wait for kijai
>>106980828https://huggingface.co/lightx2v/Wan2.2-Distill-Loras/discussions/6#68f91663a3f9e27c0ee46820KJBoss says you can ignore those warnings
>>106980828it works fine, it just has extra keys that comfy does not use for modulation
>>106980892oh yeah bro so SOUL, the grain is so AUTHENTIC AND ANALOG bro, cant wait for you to post your other gens oh yeah damn THAT treeline looks so natural and REAL omg bro I think you hit jackpot with your gen techniques, I love chroma so much bro POST MORE
>>106980892>noise, the model
what's hunyuan good for
>>106980899>>106980902Post your realistic car in forrest gen
>>106980892where's the girl?
they deleted smash cut lora
>>106980913why would I dedicate my GPU time to gen garbage? fucking retard kys with your shitty gens
>>106980916jej
>>106980923I accept your concession
>>106980916>smash cut lorawhat's that?
>>106980931do you seriously believe that garbage gen is good? get your eyes checked, I pity you
>>106978641looks like an improvement over the previous I2V lightning loras
smoothmix with new light loras when
>>106980951create your own smoothbrainmix, just merge a general NSFW lora + the new lighting. Or are you this incapable?
>>106980936See >>106980931
>>106980914bound and gagged in the trunk
>>106980951https://civitai.com/models/1995784?modelVersionId=2323420it's like a finetune of wan?
>>106980971no it's a jeetmix
>>106980024Restricting the user.
>insert tesla quote
>>106980945>Hatsune Miku appears on screen from the left and shakes hands with the blue-haired anime girlyou know what's sad, is that Wan is able to add a new character to the scene while keeping the same artistic style while Qwen Image Edit can't
>>106980994butiful
>>106981005keeek
>>106981016>>106981016
>>106981017lmao bro, this is pathetic
>>106980527>Skimming snakeoilits really good for composition for hard to achieve poses by being able to give a somewhat good outline at something like 60 CFG. then you can use the model as normal in a second stage sampler, send in the burnt image as a depth map to control net and use a fresh empty latent as not to transfer any colour. Saves time and frustration with messing about with prompts, it can also produce interesting results on lower strength.Another set of nodes i use does something slightly different pre_cfg_comfy_nodes_for_comfyui, there is also cfg_pp versions of samplers that work at low cfg typically 1 cfg. So all is not necessarily lost, don't be so quick to dismiss things you have no fucking clue about.
this is for the guy who, i assume, came from here that wanted 300 of these. here's the first.
adios, дo cвидaния