[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/g/ - Technology


Thread archived.
You cannot reply anymore.


[Advertise on 4chan]


Against My Better Judgement Edition

Discussion of Free and Open Source Text-to-Image/Video Models and UI

Prev: >>106638601

https://rentry.org/ldg-lazy-getting-started-guide

>UI
ComfyUI: https://github.com/comfyanonymous/ComfyUI
SwarmUI: https://github.com/mcmonkeyprojects/SwarmUI
re/Forge/Classic/Neo: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassicneo
SD.Next: https://github.com/vladmandic/sdnext
Wan2GP: https://github.com/deepbeepmeep/Wan2GP

>Checkpoints, LoRAs, Upscalers, & Workflows
https://civitai.com
https://civitaiarchive.com/
https://openmodeldb.info
https://openart.ai/workflows

>Tuning
https://github.com/spacepxl/demystifying-sd-finetuning
https://github.com/Nerogar/OneTrainer
https://github.com/kohya-ss/sd-scripts/tree/sd3
https://github.com/derrian-distro/LoRA_Easy_Training_Scripts
https://github.com/tdrussell/diffusion-pipe

>WanX
https://comfyanonymous.github.io/ComfyUI_examples/wan22/
https://github.com/Wan-Video

>Chroma
https://huggingface.co/lodestones/Chroma1-Base
Training: https://rentry.org/mvu52t46

>Neta Lumina
https://huggingface.co/neta-art/Neta-Lumina
https://civitai.com/models/1790792?modelVersionId=2122326
https://neta-lumina-style.tz03.xyz/

>Illustrious
1girl and Beyond: https://rentry.org/comfyui_guide_1girl
Tag Explorer: https://tagexplorer.github.io/

>Misc
Local Model Meta: https://rentry.org/localmodelsmeta
Share Metadata: https://catbox.moe | https://litterbox.catbox.moe/
GPU Benchmarks: https://chimolog.co/bto-gpu-stable-diffusion-specs/
Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-one
Txt2Img Plugin: https://github.com/Acly/krita-ai-diffusion
Archive: https://rentry.org/sdg-link
Bakery: https://rentry.org/ldgcollage

>Neighbours
>>>/aco/csdg
>>>/b/degen
>>>/b/realistic+parody
>>>/gif/vdg
>>>/d/ddg
>>>/e/edg
>>>/h/hdg
>>>/trash/slop
>>>/vt/vtai
>>>/u/udg

>Local Text
>>>/g/lmg

>Maintain Thread Quality
https://rentry.org/debo
>>
File: 1745826591666564.mp4 (462 KB, 512x512)
462 KB
462 KB MP4
>>
Blessed thread of frenship
>>
File: G0wqNYlXcAA2Pq6.jpg (3.74 MB, 3072x4096)
3.74 MB
3.74 MB JPG
posting in a seedream thread!
>>
File: 1755892069695884.jpg (647 KB, 1552x1944)
647 KB
647 KB JPG
>>
File: 1752278470727585.jpg (743 KB, 1552x1944)
743 KB
743 KB JPG
>>
>>106642335
>>106642342
très bien, first especially
>>
>>106642309

I thought about it, then it gave me an estimated time of 40 minutes and I just stopped.
>>
>>106642335
>>106642342
Is this a lora trained on whoever does those Monogatari ED's?
>>
>>106642361
https://x.com/papilioninight
>>
File: Screenshot.jpg (34 KB, 405x376)
34 KB
34 KB JPG
>>106642333
>noisedream
>>
>>106642383
Danke
>>
uh oh the chromafootschizo arrived
>>
File: 1000010674.webm (1013 KB, 720x1280)
1013 KB
1013 KB WEBM
Footfags..
>>
It's crazy how much foot fags get away with. Like, if we assume they get as much arousal out of a foot as most people do looking at a pair of boobies. They essentially get do whatever the fuck they want and get away with it. Imagine if I posted bare boobies all over this thread and it was okay. I think we should hold foot fags to the same standard. Maybe make them war a foot shaped armband so we know they're foot fags.
>>
>>106642482
good thing with the foot fetish stuff is if the girls are wearing skirts I usually get so see some cute panties. so I don't mind.
>>
>>106642514
99% of footfags are just as into legs and ass
it's the 1% that wants only stinky feet that gives the rest a bad name
>>
File: 1740748300963022.mp4 (394 KB, 640x640)
394 KB
394 KB MP4
>>
File: Wanimate_00011.mp4 (1.51 MB, 620x848)
1.51 MB
1.51 MB MP4
Jesus christ, westoids..
>>
File: 1745463271331799.mp4 (994 KB, 640x640)
994 KB
994 KB MP4
>>
>>106642714
Are the original image gens Dall-E 3?
>>
>>106642559
I remember in University, I had a class on Anatomy and all the professor would ever talk about is how there is a nerve from the feet to the genitals. Like he would bring it up every lecture. The man loved feet.
>>
>>106642648
>Average 18 year old white woman.
>>
>>106642468
those legs have just the right amount of muscle on them, literal perfection
>>
>>106642648
The chink was hideous, what do you expect
>>
File: 93b.png (254 KB, 900x806)
254 KB
254 KB PNG
Can someone post a tutorial on how to inpaint with pony/illustrious/noobai in swarmui? It just doesnt work for me, im no using a workflow but swarm itself.
>>
>>106642301
5060 ti 16gb or 5070 12gb?
correct answers only
>>
>>106642795
More vram = better. Literally nothing else comes into play, especially if the GPUs are of the same series.
>>
File: 1744251311478158.mp4 (731 KB, 640x640)
731 KB
731 KB MP4
>>106642731
Yeah, pre-cucked Oct 2023 gens
>>
>>106642800
>Literally nothing else comes into play
the 5060 will be slower with models that fit in both
>>
File: 1737639607125145.jpg (197 KB, 680x1074)
197 KB
197 KB JPG
>>106642800
That only applies to nVidia tho.
>>
>>106642811
>the 5060 will be slower with models that fit in both

A whole lot less relevant models fit into the 12gb than fit in the 16gb. That kind of shitty information might make anon buy a GPU he sorely regrets buying.
Is that what you want?
>>
>>106642714
>>106642806
Really nice, 90s arcade feel achieved

Is it supposed to be 'Rage Rust' or is it just something random ?
>>
>>106642829
it's not shitty information, retard and I wasn't telling anon YOU MUST BUY THE 5060
go back to school
>>
>>106642835
Nah, nobody in history has picked up a smaller GPU because it was a little faster than the bigger one. It was a dumb statement designed to deceive a dumb person who has no idea what GPU to buy.
>>
File: Wanimate_00015.mp4 (1.29 MB, 614x832)
1.29 MB
1.29 MB MP4
Why does the video length double sometimes? Can I find where to set the length
>>
>>106642833
No, part of the meta for getting nice boobs back then was to put "race queen" in everything. "Rust" just popped in from a detail prompt.
>>
>>106642844
>Nah, nobody in history
stopped reading here, fuck off
notice how I didn't address the anon asking the question, I addressed you and your retarded comment that "nothing else comes into play"
>>
>>106642859
Nah, you wrote that under the assumption he would see the message and implied it was relevant to his final decision. You are a deceiver and a demon.
>>
>>106642867
and you are retarded
>>
>>106642848
Oh yeah, looking nice and fluxxy. Got that plastic glow.
>>
>>106642848
if they can get rid of the slop that would be a fine product, but for the moment... meh... still on the wan 2.5 waiting room
>>
>>106642853
Heh, ok
>>
>>106642867
>>106642859
>>106642873
>>106642853
Why does the generation come out blank when I inpaint on swarm?
>>
>>106642880
I swear to god it's the strength of the controlnet but nobody ever listens to me even when I'm very very right.
>>
File: Wanimate_00015.mp4 (792 KB, 1176x544)
792 KB
792 KB MP4
>>106642848
I not 100% sure but I think it's the frame window size on the Wan Animate Embeds node. Feels like if your video goes over the default value it does another 77 regardless of how many remain in your source video
>>
File: xmas_23_08.jpg (1.43 MB, 2610x3987)
1.43 MB
1.43 MB JPG
has anyone trained a Krys Decker style lora? the ones on civit kinda suck

picrel
>>
>>106642884
>swarm
There's your problem, why the fuck would you use swarm ? It's duct tape C# over Comfy, it's crap
>>
>>106642301
how do i generate myself ai little girls? i'm on a amd raedon system on arch linux
>>
File: 1727125445145005.mp4 (1 MB, 640x640)
1 MB
1 MB MP4
>>
>>106642914
sovl
>>
>>106642899
I have dataset ready, just havent trained. Perhaps next week for Illustrious
>>
>>106642931
pls post on civit if you get around to doing it
>>
File: 1758212392259566.jpg (100 KB, 1024x683)
100 KB
100 KB JPG
>>106642902
I tried some workflows, but its just FOSS tinker tranny shit, I need a UI, not a rocket schematic
>>
>no you just need 30 workflows to get the same functionalities as swarmui trust me im a real woman
>>
>>106642986
Use one of the Forge's then
>>
File: Wanimate 00019.mp4 (3.69 MB, 740x1024)
3.69 MB
3.69 MB MP4
Original video is 10s long, it randomly added 4s behind.
>>
>>106643108
>Rubber ball turns into boxing gloves
Cool trick
>740x1024
How long did that take?
>>
>>106643121
Sometimes my generation will slow to a crawl. I have twitch and youtube running on 2nd monitor
>>
>>106643160
are you maxing out your vram?
>>
File: 1747207628403952.mp4 (512 KB, 640x640)
512 KB
512 KB MP4
>>
File: 1728575108358714.mp4 (669 KB, 640x640)
669 KB
669 KB MP4
>>106643177
>>
>>106643177
>>106643210
SNK logo on the top would be so fitting
>>
Newfag here. What kinds of mileage can I get out of a laptop 3060 and 6GB VRAM with WAN 2.2?
>>
>>106643227
None because that's a piece of shit.
>>
>>106643227
It can theorethically work with enough quantization, but seriously it's not worth it, at least get a 12gb vram card and 64gb system ram
>>
>>106643243
NTA but any meaningful difference between RAM speeds? Mostly gaming but I might upgrade to 64GB next time I do a system update.
>>
File: Wanimate_00021.mp4 (2.36 MB, 1038x480)
2.36 MB
2.36 MB MP4
>>
>>106643108
>>106643160
5090?
it's so funny while it takes you 14 mintues to gen a 1024p video
it takes me 14 minutes to genn only one high res picture lmao
>>
File: Wanimate_00020.mp4 (2.54 MB, 1244x576)
2.54 MB
2.54 MB MP4
>>106643297
4090
>>
>>106643257
No, not for AI, DDR4 or DDR5, there will be no noticeable difference, most of the time is spent squeezing data through the GPU bus
>>
>>106643312
Yeah figured as much, thanks.
>>
File: 1727188496956154.mp4 (414 KB, 560x560)
414 KB
414 KB MP4
>>106643227
>laptop 3060 and 6GB VRAM with WAN 2.2
>>
>>106643227
maybe a not so high res video per 1.5h? you may not want to do that.
>>
File: QWEN_00007_.png (2.21 MB, 768x1280)
2.21 MB
2.21 MB PNG
why the shit is qwen image so slow.. I can gen a wan video, more like 2 by the time I get a gen from qwen
>>
can you use qwen and linux yet?
>>
>>106643508
no. windows users only
>>
>>106643508
???
>>
>>106643518
fuck this
imagine making a fucking android build but no linux. lmao
>>
>>106643525
Are you retarded ? Qwen works on Linux, why the fuck wouldn't it
>>
>>106642648
WTF are you doin to Léa Hélène Seydoux-Fornier de Clausonne!?
>>
How tf can i get a lora or some shit to inpaint change an jewelry around someones neck or add
>>
anyone uses Kijai's WanAnimate example workflow ?
Where do I get "refer.jpeg, and raw.mp4"
>>
>>106643604
nigger just use your own reference image and video
>>
>>106643604
there's where u put your own video and image dummy
>>
>>106642907
1. buy a rope.
2. use it.
>>
>>106643483
nunchaku?
>>
>>106643099
I did just that, loaded a model, tried to do a gen on fresh install and it hits me with "TypeError: 'NoneType' object is not iterable"

What now?
>>
>>106643655
used the rope to get some irl little girls. Thanks for the help, King.
>>
File: QWEN_00013_.png (3.44 MB, 960x1600)
3.44 MB
3.44 MB PNG
>>106643666
Nope Q8. The nunchaku version is fast, but blurry af
>>
>>106643616
>>106643622
I did it but Idk wtf is this points editor and What Am i supposed to do with it
>>
>>106643676
then use 4-8 steps lora then
>Q8 Muh quality
fuck this jew shit.
>>
>>106643669
What model ?
>>
File: qwenshit.png (368 KB, 170x1076)
368 KB
368 KB PNG
>>106643698
Look at this wildly varying times

I don't use the lora, not a fan.
>>
>>106643687
red = negative target
green = positive target
>>
>>106643701
models/1620407/diving-illustrious-flat-anime-paradigm-shift
>>
>>106643711
>480.34s
Time to give up
>>
>>106643727
already did
not worth it
>>
>>106643726
Start with some standard model and see if it works, that way you know it's something with this weird hybrid that is a problem
>>
>>106643727
>>106643731
thats the thing, some gens are like 30-40s. some go on forever, Idk why.. utterly confusing
>>
>>106643737
what's your vram usage while genning?
>>
>>106643735
Whats weird about that one? It works on swarm and sdnext, the culprit heres forge, or my lack of understanding of it
>>
File: qwenshit2.png (181 KB, 151x572)
181 KB
181 KB PNG
>>106643737
>>106643743
again more bs - picrel

98%
>>
>>106643752
>98% (vram usage)
gee i wonder what the issue is
>>
File: 1756691837181577.png (29 KB, 267x623)
29 KB
29 KB PNG
why doesn't it work then https://files.catbox.moe/wix79k.mp4
>>106643719
>>
>>106643752
I think you are using system ram and it slow down the process
Try to leave at least 1.2-1.5GB free in the vram
>>
>>106643761
>>106643769

Oh..ok how do I fix it?

The problem I have is with time consistency. Why does one img take a certain time while another, with just a randomized seed, take a whole different time.
>>
>>106643775
1st step is to disable system ram fallback on your nvidia panel, then you will have to play with offloading stuff to gpu (using the multigpu nodes for ease of use)
>>
>>106643784
ok, Ive no idea how to go about doing the 2nd part. Shall read up
>>
File: 0920200703024-iLu7gMjTgqG.png (1.2 MB, 2230x1028)
1.2 MB
1.2 MB PNG
>>106643762
in the preview u can see the black squares are getting swapped. that's how u know the AI got the correct spots
>>
>>106643775
it seems your settings put you right on the limit of your vram. different seeds will use more or less vram so some might gen normally but others might cross the threshold where the vram is too full and it causes catastrophic slowdown. i think that's what's happening anyway
>>
File: file.png (5 KB, 198x78)
5 KB
5 KB PNG
it's good to be the king
>>
>>106643837
gen time?
>>
how many avenues does 8gb -> 16gb vram open? not interested in training/tuning
>>
>>106643837
renting is not owning
>>
File: Wanimate_00030.mp4 (2.39 MB, 1246x576)
2.39 MB
2.39 MB MP4
>>
Does ComfyUI have a built in wan animate workflow now? Not using Kijai
>>
>>106643800
https://files.catbox.moe/fpt1d6.mp4
it works but 480p quality kinda bad. Probably only good for closeup video
>>
>install comfyui
>installs on C:/appdata without asking
>finishes installing
>1 error, want to send a report?
>no explanation, please reinstall o algo
Kill all FOSS trannies
>>
>>106643961
it looks really fake
>>
Does Kijai still refuse to implement the usage of quants into his nodes
>>
>>106643961
it's fried, try lowering cfg or cnet strength
>>
>>106643967
just install the portable version. always works fine
>>
Once I use inpaint, or try to attach a pic to a loras meta, this is what happens, the pic gets fucked up, how can i solve this?

I can generate just fine in comfy and swarm, but once i edit it with inpaint (or just attach any image to metadata) this happens.

Please help
>>
>>106644027
it's Kijai's example workflow. I didn't edit anything
What is cnet strength?
>>
>>106644025
this. kijai nodes are fucking boring
>>
>>106644055
Changing scheduler to euler helps a little
>>
File: Wanimate_00033.mp4 (1.49 MB, 738x1024)
1.49 MB
1.49 MB MP4
>>
>>106644055
try a shift of 1
>>
>>106644070
workflow? i'm late to the wan animate party
>>
get the 5060 ti, or wait even longer for gpus we're never getting
this shit pushes 1280x720 in wan, pretty neato.
>>
>>106644096
me? I'm personally waiting for a 96gb cuda compatible 2k eurodollars card. my 4080S will do for now. sad vramlet face
>>
something fucking wrong with KiJai workflow
I tried Animate HF and it gave completely different (better) result
https://files.catbox.moe/n9qu8v.mp4
>>
This is the image im trying to attach on swarm's metadata editor right?
>>
>>106644132
And this is what I get, it also applies to inpainting and comfyui workflows, what the fuck is happening anons?
>>
>>106644070
can do the reverse?
Like make reference image do walking pose of the video, instead of replacing original character with ref character.
>>
Are there any hubs with real people loras apart from the archive?
>>
>>106644132
>>106644137

bro idk how to tell you, nobody here uses swarm.
it's either comfyui spaghetti masters or forge copers.
>>
>>106644144
But this applies to comfyui workflows too...
>>
>>106644139
won't u just use wan i2v for that?
>>
File: 1744038776461252.png (76 KB, 582x842)
76 KB
76 KB PNG
>>106644025
they work just fine
>>
>>106644139
Yes, you just disconnect Get_background_image and Get_mask from WanVideo Animate Embeds
>>
>>106644137
what custom node is this?
>>
>>106644155
Huh? animate can clearly do that. See move examples.
https://www.modelscope.cn/studios/Wan-AI/Wan2.2-Animate
>>
>>106644183
Cool, first time using KJ nodes in a while. What turned me off them was his steadfast refusal to implement gguf into his 2.1 nodes.
>>
>>106644095
https://github.com/kijai/ComfyUI-WanVideoWrapper/blob/main/example_workflows/wanvideo_WanAnimate_example_01.json
>>
File: Wanimate_00034.mp4 (3.26 MB, 740x1024)
3.26 MB
3.26 MB MP4
I still don't understand why random seconds was added behind for some videos
>>
so uhh... is there a model that can do amateur porn image frame? like sex act in motion, realistic, not ai-slop looking? can chroma do that or something?
>>
>>106644228
It generates in chunks of 77 frames unless you use the context options
>>
how decent is qwen as a refiner/upscaler? like how people use wan for images.
>>
>>106644189
thanks
>>
>>106644228
what cfg,shift, sampler value?
>>
File: 1752356408541466.mp4 (3.49 MB, 1248x720)
3.49 MB
3.49 MB MP4
hatsune miku reads a book at the library

cool
>>
>>106644298
>>
13.3B qwen image pruned model just dropped
https://huggingface.co/OPPOer/Qwen-Image-Pruning
>>
Using the new fine tuner for vibevoice is practically perfect cloning even in other languages.
https://github.com/voicepowered-ai/VibeVoice-finetuning

It's quite slow for realtime stuff and using low steps+cfg 3.0 only works for sort sentences, otherwise it goes "monster" mode.
So it seems like 10 steps and 1.3CFG is the best compromise for longer than 1-2 sentences
>>
>>106644326
but why? (genuine question)
is it for vramlets?
>>
>>106644326
what's the point? I have no issue running the 20b model on my 3090
>>
>>106644356
you can already do other voices if the input voice is in that language.
far more interested if this will let you control the emotion better. like a "asmr" lora or a "moaning cumslut" lora
>>
>>106644326
> The pruned model has experienced a slight drop in objective metrics.

literally worse with ZERO benefits. what the fuk m8
>>
To the 5s club members

What sampler gives the best results?
I've stayed on euler and simple combo for a while now.
>>
File: 1588629222545.gif (933 KB, 220x220)
933 KB
933 KB GIF
>>106644326
i have the image and vace models fatigue
>>
File: Wanimate_00039.mp4 (1.85 MB, 738x1024)
1.85 MB
1.85 MB MP4
>>106643676
>>
File: 1737550411701142.png (80 KB, 723x444)
80 KB
80 KB PNG
this animate shit is so slow man
832x832 each chunk takes 4minutes (x3-x4) on 3090
>>
File: 1740697341456866.mp4 (1.61 MB, 796x480)
1.61 MB
1.61 MB MP4
kek

animate tip: add green points to whatever you want to mask, add red points to stuff you dont want swapped
>>
>>106644478
lower the res to 640x480 or whatever, higher res = slower gen time

same with wan 2.2
>>
>>106644442
what gpu? how long for this 720p gen?
>>
>>106644492
*also I had to set block swap to 30, prob dont need to if you have more than 16gb.
>>
>>106644509
actually its 578x1024, im on 4090
>>
File: 1754801512288998.mp4 (1.65 MB, 796x480)
1.65 MB
1.65 MB MP4
>>
>>106644531
that's.. surprising that it only took that short for that res.
>>
>>106644326
Isn't this just for training?
>>
File: 1752343986298810.mp4 (1.48 MB, 796x480)
1.48 MB
1.48 MB MP4
with a realistic photo

literally upset
>>
>>106642629
nice
>>
>>106644539
I don't get it, does wan-animate use openpose?
>>
>>106644619
openpose and face capture for identifying/masking, seems like
>>
File: 1739608160756350.mp4 (1.66 MB, 788x480)
1.66 MB
1.66 MB MP4
not exactly haruhi but still a neat swap:
>>
>>106644478
>each chunk
you mean each step? that doesn't seem right. a 720x1040 gen with 6 steps takes me like 8 minutes total on a 3090
>>
/adt/ anon here, UPDATE: CivitAI Helper Fix for NeoForge and Classic

Important message for anon >>106636188
If you're trying to get CivitAI helper working in NeoForge

>For NeoForge branch:
1. Get the fix from yesterday's github repository post
2. Apply the checkpoint fix from >>106643964
3. Replace the VAE the code from >>106644511
4. Done, CivitAI helper should now work

>For Classic branch:
- Just use the GitHub fix as is, no modifications needed

Tested and working on my setup. Questions?
Find me in /adt/
Bye
>>
>>106644689
>>106644511 says the lora section but its the VAE section
>>
File: 1736289374471202.mp4 (1.56 MB, 788x480)
1.56 MB
1.56 MB MP4
okay, now this one's actually impressive. I think the model leans more towards realism than anime (initially).

prompt: japanese woman standing on a race track.
>>
>>106642806
had to ask just because the look was so intimately familiar. That model had such an obvious 'house style' but it's hard to pinpoint what exactly it is about it
>>
>>106642335
>>106642342
cute!
>>
>>106644503
>>106644688
Oh NVM, you're talking about wan animate? I haven't tried that yet
>>
>>106644708
nice
>>
%chances that I can run qwen with 8gb vram/32gb ram?
>>
>>106644233
Yeah Chroma can do that, but be warned that not all sex acts are equally easy to get right. e.g. doggy is gonna be all over the place for the same reason that lying in the grass is, whereas cowgirl pov is pretty easy. The difference is in how widely the source images vary
>>
File: tmpa27y7s86.png (3.51 MB, 1080x1920)
3.51 MB
3.51 MB PNG
>>106644689
Will you put this on the git page too?
>>
>>106644326
>>106644359
Yes I am a VRAMlet and I am interested in a good distill or pruned model.
I am not ashamed to admit this.
>>
>>106644708
pls consider: >>106642097
>>
Is there a way to combine or merge samplers or schedulers? Like the model merge node? I wanna make some wacky stuff
>>
Did anyone ever made a https://github.com/LeanModels/DFloat11 for vaes and text encoders? I am curious if it brings any speed up (for stuff like (um)t5 and tiled vaes for large upscales)
>>106644818
Samplers and schedulers are not weights so the answer is no.
You can just read the source code for what they do and try to write your own with combined behavior.
>>
https://www.reddit.com/r/StableDiffusion/comments/1nlybq8/wan22_animate_test/
lmao nice
>>
>>106644879
I'm far too stupid for that, however I did find these

https://github.com/BlakeOne/ComfyUI-CustomScheduler
https://github.com/BlakeOne/ComfyUI-SchedulerMixer

Looks interesting, I'll give them a shot
>>
>>106644818
i just want to point out how really fucking stupid this question is.

you can however end a sampler at x step and then continue the generation with another sampler. don't know why you'd want that but sure.
>>
cozy bread
>>
>>106644708
that is not anri okita though
>>
Can wananimate be used for static image character swap? Can Qwen do that?
>>
>>106642301
Been out of the loop for a few years. Is that Wan2.2 model only available with comfyUI? I have a 4090 and am wondering how it runs, the limitations, but the guide is pretty barebones
>>
>>106645122
Whatever smart ass
>>
>>106645176
WanAnimate is for videos, for images you can use either Qwen Image Edit or Flux Kontext.
>>
>>106644689
Based artsy and smart sister general
>>
>>106642907
join the feds and try to get yourself on chan sabotage duty
>>
>>106645334
retard or bot? I asked if anyone tested the character swap abilities of both. Not generic information about the models.
>>
File: Wanimate_00049.mp4 (2.56 MB, 740x1024)
2.56 MB
2.56 MB MP4
>>
AniStudio will get sound support next week. I'm preparing a new release.
>>
>>106645408
test it yourself?
https://www.modelscope.cn/studios/Wan-AI/Wan2.2-Animate
>>
>>106645421
Fuck off
>>
SDXL bros, rejoice nunchaku-sdxl https://huggingface.co/nunchaku-tech/nunchaku-sdxl/tree/main
>>
>>106645466
Seems like you are butt hurt.
>>
>>106645421
penises
>>
>>106645470
why? sdxl is fast enough.
this simply degrades the quality
>>
File: 1751968351363083.png (1.96 MB, 1328x1328)
1.96 MB
1.96 MB PNG
qwen image q8 is pretty good for anime, used the waiv14 banner image/prompt to test:

hatsune miku, power elements, microphone, vibrant blue color palette, abstract,abstract background, dreamlike atmosphere, delicate linework, wind-swept hair, energy ,masterpiece,best quality, On her arm is the text "01" in red text.

also if you dont add the last part it doesn't add the number to miku, for whatever reason. but it works.
>>
>>106645421
Can i finally run it then without it crashing all the time? And are the text clipping issues finally solved?
>>
>>106645463
Wow I can test is myself??? Amazing! That totally answerers my questions about if others tested it and their results. What a great discussion thread.
>>
>>106645506
Yes and yes. I noticed that model offloading didn't do anything.
>>
>>106645408
>>106645525
Kys you retarded lazy faggot.
>>
File: ghmgmgmgv.mp4 (671 KB, 768x768)
671 KB
671 KB MP4
>>
File: 1755928374072748.png (5 KB, 564x467)
5 KB
5 KB PNG
>>106642301
bros I am new to this. Can someone explain if I want to generate realistic images is it the underlying model that changes things or is it the technology that is different ( i.e: stable diffusion for anime like , some other thing for realistic )
>>
STOP BULLYING BASED ANI
>>
>>106645290
wan2gp
>>
File: 1741847228364914.png (74 KB, 1704x634)
74 KB
74 KB PNG
What is recommended for picrel?
I get artifacts 2.0 and slowmo at 1.0...
>>
Is it normal for fp8 to bring extremely little to no speed up over Q8 in 3000 series?
>>
>>106644399
>model fatigue
No such thing
>>
>>106645408
If you think a video model will perform better than a good dedicated image editing model like qwen edit then you'll have to test that yourself, we're not gonna waste our time
>>
>>106645492
gotta go fast
>>
>>106645587
Yes.
>>
>>106645408
It's new model, nigga. most people here running distill loras with it and result will be always worse.
you expect a honest answer from who exactly?
>>
>>106645583
don't use the 2.2 light loras, they are literally broken.
>>
>>106645534
Maybe listen next time 20 anons tell you about problems
>>
>julien
>>
File: 5839681376.png (937 KB, 1360x768)
937 KB
937 KB PNG
>>106645554
It's all stable diffusion
>>
>>106645784
I'm extremely busy. Animation pipeline has taken most of my time.
>>
>>106645774
What can I use to get faster gens?
>>
File: clownin around.jpg (41 KB, 788x406)
41 KB
41 KB JPG
Man this stuff is nuts, akin to gambling, kek
>>
>>106646012
You could use lightning loras but for 2.1 instead, but you could also crank up your high noise cfg to 2 or 3 and specify the speed of motion in the prompt
>>
fucking comfy, KJ chads are using wananimate left and right and comfy still struggling to make it work for native
>>
>>106645554
there is both, underlying models that kind-of change everything and also some different ways to use it though of course the "meaningful" methods are typically somewhat constrained to, like, methods that at least might give you a dog if you prompt a dog.
>>
>>106646134
OK will try, thank you anon.
>>
https://huggingface.co/nunchaku-tech/nunchaku-sdxl/tree/main
base sdxl got nunchaku before any other model that actually gets used.

are they just fully stupid?
>>
>>106646278
OH COME ON
That before wan2.2?
Or even before qwen lora support?
Do they all have adhd or something?
>>
File: 1488383322064.png (17 KB, 131x141)
17 KB
17 KB PNG
>>106646278
>sdxl
>>
File: comfypromises.png (11 KB, 837x111)
11 KB
11 KB PNG
almost one month and still comfy doesnt deliver his WAN S2V native support aka the best implementation, kek what a joke this guy is, now that hes not getting any models in advance hes getting behind, Kijai is making this guy bite the dust
>>
has anyone come up with the solution for self forcing wan always doing slowmo?
>>
File: 1746788448615238.png (1.36 MB, 896x1160)
1.36 MB
1.36 MB PNG
qwen edit is fun
>>
>>106646332
so true grandma
>>
>>106646332
doesnt kijai work for them now? meaning we will never get native implementations again.
>>
File: s2v.png (205 KB, 1523x940)
205 KB
205 KB PNG
>>106646332
?
>>
>>106646332
serves him well for going down the API route, now major new models wont send him shit kek
>>
>>106646334
just use this workflow
https://civitai.com/models/1818841/wan-22-workflow-t2v-i2v-t2i-kijai-wrapper
>>
>>106646314
>That before wan2.2?

kek, they were suppose to do 2.1 a long time ago
>>
>>106646386
I know but at least I would have understood them migrating to wan2.2 since it's superior in all fronts.
But nothing happened, it's crazy.
>>
>>106644708
>I think the model leans more towards realism than anime (initially).
So does every model released this year, minus the underbaked neta
>>
Also S2V is fucking useless because what the fuck are you gonna do with a 5 second sound? Say half a sentence?
>>
File: maid.webm (3.46 MB, 704x1280)
3.46 MB
3.46 MB WEBM
>>106645554
yes the style is controlled by the checkpoint, usually good checkpoints will focus on one style really well and not be able to do much else, multipurpose checkpoints are garbage. but for realistic i find you can only use it to generate contemporary stuff, even if you try to generate an "elf" it will make a halloween costume with plastic ears
>>
>>106646405
Listen, I wou
>>
File: 1757761762735691.png (1.28 MB, 896x1160)
1.28 MB
1.28 MB PNG
>>106646340
>>
>>106646405
yep, this anon got it
>>
why are there so many shitty nodes abstractions for all these new models? why couldn't it just be factory-styled like more competent software?
>>
>>106646457
lol
>>
>>106646457
kek
>>
>>106646505
That's inevitable because CumUI is headed towards being a bloated mess. Some devs don't bother even with cum backend but make wrappers instead (like vibecvoice nodes).
>>
Redpill me on nunchaku, I have skimmed through its paper but I got some questions:
I am on 3000 series so I should use the int4 version, correct? (I expect the fp4 performance to be ass without dedicated NVFP4 acceleration of Blackwell)
In the paper, they claim that they have chosen rank 32 as a compromise between 16-64 for overhead/quality optimal, but I see that they have rank 128 version for models available. Does anyone have a rough ballpark number/anecdote for how slower these versions are? Is it 10-20% slower for noticeably better output so worth it thing or is it 2 times slower for little difference it's worthless thing?
I expect the answer to be yes but are loras trained on standard fp16 compatible with these quants?
Lastly nunchaku needs its own comfy nodes, any BS or limitations I should be aware of?
Thanks if you respond.
>>
>>106646564
nunchaku is cope for vramlets, if you care about quality and have even 12gb of vram, just use q8 and wait
>>
>>106646564
If it's qwen just use a q6 with distill lora
>>106646609
12 is not enough for Q8. I couldn't run it on a 4070s
>>
>>106646609
nunchaku for video gen would make gen way faster for the same quality as fp8 and without the need to use lightning loras
>>
>>106646629
>for the same quality as fp8
lmao
>>
i just peaked at sdg.

oh god why are they so shit
>>
>>106646653
lmao to you, read the paper
>>
>>106646564
>Redpill me on nunchaku
it's a Q4_M quality quant
>>
>>106646676
because of censorship of the online moduls lmao
>>
File: 1753854007824244.png (695 KB, 2540x1462)
695 KB
695 KB PNG
>>106646679
>he believes papers
>>
File: gpu weights.jpg (293 KB, 2923x1176)
293 KB
293 KB JPG
I was genning and I got an error saying to lower the gpu weights, it said to lower to a save 1500mb or so or face potential burn out
>>
>>106646231
what's KJ? does anyone still use voldy?
>>
File: file.jpg (709 KB, 2482x963)
709 KB
709 KB JPG
>>106646692
show me where it's wrong
>>
File: qwenpls.jpg (356 KB, 2805x1125)
356 KB
356 KB JPG
wonder if qwen edit would be able to edit Mahiro into the first picture.
because img2img and controlnet really hate the bolt cutter.
>>
>>106646679
>b-b-but THEY SAID IS THE CAME QUALITY SO ITS TRUE I DONT NEED ACTUAL COMPARISONS
absolute state of underage newfag vramlets
>>
>>106646692
>in areas of the social sciences
Are you retarded anon?
>>
>>106646676
You are not schizo-anon.
>>
File: 1737564033894023.jpg (2.04 MB, 7961x2897)
2.04 MB
2.04 MB JPG
>>106646750
can you compare that with Q8, Q5 and fp8?
>>
>>106646762
so you can't show where it's wrong, thought so
>>
>>106646750
Post your own reproducible workflows instead of slopped cherry picked images of flux
Copechaku is a meme
>>
>>106646763
yep, you're completly braindead
https://en.wikipedia.org/wiki/Replication_crisis
>A 2016 survey by Nature on 1,576 researchers who took a brief online questionnaire on reproducibility found that more than 70% of researchers have tried and failed to reproduce another scientist's experiment results (including 87% of chemists, 77% of biologists, 69% of physicists and engineers, 67% of medical researchers, 64% of earth and environmental scientists, and 62% of all others)
>>
>>106646779
are you illiterate?
>>
>>106646801
Concession Accepted.
>>
Absolute state of poorfags I can't even
>p..please stop doing random shit please work on nunchaku wan
>>
>>106646779
Oh nice, this doesn't apply to the svdquant paper then. Thanks for the confirmation.
>>
>>106646384
ive tried this out, smaller videos are still slowmo, and i get oom for things i can gen with this work flow https://rentry.org/wan22ldgguide
>>
>>106646681
More like Q2_K_L and only for some models that quant ok, i tested it initially for flux kontext, and it was unusable the moment you needed any semblance of quality preservation
>>
>>106646814
it applies to your appeal to authority of "read the paper bro xD", brainlet
>>
File: bold move.png (216 KB, 450x270)
216 KB
216 KB PNG
>>106646814
>sure there's a 70% chance the svdquant paper is bullshit but let's gamble for the 30% instead
>>
>>106646815
what do you mean by "smaller videos"? obviously you fucked something
>>
File: 49.jpg (179 KB, 1440x1800)
179 KB
179 KB JPG
How do I lewd up this pic into a video, bros?
>>
File: 00074-3498171204.png (2.73 MB, 1248x1824)
2.73 MB
2.73 MB PNG
On this day.. I go to coomer sovngarde..
>>
>>106646840
..mostly to find a better model than nova animal because i'm kind of sick of the grainy details. that or i just need to stop being a jew and do more than 10 hires steps.
>>
>>106646805
trying not posting irrelevant stats next time
some percentage of researchers failing to reproduce another's results =/= some percentage of papers not being reproducible
>>
>>106646826
I'm curious, where do you find 70% of a math paper being wrong? It's not in any of the "studies" you cited.
And how do you know the svdquant paper is wrong?
Please share a workflow, we can test that easily.
>>
>>106646852
>how do you know the svdquant paper is wrong?
how do you know the svdquant paper is not wrong, you're the one claiming it's right, therefore you have the burden of proof, and since you have the burden of proof...
>Please share a workflow, we can test that easily.
>>
>>106646839
There are million females like this, just open up xvideos and browse up.
>>
>>106646849
>the reproduction crisis is a hoax bro
kek
>>
>>106646840
catbox me nugga
>>
>>106646870
I didn't claim anything, I'm not the anon who talked about it, but I'm tired of you posting the "70%" studies without understanding what it is nor how it doesn't applies to every paper under the sun.
>>
>>106646834
i mean low resolution. i can do 120 frame 704x1280 videos in my usual work flow, but the one you shared is screwing something up, im using the same block swapping. but it doesnt matter since it doesnt solve the issue anyways
>>
>>106646881
what I said doesn't imply the reproduction crisis is a hoax, retard
>>
>>106646890
hmm? sweety? why are you not tired of anons believing papers like it's gospel though?
>>
>>106646896
>the reproduction crisis is not a hoax, it's real retard!
>btw, check out this paper and look at those nice numbers, you got to believe them broo!!
>>
>>106646895
you shouldnt go above 81 frames, and you should use 720x1280, chain the videos if you want to elongate them with the "loop" workflow from that link
>>
>>106646907
there really should be a minimum IQ allowed on the internet
>>
>>106646839
generate a better1girl first then animate with wan i2v. or generate it all with just wan t2v.
>>
>>106646920
Concession Accepted again.
>>
I hate everything about the current AI ecosystem and it's thanks to comfyui. I can't even find a simple wan-animate workflow, even Kijai's own github is slop infested garbage with broken missing nodes that the manager can't find, not to mention completely ignoring the design sensibilities of comfyui with all that "set vae" global variable slop. WHY DID YOU NOT ESTABLISH STANDARDS COMFY YOU FUCKING PRICK. This dude just released an unfinished barebones UI using someone else's node graph library and did fucking nothing for 3 years and someone gave him SEVENTEEN MILLION DOLLARS
>Want an integer? Here sarr, let me randomize that for you
>Want to loop through a directory? Go fuck yourself
>Want to do basic math like adding two numbers? Go ahead and download someone's node pack with 6 gorillian dependencies. Oh, you only wanted the node that adds stuff? Tough luck kid, you're getting everything
>Want to use models stored somewhere else on your hard drive? There's some esoteric yaml file you have to add and then you have to make symlinks anyway because it expects a specific directory structure
FUCK YOU COMFY YOU FUCKING NIGGER
>>
>>106646926
okay retard
>>
File: 00086-1477115875.png (2.95 MB, 1248x1824)
2.95 MB
2.95 MB PNG
>>106646884
https://civitai.com/models/784543/nova-animal-xl
>cfg 5
>use any sampler but i think DPM ++ 2M Align your steps does best

>1216x832

>hires 1.5x 20steps same sampling method

>no changes to adetailer

upped my hires from 10 steps to fix the grainy details. mostly. model still suffers from extra fingies. and my hires model is something a trainer sent me a while ago so there's extra realism from that.

>>106646927
wow he's literally me, preach it from the rooftops brother.
>>
>>106646927
3/4 of these are literal skill issues
>>
>>106646927
Nothing's stopping you from forking it and fixing the shit for yourself.
>>
>>106646957
Nobody's giving me 17 million dollars either
>>
File: qwen-image-nunchaku.jpg (3.94 MB, 5924x5708)
3.94 MB
3.94 MB JPG
>>106646609
>>106646628
>>106646681
>>106646816
Is it that bad? Are these images just cherry picked? I swear I remember reading about some peopling talking about waiting for Wan version here. Were they doing it for shits and giggles?
What about speed? It's activates are in low weight, so it should be fast at least?
>>
File: file.jpg (700 KB, 2159x1437)
700 KB
700 KB JPG
>>106646907
>look at those nice numbers
I posted images retard, you can say you don't believe in them, but that's a (You) problem
>>
>>106646963
>Are these images just cherry picked?
All latest models are benchmaxxed shit for muh numbers.
>>
>>106646963
The problem with this comparison was always that it's too basic, with a huge room for error in the image, you can fuck it up during inference a lot and as long as it's vaguely a book shop of books with correct words on it, it's good.

Gen a realistic crowd of different people of different clothes/races all holding different objects engaged in battle for example or other similar complex prompts, it will shit itself.
>>
>>106646927
do your own fork if you don't like how some ui details work, submit features as patches.

comfy did not nothing but quite a lot, see the commit log
>>
>>106646963
qwen nunchaku is ass.. blurry shit
>>
>>106646968
promo images are worthless, if its so good and you're using it, you would have posted an actual replicable comparison
>>
>>106646766
Probably a worth a test for qwen I guess.
FP16/FP8/Q8/Q4/Nunchaku
>>
>>106646992
well, go ahead, let's see comparison gens using that prompt
>>
any idea for a complex prompt?
>>
>>106646957
>>106646993
>just work for free fixing everything so that cumfart can take the credit
>>
>>106647019
>no u
you're the one claiming its good, so if you are not a liar you should already have it set up and ready to go? i deleted the trash when i tested it initially and saw that it was trash
the burden of proof is on you
>>
>>106646946
thanks nugga
>>
>>106647032
>MOOOOOOOOOOOOM CUMFART OOM'D AGAIN!
>>
>>106647037
you see anon it worked great when I tested it so I only kept the nunchaku version
who is right? how can we know?
>>
>>106647064
you dont need to install anything extra to use a full sized model, it works out of the box after 1 click download
>>
>make my own nodes by fixing other people's shit
>"share your workflow anon!"
>"wtf anon I can't find these nodes! Where are they? I can't use this shit help me!"
I find this a win desu
>>
>>106647081
it would take me so long to download the big fat Qwen tho, my internet is shit
you do it
>>
>>106646910
>you should use 720x128
their git recommends 704, i assumed because its divisible by 32
>you shouldnt go above 81 frames
not an issue for a lot of videos like a girl dancing, youd only get problems with panning shots or people walking from one place to another
>>
>>106647098
>their git recommends 704
where?
>>
File: 704.jpg (87 KB, 912x380)
87 KB
87 KB JPG
>>106647133
https://github.com/Wan-Video/Wan2.2
>>
>>106647143
that is only for the Wan2.2-TI2V-5B model, which is trash that nobody should use, if you dont have much vram just stick to 14b in wan2gp
>>
>>106647143
Many such cases! (im just jumping into this tard fight for fun)
https://chimolog-co.translate.goog/bto-gpu-wan22-specs/?_x_tr_sl=auto&_x_tr_tl=en&_x_tr_hl=bg&_x_tr_pto=wapp#1280%C3%97704%EF%BC%9AWan22_720p_%E3%83%99%E3%83%B3%E3%83%81%E3%83%9E%E3%83%BC%E3%82%AF
>>
>>106647201
>>106647201
>>106647201
>>106647201



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.