[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/g/ - Technology


Thread archived.
You cannot reply anymore.


[Advertise on 4chan]


Napkin Math Edition

Discussion of Free and Open Source Text-to-Image/Video Models and UI

Prev: >>106708328

https://rentry.org/ldg-lazy-getting-started-guide

>UI
ComfyUI: https://github.com/comfyanonymous/ComfyUI
SwarmUI: https://github.com/mcmonkeyprojects/SwarmUI
re/Forge/Classic/Neo: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassicneo
SD.Next: https://github.com/vladmandic/sdnext
Wan2GP: https://github.com/deepbeepmeep/Wan2GP

>Checkpoints, LoRAs, Upscalers, & Workflows
https://civitai.com
https://civitaiarchive.com/
https://openmodeldb.info
https://openart.ai/workflows

>Tuning
https://github.com/spacepxl/demystifying-sd-finetuning
https://github.com/Nerogar/OneTrainer
https://github.com/kohya-ss/sd-scripts/tree/sd3
https://github.com/derrian-distro/LoRA_Easy_Training_Scripts
https://github.com/tdrussell/diffusion-pipe

>WanX
https://comfyanonymous.github.io/ComfyUI_examples/wan22/
https://github.com/Wan-Video

>Chroma
https://huggingface.co/lodestones/Chroma1-Base
Training: https://rentry.org/mvu52t46

>Neta Lumina
https://huggingface.co/neta-art/Neta-Lumina
https://civitai.com/models/1790792?modelVersionId=2203741
https://neta-lumina-style.tz03.xyz/

>Illustrious
1girl and Beyond: https://rentry.org/comfyui_guide_1girl
Tag Explorer: https://tagexplorer.github.io/

>Misc
Local Model Meta: https://rentry.org/localmodelsmeta
Share Metadata: https://catbox.moe | https://litterbox.catbox.moe/
GPU Benchmarks: https://chimolog.co/bto-gpu-stable-diffusion-specs/
Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-one
Txt2Img Plugin: https://github.com/Acly/krita-ai-diffusion
Archive: https://rentry.org/sdg-link
Bakery: https://rentry.org/ldgcollage

>Neighbours
>>>/aco/csdg
>>>/b/degen
>>>/b/realistic+parody
>>>/gif/vdg
>>>/d/ddg
>>>/e/edg
>>>/h/hdg
>>>/trash/slop
>>>/vt/vtai
>>>/u/udg

>Local Text
>>>/g/lmg

>Maintain Thread Quality
https://rentry.org/debo
>>
File: IMG_5117.jpg (29 KB, 616x294)
29 KB
29 KB JPG
>>106711909
>>
File: elf-hugger_00945_.png (3.04 MB, 1088x1920)
3.04 MB
3.04 MB PNG
>>106711936
>>
File: mario kart 64 it ok.jpg (90 KB, 1024x768)
90 KB
90 KB JPG
>>106711936
>>
Blessed thread of frenship
>>
>>106711946
grabs your balls, twists really hard, then pulls on them like from that video of the dog ringing the church bell
>>
File: 1739604516256632.png (2.16 MB, 1728x1344)
2.16 MB
2.16 MB PNG
this thread has kino gens and documents progress in genAI from 2022 to now
https://la*nch*n.org/vis/res/9112.html
>>
WHERE THE FUCK IS THE FP8 SCALED VERSION OF QWEN IMAGE EDIT REEEEEEE
>>
>>106711955
There isn't one, least not that I can see on hf. Same with the original image edit model
>>
>>106711955
just quant it in the loader?
>>
>>106711954
neat
>4chanxt adds thread watcher to lainchain populated with threads from 4chan
heh
>>
File: 1748185553377901.png (3.05 MB, 1728x1344)
3.05 MB
3.05 MB PNG
>>106711955
learn2google
https://huggingface.co/Comfy-Org/Qwen-Image-Edit_ComfyUI/tree/main/split_files/diffusion_models
>>
>>106711967
isn't scaled the type that runs faster on 40 and 50 series gpus? i think it needs to be manually quanted
>>
File: qie.png (3 KB, 604x25)
3 KB
3 KB PNG
>>106711955
Is this the one you're looking for
>>
>>106711978
No that is one version of nunchaku quants
>>
>>106711974
>learn2google
it's not the 20th century any more granddad
>>
The lightning lora for qwen fucking blows, I knew it was too good to be true. Almost two minutes for a 20 step output using two image inputs is aids though...
>>
>>106712005
yeah we use askjeeves now you bellend
>>
>>106711936
napt is gone forever faggot
>>
File: 1738377528568-0.png (1.15 MB, 896x1152)
1.15 MB
1.15 MB PNG
>>106711954
>if you have an android phone with the palm gesture selfie feature, point your phone's camera at this picture
lmao this actually works
>>
>>106712013
>The lightning lora for qwen fucking blows
Depends on the task. Simply stuff like removing and object it's a decent timesaver. Most other things you don't want it.
>>
>>106712132
tried the multigpu distorch nodes?
>>
File: elf-hugger_00948_.png (3.26 MB, 1088x1920)
3.26 MB
3.26 MB PNG
>>106712015
your dotsafetensor file m'lud
>>
>>106709656
>>106712132
They're probably offloading. As a general rule, the file size of the model is the minimum amount of VRAM you'll need to at least load the model. Qwen Image Edit Q8 is 21GB. That's how much VRAM it needs to load. 16GB GPU users are 100% offloading with virtual_vram_gb.
>>
Why isn't there a true multi GPU solution for comfy UI yet? I have 2 3090s and one is always sitting idle when I could have them both blasting away of a job.

I know it's possible. Why hasn't anyone implemented it? And no I'm too dumb to do it myself.
>>
>>106712162
>offloading
They could have fucking said that last thread then, I thought something was wrong with my setup.
>>
File: 00308-1116538136.png (2.74 MB, 1248x1848)
2.74 MB
2.74 MB PNG
>>
>>106712177
>"so you bought a second 3090 thinking it would magically double your gen speeds"
lol
>>
>>106712189
If you didn't know that filesize is always the minimum vram load then that's on you, fucko.
>>
>>106712177
Is it possible? I thought that the sequential nature of diffusion via iterative denoising makes true multigpu impossible, or that the model would have to be designed for parallelism in the first place, and none are.
>>
>>106712262
>>106712177
Wasn't the multigpu node allowing you to use the vram of the second gpu, just not the the core?
>>
>>106712262
Far as I know, when it comes to SD, more than one GPU is only useful for batch runs or for training.
>>
>>106712162
Would someone be so kind as to share their workflow? I've been bashing my head against this for the for past 6 hours trying to 'fix' it but if it's working the way it's supposed to be working that would certainly change things. I've tried a couple different workflows I found but they all have the same issue with the raw vram size of the Q8 model
>>
>>106712223
>so you bought a second 3090 thinking it would magically double your gen speeds
I never said that. I use it for LLMs.
>>
Realistically couldn't you send half of the latent to each gpu and then just stitch them together?
>>
I need to griftMAXXXX give me your worst /ldg/!
>>
>>106712330
ask ranfaggot
>>
In comfy is there a way to make sure two text prompt nodes have the same text?
>>
>>106712330
InvokeAI supports v-pred, chroma, Flux, Qwen, and some workflows, but it's still being ignored.
>>
>>106712336
Just plug a string node into the input of each text box?
>>
File: 1758793850791140.png (31 KB, 707x574)
31 KB
31 KB PNG
>>106712336
>>
>>106712347
>>106712357
thanks
>>
Is there any ComfyUI function that lets you merge multiple nodes? Like I have 4 VAE nodes, can't I compact them somehow? Or compact the model loading nodes into some kind of rack? So I end up with a neat little tower of load checkpoint, upscaler, VAE, LoRA, all compact?
>>
>>106712364
>Like I have 4 VAE nodes
just use one??
>>
>>106712364
Right click, convert to subgraph.
>>
im confident enough to not need samples anymore btw if you cared
>>
>>106712367
I'm a power user??????
>>
>>106712295
Just use the default comfy workflow for 2059 that you can find in comfy's browser. Q8 gguf uses slightly more VRAM than fp8_e4m3fn, so if you swap out the loader, you might need to offload very slightly even on 24GB (maybe 2-4, it'll barely affect speed).
I use fp8_e4m3fn. Even with three input images, VRAM usage maxes out at 23GB at 1 megapixel output.
>>
When I'm using Comfy is there a way to make sure the cock is lubed up before it goes into my ass? It hurts going in dry.
>>
>>106712385
>VRAM usage maxes out at 23GB at 1 megapixel output.
Exactly 23 or 23.5+? It offloads to shared memory if that happens yeah?
>>
>>106712373
Thanks Comfy. How are the investor shares going?
>>
>>106712398
Between 23 to 23.1 with three inputs. It never goes over 23.5, otherwise it'd go at a snail's pace or OOM. A single image input is 21.9. That's fp8_e4m3fn.
>>
>>106712357
Can you turn that into a list of strings and have it pick randomly, without needing third party nodes ?
>>
File: file.png (34 KB, 481x310)
34 KB
34 KB PNG
>>106712385
How do you offload? I asked AI but it gives me node options that don't exist
>>
>>106712431
Wrong loader. You need UnetLoaderGGUFDisTorchMultiGPU.
>https://github.com/pollockjj/ComfyUI-MultiGPU
Increase virtual_vram_gb.
>>
File: IMG_5119.jpg (164 KB, 1179x1387)
164 KB
164 KB JPG
>>106711940
>>
>>106712425
There are wildcard nodes but idk if they are core
>>
>>106712435
Thank you anon, it works. It takes 230 seconds for a single 1k~ image though, which seems like a long time for a single frame (compared with Wan doing 101 frames in 330 seconds sometimes)
>>
File: 1738060773076544.png (342 KB, 1564x801)
342 KB
342 KB PNG
>>106712431
I use q8 qwen edit v2 on a 4080 and thats 16gb, it should use some regular ram if it doesnt load fully into vram

all I swapped from the original workflow was the positive prompt node, didnt change the other one (dont use negative prompts for it)
>>
>>106712425
"A photo of {tag1|tag2|tag3|tag4}, high quality".
>>
>>106712471
I've got a 4090, and 25 steps at 2.5 CFG takes a little under a minute. Yeah, it's a slow model to run. Shame sage doesn't work with it.
>>
File: reiwalk.webm (3.3 MB, 1080x1428)
3.3 MB
3.3 MB WEBM
>>
File: storyboard Final_00008_.png (248 KB, 1280x720)
248 KB
248 KB PNG
how is it inspyrenet still the goat at rembg?
>>
>>106712481
You don't get the actual end prompt saved with the image with this method, also it's really bad when you want to have a large selection

It's weird how such basic functionality isn't available in core nodes, like just picking random strings from a list and sending to prompt
>>
>>106712471
The more input images you use, the slower it is as well. The lightning lora massively speeds it up, but it pretty much lobotomizes the model for anything but the simplest of prompts
>>
>>106712493
Have you tried toonout
>>
>>106712494
>You don't get the actual end prompt saved with the image with this method
You need third party nodes for that, ie ImpactWildcardProcessor. That saves both the wildcard prompt and the final prompt it picks.
>>
>>106712493
I find using flux kontext (and probably qwen image edit too) better, but slower obviously. Inspyrenet does great with single subject but if there's multiple subjects in the foreground it starts to fall apart.

>>106712510
Gonna check it out.
>>
>>106712483
use this and set it to 8 steps, 1 cfg:

https://huggingface.co/lightx2v/Qwen-Image-Lightning/blob/main/Qwen-Image-Edit-Lightning-8steps-V1.0.safetensors

in general it takes like 20-30 seconds or so. same result but faster with the lora. and you can use other loras if you want with it.
>>
File: 00165-2711366108.png (2.28 MB, 1536x1536)
2.28 MB
2.28 MB PNG
>>
File: catbox_l716l4.png (1.19 MB, 1152x896)
1.19 MB
1.19 MB PNG
>>106708528
>>106708856

A couple anons asked for the catbox to this image, here's the base before upscale and inpaint.
>>
>>106712494
you just attach another "show text" node to it then u can see what is picked
>>
>>106712510
Interesting

>>106712525
yea the speed is kinda krazy
>>
quen edit lightning v1 or quen image lightning v2 or edit v2 gens? I wonder how they compare, need to test.
>>
>>106712580
*for edit gens even
>>
the man is pointing to a laptop he his holding with his right hand, with the text "LDG" on the screen.

Qwen-Image-Lightning-8steps-V2.0-bf16.safetensors

seems fine
>>
>>106712597
looks great
>>
File: 1730063868860080.png (1.12 MB, 896x1160)
1.12 MB
1.12 MB PNG
>>106712597
helps if I attach the image.
>>
>>106712020
big if true
>>
File: 1751621902538549.png (1.13 MB, 896x1160)
1.13 MB
1.13 MB PNG
>>106712603
and this is with edit v1.

jacket seems better with the v2?
>>
File: 00450-2621792877.png (2.66 MB, 1248x1848)
2.66 MB
2.66 MB PNG
>>
File: esgbeabeabae.png (44 KB, 478x756)
44 KB
44 KB PNG
How is this for flux ultimatesdupscale settings? it still seems a bit grainy, like i could do more steps but it may also be a low denoise setting, not sure.
using flux1devnsfwunlockedQ8
>>
>>106712570
Thumbs up!
>>
>>106712620
0.25 denoise is too low. I don't use flux anymore, but I'm pretty sure it has tile controlnets for upscaling. With them, you could push 0.6-0.7 denoise without large changes to the composition. Without them, use 0.45-0.5.
>>
File: 1750828550574623.png (1.07 MB, 1024x1024)
1.07 MB
1.07 MB PNG
the girl in image1 is wearing the outfit of the girl in image2.

pretty clean haruhi. yeah, the 2.0 regular qwen image lightning might be better than 1.0 qwen edit lightning lora. need to test more though.
>>
File: 1742120271577714.jpg (789 KB, 1552x1944)
789 KB
789 KB JPG
>>
File: 1757683384535833.png (995 KB, 1024x1024)
995 KB
995 KB PNG
>>106712643
yep it seems better. using: Qwen-Image-Lightning-8steps-V2.0-bf16.safetensors

gonna try with the full size one to compare. but, it's consistent and *seems* a bit better than the v1 edit specific lora.
>>
File: 1731517927925558.png (1 MB, 1024x1024)
1 MB
1 MB PNG
>>106712666
>>
File: 1743806431390553.png (801 KB, 744x1400)
801 KB
801 KB PNG
yeah, 2.0 lora def works better than the 1.0 edit lora cause it wasn't getting this same prompt right before.

actually got the plugsuit on elegg proper this time.

Qwen-Image-Lightning-8steps-V2.0-bf16.safetensors > Qwen-Image-Edit-Lightning-8steps-V1.0.safetensors, for now.
>>
File: 1748792157409679.png (763 KB, 744x1400)
763 KB
763 KB PNG
>>106712693
and this with Qwen-Image-Lightning-8steps-V2.0.safetensors (full size)

kept the midriff from the original this time.
>>
>>106712693
Yeah, the old 2.0 for regular Qwen Image works better than the new one specifically for Edit, at least in my tests. It's not even close. How'd they fuck up the new lora so badly?
>>
File: file.jpg (797 KB, 1536x2048)
797 KB
797 KB JPG
UwU
>>
>>106712781
idk, not sure how they would train for an edit model specifically, but the regular qwen image lora should work on all images so it should work with any edit prompt too i'd guess?
>>
File: elf-hugger_00026_.jpg (940 KB, 1088x1920)
940 KB
940 KB JPG
>>106712782
>>
RNGing a few gens with euler sigmoid offset with QIE 2509 seems to give better results than the default euler simple every few gens.
LCM can also RNG into a good result but it's much worse most of the time so not worth.
res_2s bong_tangent can also be better but it's hard to say.

But it's unfortunate that there is so little difference in most of the sampler space nowadays, the main one that can be good that is quite different is res2s bong.
>>
>>106712797
>https://huggingface.co/lightx2v/Qwen-Image-Lightning/discussions/17#68c3e582372a5302cd062d88

They talk about it there.
>>
File: 00008-3695119619.png (2.75 MB, 1824x1248)
2.75 MB
2.75 MB PNG
>>
File: 1740453896977345.png (968 KB, 744x1400)
968 KB
968 KB PNG
the girl in image1 is wearing the outfit of the girl in image2.

cute! yeah im definitely sticking with the 2.0 reg qwen lightning for edit. it's more consistent and has better detail.

like for this 2b test, it wasnt getting the dress on properly it was doing a hybrid with the old lora, and was less consistent.
>>
>>106712638
sent a response earlier and it seems 4chan ate it.
thanks. though my upscales are still a little sub par. guess ill fuck with it more in the morning.
base
https://files.catbox.moe/1blh4o.png
https://files.catbox.moe/et2v65.png
upscale
https://files.catbox.moe/d11o1b.png
https://files.catbox.moe/jlmmyh.png

>>106712812
sigmoid balls
>>
>>106712820
is that from a kingdom hearts CG style lora? looks really really fucking good man.
>>
File: 1758434310133334.png (974 KB, 744x1400)
974 KB
974 KB PNG
>>106712822
chun li alpha 3 test:
>>
>>106712830
anon, pls
>>
File: 1750080922240833.png (1014 KB, 744x1400)
1014 KB
1014 KB PNG
>>106712839
chun dress alt, turned out really good imo:

2.0
>>
for the loras, is there notable difference between bf16 and full size versions? I assume the larger file is better and the smaller one is making some compromise or is missing some data.
>>
>>106712823
Try using Chroma as your upscaler at 0.45 denoise.
>>
File: 00014-4281019147.png (2.73 MB, 1248x1824)
2.73 MB
2.73 MB PNG
>>106712842
here is catbox
https://files.catbox.moe/zo7120.png
https://civitai.com/models/507472?modelVersionId=1880913
https://civitai.com/models/1585110/kairi-kingdom-hearts-1-2-illustrious
>>
>WAN
>want a fully clothed character to sway butt around calmly
>gen shits itself after 1 second and does random shit like a lunatic on drugs suddenly a dick comes out of vagina and pierces through the clothes
How do you make these hallucinations less mentally deranged?
>>
>>106712913
just roll with it bruh
>>
File: 1584566781988.png (61 KB, 212x218)
61 KB
61 KB PNG
>qwen image edit consistently tones down breast and ass size when swapping outfits
>>
>>106712925
say large or huge, it will listen
>>
>>106712925
Gotta say keep her breasts very huge, but it will still nerf them depending on the dataset because some full body outfits didnt have anyone with big tits in them
>>
File: 1756666383552597.png (1.37 MB, 992x1048)
1.37 MB
1.37 MB PNG
a bit unfair cause peripheral vision will also make you spot the sign right away, but...
>>
File: 1732076662736813.png (1.29 MB, 992x1048)
1.29 MB
1.29 MB PNG
imagine 20 years ago someone said "you can tell a computer to remove clothes and it will do it".

and yes, the 2.0 lora is more effective.
>>
File: 00021-4079227763.png (2.19 MB, 1824x1248)
2.19 MB
2.19 MB PNG
>>
File: 1754596522385279.png (1.42 MB, 992x1048)
1.42 MB
1.42 MB PNG
>>106712976
and for a silly prompt to test the model...

the woman is wearing a suit of white gundam armor.
>>
File: 1736467462498996.png (1.28 MB, 992x1048)
1.28 MB
1.28 MB PNG
replace the red shirt of the woman in image1 with the outfit of the girl in image2. she has large breasts. keep the expression of the woman in image1 the same.
>>
File: 1741908882382593.jpg (1 MB, 1744x1744)
1 MB
1 MB JPG
>>
File: 1753913428631976.png (1.33 MB, 896x1160)
1.33 MB
1.33 MB PNG
>>
File: 00560-1165251786.png (2.8 MB, 1248x1848)
2.8 MB
2.8 MB PNG
>>106713064
love these, on my wave length
>>
File: 1733019838524676.png (1.13 MB, 832x1248)
1.13 MB
1.13 MB PNG
the woman in image1 is waving hello with the anime girl from image2, seated beside her.
>>
File: 00038-1390071644.png (2.61 MB, 1152x2016)
2.61 MB
2.61 MB PNG
>>
File: 1757698737148485.mp4 (2.6 MB, 720x1040)
2.6 MB
2.6 MB MP4
>>106712902
>>
File: 1751438861375986.png (1.03 MB, 768x1360)
1.03 MB
1.03 MB PNG
>>106713094
the woman in image1 is waving hello with the girl from image2, sitting beside her.

endless possibilities with edit, then you have wan to animate stuff.
>>
File: 00044-3636589716.png (2.57 MB, 1152x2016)
2.57 MB
2.57 MB PNG
>>
File: 1750202504434983.mp4 (1.82 MB, 720x1264)
1.82 MB
1.82 MB MP4
>>106713117
>>
>>106713212
nice
>>
File: 1748346177933791.png (1.21 MB, 768x1360)
1.21 MB
1.21 MB PNG
the woman in image1 is wearing the outfit of the girl in image2.

tried swapping the outfit of ann from persona. pretty neat what you can do with these models desu. even if I wanted to do this with inpainting it'd be very hard if not impossible to do, even with controlnets.
>>
File: 1746077742940962.png (1.09 MB, 768x1360)
1.09 MB
1.09 MB PNG
a gigantic version of the girl is lying on her side at the beach, looking down at a tiny man in plate armor holding a sword. she has very large breasts.

meanwhile in anor londo:
>>
File: 1754160078131707.png (1.21 MB, 768x1360)
1.21 MB
1.21 MB PNG
>>106713250
almost looks like the demons souls armor, kek
>>
File: 1732836008243062.jpg (1.03 MB, 1416x2120)
1.03 MB
1.03 MB JPG
>>106713087
original artist is good https://x.com/tsunosame/media
>>
>>106712925
You need to tell it to keep the proportions and pose exactly the same, especially if the target outfit is on someone with smaller TNA. Use a more detailed prompt.
>>
File: 1739664696125460.mp4 (557 KB, 384x672)
557 KB
557 KB MP4
>>106713253
>>
File: 1738519962851156.jpg (827 KB, 1416x2120)
827 KB
827 KB JPG
>>
Is the interpolation inside comfyui worth it?
>>
>>106713259
>https://x.com/momomo_906
reminds me of this dude. Ain't really the exact style but also it is the same aesthetic. working on a lora, I feel like it will be hot
>>
>>106713332
good for wan gens, 16 to 32 fps with film vfi is much smoother
>>
>>106713332
FILM is still the best in my opinion. I'm saying that after trying every other option, including Topaz Video's commercial interpolation.
Only downside is that it's quite slow.
>>
>>106713341
nice find. do share gens when you finish it
>>
File: 1754501588512867.mp4 (805 KB, 640x480)
805 KB
805 KB MP4
>>
File: 1756733428184349.mp4 (1.29 MB, 640x480)
1.29 MB
1.29 MB MP4
>>
Wait, do I need this Get Image node..? I think I kept it from another workflow.

>>106713343
>>106713355
I think I can spot a small difference between rife and film vfi. Also the interpolation helps a lot with the too fast movements I often get.
>>
>>106713456
RIFE is fast, but it had problems with fast motions from what I recall, artifacts and some blur. There's videos on youtube comparing various interpolation methods that show it.
>>
File: 1732163714530647.mp4 (978 KB, 672x480)
978 KB
978 KB MP4
but...I worked for blizzard...
>>
>>106713483
I wonder how history would be different if he didn't just roach out like that in diremaul.
>>
File: 1744421496897271.png (829 KB, 848x1232)
829 KB
829 KB PNG
the girl in image1 is wearing the outfit of the girl in image2.
>>
File: 1736874533699596.png (698 KB, 744x1392)
698 KB
698 KB PNG
>>106713516
remove the black dress of the girl in image1. the girl in image1 is wearing the outfit of the girl in image2.

2b, but casual clothes ann from persona 5:
>>
File: 00226-2501199323.png (1.98 MB, 1248x1848)
1.98 MB
1.98 MB PNG
>>106713568
>completely change the style into something interesting
>an alcoholic or something, move past a generic img2img of the prior input
man ai is amazing
>>
File: 1737679356288766.jpg (546 KB, 1024x1504)
546 KB
546 KB JPG
I've never made an honest effort at attempting to solve wildcards even though I've always wanted to since that cathode anon taught me about nestled wildcards. Even before that I wanted to make some but they've always been half assed.
>>
>tfw I figured out Set and Get nodes all by myself
>>
File: 1730024829010723.jpg (523 KB, 1024x1504)
523 KB
523 KB JPG
>>106713609
but I'm gonna do it right now anons
I'm making some wildcards!!!!
>>
>>106713609
>>106713630
I have ones for facial expressions, composition, poses, and focus
>>
>>106713629
We're all very proud of you.
>>
>>106713638
Thank you.

How can I make them smaller though? Right now it shows the Constant as well as the title. It's like 3 times larger than it needs to be.
>>
File: 1758120299887804.jpg (693 KB, 832x1216)
693 KB
693 KB JPG
>>106713635
Yeah I've got a bunch like that too but the possibilities should be in the zillagorillions, with many layers of nestled concepts that work. So then I can just use 1 wildcard and have it produce infinity unique characters. And right now I'm thinking about equipment rather than those things.

For focus do you mean like close-ups and stuff like that?
>>
>>106713650
All the tags under "Body parts" https://donmai.moe/wiki_pages/tag_group%3Afocus_tags
Close-up would fall under composition https://donmai.moe/wiki_pages/tag_group%3Aimage_composition
I use all four cards except when I'm dialing in the look
>>
File: 1758655333593780.jpg (819 KB, 1024x1504)
819 KB
819 KB JPG
>>106713703
>armpit focus
>ass focus
naruhodo
fine taste
>>
File: ComfyUI_01432_.png (1.4 MB, 1008x1448)
1.4 MB
1.4 MB PNG
>>
File: ComfyUI_01433_.png (2.01 MB, 1120x1120)
2.01 MB
2.01 MB PNG
>>
>>106713781
>>106713781
>My boob Is coo 'me aod it tierte...
>>
File: ComfyUI_01434_.png (2.33 MB, 1368x912)
2.33 MB
2.33 MB PNG
>>
How the fuck are people doing first frame last frame long videos without any degredation in quality?
>>
>>106713862
context nodes.
>>
File: ComfyUI_01438_.png (966 KB, 1024x1024)
966 KB
966 KB PNG
>>
>>106712214
nice
>>
>>106712386

ber and chiken wings?
>>
I think I'm done with this hobby. Now that qwen has pulled out of open source we're basically dead in the water for the next 4-6 months.
>>
>>106712486

who are you if rei talks to you
>>
>>106713936
Chroma slow gen speed is killing me
>>
File: 1727919491447146.png (1.7 MB, 1024x1024)
1.7 MB
1.7 MB PNG
https://xcancel.com/TencentHunyuan/status/1971805621738700947#m
>omg guyz it can do text
I hope it does, it's a fucking 80b model
>>
>>106714007
I can't stand how they're hyping this model up. It looks so slopped it's unreal. It seems like diminishing returns stack out of control after 12B.
>>
File: loooool.png (1.28 MB, 1373x1303)
1.28 MB
1.28 MB PNG
https://xcancel.com/jampani_varun/status/1971599630187258024#m
Who wants a distilled SD3.5?? please clap...
>>
>>106714023
it's more like they still don't understand that using a shit synthetic slopped training dataset to your model will only produce a slopped model, garbage in, garbage out
https://en.wikipedia.org/wiki/Garbage_in,_garbage_out
>>
>>106714026
SAI the underdog completely assblasting the competition. Xi quakes in feat.
>>
>>106714026
I actually feel anger over the fact they are still around at this point.
>>
File: ClipartKey_1972038.png (505 KB, 698x673)
505 KB
505 KB PNG
Best stable diffusion front end to install on linux?
>>
>>106714026
>SketchX
>https://huggingface.co/SketchX
>The ultimate vision for SketchX is to understand how seeing can be explained by drawing. In other words, how better understanding of human sketch data can be translated to insights on how human visual systems operate, and in turn how such insights can benefit computer vision and cognitive science at large.
>SketchX has been actively investigating all aspects of sketch research since 2012. Problems we study range from conventional tasks such as sketch recognition and sketch synthesis, to those pioneered within, e.g., fine-grained sketch-based image retrieval and memory-aware forensic sketch analysis.
Inner essing.
>>
>>106713793
lmao
>>
>>106713793
>>106713807
ngl, she's really beautiful, in another universe she would've thrived as a model, not as a fucking cumdump...
>>
>>106714023
I honestly don't care about slop as long as prompt following is improved. Slop can be finetuned away prompt following can't be improved unless you spend a lot of money.
>>
>>106714082
>Slop can be finetuned away
I've heard this shit since the release of Flux more than a year ago, can this meme just leave? It's obvious no has any idea how to bring back sovl to a model, I hope you can transform flux into midjourney but good luck for the one that manages to do that, there's a furry fuck that blasted his teeth putting 140k dollars in it
>>
>>106713355
film shifts colors a little compared to topaz but its not a big problem
>>
>>106714077
>ngl
>>
>>106714007
>That lettering
Their datasets are poisoned by openai watermarking beyond repair, they should throw them all away and start from scratch.
>>
SDXL forever
>>
>>106714077
I agree. She is extremely beautiful and unique as well. And it's not like a conventional beauty either. She's uniquely beautiful.
>>
>>106714119
>they should throw them all away and start from scratch.
true that, and next time they should go for a 200b model, just one more billion parameter bro...
>>
>>106714119
>Their datasets are poisoned by openai watermarking beyond repair
I don't get China, they have hundreds of millions of slaves ready to annotate captions proprely, and they're as lazy as as your average lora maker by using synthetic slop to finetune their shit, that's crazy
>>
>>106714082
lol, just like all those millions of unslopped hidream and hunyuan 2.1 finetunes, right? chroma furry wasted 150k and still had to train at 512x512 because of funds. nobody is finetuning this 80b shit
>>
>>106713296
me in the armor
>>
>>106714007
nice calendar lmao
>>
>>106714082
>Slop can be finetuned away
Yes and I usually I would agree, but 80B parameters at this point. There is a point where the model gets frankly unrealistic to expect some rando to finetune.
I think we crossed that long ago before or around qwen.
Even if it can be fine tuned it's way too big for anyone to reasonably expect a decent fine tune to just happen.
>>
>>106714168
>There is a point where the model gets frankly unrealistic to expect some rando to finetune.
>I think we crossed that long ago before or around qwen.
I'd say even for Flux it's ridiculous, look at lodestone, he cut some shit to get a 9b model, and he spent 150k to "save" it (that took him half a year and he failed)
>>
>>106714179
Exactly. It's unreasonable. I don't think I could even train LoRAs for this hunyuan model and I have 2 3090s. Qwen was a stretch at a decent rank and resolution.
>>
>>106714154
It looked like they hacked the base model, but things only got a bit better, and in the end it still couldn’t overcome the potent poison the Flux developers planted for open-source use.
If you try a model without that kind of poison, like Wan2.2, you’ll see how compliant it is.
That said, there’s a high chance the core of his kink is satisfied. I’m not interested in that part, so I haven’t tested it.
>>
>>106714194
>If you try a model without that kind of poison, like Wan2.2, you’ll see how compliant it is.
China is able to make good video models only because there's not enough synthetic video slop to poison it, so they had no other choice but to work hard and caption real videos, at some point I expect them to be lazy on that front as well and use veo3's video to make their next models, this is grim lol
>>
File: 1751556544883263.png (1.34 MB, 1024x1024)
1.34 MB
1.34 MB PNG
>>106714007
>Prompt 4: Solve the system of equations 5x+2y=26, 2x-y=5, and provide a detailed process.
I don't think people realize how much VRAM it's gonna eat, not only it's a 80b model, but it's an autoregressive model lmao
>>
>>106714211
It's already starting. Look at 2.5, the outputs look... bad?
>>
>>106714219
>not only it's a 80b model, but it's an autoregressive model lmao
Maybe it will just be a bit more intensive than an 80b LLM
>>
Chinabros...
>>
>>106714234
>Maybe it will just be a bit more intensive than an 80b LLM
Sounds awful.
>>
>>106714230
when you listen to the sound of wan 2.5 it sounds like the compressed sound of veo 3 (which is already a compressed sound) so yeah, they're already poisoning their models with slop, Alibaba and Tencent will go nowhere until they decide to make effort on a real dataset once and for all
>>
>>106714094
Look just cause you like gpt piss filter and blur dreams or MJ neon retro forever doesn't mean everyone does. Fucking hate those models and their gens with a passion, Lora+ any local model looks tons better.
>>
>>106714252
I mean people run deepseek r1 without issues you just need to quant it
>>
>>106714264
>just one more lora bro
that sounds more like cope than "just see you guys, some guy will fall into the sky and finetune Flux into the next Seedream" desu
>>
>>106714267
Running something and running something at speed that makes it worth it aren't the same thing.
>>
>>106712902
thank you anon you're a leggendd
>NEW kairi lora
ah that also explains why it looks so good.
>>
File: 1741430793227348.png (112 KB, 2063x539)
112 KB
112 KB PNG
>>106714234
>Maybe it will just be a bit more intensive than an 80b LLM
just look at that autoregressive 7b model, now imagine you're multiplying that with 10
https://github.com/Alpha-VLLM/Lumina-mGPT-2.0
>>
>>106714269
I fucking hope not, if they made flux into blurdream fest I would be livid. I want actual details on my gens not sand grain
>>
File: Seedream 4.0.jpg (2.78 MB, 2868x3824)
2.78 MB
2.78 MB JPG
>>106714287
>sand grain
what grain?
>>
Here's what happens next
>Here's the grain
>Uhm actually here's a shot from a dslr camera it just looks like that.

I'm from the future you see.
>>
>>106714291
>shows blurry example to disprove how blurry it is
Damn anon...you got me
>>
>>106714300
>anon learns about bokeh on cameras
>>
>>106714291
this was my 9/11
>>
>>106714310
>>106714298
Like clockwork lmao
>>
>>106714315
I mean, is he wrong though? that's how cameras behave, now you're blaming the model from emulating cameras well?
>>
>>106714324
One might argue it overdoes the effect making people look overly fuzzy.
>>
>>106714298
>dslr camera
I'll take dlsr camera images over synthetic slop on the dataset, yes please
>>
>>106714331
No just people we had a dude spam painting gens and it had the same fuzziness, the overtrained on this weird bokeh effect.
>>
>>106714334
Looks like hdr slop instead dlsr photo
>>
>hdrslop
>dlsrslop
>syntheticdataslop
>censorshipslop
>slopslop
i'm tired boss
>>
Some anons will forever have the "grass is greener on the other side" syndrome
>>
File: 00117-1438135486.png (1.83 MB, 1536x864)
1.83 MB
1.83 MB PNG
>>106714354
but what anon doesn't know is that is true!
>>
>>106714354
Ambition will always be a good thing, that's humanity in a nutshell. If we had settled for mediocrity, we never would have gotten past inventing fire.
>>
>>106714298
>I'm from the future you see.
Based time travelerbro what the the future for local models look like
>>
>>106714349
This post is pure tiredslop
>>
>>106714354
let's not pretend we're nitpicking shit from almost flawless models, they are far from perfect, and I'm being nice here
>humans look like plastic
>the bokeh is too strong
>it only knows Miku and Trump
>it knows 0 artist styles
>doesn't know NSFW
there's a lot to complain about
>>
>>106714375
you are what you slop
>>
>>106714282
make good use of it. same lora creator made other KH girls.
https://civitai.com/collections/10035609
>>
>>106714349
>>106714375
>>106714381
I'm getting complainslop fatigue.
>>
>>106714370
SDXL is still the main model for anime.
Wan 2.2 gets another official tune from Qwen, I'll let you guess the name, but we don't get 2.5.
Hunyuan image 3.0 is posted about like twice on reddit before never being mentioned again.
>>
File: 00026-2513190305.png (3.11 MB, 1824x1248)
3.11 MB
3.11 MB PNG
>>106714354
coincidentally enough, the same anons also have mile wide anuses. Many such cases!

>>106714369
>>106714375
>>106714381
Very true(slop)!

>>106714383
>he did an aqua one as well
looks like im going into an early AM's cum coma
>>
>>106714388
>Hunyuan image 3.0 is posted about like twice on reddit before never being mentioned again.
the hype is already dead, they already know it's a 80b model kek
>>
>>106714391
ToT
>>
this is why chroma is the best, everything else is slop
>>
>>106714388
Oh I forgot. A T2I model will be released with native controlnet support preprocessor. The catch is that it will work like wan 2.2. There will be a low res model that convers broad stroke details of the image of the model trained on low resolution images with detailed tags that will go to a second model that will then refine the output at a higher resolution.
>>
>>106713894
Your blender looks buggy
>>
File: do it.png (1020 KB, 1160x900)
1020 KB
1020 KB PNG
>>106714412
>Oh I forgot. A T2I model will be released with native controlnet support preprocessor.
ok, tell us more
>>
File: 00035-1147290342.png (3.08 MB, 1824x1248)
3.08 MB
3.08 MB PNG
>>106714406
heres a better one with non fucked fingers (forgot to turn on detailer)
https://youtu.be/QJlBRjPVWqA?si=ZXZWYOUYUi8yYGG-
>>
>>106714419
Not much more to explain it's trained on millions of lower resolution images focused on image composition and a second model focused on higher resolution textures and details.
>>
>>106714412
>>106714439
like Stable Cascade?
>>
>>106714444
Yeah sure, why not. Nice digits.
>>
>>106714409
I wish it was faster
>>
>>106714455
same for you, checked
>>
>>106714427
>heres a better one with non fucked fingers
She has three hands
>>
File: the sovl is gone.png (1.22 MB, 1080x906)
1.22 MB
1.22 MB PNG
>>106714409
chroma lost its sovl at v30 when he decided to go for some low steps mumbo jumbo to please his discord vramlet cult
>>
>>106712177
>I know it's possible.
>I'm too dumb.

really makes you think.
>>
File: 00044-2364833312.png (2.98 MB, 1824x1248)
2.98 MB
2.98 MB PNG
>>106714464
i am going to kill myself for not noticing the extendo hand god DAAAAAAAAAAMN
>>
>>106714468
can't you just keep using v29?
>>
>>106714381
yep. my measure of a model is how well it trains and how little its innate bias influences training.
complainers are the same everywhere. what they're really upset about is their often self-imposed lack of control.
>>106714468
v29 still exists. chroma has 50+ sovls.
>>
>>106714485
v29 has sovl but the anatomy is fucked, that was the point of going for 50 epochs, to get images that got the anatomy right (and it never happened lol, I guess he needed more epochs and training on bigger resolutions but he hasn't the compute for it so I can't blame him for that)
>>
I'm glad we're allowed to talk about Chroma and mention it's also deeply flawed in the same breath now.
>>
>>106714490
>chroma has 50+ sovls.
it has 29 sovls, after that it's as slopped as Flux Schnell (which is ironic because that was his goal to get rid of the flux slop)
>>
File: 00059-3098973751.png (3.16 MB, 1248x1824)
3.16 MB
3.16 MB PNG
>>106714490
>>106714491
>>106714505
so did chroma actually git gud or is it still a meme?
>>
you might want to readjust you adetailer prompts and negative prompts. Also be specific with hand placement of arms and hands to reduce chances of deformities.
>>
File: 1744205122358221.png (690 KB, 1200x1565)
690 KB
690 KB PNG
>>106714137
>hundreds of millions of slaves
and they better use them now before it's too late
>>
File: ComfyUI_18620.png (3.44 MB, 1152x1728)
3.44 MB
3.44 MB PNG
>>106714026
>distilled SD3.5
I didn't use it beyond playing with it for a day, but I don't recall it being that heavy to where it would require low-step distillation (I do have a 4090 though).

>>106714468
Wouldn't he be able to easily go back and branch off of v29?
>>
>>106714664
>Wouldn't he be able to easily go back and branch off of v29?
he doesn't have money anymore so no
>>
>>106714310
>>106714324
thats not how cameras behave, luddite kid, thats how post processing contrast boost on dogshit cameras behave to make the shot more dramatic instead of actually realistic, retard
>>
>>106714679
>thats not how cameras behave
you're wrong
>>
OP is a faggot

why did you remove wan21 info?

https://rentry.org/wan21kjguide

while your wan2.2 info is incomplete
>>
>>106714705
Use case of wan 2.1 over 2.2?
>>
>>106714490
>my measure of a model is how well it trains and how little its innate bias influences training.
a great model doesn't need to be trained after release, change my mind
>>
>>106714719
>a great model doesn't need to be trained after release
This is true, but it also doesn't exist so you're using an impossible and as of yet unfulfilled standard to judge a model.
>>
>>106714683
thanks for that great argument and confirming that you are a low iq subhuman kid, cheers
>>
>>106714725
model trainers lack ambition. it's just benchmark chasing for funding, that's it. if they had incentive or balls they would just make a good fucking model
>>
yu rack disapren shamefur dispray
>>
Weirdly enough
I used the new Qwen edit 2059 base and it worked even tho I only have a 12 gig laptop 4080
>>
>>106714735
When that model exists I'll agree. Until then we have to live with good enough and trainable.
>>
>>106714729
>great argument
what's your argument instead of saying "nuh uh" saar?
>>
>>106712913
why would you just make that up?
>>
File: ComfyUI_01554_.jpg (413 KB, 1664x2432)
413 KB
413 KB JPG
I love these dolphin jumscares when genning for dolphin shorts lol
>>
>>106714759
my argument is human eyes look at reality without an insane post processing contrast boost like your dogshit slopped model has that cant be removed
>>
>>106714765
I wish it was just made up bs.
Wouldn't have to complain here about this random shit generator.
>>
>>106714713
absolute destitutefags swapping models off an HDD?
>>
>>106714804
loli aside, do women know dolphin shorts are arousing as fuck?
>>
>>106714362

GO BACK TO YOUR CONTAINMENT GENERAL
>>
>>106714840
yes they know it.
they also know when their nipples are peaking through their clothes.
they also know when you can see their pussy lips through their skin tight yoga pants.
>>
>>106714914
god i love to hatelove women so much it's unreal
>>
>>
>>106714969
did you genn for the plate to be pushed down by her tail?
>>
>>
>>106714719
>a smart boy doesn't need an education, change my mind
we can always improve and specialize. if you've never hit the limits of a model or text encoder, that's a fault of the wetware.
>>
When Chroma was the only option, I thought it was great, but then Wan2.2 came out and it became clear Chroma was basically doomed to fail at training—it only made an impression as an open-source “poisoned” model and, in practice, it can’t compete at all with the “normal” success you get by training from a clean, plain Wan2.2. Ironically, Chinese companies then started releasing “normal” base models one after another, and the Chroma guy ended up stuck, unable to move on from his own creation. That said, there still aren’t many models that handle /h/, so in that sense Chroma is among the top.
>>
>>106715060
feet?! bruh.. when the crotch is right there
>>
>>106714007
>>106714026
GOD JUST PAY FOR AN IMPROVED DATASET ALL THIS SHIT LOOKS THE SAME, QWEN, HUNYAN, THIS.
>>
>>106715060
T5 is trying to keep their panties from showing.
>>
>>106715098
>GOD JUST PAY FOR AN IMPROVED DATASET
this, why the fuck are they wasting tens of millions for a gozillion parameters model, if the dataset is shit, the model will output shit, it's basic stuff
>>
>>106715060
how is this in any way appealing?
you genuinely need to have something different in your brain to like feet baka.
>>
>>106715114
nta but I just like the upskirt part when it comes to feet stuff.
>>
>>106714804
KYS
>>
File: 1738024090573023.png (93 KB, 359x331)
93 KB
93 KB PNG
>>106715114
idk dude, women's feet make my pee pee hard, it is what it is
>>
>>106715106
it's still early, yet. sometimes we have to make something to see what happens. for instance, does training always scale linearly with resources?
>>
File: lmaoo.png (2.26 MB, 1341x1518)
2.26 MB
2.26 MB PNG
https://xcancel.com/bdsqlsz/status/1971871215846674942#m
>80b parameters and it looks like this
>>
>>106715114
our feet are cobbled together from 20 bones stuck together like glue. Tall people regularly have more foot injuries in sports. women are obsessed with height and our feet suck ass. like put 2 and 2 together dude. You thin girls like tall guys and guys like short girls is a fucking coincidence? We're regulating our height dumbass.

imagine a woman with massive bunions and hairy weird long toes that are mangled as fuck? You wanna fuck her?
>>
>>106715156
suddenly the rumors of it being closed source are sounding a lot nicer right about now
>>
File: 1736876785676607.png (879 KB, 1309x716)
879 KB
879 KB PNG
>>106715156
even Chroma has better hands, and it has 10x less parameters, embarassing
>>
Train the base model thoroughly and invest heavily.
Quality isn’t about how pretty it looks; it’s about the dataset’s quality and quantity.
Which LLM you use might be the most important factor.
>>
>>106715159
yeah judging by this post foot-faggotry is definitely some kind of mental illness
>>
Is it possible to get Wan 2.1 LoRA's to function correctly in 2.2? I've got some sweet twerking LoRA's that had great looking physics, but I can't get them to work in 2.2, not at 1 strength and not at 3. Do they need to run with both low and high or just one? Is there some trick to running them or do they need to be retrained or something?
>>
>>106715162
>pircel
I used to look up to that man, why did he turn out to be such a giant crying pussy?? :(
>>
>>106715172
I think it depends on what model the 2.1 lora was trained on (t2v or i2v). And I think i2v trained loras are basically fucked in 2.2, or maybe it's the t2v ones
>>
>>106715172
I ran a few tests, wasn’t satisfied, and immediately started training.
>>
Currently we have:
/ldg/ local diffusion general
/sdg/ stable diffusion general
/adt/ anime diffusion thread
/de3/ dalle 3 cloud AI general

I understand local, I understand cloud gen AI because it's the opposite. The anime one because anime website.

Can someone explain /sdg/?
>>
>>106715098
>JUST PAY FOR AN IMPROVED DATASET
you know what would be better than a leaked model, a leaked dataset, imagine every companies having access to a great quality model, it would even the field so much, the quality of the data is always the most important part
>>
File: jordan peterson skyrim.jpg (458 KB, 1682x1080)
458 KB
458 KB JPG
>>106715181
>giant crying pussy
*sniffle* "WATCH IT BUCKO. BENZOS ARE NO JOKE. NOW THAT, THAT IS CHAOS.

Agreed though, he was influential in my early twennies.
>>
File: file.png (2.81 MB, 832x1488)
2.81 MB
2.81 MB PNG
>>106715156
hmm... well, there are paintings like that
>>
File: 1750498265947691.png (718 KB, 2209x329)
718 KB
718 KB PNG
>>106715210
>there are paintings like that
even picasso make better faces lol
>>
>>106715195
the S in sdg is for schizo

you can see it once you pay a visit there.
>>
>>106715195
/sdg/ is a containment thread for a handful of anons who post really terrible gens and are too socially awkward to just get on a discord chat together
>>
>>106715195
We should ask to shut down the /sdg/. They post gens from Chroma, SDXL, Dalle, and OpenAI without focusing on any specific style. That thread should be closed and users can move to either /de3/ or /ldg/ instead.
>>
>>106715172
Depends on the lora, but most 2.1 motion loras look like shit in 2.2 imo. If they're your own loras, retrain, if not, hope for the author to update it.
Also, this is a pretty good twerk lora for 2.2 :
>https://civitai.com/models/1953874?modelVersionId=2211367
>>
>>106715195
>Can someone explain /sdg/?
asylum containment board, without that our general would be doomed by the schizos
>>
File: 1742400418817926.png (1.52 MB, 832x1216)
1.52 MB
1.52 MB PNG
so this is the power of v-pred... WOWZERS!!!
>>
If I want to shave my head bald and become a degenerate coomer generating AI porn locally what should I be aiming for in terms of pc parts if im setting up a rig from scratch reasonably priced but willing to go $2000+
>>
>>106715268
well if you're really going full picrel with a budget like that, the obvious choice is a threadripper/5090 system.
>>
>>106715169
uh oh! looks like your defense mechanisms kicked in when they realized they were retarded. Good luck with you're little testosterone fueled trip retard.
>>
>>106715275
>threadripper
no, it's cheaper to go epyc 9005 SP5 series, you can find used stuff really easy on ebay too. threadripper is better for gaming but do you game lmao?
>>
>>106715275
>a budget like that
>$2000+
it's not 2016 anymore anon...
>>
dont worry bros the chink 112gb HBM ram gpu will save us
trust the plan, two more weeks
>>
The CPU is just a bonus; in practice, the maximum RAM capacity is all that matters.
>>
>>106715288
>the epyc is literally $1k over his budget for the entire computer
what did he mean by this?
>>106715291
oh you know what you're right, a $2,000 budget today is way different from back then.
good luck dude you might get a used 4090 then kek
>>
>>106715275
>>106715288
>>106715291
I legit serious I can deal with slower generation and stuff but the first step is having a rig that can handle generating to begin with
>>
I can't figure out what the set node in the middle is. When I hover over it says it's the node at the top, which it clearly isn't.
>>
File: 1558317559576.png (25 KB, 395x474)
25 KB
25 KB PNG
>take six months break from genning
>come back
>the top anime model is still illustrious
>there's still no good realism or artistic model, everyone is still using either flux or chromakek
>>
>>106713871
NTA but can you post a workflow for it, is it using kijai's workflow? I tried with the native comfy setup and I couldn't get i2v to work wihtout setting the context length to the length of the whole video which seems pointless
>>
>>106715309
nothing changed in a year, I'm still waiting for the replacement of flux dev (for video better it's way better thanks to wan though)
>>
>>106715304
>slower generation
shit nigger it's not 2022-3 anymore, even the 3000 series will get you started on the fast lane. but i can personally with my new bias recommend the 5060 ti 16gb, but with your budget, just go 4090 or something.
only cards that can't handle generating are literally gtx and every AMD/Intel(rip lol) card.
>>
>>106715306
>kijai
LOL
>>
File: WAN2.2_00098.mp4 (2.7 MB, 960x544)
2.7 MB
2.7 MB MP4
>>106715275
your wife is calling for u
>>
File: 00007-917646120.png (1.05 MB, 1280x768)
1.05 MB
1.05 MB PNG
>>
>>106715304
let's start over. so what do you have right now, and what do you what to be able to do?
>>
File: kek.png (401 KB, 768x1280)
401 KB
401 KB PNG
So this is the power of 80b...
>>
>not having at MINIMUM 24gb

you're are become the regret
>>
>>106715315
>5060 ti 16gb
Thats reasonable and I can adjust my budget since I kinda was just lowballing since I want something that can do the work but Im not entirely up to fucking fully commit to it. Is there anything else you would suggest to throw in there? Im essentially building from scratch so any other suggestions would be appreciated since I probably should have just posted this in the pc building general but wanted I wanted suggestions from the source
>>
File: WAN2.2_00100.mp4 (3.78 MB, 960x544)
3.78 MB
3.78 MB MP4
what's the general consensus on thick women?
>>
File: 1741576248881613.png (1.95 MB, 3334x1122)
1.95 MB
1.95 MB PNG
>>106715336
the jpeg artifacts are insane, maybe they're not using a vae anymore?
>>
>>106715336
When you get that level or low detail/artifacting around the eyes that close up, you know you fucked up
>>
>>106715347
Honestly also take advice from this fella >>106715334 because that's a reasonable place to start too. if you're not about committing fully yet i'd even say go slightly lower end than that card and if you want more speed, move up tiers. save yourself that few extra hunned bux at the start.
pcpartpicker is REALLY helpful for making decisions on these things too.
>>
>>106715348
>what's the general consensus on thick women?
I don't like it, I find fat females ugly as fuck
>>
>>106715318
Let me guess, it was the reason why my gen was broken?
>>
>>106715336
chinese can't into image models
well established fact
>>
File: aegeadbgvaedgbae.png (68 KB, 336x213)
68 KB
68 KB PNG
>>106715349
aye dios mio santa maria this CAN'T be real
>>
>>106715348
she's fat. not thicc.

thicc is having a healthy amount of bf% but having large thighs.
landwhales have adopted the term "thicc" to justify their bad lifestyle and dying bodies.

do not fall for their tricks.
>>
>>106715375
well thats what Im trying to wrangle wan to do. It isnt listening sadly
>>
>>106715358
>>106715334
well I have nothing redeemable hence why Im looking at doing an entire new build
>>
>>106715349
I think it looks more like low bitmap than jpeg artifacts.
>>
>>106715378
obese, fat etc in negatives (in wan nag)
and use stuff like "slender with large thighs/hips" or something like that. desu is easier using i2v for that because wan is dumb
>>
>>106715336
look at the filesize. It seems it's just a low bitmap conversion of an original image.
>>
>>106715384
a build that is future proof will set you back way the fuck more than 2k. absolutely do not listen to the vramlets. i'm not memeing, do not get anything under 24gb vram ideally with minimum 64gb ram.
your ideal system is a 5090 and whatever cpu you want. cpu literally does not matter at all for this.
i promise you: if you get a 16gb vram card you WILL regret it. i have a 24gb card and can run everything without having to fuck around with pissy little quants.
>>
>>106715414
>you WILL regret it
psh speak for yourself bitchboooyy. you will be crying when your ancient card gets less and less speedup feature support.
>inb4 he unironically paypigged for the tesla rigs in 2023 and still hasn't gotten over it
>>
>>106715224
It's worse than that, they have a discord as well but also poison the thread with their obnoxious behavior for attention.
>>
>>106715426
..what the fuck are you talking about? idk who or what you are refering to.

did you even remotely understand what i wrote? jesus fuck.
>>
File: WAN2.2_00106.mp4 (3.31 MB, 960x544)
3.31 MB
3.31 MB MP4
this is surprisingly good.. the lotion moves right
>>
>>106715268
get a 5070 ti to start. The 5060 ti is too slow and is only for sdxl really- it's great for budget friendly ability to do ai shit, but the 5070ti is more than twice as fast for ai and your budget suggests you want something decent. The speed of the 5070 ti will allow you to fuck with new image models like qwen edit, wan video, and run nice llm's like glm air fine.

in a year, sell it and get the new 24gb cards probably.
>>
File: G12G2r6XcAAZ4XL.jpg (180 KB, 1024x1024)
180 KB
180 KB JPG
>>106715336
Another one. It's cooked.
>>
>>106715456
someone make a starter kit lmao
>>
>>106715456
All the example images they've released are 1 megapixel as well. Instead of a retarded 80B model nobody can run, and which looks like shit, how about an 8B model that can do 2 to 4k?
>>
>>106715456
>180 KB
Does it really output 50% qual jpegs or are you converting it?
>>
>>106715384
Honestly, there's some truth to >>106715414 . The best option for citizen plebs is a 5090. The next best options is any Nvidia card with 24 GB of VRAM. Beyond that and then you'll start having to limit yourself and rely on cope shit. Also, upgrading a GPU hurts a lot more than upgrading ram, the GPU should be your primary focus.
>>
>>106715473
This is what Hunyuan themselves posted on X.
>>
>>106715456
>80b model
>jacket has buttons
> skirt too long
80b model
>jacket too long
>handbag strap far longer
80b model

this is a 80b model.
>>
>>106715468
>Instead of a retarded 80B model nobody can run, and which looks like shit, how about an 8B model that can do 2 to 4k?
that's too smart of an idea anon, the Tencent engineers are dumbfucks lol
>>
>>106715483
There's a distinct yellow sheen to every image too...
https://x.com/TencentHunyuan/status/1971805621738700947
>>
>2025
>still cropping your data set
>>
File: kek.png (95 KB, 220x298)
95 KB
95 KB PNG
>>106715485
>this is a 80b model.
it definitely is one of the models of all time
>>
File: fucking NPCs.png (292 KB, 1142x1366)
292 KB
292 KB PNG
>>106715495
>https://x.com/TencentHunyuan/status/1971805621738700947
wtf is wrong with those people??
>>
>>106715514
kek
>>
>>106715514
probably bots. people who comment in x are subhuman (literally)
>>
>>106715514
>>106715523
why doesnt anyone just ask for nsfw directly.
>>
>>106715531
You are not getting a nsfw model ever again lol
>>
>>106715538
damn shame
>>
>>106715329
animating this did not do me any favors
>>
File: 1750346897113355.mp4 (3.74 MB, 1056x512)
3.74 MB
3.74 MB MP4
>>106715531
>>106715538
this, vu will own nothing, vu will be happy
>>
>>106715485
Not to mention the fucking ZIPPER it created, despite the fact it's a button up jacket. Fucking lol.
>>
>>106715551
wheres the bj anon when we want him
>>
>>106715549
Not him or you, but I'm guilty of the same sin...
>>
>>106715275
a 5090 costs $2500+ by itself, retard
>>
How long and how much would it take to chromify qwen with SRPO?
>>
>>106715566
Now make it twerk
>>
File: file.png (2.87 MB, 3315x1721)
2.87 MB
2.87 MB PNG
How do you prompt Qwen to replace the character in one image with another? Every time I've tried it has failed miserably.
>Let the woman in image 2 replace the character in image 1. Let her wear the outfit from image 1 and use the pose of image 1.
>>
>>106715579
afaik, you can't do it reliably. Same problem with style transfer, ie "Change image 1's style to match image 2's style".
>>
new
>>106715652
>>106715652
>>106715652
>>106715652
>>
good sheit

https://civitai.com/models/1988265?modelVersionId=2250722
>>
File: 1756833655086398.jpg (145 KB, 1103x1464)
145 KB
145 KB JPG
>>106712561
thanks anon, was looking forward to this link
>>
File: file.png (3.02 MB, 832x1488)
3.02 MB
3.02 MB PNG
>>106715336
are they also doing pixel space?



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.