[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/g/ - Technology


Thread archived.
You cannot reply anymore.


[Advertise on 4chan]


GDP of India Edition

Discussion of Free and Open Source Text-to-Image/Video Models and UI

Prev: >>106568960

https://rentry.org/ldg-lazy-getting-started-guide

>UI
ComfyUI: https://github.com/comfyanonymous/ComfyUI
SwarmUI: https://github.com/mcmonkeyprojects/SwarmUI
re/Forge/Classic/Neo: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassicneo
SD.Next: https://github.com/vladmandic/sdnext
Wan2GP: https://github.com/deepbeepmeep/Wan2GP

>Checkpoints, LoRAs, Upscalers, & Workflows
https://civitai.com
https://civitaiarchive.com/
https://openmodeldb.info
https://openart.ai/workflows

>Tuning
https://github.com/spacepxl/demystifying-sd-finetuning
https://github.com/Nerogar/OneTrainer
https://github.com/kohya-ss/sd-scripts/tree/sd3
https://github.com/derrian-distro/LoRA_Easy_Training_Scripts
https://github.com/tdrussell/diffusion-pipe

>WanX
https://comfyanonymous.github.io/ComfyUI_examples/wan22/
https://github.com/Wan-Video

>Chroma
https://huggingface.co/lodestones/Chroma1-Base
Training: https://rentry.org/mvu52t46

>Neta Lumina
https://huggingface.co/neta-art/Neta-Lumina
https://civitai.com/models/1790792?modelVersionId=2122326
https://neta-lumina-style.tz03.xyz/

>Illustrious
1girl and Beyond: https://rentry.org/comfyui_guide_1girl
Tag Explorer: https://tagexplorer.github.io/

>Misc
Local Model Meta: https://rentry.org/localmodelsmeta
Share Metadata: https://catbox.moe | https://litterbox.catbox.moe/
GPU Benchmarks: https://chimolog.co/bto-gpu-stable-diffusion-specs/
Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-one
Txt2Img Plugin: https://github.com/Acly/krita-ai-diffusion
Archive: https://rentry.org/sdg-link
Bakery: https://rentry.org/ldgcollage

>Neighbours
>>>/aco/csdg
>>>/b/degen
>>>/b/realistic+parody
>>>/gif/vdg
>>>/d/ddg
>>>/e/edg
>>>/h/hdg
>>>/trash/slop
>>>/vt/vtai
>>>/u/udg

>Local Text
>>>/g/lmg

>Maintain Thread Quality
https://rentry.org/debo
>>
blessed thread of sage attention
>>
Does "duplicate characters" in neg prevent the people cloning? Or does that not exist in the dataset?
>>
>>106571097
clone is a danbooru tag, duplicate characters is not
>>
>>106571092
Cursed thread of python dependencies
>>
>>106571106
I am talking about NL models.
>>
>>106571076
Keep going...
>>
File: 00276-1803632336.png (3.36 MB, 2480x2048)
3.36 MB
3.36 MB PNG
New thread
Chroma is doing a lot of good stuff without extra work
>>
>>106571097
it usually won't do anything

if you want to define multiple characters that works vastly better on wan, qwen, hidream and also better on chroma, flux and so on
>>
>>106571113
i'll post the best ones on /b/ tomorrow

but i'm starting to get seriously fed up with wan doing adult anatomy all the time unless the face is present in the starting frame. all the girls that actually look the age i prompted for have weird super long adult legs

this is literally going to be solved in one or two generations of local video so im not going to kvetch about it. i'm still shocked how difficult it is for wan to do braces when hunyuan made them without a second thought without the word "orthodontic" needed
>>
>>106571120
No I mean sometimes when the model overshoots and start duplicating the the features in the positive prompt and starts schizoing another entire characters out of thin air even if you didn't prompt for it. And I use chroma.
>>
>>106571136
>when the model overshoots and start duplicating the the features in the positive prompt
if you mean extra feet when you prompt for feet and stuff like then then you either have to lower the weight of those tokens or if you can't do that expand your prompt so that the total % of the prompt being the token thats overshooting is lower
>>
Imagine attempting to explain the problem with your gen rather than actually posting it and/or a catbox
>>
that is not a college I would attend
>>
>>106571136
getting other characters maybe can be reduced with "solo" positive or negative "multiple views" "multiple girls" "multiple boys", also depends on the model tho

and your sampler and scheduler may also influence this quite a lot
>>
File: 1735935439461400.png (584 KB, 496x720)
584 KB
584 KB PNG
>>106569466
https://files.catbox.moe/cstl4k.mp4
>>
File: 125127627234.jpg (3.97 MB, 3072x3072)
3.97 MB
3.97 MB JPG
>>
File: 00282-1803632336.jpg (850 KB, 2480x2048)
850 KB
850 KB JPG
>>
shit baker
>>
File: table.png (2.93 MB, 1512x843)
2.93 MB
2.93 MB PNG
Finally, autistic wood.
>>
File: 00016-1720590611.png (2.98 MB, 1248x1848)
2.98 MB
2.98 MB PNG
>>
i'm doing img2img by vae encoding an image, but my current workflow doesnt have any node where i can set a denoise value. is there some node i can hook into my sigma to set the denoise instead? i really dont want to change out my scheduler.
>>
>>106571187
I was thinking about this.

Has anyone compared the last version 3.0 of NetaYume, its better or its more the same impredictable thing?
>>
>>106571187
i think this is the first time i've actually seen the eye of horus on a human being and not as a hieroglyph
>>
File: 0_00086_.mp4 (939 KB, 1024x576)
939 KB
939 KB MP4
Can you imagine if we had this technology in 2006?
>>
>>106571194
What shitter nodes are you using where a sampler or scheduler don't have a denoise value?
>>
>>106571194
??????????
Speak in human please
>>
>>106571194
bit ambiguous (more ways to use the words you are using) but kj/inspire both have noise/sampler nodes that do that
>>
File: 1741755162568515.png (22 KB, 482x250)
22 KB
22 KB PNG
>>106571206
>>106571205
>>106571204
im using this one
>>
>>106571167
lol nice
>>
>>106571201
>Can you imagine if we had this technology in 2006?
idk about normies but pedos literally wouldnt care because the internet, google images etc was fully uncensored during this time
>>
>>106571194
NeoForge user here, never touched Comfy nodes. Your post is pure meme material anon.
>>
>>106571208
fuck you
>>
>>106571208
Regular scheduler node has beta and denoise.
>>
>>106571221
k....
>>
>>106571194
Ehemmmmmmm
>>
>>106571228
you asked, i imagined, faggot
>>
>>106571221
I meant when memes were actually funny, before you were born.
>>
File: ComfyUI_00237_.mp4 (781 KB, 832x1232)
781 KB
781 KB MP4
>>106571187
glad to see you back
>>
>>106571201
Can you imagine if we had this technology in 1965 ?
>>
File: w.mp4 (2.18 MB, 640x1144)
2.18 MB
2.18 MB MP4
>>106571201
we wouldn't have had the hardware then.

it's just as needed now to for example lampoon all the authoritarians that want to mess with the internet, they're everywhere - even in the EU and USA

or you can just make hot 1girls or hot 1boys
>>
>>106571251
>people born in 2006 are 19 now
this would strike me as crazy but my favorite starlet is 18 now so i have already been hit with this realization

and memes are still funny, you're just unironically a faggot lol
>>
what's with this weird pedo posting? what causes this?
>>
i'd take pedos over indian shills any day
>>
>>106571278
are you lost discord zoomie?
>>
>106571276
probably the most pathetic post itt, congrats
>>
>>106571286
i mean, its not like i cared about any of her content since 2019/2020, i'm not a simp

but she's a cunny staple so of course i heard when she finally aged out completely
>>
>106571289
probably the most based post itt, congrats
>>
sametroon
>>
File: ComfyUI_00236_.mp4 (1.11 MB, 480x832)
1.11 MB
1.11 MB MP4
>>106571259
i'm always here
>>
File: 1735399762913001.png (625 KB, 480x720)
625 KB
625 KB PNG
>>106569484
https://files.catbox.moe/63ayjl.mp4
>>
>finish training
>first result: character comes out perfectly, 1:1
>second result: character comes out unrecognizable
>becomes a russian roulette of gens
what the fuck am i doing wrong? is it the dataset? should i only train from concept art and only include them in an expressionless state?

i think them showing emotion or using different facial expressions is fucking it up but im not even sure anymore. any general tips for onetrainer? this is for SDXL/Illustrious and i'm using the standard preset
>>
>>106571356
Did it ouput you several backup files from various stages of training? Try those.
>>
>>106570600
kek can you imagine the artfags seething to this?
go download some twitterfags art and generate vids of it being drawn by someone else through ai
>>
>>106571356
too little information to know
>>
>>106571376
>Try those
alright, i'm gonna go to sleep and then i'll mess around with it in the morning, i'm mad as fuck right now because of the gacha quality of my lora.
>>
File: 0_00095_.mp4 (735 KB, 832x480)
735 KB
735 KB MP4
lol at the tranny in the back
>>
>Do not use my works without permission!Use in AI training is also prohibited.
hahahahahhahahaha
>>
File: Výstřižek.png (2.18 MB, 1033x786)
2.18 MB
2.18 MB PNG
>squishing cheek
rraaah, entire gen fucked
>>
File: oc.png (64 KB, 360x334)
64 KB
64 KB PNG
>>106571447
my works btw
>>
File: 0_00097_.mp4 (762 KB, 832x480)
762 KB
762 KB MP4
>>106571447
I have no respect for artists. Seeing that kind of stuff just makes me want to ape it more.
>>
Death to artists, coomers, and devs
>>
>>106571437
Cursed gen
>>
>>106571491
rocketnon??
>>
god bless the API
>>
File: 0_00099_.mp4 (807 KB, 832x480)
807 KB
807 KB MP4
>>106571499
you forgot jannies
>>
File: 0_00074_.mp4 (644 KB, 832x480)
644 KB
644 KB MP4
>>106571516
no just some guy who loves titties and doesn't have creative prompts. HH
>>
>>106567568
nice
>>
File: 1728497339765475.png (725 KB, 720x720)
725 KB
725 KB PNG
>>106569406
https://files.catbox.moe/uwbghi.mp4
>>
>>106571659
>https://files.catbox.moe/uwbghi.mp4
That thing is as thick as a baby's arm.
>>
>>106571671
I think you mean leg
>>
Has anyone experimented with running the lightning loras for either high or low pass separately?
>>
File: 1743631464303502.png (446 KB, 432x720)
446 KB
446 KB PNG
>>106570263
https://files.catbox.moe/7k08mz.mp4
>>
>>106571760
>https://files.catbox.moe/7k08mz.mp4
Heh, ShoeGiveHead.
>>
File: 1747201722383700.png (1.01 MB, 720x720)
1.01 MB
1.01 MB PNG
>>106569479
https://files.catbox.moe/igk7v6.mp4
>>
>>106571907
is she like 3 foot tall or is the guy like 9 feet tall lmao
>>
File: 00186-2768897287.png (2.98 MB, 1248x1761)
2.98 MB
2.98 MB PNG
>>
>>106571671
for you
>>
>>
>>
>>
File: ._00022.mp4 (646 KB, 480x832)
646 KB
646 KB MP4
please stop taking screenshots and capture full motion video like a real human bean
>>
>>106571958
>>106571965
>>106571973
>>>/r/
>>
Can you run comfy with a rented GPU? Do you pay for raw hour or hour of gen time? Also how much data do you send the rented gpu? I only have slav internet.
>>
>>106572042
>comfy with rented GPU
Better to just use python and colab (free or paid)
Paperspace another decent alternative.
>>
https://files.catbox.moe/ga05fb.mp4
>>
>>106572083
I'd rather use my existing WFs and models.
>>
File: 0913143040785-iwCNfCZ9tOl.jpg (501 KB, 2492x1792)
501 KB
501 KB JPG
I tried using Wan2.2 i2v 3 steps workflow downloaded from civitai to fix slow mo issue. What the hell is changing the artstyle and face so drastically? The .gguf model?
>>
>>106572127
What's in your prompt? Have you specified the artstyle?
>>
>>106572042
sure, did runpod for a minute. storage is the issue, you have to script out the instance and download the models/loras when you want to gen. pain in the ass.
>>
File: 1727283203992291.png (407 KB, 416x720)
407 KB
407 KB PNG
>>106570598
https://files.catbox.moe/vaxxp5.mp4
>>
>>106572133
nope
"she begin walking towards the camera continuously while the camera follows her, she walks with great confidence, her big breasts jiggles with each stride"
The workflow I have been using output, artstyle and face retained
>>
>>106572084
interesting how the """sfw""" version is far more erotic
>>
>>106572146
>https://files.catbox.moe/vaxxp5.mp4
Wow. Subtext of that is grim.
>>
3 step workflow, artstyle and face changed
>>
Is there a way to pick and choose the mp4 codecs in comfy?
>>
File: file.png (129 KB, 1280x991)
129 KB
129 KB PNG
>>
>>106572173
like this?
>>
>>106572146
lmao
>>
>>106572181
Nice. But is there also a way to control the bitrate?
>>
Blazing hot take: If you expect any semblance of quality out of something as ludicrously low as 4 steps, you are a fool.
>>
File: 1756187835924680.png (3.33 MB, 1152x1728)
3.33 MB
3.33 MB PNG
>>106572187
couldn't you just output the frames and ffmpeg them yourself
>>
>>106572187
the "crf" is how u control picture quality
>>
>>106571076
>orthodontics
pls share
>>
I love this workflow.

https://files.catbox.moe/1fxkbm.mp4
>>
File: 00064-2065459701.png (1.79 MB, 896x1152)
1.79 MB
1.79 MB PNG
>>
>>106572236
Kek that's so good
>>
>>106572194
True, but I figured there could be a node for it since comfy is meant to be the customizable way of genning.

>>106572197
How does it work? It could increase quality, but the default mp4 codec could be trash.
>>
File: 1756943164051454.png (1.56 MB, 912x1144)
1.56 MB
1.56 MB PNG
>>
>>106572311
>balenciaga
>>
Do I hook this sage attention node up in between the load diffusion model modelsamping nodes for it to work?
I also saw a video and lost it, but it allowed you to pick which sage attention version to pick. What's the default setup using, if fully updated?
>>
File: Chroma-ComfyUI_00110_.png (1.3 MB, 1024x1024)
1.3 MB
1.3 MB PNG
>>
File: Chroma-ComfyUI_00111_.png (1.31 MB, 1024x1024)
1.31 MB
1.31 MB PNG
>>
File: Chroma-ComfyUI_00126_.png (1.14 MB, 1024x1024)
1.14 MB
1.14 MB PNG
>>
>>106572042
>>106572109
A proper cloud use case for Comfy, rather than focusing on getting cloud models running, find a way for cloud users to use Comfy.
>>
File: Chroma-ComfyUI_00120_.png (1.05 MB, 1024x1024)
1.05 MB
1.05 MB PNG
>>
>>106571086
>Discussion of Free and Open Source Text-to-Image/Video Models and UI
should be "Discussion and development of local image and video models"
its not "open source" its open weight, anyway
>>
>>106572236
She sounds like frieza
>>
File: 1729070403288818.png (539 KB, 720x544)
539 KB
539 KB PNG
>>106572236
https://files.catbox.moe/jhrqb2.mp4
>>
>>106572173
You can like the other anon showed, but if you're sharing mp4s that aren't h264 you're kind of a fag desu
>>
>>106571194
If you are using an advanced or custom ksampler, the "denoise value" is just what % of the total steps to start at. So just increase the starting step higher to keep more of your input latent image.
>>
>>106572335
Oh my god sage attention wasn't even installed.

Ok, so do I need to hook this up for both high and low noise?
>>
File: 1740191303699266.jpg (9 KB, 253x268)
9 KB
9 KB JPG
>>106572478
beat me to it
>>
>>106572229
I'll make a catbox gallery in like 16 hours of the best braces stuff I've been able to make so far, and since multiple anons are interested in this beyond the normal pedo interest I guess I'll link it here instead of just on /b/ as usual
>>
>>106572192
>If you expect any semblance of quality out of something as ludicrously low as 4 steps, you are a fool
One day before self-forcing came out, if you told anons ITT that it was possible you would be called a gorilla retard

I agree that 4 steps is too low but the fact that just going up to 6 basically resolves 80% of the problem is still crazy

There's still so much low hanging fruit
>>
>>106572376
I completely forgot cute girls take swim classes and then shower off the chlorine after. Thanks for reminding me about this anon

I don't have anything to offer you except a reminder that young girls love going to amusement parks. People often forget this when considering settings/backgrounds for their girls
>>
>>106572361
Share pls, j conn the prettiest woman ever
>>
Wan 2.2 doesn't actually work with negative prompts? I'm watching so many videos and out of like 30 watched, only one has mentioned it.
>>
>>106572516
Yeah you need to put a node in between the load model and load Lora sections for both the high and low
>>
>>106572568
It works with negative prompts, but adherence is meh. NAG makes it better, but if you are using lightning your cfg is 1 anyways so negative prompts don't actually have an effect
>>
File: 1559068629709.jpg (70 KB, 876x1024)
70 KB
70 KB JPG
>flan works better than GNER on super high cfg setup
the fucking furfags were right all along
>>
File: Chroma-ComfyUI_00142_.png (1.64 MB, 1024x1024)
1.64 MB
1.64 MB PNG
>>106572556
Sure anon

jennifer connelly
https://files.catbox.moe/1oyen4.safetensors

ellen page
https://files.catbox.moe/d9c9xc.safetensors

I only trained using 'rcng' as caption, so use that as a 'trigger' for extra punch, these were trained on Chroma v49, but they should work fine for Chroma1-HD
>>
>>106572629
>I only trained using 'rcng' as caption
not even "a photo of a woman" ? interesting
>>
>>106572629
TY for sharing. JConn is dreamy
>>
>>106572575
I guess I should refrain from using lightning loras when doing the final renders then, or does it change the generation/motion too?

>>106572570
Thanks. I got a lot to test now.
>>
>>106572643
When you train something the model knows VERY well, like a human, you can just put a nonsense string and it will train perfectly well since it can easily deduce the pattern, as in the same person in all images.

I could use their actual names, but since these are primarily for testing purposes, it makes it easier when test prompting to just use the same 'trigger'.
>>
why is wan 2.2 so shit for twerking, slow motion garbage no matter what i try. prompting, 2.1 loras, shitty speed hacks, she just sits there shivering like shes cold. the only lora for 2.2 is called "slow motion twerking" kek
>>
>>106572656
>I guess I should refrain from using lightning loras when doing the final renders then, or does it change the generation/motion too?
I have no idea what you asked me. If you use lightning loras, forget about negative prompts

>>106572683
Obviously a skill issue when jiggling videos with wan 2.2 are posted literally every day here
>>
>>106572682
I'll have to take a mental note and test it again. I just faintly have a memory that even lightly tagged celeb lora would be more flexible. That being said you loras work so god damn well that there's nothing to argue about, just one way more to train lora.
>>
File: Chroma-ComfyUI_00163_.png (1.2 MB, 1024x1024)
1.2 MB
1.2 MB PNG
>>106572651
No problem, the quality is not peak though since I exclusively went for her 80s images and those aren't exactly high quality, mostly scans from magazines it seems.
>>
>>106572704
who all have you trained on?
>>
>>106572695
you mean that tame shit that gets posted? thats easy to do. 2.1 remains supreme.
>>
>>106572704
Those japanese ads have great aesthetic. That could be a style lora. Again ty for share
>>
File: Chroma-ComfyUI_00168_.png (1.41 MB, 1024x1024)
1.41 MB
1.41 MB PNG
>>106572701
Most likely you will get a slightly better result with some 'light' descriptions as you put it, particularly if there's some things in the images that the model might have a bit of problem grokking.

That said the Chroma model is surprisingly good at picking up even small details without them being prompted.

The images here are tests against a Miles Aldridge style lora I just trained, for that lora I did use 'natural language captioning' since otherwise it's hard for the model to understand what to learn since it's not a clear pattern.

Here's the Miles Aldridge lora if anyone wants to play with it:
https://files.catbox.moe/dll1yn.safetensors

The prompt for this image:
rcng pale voluminous ginger hair makeup coat scarf curtain tiled wall detailed highest quality

negative:
low quality, ugly, unfinished, out of focus, deformed, disfigured, blurry, cropped

steps: 25
cfg: 4.0
sampler: dpmpp_2m
scheduler: beta
>>
https://xcancel.com/TencentHunyuan/status/1966483962370654388#m
>they're still not tweeting about SPRO
kek, they really believe their method is a nothingburger...
>>
File: the poorest localfag.png (190 KB, 736x847)
190 KB
190 KB PNG
https://www.reddit.com/r/StableDiffusion/comments/1nfo5rn/opensource_cost_more_to_create_with_than_using/
>jeet complaining that local is for richfags
that's right
>>
File: Chroma-ComfyUI_00161_.png (1.02 MB, 1024x1024)
1.02 MB
1.02 MB PNG
>>106572726
It's a lora actually, it's Miles Aldridge style, you can download the lora here:
https://files.catbox.moe/dll1yn.safetensors

Also forgot, here's Emma Watson
https://files.catbox.moe/kpfn34.safetensors
>>
>>106572742
>>106572757
Damn dude thanks. Gotta test em later with that 2k model
>>
>>106572757
>It's a lora actually, it's Miles Aldridge style, you can download the lora here:
how many images did you use for it? the dataset
>>
Test
>>
C-can a 1080ti and 7700k gen this stuff?
>>
>>106572851
N-no.
>>
File: Chroma-ComfyUI_00188_.png (1.18 MB, 1024x1024)
1.18 MB
1.18 MB PNG
>>106572772
You're welcome, take note though that I trained against (and generated with) Chroma v49, it's my favorite of the Chroma versions.

So you might not get as good results if you are using another model, but when I tried some of these loras with Chroma1-HD they looked fine.

>>106572802
This was just a test, so only 40 images.
>>
Chroma loras are mostly interchangeable. Hell even some flux ones work. But I'd train them on Base.
>>
File: Jenny Gets Adjusted.webm (3.92 MB, 852x1280)
3.92 MB
3.92 MB WEBM
>An octopus steals the woman's shirt.
Nailed it!
>>
File: ComfyUI_temp_cbxpk_00070_.jpg (1.03 MB, 1920x1280)
1.03 MB
1.03 MB JPG
>>
>>106572920
>>106572896
>>106572757
>>106572742
Anime website fuck off to reddit
>>
>>106573017
trani melti
>>>/g/adt
>>
>>106573017
(You)
>>
File: Chroma-ComfyUI_00191_.png (1.3 MB, 1024x1024)
1.3 MB
1.3 MB PNG
>>106573017
Your faggotry has no power over me!
>>
File: hmm.jpg (12 KB, 417x201)
12 KB
12 KB JPG
What did I miss since flux kontext?
Work kept me away from /ldg/
>>
File: Chroma-ComfyUI_00206_.png (1.06 MB, 1024x1024)
1.06 MB
1.06 MB PNG
>>
>>106573107
We got a vace model for wan 2.2. But it's not from the actual vace team, but kind of is?
>>
>>106573107
>What did I miss since flux kontext?
Wan 2.2, Qwen Image, and Qwen Image Edit, and light2x (that let's you make wan videos with only 4 steps) the rest if a bunch of nothingburgers (SPRO, HunyuanImage...)
>>
File: thanks.jpg (29 KB, 746x512)
29 KB
29 KB JPG
>>106573116
>>106573123
>>
Does non-specified high or low noise loras still need to be linked to both the high and low noise models?
>>
File: Chroma-ComfyUI_00214_.png (1.19 MB, 1024x1024)
1.19 MB
1.19 MB PNG
>>
>>106573149
I use it like this :
wan2.2 high noise -> node with 2.1 loras -> node with high noise 2.2 loras -> high noise sampler
wan2.2 low noise -> node with 2.1 loras (clone) -> node with low noise 2.2 loras -> low noise sampler
>>
File: 1757725613935635.webm (2.93 MB, 640x640)
2.93 MB
2.93 MB WEBM
hello sirs
kindly prompt video2video with this source
>>
>>106573168
>>106573109
>>106573044
>>106572757
>>106572704
damn cool anon
>>
File: AnimateDiff_00015-1.mp4 (3.76 MB, 500x406)
3.76 MB
3.76 MB MP4
>finally get some nice results
>the gens now have this distorted flash at the end of it

God damn it, please tell me it's fixable. I'm using a FFLF workflow, sageattention, lightning, various loras.
>>
>>106573181
Is she getting fingered or sumthin??
>>
>>106573190
Nooo, just exhausted on the exercise bike.
>>
>>106573194
some bike bruh

all gens have that flash? thats new. Ive done FFLF stuff, mostly its just color issues for me.
>>
>>106573181
just trim those frames off the end with split images node
>>
>>106573170
>2.2 loras

Don't use these. I have no idea how this information keeps getting forgotten but the 2.2 light LoRAs are basically broken. The 2.1 ones work fine and don't kill motion/blow out colors.
>>
>>106573200
I'm backtracking the things I did this morning, updated sage attention and swapped to Video Combine instead of Save Video node. But yeah, large color shift or a flash at the end.

>>106573210
Can't do that for loops.
>>
>>106573238
I meant wan 2.2 loras in general, I do not use light loras, they make everything slow motion and look lower quality.
I can wait.
>>
>When you're in the vace workflow business and baby, business is good.
>>
>>106573288
>I do not use light loras
How long are your gens? Without those it takes 5-10 times longer to gen
>>
File: ComfyUI_temp_lciqh_00018_.jpg (1.16 MB, 1152x2016)
1.16 MB
1.16 MB JPG
>>
>>106573305
nta, I think an outright refusal to use any speedup LoRAs, especially for something you'd post here is a retards game. But there are places you can take skimp on it. Mainly the high noise model. You can even crank it down quite low on the low noise model and still get good results with a couple of extra steps.
>>
>>106573305
15-20min depending on whatever I play with in the sampler or lentgh of the video in 720p.
I just do other stuff when it gens.
Though I'm still waiting for nunchaku to release, it will make this so much easier.
>>
File: Chroma-ComfyUI_00230_.png (1.47 MB, 1024x1024)
1.47 MB
1.47 MB PNG
>>106573174
Thanks man, you can make them yourself if you want, I've uploaded all the loras I used for the images I posted as well as some prompts

Here's the one for this image

prompt: rcng curled hair headband makeup pink polka dot dress inside old car broken window cleavage detailed highest quality

negative: low quality, ugly, unfinished, out of focus, deformed, disfigured, blurry, cropped

Used two loras, the person lora at 0.8-0.9 strength, and the Miles Aldridge lora at 0.7

Oh, and here's the lora for abigail shapiro:
https://files.catbox.moe/co70vz.safetensors

Have fun!
>>
>>106573334
>I'm still waiting for nunchaku to release
Heh, too bad, kid. Looks like Some random model nobody will ever use just released. Back of the line.
>>
File: AnimateDiff_00021.mp4 (1.05 MB, 480x928)
1.05 MB
1.05 MB MP4
I'm just confused now. The flash/color shift seem to have fixed itself after I swapped back and forth between the CombineVideo and CreateVideoSaveVideo nodes.

Should I match the fp8/fp16 clip when using either diffusion models?
>>
>>106573342
Yeah it's like their devs all have ADHD or something.
>>
>>106573343
If you have the hardware for it, then run full size everything. Also fp8 is dogshit compared to q8
>>
>>106573358
>Also fp8 is dogshit compared to q8
Finally someone agrees with me.
>>
>>106573358
Fp8 scaled is equivalent to q8.
>>
>>106573374
>equivalent
No, it's not. It really isn't.
>>
>>106573358
It's so strangely difficult to find all the available models that I never bothered to try something different than what's been recommended in a majority of guids.
>>
File: ComfyUI_temp_ukpep_00001_.jpg (1.12 MB, 1152x2016)
1.12 MB
1.12 MB JPG
pic related Chroma 2K upscaled by Base
>>106573307
Chroma 2K upscaled by 2K

Both 0.70 denoise. 60 steps.
>>
>>106573376
It is for the scaled version.
>>
>>106573392
No, it's not. Do a simple A/B test and see for yourself.
>>
>>
>>106573410
wheres the blowjob anon. someone needs to put a dick in this shill's mouth
>>
>>106573410
that's slopped as fuck, at this point I would simply use Nano banana to make an image of him then i2v with Wan
>>
File: 2859849964.png (2.56 MB, 1248x1824)
2.56 MB
2.56 MB PNG
>>
File: 1727364677437103.png (1 KB, 82x35)
1 KB
1 KB PNG
>>106573426
I'm busy
>>
>>106573426
>someone needs to put a dick in this shill's mouth
nigga that's gay!
>>
>>106573470
I see youre not familiar with >>106573463 his work. A true artist.

NO ONE IS SAFE.
>>
>>106573434
Yeah it seems this "fun" vace deep fries the outputs.
>>
This is an anime website, I have priority and the right to tell all 3DPD fans to fuck off.
>>
>>106573568
>>>/g/adt/
>>
>>
>>106573586
what part of that required vace?
>>
>>106573591
None. I'm just playing with the model and posting it here for all to see. Maybe stop being such a salty vagina?
>>
now you've done it
>>
This new vace sucks, I get better results with 2.1 VACE unfortunately.
>>
File: ComfyUI_13327.png (2.46 MB, 1080x1440)
2.46 MB
2.46 MB PNG
>>106573568
Jenny's cuter than any cartoon character I've ever seen.
>>
>>106573617
It's not really vace. It's FUN vace. Basically vace trained by a team that didn't do the original vace. I think there is some fun shit in there too. I usually just avoid fun models as they're genuinely just bad.
>>
WOW I'm stupid. I thought all 2.2 lora's that needed both high and low came down in zips. so when it only showed one I'd copy and paste it and use the high on low again.. just now i noticed there was links near the top that had the low noise attached. I'm learning! ^_^
>>
>>106573410
Furkan? More like Glurkan.
https://files.catbox.moe/hycrkv.mp4
>>
>>106573619
I sure hope 'Jenny' is the name of one the horses in this image
>>
>>106573648
I dare you to send this to him on twitter.
>>
>>106573648
Battle of kosovo 1389 toughest turkish soldier battling against the weakest slavic spear colorized
>>
>>106573648
This is just how turkish men greet eachother, nothing to see here
>>
Any anon not using lightning loras for wan2.2?
What sampler and scheduler do you use for best result?
>>
>>106572757
>>106573338
very nice and only 125mb. do you have more bro?
>>
>>106571272
what breast lora is this?
>>
>>106573671
dpm++sde desu

Unless you're using native, then I dunno.
>>
>>106573683
probably this
https://civitai.com/models/1944129/slop-bounce-wan-22-i2v
>>
>>106573685
Thanks, what scheduler anon?
Simple?
>>
>>106573689
Shrimple
>>
>>106573693
Alright, I'll try that, thanks.
>>
>do i2v of a foxgirl with wan
>wan transforms her into a full blown anthropomorphic fox
cursed model
>>
>>106573300
they've made wan 2.2 vace finally?
>>
>>106573701
did you prompt "the foxgirl does x"?
>>
>>106573709
yea pretty much
>>
>>106573703
Kinda? It's by a different team who did the original vace. They just used the same training method and are also from alibaba.

It kind of deep fries stuff though,
>>
>>106573678
Yeah I have lots more, but I have to go through them to find the best epochs, I'll likely be uploading Christina Hendricks next since there was an anon requesting her, again just have to see which epoch to pick. Anyway, that will be later since I'm about to hit the sack.
>>
File: wan22_light21_00197.mp4 (1.7 MB, 720x720)
1.7 MB
1.7 MB MP4
>>106573712
just write girl
>>
File: kek.png (482 KB, 750x1000)
482 KB
482 KB PNG
>>106573648
the funny thing is that I have suspicion he's lurking there, I can imagine the fumes comming out of his ears lmao
>>
>>106573727
how the hell did you avoid the model making her endlessly yap?
>>
>>106573750
I really don't think he is. I'd like to see your proof.
>>
do wan 2.1 and 2.2 share the same loras in i2v?
>>
>>106571187
>>106571259
>>106571314
this sounds like avatartranny talk
i'm gonna have to kindly ask you to leave
>>
>>106573770
no? but you can use 2.1 loras with 2.2. might have to increase the strength on the high model
>>
>>106573791
You don't have to use the LoRA on the high model.
>>
>>106573776
excuse me, ran is an "avatartranny" and a founding member of this thread, therefore /ldg/ is an avatartranny thread
>>
>>106573791
do I need to apply loras on both high and low models?
>>
>>106573795
care to explain?
if you use a 2.1 motion lora, you would need it in high noise lora of 2.2
>>
Is there a working black dick lora for wan2.2 it keeps making weird stumps. hell even brown would be okay.
>>
i wish i could play in mud with comfy
we could build a castle together or something
>>
>>106573806
yes, I don't what that other anon is talking about
>>
>>106573801
whatever you say, avatartranny
>>
>>106573809
use a dick lora then ask for any color you want
>>
>>106573809
Brother Xi says go be brown somewhere else, hapa century is upon us.
>>
>>106573806
nta but I usually go with 2.1 loras used in both high and low
it works fine
>>
>sage attention stopped my pc from freezing up when offloading to the ram

Very nice.
>>
>>106573823
so a dick lora and a blowjob lora? does the order matter or nah? I think for this one I specifically used the bbc lora for 2.2 and they kept coming out like this.
https://files.catbox.moe/wz3piu.mp4
>>
>>106573839
Yes, unless there is a "black dick blowjob" lora, then use that.
Share your wf anon.
>>
>>106573837
> sage attention
what is that?
>>
>>106573839
>so a dick lora and a blowjob lora
sure

>does the order matter or nah
lora order doesn't matter, you can even use lists
what matters is their relative weights

>I think for this one I specifically used the bbc lora for 2.2 and they kept coming out like this
either your workflow is broken, or the lora itself is broken
>>
How can I add denoise or creativity to the refinement process in swarm in txt2img??
>>
File: screenshot.1757719812.jpg (239 KB, 661x907)
239 KB
239 KB JPG
Got tired of not being able to play games while WAN gens, so I threw my old 1070 back in. Finally I have my life back
>>
>>106573859
I'm using wan2gp so no workflow only web interface. I couldn't install comfyui. Well I installed it and set it up but i couldn't gen because i kept getting the same vram error. I only have 8 and 64 system ram
>>106573867
Yeah it might be something on my end.
https://files.catbox.moe/wz3piu.mp4
https://files.catbox.moe/ljcm7a.mp4
white dick works fine
https://files.catbox.moe/0hi268.mp4
>>
File: screenshot.1757768249.jpg (127 KB, 519x461)
127 KB
127 KB JPG
>>106573881
whoops
>>
>>106573884
>wan2gp
oh, yeah can't help on that sorry anon, I don't use wan2gp
>>
>>106573895
it's okay I'm going to probably try comfyui again soon. I'll fresh install everything, including git/python etc
>>
>>106573884
try a combination of the oral insertion lora and/or bbc service lora and/or wan 2.1 penis lora on low noise
>>
>>106573918
I'll try that next. also wtf lmao
https://files.catbox.moe/gazhhx.mp4
>>
>>106573839
try "with red tip" or "red glans"
>>
>>106573928
Pretty sure this is from one of the leet speak loras that does this shit whenever the dick isn't in the image from the beginning.
>>
File: ComfyUI_temp_tyvuu_00046_.png (2.98 MB, 1152x1152)
2.98 MB
2.98 MB PNG
https://github.com/Nerogar/OneTrainer/issues/985
Is this gonna fuck up my lora training? Should I wait for fix?
>>
>>106573935
I'll give that a shot
>>106573939
Yeah its this jfj one https://civitai.com/models/1497390/deepthroat-blowjob-wan-2x-i2v-and-t2v
and yeah i think it'll keep doing this if the dick isnt in the frame.
https://files.catbox.moe/q3a65c.mp4
>>
>>106573862
Performance stuff. Was a bitch to get it working. Found a script someone made to install it in the end.
>>
File: w.mp4 (2.03 MB, 640x1144)
2.03 MB
2.03 MB MP4
>>106573683
Yes >>106573688 is correct. I had linked it in the other post.
>>
>I had fp16_fast enabled in wanvideo model loader this entire time
SNAKE OILED AGAIN
>>
>>106573959
duh, you need oral insertion
>>
Stop using the faggot kijai's WAN nodes. He's forcing scaled models and text encoders down your throat
>>
scaled models are good
>>
File: thinking.jpg (68 KB, 1234x1186)
68 KB
68 KB JPG
>fp8 scaled is fine
>no it's shit, use gguf
>>
>>106574027
Download both, do some a/b tests and decide for yourself.
>>
>>106574027
if you can run Q8, always go for Q8, as simple as that
>>
>>106573884
>wanting to create blacked videos
Have you considered chair and rope instead?
>>
>>106573881
>he doesn’t just have two separate computers
Anon honey, get it together.
>>
>>106574042
No thanks you thin skinned faggot, they will be for personal use. simple as.
>>
>>106574027
it can be ok. but on average and across models chances are high that the Q8 GGUF is better
>>
>>106574027
never seen any difference between q8 and scaled fp8
>>
File: AnimateDiff_00355.mp4 (3.55 MB, 1040x690)
3.55 MB
3.55 MB MP4
>>106573968
SLOPPA BOUNCE
>>
File: 1458444987906.jpg (141 KB, 534x820)
141 KB
141 KB JPG
WHERE THE FUCK DO I DOWNLOAD THE Q8 MODEL?! GOOGLE GIVES ME FUCKING JACK SHIT
>>
oh maybe moving in the right direction.
https://files.catbox.moe/3szozy.mp4
https://files.catbox.moe/l0c0hu.mp4
also kek at detachable penis
>>
>>106574064
if a woman's body doesnt look at least like this, she is worthless
>>
>>106574079
often the gguf models are on huggingface, I don't know which model you're looking for tho

it also wouldn't be hard to create a Q8 GGUF from an original model for some model types
>>
what do you use for cfg in high and low noise for wan?
I go with 3.5 then 1 for low noise
>>
what's the ETA on nunchaku wan 2.2? these lightning loras are shit in very different ways
>>
>>106574146
soon™
>>
Does this power lora node require one each for high/low, meaning you also need two clip nodes? And both of them goes into the prompt nodes? Or does it just require one which is shared between the two parts?
>>
File: WAN_FINAL_00001.mp4 (1.15 MB, 480x800)
1.15 MB
1.15 MB MP4
>>
ATTENTION, LATEST COMFY UPDATE BREAKS KJNODES. IT WAS JUST UPDATED WITH THE FIX 3 HOURS AGO.

https://github.com/kijai/ComfyUI-KJNodes
https://github.com/kijai/ComfyUI-KJNodes
https://github.com/kijai/ComfyUI-KJNodes
>>
>>106574203
>pulling
>>
>>106574146
Can be next weeks, can be in 6 months.
My guess is 6 months.
>>
File: screenshot.175777069311.jpg (1.56 MB, 3715x883)
1.56 MB
1.56 MB JPG
>>106574048
I have two servers. I could throw the gpu in one of them but so far i see no issue with having both gpu's in my desktop. I can play my games at full fps + gen

its great!
>>
>>106574242
Does she apply enough hydro herself for Lauma?
>>
>>106574254
yes, she's probably the best user for lunar bloom with lauma.
>>
File: grosclip1.webm (1.37 MB, 1080x1920)
1.37 MB
1.37 MB WEBM
>>106571086
what's wrong with this 'neighbours' link section? just a pile of degenerate loli porn crap
where the fuck am I supposed to share my videos of obese white men getting biomechanically fused into car chassis again?
>>
>>106574259
here
>>
>>106574203
seems to be all fixed from what I'm seeing?
>>
>>106573942
Please respond
>>
>>106574268
yes, but you need to git pull for the fix. im just letting people know which node was causing the problem
>>
is there an easy remote way to train wan loras yet
>>
>>106574171
>Does this power lora node require one each for high/low
yes
>>
>>106574295
>rent pod with 4090
>train
>>
>>106574278
No, dont use embeddings
>>
>>106573918
>>106573935
>>106574015
AH! Thank you so much! Finally got it working
https://files.catbox.moe/k8mkbk.mp4
changing the strength of insertion lora to .3 on high and low has done the trick. now to see if i can get the glans right and it's set, thanks alot anons!
>>
>>106574171
You don't need to connect clip
>>
Is there a ComfyUI extension with InvokeAI's Canvas feature? I want Canvas in ComfyUI without downloading InvokeAI.
>>
>>106574305
>>106574404
The duality, devilish.
>>
>>106574171
no clip needed
>>
>>106574438
first anon is saying you do need two power lora loader nodes. one for high, one for low. second anon is saying you dont need to connect the umt5_xxl clip to the power lora loader.
>>
>>106574438
read both posts again with (((reading comprehension:1.4)))
>>
Has anyone tried making a realistic person and using it with an anime style lora?
>>
File: comfy2.jpg (727 KB, 1280x1280)
727 KB
727 KB JPG
>>
>>106574594
thats been done a thousand times
>>
File: AnimateDiff_00038-1.mp4 (3.96 MB, 480x928)
3.96 MB
3.96 MB MP4
Huh, you need different workflows for light loras?
This is without them.
>>
>>106574662
>you need different workflows for light loras
no? you just load the loras, set CFG to 1 and lower the steps.

also what the fuck is going on in that
>>
>>106574719
ahhh wrong thread show mercy jannie-san
>>
>>106574719
MR VILE AD-SKENAZI SHEKELSTEIN GET DOWN
>>
Schizo idea:
ComfyAnon must to fine tune an open weight LLM with a very specific purpose:
to master workflows.
The model is trained to do three things perfect, Interpret and explain json workflows,
generate complete workflows from user's plain lanuage prompts, debug and fix broken or inefficient workflows.

What do you think? It could be the perfect solution!
>>
>>106574743
More trouble than it's worth.
>>
>>106574743
Or you can just make your own.
>>
>>106574743
More like retard idea. Nodes change and their code gets updated all the time, dumping the docs into context is enough.
>>
I've seen some loras on civitai that claim they did it with one image. anybody here done that?

i assume its the same image over and over again but in different angles and flipped
>>
File: ComfyUI_00143_.png (2.72 MB, 1080x1328)
2.72 MB
2.72 MB PNG
Get baking the next one!
>>
>>106574768
model????
>>
Why does ComfyUI filter so many people?
You wanna use cutting edge latest tech, just learn it.
Or write your own stack you clearly have an idealised approach.
It's not comfy at all, no, but it's not even that tough to use. This is like OOT Water Temple levels of victim complex.
>>
File: 4206983601.png (1.14 MB, 896x1152)
1.14 MB
1.14 MB PNG
>>106574768
My, what lovely thighs.
>>
>>106574791
>Why does ComfyUI filter so many people?
I want to do adetailer inpainting and controlnet and regional prompting, did you see the worfklows for that? or do you just proompt and click generate and wait 1hour for a qwen result?
>>
File: ComfyUI_00156_.png (2.69 MB, 1080x1328)
2.69 MB
2.69 MB PNG
>>106574788
its Qwen
>>
>>106574816
yeah true all of those experiences are dogshit on comfy
but I never inpaint on comfy, I use it exclusively for new models like Wan or Kontext
are you inpainting on Qwen or something?
Not even nuForge is any use there I think
but yeah tools for the job and all that, there is absolutely no argument to be made to use one tool for all jobs
>>
>>106574693
It does make sense, but this setup breaks once the light lora isn't running. I upped the cfg and steps and it's still broken.
>>
>>106574816
stick to forge/neo for shit like that. Noodles are a menace to deal with for
>>
>>106574830
Exactly. NeoForge is also bad when it comes to inpainting, outpainting, and regional prompting. InvokeAI also feels years behind in terms of v‑pred implementation support, plus it’s not portable. I find generating complex scenes tricky with pretty much any UI.
>>
>>106574858
will we ever have like a standalone app/software implementation off all this things moving over comfyshit
>>
File: comfy334.jpg (1.09 MB, 1280x1280)
1.09 MB
1.09 MB JPG
>>
I tried to gen at 1080x1920, and the video barely has any motion.
Does the Shift value have to go up exponentially with the increase of base resolution?
>>
>>106574921
asianposters cant stop winning, catbox?
>>
>>106575022
No, wan is completely broken above 720p, since it was trained at this resolution.
>>
File: 1750070905479738.jpg (68 KB, 700x394)
68 KB
68 KB JPG
>>106575022
>>
>>106575022
>gen at 1080x1920
anon...
>>
>>106575022
How long did it take lol?
>>
>>106574794
How are these pants called? I want to abuse it in my prompts from now on.
>>
NeoForge gods, 2 things:
1)
Using ControlNet Gaussian Blur at 0.01 in txt2img works like img2img but keeps txt2img perks (hires fix, add‑ons).
You can remove the img2img BLOAT tab.

2)
You can add # comments in prompts. They’re ignored by the model, letting you label/organize prompts without affecting output.
>Example positive prompt:
# Scene setup
A vast medieval battlefield at dawn # environment
highly detailed painting # style
cinematic lighting # mood

# Focus
knights in shining armor,
dragons flying overhead,
siege weapons in the distance,
>>
>>106575032
>>106575044
>>106575051
Damn, the quality looked so good.

>>106575075
At 10 shift, strangely just 2minutes, but just the head was moving. So I went to 30 and that took 12min.
>>
>>106575083
nta but probably leather corset shorts, high waisted or something like that
>>
File: ComfyUI_00192_.png (2.67 MB, 1080x1328)
2.67 MB
2.67 MB PNG
>>106574921
way too noisy anon
>>
File: Chroma2k-test_00005_.jpg (581 KB, 1440x1632)
581 KB
581 KB JPG
>>
File: ComfyUI_00194_.png (2.66 MB, 1080x1328)
2.66 MB
2.66 MB PNG
>>106575128
>>
File: 2309063243.png (1.11 MB, 896x1152)
1.11 MB
1.11 MB PNG
>>106575083
>>106575122
"Leather booty shorts with lace-up details" is what I used.
>>
>>106574597
Model?
>>
File: 1838541782.png (1.58 MB, 1248x1824)
1.58 MB
1.58 MB PNG
>>
File: ComfyUI_00209_.png (2.7 MB, 1080x1328)
2.7 MB
2.7 MB PNG
>>106575173
your prev gen was perfect
>>
Schizo idea2
You know what would be really cool? A tree view mode for our gens.
That way, when you’ve done lots of inpainting or img2img, you could trace how each image was developed, branch by branch, and pick exactly the version you want by seeing its whole “evolution path.”
>>
File: 2898715054.png (1.18 MB, 896x1152)
1.18 MB
1.18 MB PNG
>>106575259
Thanks, yours look good too.
>>
Why is ComfyUI so complicated? I just want to goooooon
>>
>>106575298
you can vibe code this in a weekend
>>
>>106575328
Ha! You know what? I'm a coomer, I don't know nothing about vibe coding.
>>
>>106575298
@haoming02
>>
What's the best wan 2.2 guide? The one in rentry says old.
>>
>try a sfw i2v
>it barely moves
>attach nsfw lora
>it starts to behave like it should

Are there loras for sfw use? Just generic stuff, like how there's detail loras for image genning.

>>>/d/11376267 Finished this earlier, first successful loop. This is one of the few best generals on the site. Who thought AI would bring people together?
>>
next
>>106575437
>>106575437
>>106575437



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.