[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/g/ - Technology


Thread archived.
You cannot reply anymore.


[Advertise on 4chan]


Manhattan Project Edition

Discussion of Free and Open Source Text-to-Image/Video Models and UI

Prev: >>106435682

https://rentry.org/ldg-lazy-getting-started-guide

>UI
ComfyUI: https://github.com/comfyanonymous/ComfyUI
SwarmUI: https://github.com/mcmonkeyprojects/SwarmUI
re/Forge/Classic: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassic
SD.Next: https://github.com/vladmandic/sdnext
Wan2GP: https://github.com/deepbeepmeep/Wan2GP
AniStudio: https://github.com/FizzleDorf/AniStudio/tree/dev

>Checkpoints, LoRAs, Upscalers, & Workflows
https://civitai.com
https://civitaiarchive.com/
https://tensor.art
https://openmodeldb.info
https://openart.ai/workflows

>Tuning
https://github.com/spacepxl/demystifying-sd-finetuning
https://github.com/Nerogar/OneTrainer
https://github.com/kohya-ss/sd-scripts/tree/sd3
https://github.com/derrian-distro/LoRA_Easy_Training_Scripts
https://github.com/tdrussell/diffusion-pipe

>WanX
https://rentry.org/wan22ldgguide
https://github.com/Wan-Video
https://alidocs.dingtalk.com/i/nodes/EpGBa2Lm8aZxe5myC99MelA2WgN7R35y

>Chroma
https://huggingface.co/lodestones/Chroma1-Base
Training: https://rentry.org/mvu52t46

>Illustrious
1girl and Beyond: https://rentry.org/comfyui_guide_1girl
Tag Explorer: https://tagexplorer.github.io/

>Misc
Local Model Meta: https://rentry.org/localmodelsmeta
Share Metadata: https://catbox.moe | https://litterbox.catbox.moe/
GPU Benchmarks: https://chimolog.co/bto-gpu-stable-diffusion-specs/
Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-one
Txt2Img Plugin: https://github.com/Acly/krita-ai-diffusion
Archive: https://rentry.org/sdg-link
Bakery: https://rentry.org/ldgcollage

>Neighbours
https://rentry.org/ldg-lazy-getting-started-guide#rentry-from-other-boards
>>>/aco/csdg
>>>/b/degen
>>>/b/realistic+parody
>>>/gif/vdg
>>>/d/ddg
>>>/e/edg
>>>/h/hdg
>>>/trash/slop
>>>/vt/vtai
>>>/u/udg

>Local Text
>>>/g/lmg

>Maintain Thread Quality
https://rentry.org/debo
>>
Baker added my gen to the collage, but the one I hated rather than the one I liked.
>>
>page 1 baking
>>
>>106438088
I flunked out of collage.
>>
File: WanVideo2_2_I2V_00245.webm (2.11 MB, 720x1248)
2.11 MB
2.11 MB WEBM
Gomennasai
>>
>>106438005
>>106438066
if first link guy is still here, i replied to that in the second link in case you didn't see it at tail end of last thread
>>
>more smoking
fucking based
>>
File: 1741475581851197.mp4 (1.36 MB, 640x640)
1.36 MB
1.36 MB MP4
2 strength high, 1 low with kijai 2.2 lora seems to give good motion, leave low alone and tweak the high value I guess.

https://huggingface.co/Kijai/WanVideo_comfy/tree/main/Wan22-Lightning

also, isn't it neat how good wan reflections are?
>>
File: 1738357217622606.mp4 (1.4 MB, 640x640)
1.4 MB
1.4 MB MP4
>>106438126
I said cuts a slice, thats a big one miku...
>>
>>106438103
Don't want to hate on the baker because I appreciate the effort but ever since he made the collages animated I tend to not open or care about it because it is a pain to look at. Too much moving shit and it is not as convenient to zoom and pan as it was with a single image.
>>
>>106438088
he only adds his own toilets and cringeposts, best not to take it too seriously
>>
File: FluxKreaOutput9992.png (2.68 MB, 1248x1824)
2.68 MB
2.68 MB PNG
>>
>>106438155
Actually I'm the toilet poster and I am not the collage maker.
>>
>smoking
giga cringe
>>
>>106438155
I doubt he posted the toilet because if he hates chroma why did he add my chroma gen.
>>
>>106438159
Man this woman is a prime example of why I prefer Asian women.
>>
I think the baker favors a few posters he likes and avoids a few posters he doesn't like and other than that is basically random.
>>
Is there anyone doing actual research on finetuning prompts for flux and wan? Im currently using wan2.2 and its hard to find people to talk about good prompting when there is so much prebaked shit outputting the same glossy garbage
>>
File: 1726961582996362.webm (2.37 MB, 640x640)
2.37 MB
2.37 MB WEBM
yep, 2 high 1 low (2.2 lightx2v) seems to work well.

isn't it wild we have all these good local models (wan, noob/illustrious, qwen, etc), making a clip like this would cost me money/tokens on an AI SAAS site and even then if you fuck up you have to pay more.

not with local. you could batch generate all day and all you really pay for is your pc, and electricity.
>>
>>106438184
I too believe that the OP is a human with preferences
>>
>>106438110
Thanks, good info
>>
>>106438199
Yeah it sounds obvious but some people have him pegged as a schemer posting his own slop.
>>
>>106438184
I have definitely fallen out of grace with a baker on occasions. I don't know what they like but I know what they don't.
>>
File: 1738033129065328.gif (1.42 MB, 256x320)
1.42 MB
1.42 MB GIF
since most of x.ai got leaked by their ex employee do you guys think this might mean we get open source grok companion rip offs in the future? I'd love to have a customizable desktop ani style assistant
>>
>>106438208
having their own gen picked just once should dispel that myth
...oh.
>>
>>106438217
>since most of x.ai got leaked by their ex employee
Yeah straight to openAI. So he open sourced the closed source to another closed source competitor.
Committed corporate espionage in the dumbest way possible and literally nobody benefits. What a retard.
>>
>>106438217
there's no value in the software, only the weights
>>
File: 1753391355018704.mp4 (806 KB, 640x480)
806 KB
806 KB MP4
the man walks to the tent and zips it up (from a previous gen)

the shadow casting from the fire is so cool, if a bit exaggerated. wan is amazing.
>>
File: 1755157248497687.png (1.75 MB, 1328x1328)
1.75 MB
1.75 MB PNG
>>
>>106438217
Clickbait and no.
>>
File: 1736093445239135.jpg (92 KB, 500x375)
92 KB
92 KB JPG
>>106438217
99.9% chance this is just musk coping about chatgpt being better than grok and throwing shade at the guy who switched sides
>>
File: 1755199027480294.mp4 (677 KB, 640x480)
677 KB
677 KB MP4
The man in green casts a fishing rod into the nearby lake.

neat
>>
my company is forcing us to use copilot to improve effective and streamline shit and I actually like to far which feels weird
>>
>>106438290
uhh, GPT5 sucks dick, and Sam Altman is a retard

even Google is beating openAI now
>>
>>
>>106438305
And grok sucks worse, by a lot
Also I've got subscriptions to both
Gpt thinking is still way better than Gemini pro
Gemini is much more willing to just lie to you
>>
File: ComfyUI_05510_.png (779 KB, 1280x720)
779 KB
779 KB PNG
>>
>>106438321
grok will win because Elon is the world's richest man and wants to rape Altman for being a con man with his "non profit".
>>
>>106438343
I like slop and I’m tired of pretending I dont
>>
>>106438349
Elon seems too busy making goonbots to take on anyone.
>>
>>106438378
waifus drive innovation. look at noob and illustrious. pony used to be inconsistent with anatomy, although mostly good. now, I can get a perfect miku every time without a controlnet. now, you can use controlnets as intended: for posing/lineart or whatever, instead of fixing hands.
>>
>>106437937
Qwen is economical to train with EQ-VAE and TREAD. Really though, the real next qualitative step for local is using EQ-VAE, TREAD, and maybe another couple order of magnitude optimizations to build our own models from scratch.

When model training is optimized enough for local base model training, there will be an explosion of experimental models leading to even more optimizations, standards will emerge, and local will win eternally.

It's closer than you think. This guy used EQ-VAE+TREAD on his local trained model:
https://huggingface.co/KBlueLeaf/HDM-xut-340M-anime
https://xcancel.com/KBlueleaf/status/1950989468175147112#m
>>
File: 1753324352153362.mp4 (892 KB, 640x480)
892 KB
892 KB MP4
he did not like that burger...
>>
>>106438369
thats art, not slop tho
>>
>>106438343
What model lets you dictate what words will appear on clothing?
I know Flux does, but what else?
>>
>>106438426
Technically all of them. The success rate tho, is another argument.
>>
>>106438426
Pretty much any non XL model really
>>
>>106438426
chroma
>>
I get into the collage all the time, but I think it's mostly due to not spamming every single thread with my gens. Like, you have an idea, gen it alone. Share it later. And then get a new idea.
>>
>Chroma anons have a normal discussion
>Some Qwen anon turns it into shitflinging contest

This is why we can't have good things.

Just try to grasp why anyone would be so offended by anyone using a model that they personally enjoy.

When did models stop being personal choice, rather than a brand?

Oh, right. It's a troon. Very likely coming from /sdg/ and the like.
>>
>>106438454
>Someone calmly points out issues with Chroma
>Completely lose your shit and begin pear clutching like the model is your mother or something.
>>
I made a slop vid for tiktok but it's at zero views and I'm pretty sure it's not being allowed through on the server side
>>
>>106438472
Does this look like /ttsg/ to you?
>>
>>106438467
>Calmly

Repeating the same shit for the past 100+ threads any time someone mentions anything positive about Chroma to the extent of hijacking every thread where there's any activity is not "calmly".
>>
>>106438491
Maybe freaking out every time someone says something negative about Chroma isn't the play?
>>
>>106438445
I get into the collage all the time and I never try to get in the collage, I actually dont care if I make it in or not.
I just share a cool pic or video I made and it usually ends up in there.
>>
File: WanVideo2_2_I2V_00246.webm (1.15 MB, 720x1248)
1.15 MB
1.15 MB WEBM
>>106438159
>>
>>106438496
>Freaks out that someone is calmly discussing Chroma
>Gets mad when someone calls him out in the same wag
>>
>>106438507
All I'm saying is, most models take the critique and laugh it off. Chroma can survive without you white knighting for it.
>>
File: 1737239675918716.png (1.51 MB, 1120x1440)
1.51 MB
1.51 MB PNG
qwen is nuts
>>
>>106438500
I too am a hipster collage-r who gets in while not caring.
>>
If I have to wait two minutes for a gen, I'm not going to choose a model that cannot generate naked people by default. And in the next 10 years, either some Chinaman is going to invent a way to give everyone 256GB of VRAM or they are going to ban personal non-"gaming locked" GPUs all together.
>>
>>106438526
I'm not talking about the model. Discussions on this thread in general is at an all time low. The quality of the posts is shit because of people like you.
>>
>>106438500
even postcard has gotten in, if you cant get it you need to git gud
>>106438308
>>106438505
diaperposting & nsfw should be insta-3day though, you guys are trying to destroy the general
>>
>Trani trying to sneak his shitty UI in OP again
Can you please fuck off you schizo faggot loser
>>
>>106438174
to each their own lol (I like ladies from all races personally DESU)
>>
>>106438562
You literally cannot help yourself from responding to anything I say to you.
>>
>>106438481
Yeah, it's not?
>>
>>106438565
bruh wtf, does anyone have a link to where the 4chan rules state or imply that diapers are bannable??
>>
File: 1753470285145771.mp4 (946 KB, 640x480)
946 KB
946 KB MP4
there, finally got a table flip.
>>
>>106438505
nice
as the genner of the original pic I approve of all versions of her quite frankly lmao
>>
>>106438586
He's becoming Baron Trump
>>
>>106438571
him or his boyfriend are baking the second the thread hits 310 replies to avoid me removing it. lel
>>
S-stop ruining the quality of the general by discussing your opinion on models r-right now!
>>
>>106438454
>intentionally ignoring the multiple people who clearly have balanced views on the pros and cons of both models
>>
>>106438601
It's funny that he keeps doing this and then pretends to be a regular whenever newfags only to get ejected by anon because of his autism and then trying to assert himself as a figure of authority in the general. Even the anime thread told him to kick rocks
>>
File: ComfyUI_16433.png (2.88 MB, 1200x1600)
2.88 MB
2.88 MB PNG
Did you guys create any Frankenstein (monster) gens?
>>
>>106438583
All our threads are either in /hc/ for 3dpd or /aco/ and /d/ for pics.
t. diaperfag
>>
>>106438626
are you sure you want my failed gens, made with fast chroma? kek
>>
>>106438635
ok, but diapers are just normal objects... yeah diaper porn should be bannable, but not just any diaper
>>
>>106438617
You can have your views, state them and go on about your gens. But that's not what you're doing. You make sure you complain about anons discussing Chroma.
>>
>>106438558
>they are going to ban personal non-"gaming locked" GPUs all together
I hope they do this so that people go crazy trying to find ways to make models more efficient on shitty hardware and we finally break free of the yoke of Nvidia
>>
>>106438674
NTA, but I don't think you are mentally capable of reading a critique of chroma either exaggerated or in earnest in good faith.
It's clear you've formed an unhealthy parasocial relationship with the model.
>>
>>106438620
did they? he's in their op as well
>>
>>106438583
I literally refuse to answer this question (bisgustin)
>>
>>106438583
local diffusion everyone
>>
File: free.png (884 KB, 896x1152)
884 KB
884 KB PNG
>>106438343
How do you get the AI to understand that you want the drawing to look like a rough, unfinished, partially-colored sketch made by someone who isn't quite a professional yet but is beyond the level of being a beginner/amateur, in a style that is kinda halfway between western and anime, like a western artist trying to imitate an anime style?
>>
File: FluxKreaOutput9993.jpg (2.77 MB, 1664x2432)
2.77 MB
2.77 MB JPG
>>106438626
now I have
>>
File: file.png (306 KB, 1854x847)
306 KB
306 KB PNG
i read somewhere that some manage to get 20s/it with a 3060 12gb and i was wondering how to do the same?
i currently do 31s/it
i have 55.5gb of ram (it's a vm)
i just use --use-sage-attention --fast, i don't use any other flags
>things i tried:
- i tried --disable-smart-memory with either --lowvram or--novrm
but the peformance is always the same, 31-33s, so i removed them
- I also tried UNETLoaderDisTorch2MultiGPU to manually offload to ram, increased virtual vram in 1gb steps and it stopped running out of memory at 9gb while printing this:
Device Layers Memory (MB) % Total
--------------------------------------------------
cuda:0 (<0.01%) 283 3.91 0.0%
cuda:0 134 4428.59 32.5%
cpu 271 9196.57 67.5%
it also performed at 31s/t, so i stopped using it and just use the model loader in pic related
>>
>>106438644
Show me what you got!
>>
>>106438756
use an anti-slop lora for more natural drawn lines and specify amateur sketch in the prompt
might need to also specify color, "sketch" prompts often end up as lineart in my experience.
>>
>>106438674
nah i have the exact opinions i referred to in the comment you're replying to lmao
>>
>>106438715
Point me to a single critique that isn't made in bad faith to shut down all discussion about Chroma because the argument boils down to "it sucks, X is better for my niche reasons".
>>
>>106438756
use base Pony V6 with like DDIM or some shit
*rimshot noise*
*canned laughtrack*
*BEGHAHAHA anon maybe*
>>
File: ComfyUI_00075_.png (3.82 MB, 1464x2144)
3.82 MB
3.82 MB PNG
>>
File: file.png (34 KB, 894x146)
34 KB
34 KB PNG
>>106207594
>i have a workflow that works on 3060 12gb + 64gb ram
>https://litter.catbox.moe/kum8tt4727pbcjrb.json
anon could you please reupload your workflow?
>>
>>
>>106438793
It shouldn't matter if the argument is made in bad faith or not, very rarely do people argue in good faith on this platform. Most people can understand that and pick apart the pieces of substance from what the use is trying to say. Your autism clearly prevents you from doing that.
>>
>>106438884
>bar \(place\)
>home
ahhh
>>
>>106438884
me on the right
>>
File: 00102-1411743791.png (3.32 MB, 1368x2000)
3.32 MB
3.32 MB PNG
>>
>>106438906
bro that bar \(place\) in particular is in like every fucking model lmao
>>
File: 1726631166943819.mp4 (1.62 MB, 640x480)
1.62 MB
1.62 MB MP4
the car with lights on flies off a ramp, high into the sky above Tokyo.

just needs eurobeat
>>
>>106438756
copy your post, go to claude.ai, paste it there and i promise you will receive an adequate answer
>>
File: AnimateDiff_00248.mp4 (1.97 MB, 928x720)
1.97 MB
1.97 MB MP4
>>
Any tips for prompting diverse breast shapes and sizes?
>>
>>106439015
use english language
>>
>>106439015
>>>/d/11365586
>>
>>106439018
no one needs to see that
>>
File: 1750318036166311.mp4 (1.81 MB, 720x1040)
1.81 MB
1.81 MB MP4
>>106438823
>>
>>106438716
He has been aggressively trying to put his software in every general and from the looks of it the previous posters were upset about it and might have given up knowing how they love to squat in threads to push this garbage.
>>
>>106438369
Finally a based poster
>>
File: 1734957952914478.mp4 (2.2 MB, 912x720)
2.2 MB
2.2 MB MP4
>>106438884
>>
>>106439113
Excuse me while I kms
>>
>>106439113
nice. Vidya posters kinda propping up the last few threads with their takes on people's images IMO
>>
File: PonyaaaackV7Output.png (1.56 MB, 832x1216)
1.56 MB
1.56 MB PNG
>oh, Pony V7 looks a bit rough if you don't use artist tags? That's fine, there's been lots of models like that in the past
>no, Pony V7 looks a bit rough if you don't use the arbitrary made up "style clusters" which literally have names formatted like "style_cluster_99921"

Good thing the Pony V7 prompt enhancer DOESN'T output these clusters, but rather textual descriptions of style that it doesn't seem as though the model was actually trained on!

Prompt (literally generated by Pony's own prompt enhacer, again):
`A human female Kirisame Marisa from Touhou, with yellow eyes, blonde long hair, and a braided sidelock, looks up. She sits outdoors on grass, with a tree, sky, and clouds in the background. She wears a witch hat with a white hat bow, a black tabard, a white turtleneck sweater, a red neck ribbon, a black skirt, a waist apron, white bloomers, and ankle boots. Full shot, digital illustration, anime style, cell shading, vibrant colors, soft highlights, warm and cool color scheme, light from above, detailed background.`

The output is pic attached. Apparently "cell-shaded digital anime" means "Just Fucking The Same Old Generic Watercolorey Base Pony Style"
>>
File: solder-dram-togpu.webm (3.5 MB, 1248x720)
3.5 MB
3.5 MB WEBM
i figured out how to solder ram to my gpu!!!
>>
>>106439211
>>no, Pony V7 looks a bit rough if you don't use the arbitrary made up "style clusters" which literally have names formatted like "style_cluster_99921"
I'm in awe. Bewildered, even.
>>
>>106439221
shouldn't your eyes be all narrow?
>>
>>106439211

No exaggerating, base v6 still looks better.
>>
>>106438775
if you want it to look like garbage I wouldn't suggest forcing speedups
>>
File: 1732372731082079.mp4 (1.84 MB, 640x480)
1.84 MB
1.84 MB MP4
>>106438962
>>
File: AD_00009.mp4 (2.44 MB, 1440x912)
2.44 MB
2.44 MB MP4
i guess i need to do 3:2 if the background color changes
>>
>>106439278
kek
>>
>>106438071
Its taking 30 minutes to generate a video from text using Wan2.2 on a 3090, is that normal?
>>
>>106439319
Cant answer that without more information
>>
>>106439319
if you're genning at 20 steps or some shit without the lightx2v lora then yes. with lightx2v it's only a couple minutes
>>
>>106439329
>>106439331
My bad meant to say image to video. Will the lightx2v help with image to video?
>>
>>106438321
I can use Gemini 2.5 Pro for free on AI Studio with essentially no limitations up to LITERALLY a million tokens, indefinitely, this fact alone makes all of OpenAI's offerings look utterly ridiculous in comparison.

Like it's simply not my problem that OpenAI doesn't have infinite Google money to run their best models maxed out for free for everyone, why would I as an average enduser care about anything beyond the end result?
>>
>>106439331
>if you're genning at 20 steps or some shit
That might be one reason why, looks like it defaulted to 30 steps in the webui
Using Wan2GP through pinokio
>>
>>106439356
>Using Wan2GP through pinokio
How can such a short sentence make me feel physically ill TWO times?
>>
>>106439356
it's 4-6 steps with lightx2v, so significantly faster. i don't know how it works with wan2gp though
>>
File: WanVideo2_2_I2V_00248.webm (325 KB, 1248x720)
325 KB
325 KB WEBM
>>
File: 00221-635349957.jpg (468 KB, 1664x2432)
468 KB
468 KB JPG
>>106439031
>fat mercy drinking
this is the content I am looking for
>>
>>106439367
herc looks like bogie there. perfect.
>>
I've been checked out a while. What's the best speedup available for wan 2.2? I haven't checked since 2.2 first came out.
>>
File: AnimateDiff_00249.mp4 (2.75 MB, 720x912)
2.75 MB
2.75 MB MP4
>>106438756
>>
File: 1755103573467323.mp4 (638 KB, 480x720)
638 KB
638 KB MP4
>>106439385
>>
>>106439458
It still is the light 2.1 LoRA.
>>
>>106439491
Thanks. What are the ideal numbers for the high/low noise?
>>
>>106439484
oh fuck I laughed, nice one
>>
File: WanVideo2_2_I2V_00249.webm (3.23 MB, 1248x720)
3.23 MB
3.23 MB WEBM
>>
>>106439362
kek
>>
File: 1728676438439497.mp4 (1.32 MB, 640x480)
1.32 MB
1.32 MB MP4
the car with lights transforms into Miku Hatsune who runs on the road.

not what I expected
>>
>>106439521
>sir this appears to be a 2080ti and my client clearly sold his soul for a 6000 blackwell
>>
>>106439510
If you were to ask me?
4 high without the LoRA and 4 Low with the LoRA.
If you only care about making slop, 2 each with the LoRA.
>>
is 2.2 in a good spot now or should I stick to 2.1? right now it takes me about 5 minutes to make an 8 second 720p vid on my 5090 using that one workflow from that one rentry
>>
File: autismtest.webm (917 KB, 1248x720)
917 KB
917 KB WEBM
>>
>>106439585
idk how you can both have a 5090 and be unable to determine the difference in quality and whether it's worth it or not yourself.
>>
>>106439620
well I don't wanna download 9001gb of stuff only for it to be a meme
if there's enough of an improvement to warrant it then i'll do it
>>
>>106439629
2.2 is basically an improvement in every sense.
>>
>>106439545
Thanks again anon, I'll give it a try tonight.
>>
>>106439629
so you have a 5090 but you don't have good internet?
>>
>>106439633
alright, i'll check it out. i remember reading complaints about the motion and stuff, but if the eggheads figured it all out and the workflows and loras are all hammered out then... might as well try it

>>106439641
my internet is fine, but i dont wanna spend an evening getting it all running for nothing
>>
>>106439641
being rich won't you better speeds if all you have around are shitty isp's
>>
File: WanVideo2_2_I2V_00251.webm (1.88 MB, 1248x720)
1.88 MB
1.88 MB WEBM
You win some you lose some
>>
>>106439662
>being rich won't you better speeds
This is highly dependent on how rich we're talking.
>>
>>106439466
hahaha what the fuck
it's silly that she started crying over that
>>
>>106439693
I doubt a billionaire would browse this thread and talk about 5090's
>>
>>106439682
>this isn't the cheese pizza I wanted...
>>
>>106439694
insert rando creepanon like "something something lemme tell ya' boy, I'd be given er' something right there to be blubberin' bout, OHHH yessir"
>>
File: 1755919617689503.mp4 (591 KB, 640x480)
591 KB
591 KB MP4
the man in the green jacket throws a stone into the lake nearby causing a large splash.

neat reflections, used wan on my pixel qwen edit gen
>>
Does wan2.2 loras not need clip or is it just unet only?
>>
>>106439710
It’s you, you’re the creep anon
>>
>>106439721
no need for clip
>>
File: 1735533974898376.mp4 (701 KB, 640x480)
701 KB
701 KB MP4
>>106439720
a large bear appears from far away and chases the man in the green jacket to the right.

he doesnt seem to mind
>>
File: WanVideo2_2_I2V_00252.webm (3.7 MB, 1248x720)
3.7 MB
3.7 MB WEBM
>>
File: KREA'D.png (2.75 MB, 1824x1248)
2.75 MB
2.75 MB PNG
>>106439724
f. I'll make up for it with reimagining the Free Use Slut, this will help a lot
>>
File: 1741452663829617.mp4 (996 KB, 640x480)
996 KB
996 KB MP4
aloha snackbar!
>>
>>106439763
can you please stop copying my prompt ideas?
>>
>>106439784
>free use
>>
File: 1726595019494079.webm (1.37 MB, 480x672)
1.37 MB
1.37 MB WEBM
wan did a good job with kikuri

the anime girl drinks from the drink box in her hand.
>>
As someone who used to be really into anime but hasn't watched it in years. I always wonder who the fuck these people are.
>>
File: 1753820273503527.webm (2.45 MB, 480x672)
2.45 MB
2.45 MB WEBM
walking out the door, needed more time but it's neat how it maintained the outline style too.
>>
File: AnimateDiff_00250.mp4 (3.05 MB, 1280x720)
3.05 MB
3.05 MB MP4
>>
>>106439844
neat
>>
File: WanVideo2_2_I2V_00253.webm (2.71 MB, 720x1248)
2.71 MB
2.71 MB WEBM
>>
>>106439844
awesome!
>>
File: sdga.png (865 KB, 896x1152)
865 KB
865 KB PNG
>>106439784
I've been prompting free use sluts all day and you can't stop me
>>
File: 00003-2727256897.jpg (311 KB, 1728x1344)
311 KB
311 KB JPG
>>
>>106439784
I wouldn't say it's copying as much as openly, directly, blatantly feeding your pics to Gippity and or Gemini and then giving the resulting prompt to Krea lol
>>
File: WanVideo2_2_I2V_00254.webm (509 KB, 1248x720)
509 KB
509 KB WEBM
>>106439844
>>
>>106439491
>>106439545

Is it the 480p or 720p version that works?
>>
File: WanVideo2_2_I2V_00255.webm (440 KB, 1248x720)
440 KB
440 KB WEBM
>>106440008
Without the moon face
>>
File: 1746905029451356.webm (633 KB, 480x672)
633 KB
633 KB WEBM
well there you have it, this is how cats dance according to AI.
>>
>>106440070
ima be real. I don't think it makes a difference.
>>
File: 1756589460883765.png (1.38 MB, 768x1344)
1.38 MB
1.38 MB PNG
>OUR HATRED IS PURE
>OUR VENGEANCE IS JUST
>>
>>106440081
been a while, i missed you
>>
File: ComfyUI_00006_.png (1.23 MB, 1152x896)
1.23 MB
1.23 MB PNG
MIGU here. Why isn't your video feed working?
>>
>>106440176
Can you make a miku in Starcraft unit portrait style?
>>
File: 1750252915354476.webm (984 KB, 480x672)
984 KB
984 KB WEBM
so is 3 str high, 1 str low with 2.1 i2v lora still best?

seems good to me:
>>
Did Layer Diffuse ever go anywhere? I played around with it a year or so ago and it worked okay with SDXL. Does anything like it exist for later models like Flux?
>>
File: WanVideo2_2_I2V_00257.webm (2.46 MB, 720x1248)
2.46 MB
2.46 MB WEBM
>>
>>106440310
>so is 3 str high

Another one lost to the 3 on high meme. 0 on high.
>>
>>106440312
the snakeoil fountain kinda dried up
>>
>>106440325
>"How dare you!"
>>
Did that last comfy commit that said it fixed memory on windows actually work? I'll pull if it did.
>>
>>106440384
feels like four updates per day at this point
>>
>>106440332
how come no lora for the high noise part? just curious why.
>>
>>106440435
LoRA kills motion and all that is needed for the motion is 4 steps. The actual ratio to high noise vs low noise steps without a LoRA has extremely few high noise steps to begin with.
>>
>>106440325
I want to breed Greta so badly it physically pains me every time i see her and know that she is not mine.
>>
File: 1726187992203329.webm (1.16 MB, 480x672)
1.16 MB
1.16 MB WEBM
I said jumps through a red hula hoop...

nevertheless, stunt complete
>>
File: 1753227966160794.webm (1.34 MB, 480x672)
1.34 MB
1.34 MB WEBM
>>106440481
>>
File: 1743691688906233.jpg (75 KB, 1024x683)
75 KB
75 KB JPG
what in the FUCK am i doing wrong to get GREEN TRIANGELS
```
=== STEP 8 DEBUGGING ===
Transforming
Pre-transform debugging:
latents: torch.Size([1, 256, 64]), torch.bfloat16, nan=False
timestep: torch.Size([1]), torch.int64, values=tensor([8], device='cuda:0')
guidance: torch.Size([1]), torch.bfloat16, values=tensor([3.2969], device='cuda:0', dtype=torch.bfloat16)
prompt_embeds: torch.Size([1, 512, 4096]), torch.bfloat16, nan=False
pooled_prompt_embeds: torch.Size([1, 768]), torch.bfloat16, nan=False
txt_ids: torch.Size([512, 3]), torch.bfloat16
img_ids: torch.Size([256, 3]), torch.int64, nan=False
Transformed Latents Debug:
min=-29.625
max=30.375
mean=-1.2890625
std=11.375
NaN=False Inf=False
GPU Memory: 0.03 MB

Shape: torch.Size([1, 256, 64])

Time elapsed: 13.163323999993736s
=== END STEP 8 ===

Diffusion complete. Final latent shape: torch.Size([1, 256, 64])
DIFFUSED latents: shape=torch.Size([1, 256, 64]), mean=-1.289, std=11.375
Processing VAE
VAE Input Latents: Shape=torch.Size([1, 256, 64]), Mean=-1.2890625, mean=-1.289, std=11.375
Latents stats: mean=-1.289062, std=11.375000, shape=(1, 256, 64)
BEFORE reshape: torch.Size([1, 256, 64]), mean=-1.289, std=11.375
Calculated: channels_per_patch=16, spatial_size=16
Final VAE latents shape: torch.Size([1, 16, 32, 32])
Latents shape after prepare: (1, 16, 32, 32)
After scaling: mean=-3.891, std=31.500
Latents shape after scaling and shifting: (1, 16, 32, 32)
VAE output before normalization: mean=-1.086, std=0.652, min=-3.062, max=0.652
Final image tensor: mean=28.375, min=0.0, max=211.0
Decoded 1 image(s)
Saving Output
```
>>
just getting started with video. using the suggested rentry workflow at 24gb, but everything comes out burnt to a crisp. there is movement, but clearly something is wrong.
>>
>>106440541
Show the crisp
>>
File: 1734100254941210.mp4 (580 KB, 640x656)
580 KB
580 KB MP4
What were they thinking
>>
Is there a way to get lightx2v to work in the stock comfy wan workflow?
>>
>>106440581
its literally part of the default template, what the fuck?
>>
>>106440581
I am genuinely confused what is the issue here? Maybe a screencap of your workflow?
>>
File: WanVideo2_2_I2V_00004.mp4 (1.23 MB, 480x832)
1.23 MB
1.23 MB MP4
>>106440551
this is just using the default "girl walks to camera" prompt in the rentry. im not seeing any errors, and the only thing i changed was disabling "merge loras" as it was telling me
>Set LoRA node does not use low_mem_load and can't merge LoRAs, disable 'merge_loras' in the LoRA select node.
>>
is there something like regional upscaling?
when I upscale a regional prompted image it usually mixes both regions or even changes some objects entirely.
>>
>>106440581
You click the Lora that is already there and ctrl B
>>
>>106440592
Post image of workflow
>>
>>106440592
It looks to me like too many steps on the high noise model.
>>
>>106440583
It's not there for me. I might be using some busted old one.
>>
Thoughts on sneedvr2?
>>
File: wan workflow.png (1006 KB, 3698x1860)
1006 KB
1006 KB PNG
>>106440602
it's the workflow provided here: https://rentry.org/wan22ldgguide#ldg-wan-22-install-and-optimization-guide
i don't think i changed anything. i'm just getting started with it
>>
>>106440630
Well turn off the quantization on the model for one. It's already quantized.
>>
File: 1726425624668796.mp4 (315 KB, 816x448)
315 KB
315 KB MP4
>>106440630
Just use the template workflow in Comfyui first and see if that works first
>>
>>106440310
2-3 on high noise, 2-3 on low noise. Don't listen to the clown >>106440442 telling you to go no lora on high noise, it'll take more than 4 steps for that to cause no ghosting on the final output. You'll need to hack together a 3 sampler schizo workflow for that to get decent results and kills the point of using a speedup lora in the first place.
>>
File: 1422051675499.jpg (39 KB, 250x346)
39 KB
39 KB JPG
Can you use the S2V Wan model as a plain T2V? I'm tired of the dual model solution
>>
>>106440648
>telling you to go no lora on high noise, it'll take more than 4 steps for that to cause no ghosting on the final output.
stfu. I've posted a bunch of gens in this thread on 4 steps with no LoRA and it works fine.
>>
Illustrious 3.5 when
>>
>>106440658
In fact. To add to this, what I'm saying isn't unknown information either.
>>
>>106440675
only $5mil left to go!
>>
>>
File: cui - Copy.png (309 KB, 2088x1012)
309 KB
309 KB PNG
>>106440591

Here you go.
>>
>>106440714
Oh, yeah the LoRA node is not there. Just ass a Load Lora model only node between the model and the shift nodes. The purple wire.
>>
File: 1748271986204952.png (361 KB, 2088x1012)
361 KB
361 KB PNG
>>106440714
Add Lora loaders in the red circles
High and low noise
>>
>>106440810
Or just low noise at 4 steps on high without the LoRA.
>>
>>106440723
Thanks anon, appreciate it.
>>
File: WanVideo2_2_I2V_00262.webm (1.61 MB, 720x1248)
1.61 MB
1.61 MB WEBM
>>
I just downloaded invoke AI and now I'm trans
>>
>>106440976
continue
>>
Hopefully we'll get a good local music gen model next month or I'll kms
Cloudfag dominance is even greater there, it's not fair
>>
>>106440810
>>106440812

Really appreciate the help. I'm guessing the Lora can stay at 1.00 strength?
>>
>>106441225
I personally do .9 on the loras and 5 steps in high 3 steps in low
Otherwise yeah
>>
>>106438980
awesome
>>
File: 1744663259716931.png (23 KB, 1400x196)
23 KB
23 KB PNG
I'm already struggling to make videos gen in normal speed, and now people add slow motions loras to wan...
>>
>>106440810
Different anon. My videos are all rainbowy static. I'm testing ideas at 480x480 with a Q8 quant. Any idea what could be causing it?
>>
>>106441526
>Any idea what could be causing it?

wrong workflow setup?
just maybe
>>
>>106440592
Looks like she's struggling to contain herself
>>
>>106440078
Lol great
>>
>>106439539
The tofu run has really changed since I was a boy.
>>
File: ouch.mp4 (2.1 MB, 832x640)
2.1 MB
2.1 MB MP4
So a gyaru walks into a bar.
>>
File: cui2 - Copy.png (428 KB, 2115x1010)
428 KB
428 KB PNG
>>106441550
Mine looks like this. Not sure I screwed up.
>>
>>106441435

You seem to have missed Hunyuan where video gens were accelerated to autistic level with no remedy to fix it
>>
>>106441701
Well wan loras have the exact opposite problem.
>>
>>106441693
CFG too high

take this vid as ref

https://www.youtube.com/watch?v=Tqf8OIrImPw
>>
>>106441709
you can always speed up a slomo video without losing quality by setting actual fps higher

for a vid running at 16 fps and being too fast, you will have to interpolate frames in between
>>
>>106441693
Also you have CFG > 1 on the low noise pass. It should be 1 since you have the LoRA active on that pass.
>>
>>106441732
Yes, but then my 5s video is 2.5s...
>>
>>106441652
>that smile
Wife material
>>
>>106441614
>what is this spiritual pressure?!?
>>
>>106441652
You're the same Anon that did the Chie videos right? Damn clean.
What's your sampler settings? Any other LoRAs?
>>
>>106441783
use the last frame as input and proceed from there
>>
>>106441716
nta but 1.0?
I use 4.5 for high and 1.0 for lora.
>>
File: ComfyUI_WAN_003.mp4 (1.16 MB, 384x672)
1.16 MB
1.16 MB MP4
Finally figured out a wf to go from 17m, to 2.2m hehehe
>>
File: elchocrano.mp4 (1.92 MB, 832x640)
1.92 MB
1.92 MB MP4
>>106441876
>17m to 2.2m
Good for you anon.
>>
Chroma training support has been added to OneTrainer, Qwen is next to be added.
>>
>>106441876
What tricks?
>>
>>106441886
thank you kind sir, some unhelpful fucks here were mocking instead of just giving a few tips. was really disappointed.
>>
>>106441870

what ever is true, that anon's symptoms (fried colors) are consistent with high CFG
>>
>>106441891
nta

a special attention lora
>>
File: ComfyUI_WAN_0011.mp4 (813 KB, 384x672)
813 KB
813 KB MP4
>>106441876
>>106441891
2.1 lightx and got triton workin right
>>
>>106441950
And now throw in Rife Tensorrt for *free* interpolation.
>>
>>106441886
what the fuck is wrong with you?
>>
>>106441950
>got triton workin

on 30xx or 40xx/50xx ??

I could not see any speed-up for my 3090 with triton (15% do not count)
>>
File: Pyio 🐥 Pyio.mp4 (3.43 MB, 420x640)
3.43 MB
3.43 MB MP4
>>
File: ComfyUI_WAN_0015.mp4 (730 KB, 384x672)
730 KB
730 KB MP4
>>106441972
haven't heard of that. Ill check it. Still figuring out video gen,
>>106441983
5080
>>
>>106441734
>>106441716
>>106441870

I see. Should I just copy what's in the video for the most part? Does the resolution have any impact on it?
>>
File: ComfyUI_WAN_0019.mp4 (847 KB, 384x672)
847 KB
847 KB MP4
>>106441997
>>106441996
cute!
>>
>>106441996
Kawaii

Will Rumiko Takahashi ever go ut of style ?
>>
>>106441996
>>106441997
>>106442044
>>106441950
These images belong to Anime Diffusion General and not here.
>>
>>106442080
fuck off faggot
>>
>>106442080
go cry about it loser
>>
>>106441526
anon, the ai is just generating your spawn of satan being vaporized by God, as it should be
>>
>>106442080
Go be gay elsewhere
>>
>>106442003
>Should I just copy what's in the video for the most part?

Just as every sane person would do

this guy's parteon is free
https://www.patreon.com/theartofficialtrainer

the post to that video
https://www.patreon.com/posts/wan2-2-is-out-135140419
>>
File: ComfyUI_WAN_0026.mp4 (1.07 MB, 384x672)
1.07 MB
1.07 MB MP4
>>106442044
>>
>>106442044
why did it change how her eyes look
>>
>>106442194
I know :/ I tried a few gens, this was the closest to the original
>>
InvokeAI status?
>>
>>106442249
Revoked
>>
Why is ComfyUI no longer giving me preview images? I have the thing set in manager.
>>
>>106442257
try reinstalling videohelpersuite
>>
>>106442273
I'm not even talking about video, regular image previews aren't even showing. I am trying to test out Chroma, and I know it worked before on my same workflow for v48 a month ago
>>
>>106442257
restarting my browser fixed it for me last time, if this isn't from pulling at least.
>>
>>106440592
that is insane spiritual pressure
>>
>>106442158
Got it working, thanks you guys.
>>
>>106442338
Remember to share your gens.
>>
>>106442338

glad to hear
>>
What do you think of Ani successfully convincing the /adt/ baker to include AniStudio in the OP?

>>106439892
>>
>>106442365
I think you should stop dramafagging and contemplate your choices in life.
>>
File: Kyou-Strip_WM.webm (3.94 MB, 600x840)
3.94 MB
3.94 MB WEBM
Kyou no!
>>
>>106442365
it means comfui's days are numbered so it kinda makes me happy
>>
>>106442403
it's great minus the name floating shit, and man does it look like a pain to do with wan
>>
>>106442403
Is there a workflow for this?
>>
>>106442403
>watermarked sloppa
Everyday we reach new lows
>>
>>106442413
no wait this concept is actually genius
imagine gooning and you finish the moment it hits the corner
>>
>>106442417
probably something like :
wan wf 0s-5s -> extract last latent -> wan wf2 5s-10s -> etc
what surprises me is how seamless it looks, I have no idea how they did that, my results are just random because the model doesn't know the motion direction with just an image
>>
>>106442417
Yes, the one in the rentry guide.

>>106442413
it has been brought to my attention that someone of indian origin has been trying to profit from the gens i have posted on 4chan, hence the watermark.
>>
>>106442450
https://www.youtube.com/watch?v=5mGuCdlCcNM
>>
>>106442460
lord have mercy.
>>
>>106442450
you can only cum when it hits the corner
>>
>>106442455
Two things.
1. I do some adjustments in Adobe Premier Pro to help make the transitions more seamless.
2. I generate absolutely massive loads of videos per 5-second interval. I rename the good candidates with numbers denoting how good they are. Then I analyse each good one and combine the best candidates into a longer video.
>>
>>106442489
>1. I do some adjustments in Adobe Premier Pro to help make the transitions more seamless.
what do you do anon? get rid of the blurriness?

>2. I generate absolutely massive loads of videos per 5-second interval. I rename the good candidates with numbers denoting how good they are. Then I analyse each good one and combine the best candidates into a longer video.
now it makes sense, you basically bruteforced the issue by generating so many variations and then checking each one lol
I can respect that, and maybe at some point I'll have the courage to do that when I'm tired of 5s loops
>>
I installed SwarmUI, how do I cool zone focused refining I seen people do? I feel like I am doing something in-complete with my generations.
>>
>>106442365
trAnies all really lead the most miserable life in existance despite living in the modern age, grim
>>
fresh
>>106442596
>>106442596
>>106442596
>>106442596
>>
>>106438980
dayum thats a nice aesthetic
>>
>>106442600
rebake

>>106442960
>>106442960
>>106442960
>>
>>106441526
steps too high and maybe your tv2 fastxl lora (or whatever its called) (4 step) is the old version and umm your second ksampler is only doing 0 steps.
This is all presumptive on you cannabalising the default workflow to improve it.



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.