[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/g/ - Technology


Thread archived.
You cannot reply anymore.


[Advertise on 4chan]


Discussion of Free and Open Source Text-to-Image/Video Models and UI

Prev: >>106669789

https://rentry.org/ldg-lazy-getting-started-guide

>UI
ComfyUI: https://github.com/comfyanonymous/ComfyUI
SwarmUI: https://github.com/mcmonkeyprojects/SwarmUI
re/Forge/Classic/Neo: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassicneo
SD.Next: https://github.com/vladmandic/sdnext
Wan2GP: https://github.com/deepbeepmeep/Wan2GP

>Checkpoints, LoRAs, Upscalers, & Workflows
https://civitai.com
https://civitaiarchive.com/
https://openmodeldb.info
https://openart.ai/workflows

>Tuning
https://github.com/spacepxl/demystifying-sd-finetuning
https://github.com/Nerogar/OneTrainer
https://github.com/kohya-ss/sd-scripts/tree/sd3
https://github.com/derrian-distro/LoRA_Easy_Training_Scripts
https://github.com/tdrussell/diffusion-pipe

>WanX
https://comfyanonymous.github.io/ComfyUI_examples/wan22/
https://github.com/Wan-Video

>Chroma
https://huggingface.co/lodestones/Chroma1-Base
Training: https://rentry.org/mvu52t46

>Neta Lumina
https://huggingface.co/neta-art/Neta-Lumina
https://civitai.com/models/1790792?modelVersionId=2122326
https://neta-lumina-style.tz03.xyz/

>Illustrious
1girl and Beyond: https://rentry.org/comfyui_guide_1girl
Tag Explorer: https://tagexplorer.github.io/

>Misc
Local Model Meta: https://rentry.org/localmodelsmeta
Share Metadata: https://catbox.moe | https://litterbox.catbox.moe/
GPU Benchmarks: https://chimolog.co/bto-gpu-stable-diffusion-specs/
Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-one
Txt2Img Plugin: https://github.com/Acly/krita-ai-diffusion
Archive: https://rentry.org/sdg-link
Bakery: https://rentry.org/ldgcollage

>Neighbours
>>>/aco/csdg
>>>/b/degen
>>>/b/realistic+parody
>>>/gif/vdg
>>>/d/ddg
>>>/e/edg
>>>/h/hdg
>>>/trash/slop
>>>/vt/vtai
>>>/u/udg

>Local Text
>>>/g/lmg

>Maintain Thread Quality
https://rentry.org/debo
>>
>>
File: file.png (389 KB, 1228x622)
389 KB
389 KB PNG
>>106672246
neat
>>
>>106672283
>cfg 1
QIEP is distilled?
>>
File: file.png (1.03 MB, 1805x718)
1.03 MB
1.03 MB PNG
>>106672283
cool, even. despite my stupid typo

>>106672319
i'm using the 4step lora, anon.
>>
>>106672325
>i'm using the 4step lora, anon.
and it's working on the new version? nice
>>
>>106672334
qiep is just a simple finetune, there are no huge architectural changes. all loras still work with it.
>>
>>106672283
Can it make a more general shape depth map that can be used as a template or is it only doing 1:1?
>>
File: comfyui workflow.png (785 KB, 2782x2092)
785 KB
785 KB PNG
>>106671842
Here's a pic of my workflow, it's not pretty and a lot of it is unused but this is how the sausage is made when I gen Chroma.

Also there's one error in there because iirc I use beta sampler with uni_pc_bh2, not ddim uniform.

As I said in the last thread, in case you didn't see my brief reply there, I have 12gb of vram so if you're at 8 you might need to make further compromises on the model and t5 encoder
>>
File: imgaug.png (255 KB, 849x763)
255 KB
255 KB PNG
Any of these benefical or should I keep it default?
>>
>>106672493
also I'm a retard for calling them "groups of samplers" when I meant to say "groups of nodes corresponding to two different choices of sampler"... or something like that...
>>
>>106672494
Random flip should almost always be on.
No idea what crop jitter is.
Color stuff shouldn't be enabled for character and most style loras. Only in concept loras, maybe.
>>
>>106672246
>>106672283
Sorry, I should have posted the prompt I got from another anon.

"Convert the character in the provided image into a fully 3D Nendoroid-style figure. The model should have a large head, small body, and chibi-style proportions while keeping recognizable features and outfit details. Include detailed textures and subtle shading to match the classic Nendoroid look. Pose the character dynamically on a small display base, with soft, even lighting and a neutral background. Render in high-quality 3D, focusing on making it look like a real Nendoroid collectible, not a flat illustration."
>>
File: ComfyUI_06470_.png (2.1 MB, 2560x2560)
2.1 MB
2.1 MB PNG
>>106672581
Sorry again, I'm drunk.
I used Comfy UI for the image and then Gemini/ChatGPT for the nendroid gen.
>>
File: 2627137246.png (2.32 MB, 1824x1248)
2.32 MB
2.32 MB PNG
>>
Is there anything like ic light but for the newer image models? Surely there must be something out there where I can take an existing image and modify the light to a high degree
>>
>>106672590
Fantastic composition
>>
>>106672512
>>106672493
tf is wrong with me today. Beta scheduler*. Fuck me.

I'm sure I wrote something else completely wrong too but I'm not posting any more corrections. Use my workflow at your own risk lol
>>
>>106672608
The edit models can do that
>>
>>106672648
Does it have a spline editor or those directional overlays where I can change the direction of the lighting? The ones I seen so far (kontext in this case) just changes the saturation, brightness and hue
>>
>>106672677
I know you can specify the direction of like (lit from above, lit from upper left side etc.) but not sure if you can take it a step further with adding light source. Maybe worth testing
>>
File: 59026242.mp4 (3.85 MB, 1152x800)
3.85 MB
3.85 MB MP4
>>106672616
I thought so too.
>>
How much loss is too much loss for lora training?
>>
>>106672754
Loss is a meme value for diffusion models.
Noise injection fucks with it.
Ignore it.
>>
File: Untitled.png (866 KB, 1043x767)
866 KB
866 KB PNG
>>
>>106672795
what hath god wrought
>>
>>106672795
>on the left
what?
>>
>>106672809
The position of the images in the workflow does not correspond to the actual position of the images as the model sees them.
>>
File: ComfyUI_01108_.jpg (314 KB, 768x1344)
314 KB
314 KB JPG
>>
File: Wanimate_00082.mp4 (2.39 MB, 750x1024)
2.39 MB
2.39 MB MP4
>>
>>106672879
It's a shame about her ass.
>>
>>106672879
Does this only work with pose CN? can it use depth or canny?
>>
File: ComfyUI_01221_.png (1.09 MB, 832x1248)
1.09 MB
1.09 MB PNG
>>
>>106672879
I enjoy how her tits are the best part both of the gen and of her body
>>
>>106672887
Lookin' good boss
>>
>>106672887
doesn't look like him at all and it's ultra slopped, I think I'll pass on that one, I'll wait for the next month and see if they improved the model further
>>
>>106672893
Well I did use a slopped image from his a reference. It's likely double slopped.
>>
>>106672893
>doesn't look like him at all
he can afford much more cosmetics than a hair transplant with his pateron bux
>>
>>106672899
kek
>>
>>106672899
Do turks go to turkey for hair transplants or is there an even shittier equivalent of turkey that turks use to get hair transplants?
>>
https://huggingface.co/QuantStack/Qwen-Image-Edit-2509-GGUF/blob/main/Qwen-Image-Edit-2509-Q8_0.gguf
FINALLY, can we run it on ComfyUi though?
>>
>>106672915
Does it use the same clip as the previous edit model?
>>
>>106672918
yes
>>
>>
just how bad will Qwen-Image-Edit be at Q2_K?
>>
>>106672915
why not? qie+ was already merged in, it's not in stable yet thoughbeit
>>
File: 1732145457058015.png (145 KB, 937x970)
145 KB
145 KB PNG
BROS???? NUNCHAKU GODS!?!?!
>>
>>106672966
The wan never meme is true.
>>
File: ComfyUI_01225_.png (1.12 MB, 872x1200)
1.12 MB
1.12 MB PNG
>concept bleed
>Ultraslopped
>Barely looks like reference.

Qwen edit bros?
>>
File: AnimateDiff_00001.mp4 (1.16 MB, 512x480)
1.16 MB
1.16 MB MP4
I'm starting to wonder if the fact that the image source being AI made is fucking with the FFLF i2v getting so much colorshift.
But the shift also only happens when it is FFLF, if just First frame there's no colorshift.
With or without the Color Match node, the shift still happens.

For these loops I make with old reactionimages, the shift doesn't happen.

And If I remove the light loras all together, the results come out broken at 20steps, cfg4
>>
>>106672979
Did you set the noise argument too high? This isn't really an issue with kij's nodes.
>>
>>106672879
cool, catbox?
>>
>>106672977
is it worse than the original QIE?
>>
File: test.webm (927 KB, 1024x1024)
927 KB
927 KB WEBM
>>106672979
for example. Seems to work fine?
>>
File: wan.jpg (39 KB, 658x657)
39 KB
39 KB JPG
>>106672966
>distracted yet again by another new release for the 7th time

kek, holy shit, I give up
>>
>>106673010
There's no point when new Wan is around the corner anyway
>>
>>106672971
my understanding is that the nunchakuing process is hard only once per architecture, so they should be able to deliver qie2 in a speedy manner (it's still qwen_image arch).
Hopefully for wan the first hurdle will be their 1st implementation of the arch on their quants, then we'll get speedy releases too.
>>
File: unnamed (3).png (817 KB, 1024x585)
817 KB
817 KB PNG
>>106673003
Worse? No and all things considered, Nano banana seems to do just as bad. I think it's just a hard prompt.
>>
File: ComfyUI_01223_.png (933 KB, 872x1200)
933 KB
933 KB PNG
Using the depth map as the reference. It's super faithful to the depth map.
>>
>>106672986
>>106673006
I'm afk atm, are the workflows in the custom nodes folder? If not, mind sharing?

I used to be able to have no color shift, but then it stopped working.
Found a new workflow that was twice as fast and no colorshift, but then it stopped working again.
Goonrot, perhaps.
>>
>>106672950
Likely troll question but almost every single AI model of any kind is dogshit at Q2.
Very limited, if any, use cases.
>>
File: ComfyUI_01229_.png (1.27 MB, 1024x1024)
1.27 MB
1.27 MB PNG
>>
File: WanimateCollage_00011.mp4 (2.3 MB, 1130x896)
2.3 MB
2.3 MB MP4
Seek the elden ring, become the elden lord...

Probably needs to increase mask grow size, otherwise voluminous headed characters gets squished inside the bounding box.
>>
>>106673056
copequant bros...
>>
I have seen the word troll used three times in these threads today. We have a new user amogus.
>>
File: USO.jpg (67 KB, 1169x823)
67 KB
67 KB JPG
>>106673014
>>106673016
In regards to radial attention as well, here's what jumped the queue

>fusionX lora
>ligthx2v
>flux kontext
>flex krea
>qwen (since its release)
>pic rel apparently

Until people demand it, wanchaku is dead despite image models being easier to implement
>>
>>106673118
there's also qwen loras support, there's actually a PR open right now. I can't build the wheel so I can't try, but qwen bros are eating very good.
>>
File: ComfyUI_01232_.png (863 KB, 1024x1024)
863 KB
863 KB PNG
Interestingly for single images it does a really good job at preserving the identity. I think 3 images is just too much.
>>
File: ComfyUI_01237_.png (651 KB, 784x1328)
651 KB
651 KB PNG
>>106673130
Single image with pose controlnet preserves the output pretty well desu.
>>
File: WanimateCollage_00013.mp4 (2.41 MB, 1130x896)
2.41 MB
2.41 MB MP4
>>106673085
>>
>>106673202
Why was https://rentry.org/wan22ldgguide removed from the OP? Had to find it in the 'chives
Fucked my comfyui install and it's basically the only retard proof way of getting triton+sage installed
>>
>>106673216
Meant for OP >>106672240
>>
File: 1743110616128645.mp4 (855 KB, 720x720)
855 KB
855 KB MP4
>>106673130
>>
File: cope.png (90 KB, 320x320)
90 KB
90 KB PNG
>>106673126
Yeah, suppose image bros gotta eat too. Just hope this new wan will allow for 30 second gens
>>
>>
>>106673238
I bet that really hurts desu.
>>
File: ComfyUI_temp_irbuh_00004_.png (3.51 MB, 1440x1440)
3.51 MB
3.51 MB PNG
>>
Ok... everywhere I go, everyone is raving about how Illustrious is sooo much better than Pony. I have spent an ungodly amount of time trying to make it work for me, and every time I end up going back to my usual setup (Incursio's Meme Diffusion 2.7 + Incase Style + Vixon's Pony Styles - Gothic Neon + Vixon's Pony Styles - Emote v1.0 + Deep Negative Pony + Pony PDXL Negative Embeddings Photoreal)
).

Illustrious models invariably give me these washed out, pastel images, and all the random (not named) people they generate are so incredibly boring.

I haven't been able to find anything that matches the skin texture, the sense of a real musculo-skeletal framework underneath, and just generally vibrant, interesting and pretty random characters my combo generates.

And yet I've been using it for sooo long. Surely there must be something better somewhere after all these years. What am I doing wrong anon? Help me graduate my noobdom.
>>
it helps if I post in the active thread lol

I am in Japan now. most anistudio work while I'm here will just be cmake and splitting things off into shared libs. sorry I haven't been active on the repo recently but I'll be back at it. wish me luck with softbank fundraising!
>>
>>106673307
Be careful with getting money from softbank, they have the touch of death.
>>
another fizzlejeet failure...
>>
>>106673330
if anything they spend tall on the stupidest shit like openai and we work then get burned
>>
Spent last 8 hours gooning and ejaculated 4 times to AI slop.
Good night anons, I am going to bed now.
I will probably feel like shit tomorrow and my dick will hurt, but we will see.
>>
>>106673307
I've been working on the repo in the meanwhile.
>>
File: 1732389909833471.png (28 KB, 705x246)
28 KB
28 KB PNG
didnt realize new thread. make sure you update to this node or you get random noise.

qwen-image-edit-remove_clothes.safetensors still works if you want to do that.

https://files.catbox.moe/y5y946.png
>>
File: 1739803982557823.png (1.04 MB, 880x1176)
1.04 MB
1.04 MB PNG
>>106673460
and now the safe version:

the japanese woman is holding a sign saying "LDG". keep her expression the same.

kept her face the same, old edit would sometimes change the faces so this is better.
>>
https://x.com/Ali_TongyiLab/status/1970401571470029070
Some goober on reddit said 10 seconds for 1080 gens
>>
File: 1731535105259137.png (1.08 MB, 832x1248)
1.08 MB
1.08 MB PNG
>>106673464
the japanese woman is holding a sign saying "LDG" with her right hand. keep her expression the same.
>>
>>106673473
mind you her right arm was behind her back in the original photo so it did a good job, it's definitely a step up over v1.
>>
>>106673473
Can you imagine if you had like a time machine to a year and half or so ago and showed people this? They'd flip their shit.
We are truly spoiled.
>>
https://x.com/RuihangChu/status/1970390357738430839
Now the fucking amerimutt retard that always mention that the source for wan 2.5 was "JUST TWITTER BRO", gonna kill himself right?
Or at least should give more of his tip to Israel.
>>
>>106673488
best edit model, best video model, noob/illustrious for anime, open source is eating well
>>
File: 1750348421043528.png (986 KB, 832x1248)
986 KB
986 KB PNG
added another image the two women are waving hello. keep their expressions the same. keep their expression the same.

ZERO image stitching just connect a load image node to image2 of the new text node.
>>
File: 1738857124943139.png (1.11 MB, 1024x1024)
1.11 MB
1.11 MB PNG
the two anime characters are sitting at a desk in a classroom in japan. keep their expression the same. the character with white hair has a black blindfold on.
>>
File: media_G1hJq6oXMAATk_W.jpg (516 KB, 643x4096)
516 KB
516 KB JPG
>>106673490
Exciting
>>
File: 1754897951478361.png (1.05 MB, 1024x1024)
1.05 MB
1.05 MB PNG
did a neat upscale of OG miku too:
>>
the problem with the new QIE is that it's actually much too obstinate about keeping things the same. it will exactly clone parts of the image instead of conforming them to the new setting.
>>
File: 1737899009670939.png (827 KB, 1176x880)
827 KB
827 KB PNG
seems to be a good upscaler, this was from a pixelized image:
>>
man my comfy install is ancient... I wanted to upgrade to whats it called... pytorch 2.8 and the new triton.
Is it worth the hassle?
>>
>>106673460
>qwen-image-edit-remove_clothes.safetensors
uhmm link? all the fucking loras have been nuked on hf and civitai
>>
>>106673565
At least make a backup.
>>
File: 1757774303568533.png (797 KB, 1176x880)
797 KB
797 KB PNG
>>106673556
the blue hair anime girl and the red hair anime girl are having a hug.

definitely better than before imo.

>>106673570
need time to upload this sites upload speed is kinda shit, 1s.
>>
File: 1734182230898742.png (931 KB, 1176x880)
931 KB
931 KB PNG
>>106673585
the blue hair anime girl and the red hair anime girl are dancing and smiling.
>>
>>106673465
>10 seconds, 1080 gens
let's hope the gen time is manageable
>>
>>106673543
Out of curiosity, are you using the lightning LoRA? I found it ruins a lot of its ability to change up styles. CFG also needs to be like 4.
>>
>>106673465
The post is gone :O
>>
>>106673588
rei is a doll, she cant smile
>>
File: 1739096880715584.png (908 KB, 1176x880)
908 KB
908 KB PNG
it thought rei was tanned cause of the shading kek, need to specify light skin
>>
>>106673608
>long hair
I hate this with QIE, the fucking bleeds, I gotta gacha gens or change wording slightly
>>
Try QIE with controlnets. They work shockingly well. Like they follow the controlnets perfectly without frying the image.
>>
>>106673307
*yawn*
>>
>>106673570
https://filebin.net/ex6rjtj8rot8vi7u/qwen-image-edit-remove_clothes.safetensors
>>
>>106673623
thanks bro
>>
File: wan2.1_t2v_training.png (1.37 MB, 1887x1877)
1.37 MB
1.37 MB PNG
Can any one whos trained a wan 2.1 t2v lora confirm this guide is accurate? I fed AI information of what apparently worked for others who have 16gb vram and 32gb ram on a small video data set. The aim is to train a t2v model and hope it works on i2v in the end (this also apparently works). Never used musubi but its already installed and ready.

>durrry why 2.1

Because

>>106673536

Sweet
>>
So now they deleted the announcement are we getting wan 2.5 tomorrow or not?
>>
>>106673677
wait nvm, it's back

https://x.com/Alibaba_Wan/status/1970419930811265129
>>
>>106673677
>>106673685
https://xcancel.com/Alibaba_Wan/status/1970419930811265129

>please escape 5 second hell
>please escape 5 second hell
>please escape 5 second hell
>>
File: 1730591887315245.png (1.09 MB, 1176x880)
1.09 MB
1.09 MB PNG
>>106673608
the blue hair anime girl and the red hair anime girl are eating ice cream at the beach in Japan.
>>
>>106673707
the problem is they look like temu versions of who they're supposed to be.
>>
>>106673584
had to reinstall nunchaku, now in the process of building the flash attention wheel (luckily I'm also a dev and have the cuda sdk + all the build tools already installed)
>>
>>106673719
you need to specify to keep the expression the same, I didn't.
>>106673473
for example is 1:1
>>
>>106673729
Try it without the lora.
>>
File: 1755115048725358.png (1.39 MB, 920x1128)
1.39 MB
1.39 MB PNG
the blue hair anime girl is relaxing in a hot spring in Japan. keep her expression the same.

that's pretty good
>>
File: 1735398932607116.jpg (75 KB, 783x960)
75 KB
75 KB JPG
>>106673761
original:
>>
>>106673765
also what's neat is the model got the bottom half of the plugsuit right despite me not specifying rei, or evangelion.
>>
>>106673765
>>106673761
ok but can you make her cum?
>>
>>106673792
not on a blue board, but you can do it anon
>>
Wtf is chroma-2k-qc?
>>
Wan 2.5?
What about 2.3, 2.4?
>>
We almost went a whole thread without mentioning Chroma.
>>
File: 1730701477509295.png (1.36 MB, 920x1128)
1.36 MB
1.36 MB PNG
>>106673761
that was the fp8 one, this time I tried Q8 (was downloading)

in general isnt fp8 close to q8? all I know is q8 is closer to fp16 in general.
>>
File: 1749207024970989.png (1.39 MB, 920x1128)
1.39 MB
1.39 MB PNG
>>106673839
>>
File: oh nooo.jpg (88 KB, 681x554)
88 KB
88 KB JPG
Wansisters, I dont feel so good...

https://xcancel.com/T8star_Aix/status/1970419314726707391
>>
>>106673860
I can't hate them because Wan 2.2 is already a great model but that is very disappointing
>>
>>106673860
I'm done.
>>
File: 1727916248994220.png (1013 KB, 1120x928)
1013 KB
1013 KB PNG
the blue hair anime girl is at the beach in Japan and holding a sign saying "LDG" in stylish black text. Keep her expression the same.

diff image, didnt change the face at all, success
>>
>>106673860
Who is this guy though. It just says "comfyui tutorials" on his profile. Is he someone of note?
>>
>>106673860
Why is comfyui listed as a partner here >>106673536
They've got Comfyui people on the English panel too
>>
>>106673860
The fingolian is already going damage control mode in the discord, considering he's on comfy's payroll, had early access to some recent wan models and that he's literally listed in their hugginface dev group. I'm gonna say that sadly it does look like api only at least for the near future.
>>
>>106673921
Beggars can't be choosers
>>
File: 1745575350032412.png (841 KB, 848x1232)
841 KB
841 KB PNG
the anime girl is holding two silver pistols and is pointing them in opposite directions. keep her black blindfold the same.

yep it's qwen edit v2 time. this is with the q8 version.
>>
Explains the complete lack of hype I suppose, just a couple vague tweets and discord messages
>>
File: 1730063971147510.jpg (59 KB, 736x1065)
59 KB
59 KB JPG
>>106673966
source image:
>>
File: 1732037485230571.png (815 KB, 848x1232)
815 KB
815 KB PNG
>>106673972
>>
>>106673860
its over
>>
>>106673860
No matter how much you pay me I am not using the API
>>
File: 1731401858182501.png (765 KB, 1120x928)
765 KB
765 KB PNG
change the location to a computer lab. the character on the left is typing on a laptop, keep his face and expression the same. the green cartoon frog is unchanged, with the same pose. A sign saying "all you can prompt" is on the wall in scribbled black text.

pepe edit success
>>
>>106673891
Does a lot of comfyui workflows on civit I think and also see >>106673921

Calling it now

>free version: 5 second lock
>api version: no lock, faster
>>
>>106674025
api first is DOA, 2.2 is fine anyway. why pay for 5 seconds more when you can stitch 2 clips.
>>
>>106673860
I WILL NOT use API nodes!
I WILL NOT have my "barely legal teenage girl sucking dick" prompt sent directly to a company just so they can forward a tip to authorities!
I WILL NEVER use it!
>>
File: 1754627604077850.png (847 KB, 1120x928)
847 KB
847 KB PNG
>>106674023
Add Miku Hatsune behind the green cartoon frog, who is pointing and laughing at the character on the left.
>>
>>106673307
So a nothingburger like always
>>
>>106673860
>>106673921
>>106674025
Reminder that Comfy is also pushing their integrated "SAAS" and servers and they work hand and hand with WAN and Nvidia.
>>
>>106674028
Mostly yeah. True, if you're doing something that requires repetitive motion, the context nodes also work well for 10 - 15 sec gens, depending on hardware of course.
>>
Was it Hunyuan that made their 720p model API and only open sourced the 480p? I really hope Wan don't go down that route
>>
File: 1756866416125901.png (827 KB, 1120x928)
827 KB
827 KB PNG
>>106674040
background swaps also follow the style better in the new version, nice.
>>
File: 1732312381004051.png (891 KB, 1024x1024)
891 KB
891 KB PNG
the character is pointing at the camera and smiling.
>>
File: Wanimate_00095.mp4 (2.84 MB, 744x1024)
2.84 MB
2.84 MB MP4
>>
>>106674076
They also went API only for their 3D models. We're fucked.
This will just be just the first in a long line of API only shit models. Qwen Image Edit was a bone they threw to us because they knew we would be mad.
>>
File: 1754213402202029.png (873 KB, 1024x1024)
873 KB
873 KB PNG
>>106674089
the character is holding a sign that says "LDG" in black text with his right hand, and giving the thumbs up with his left hand.

gj carlos
>>
File: Capture.png (28 KB, 623x246)
28 KB
28 KB PNG
>I'll be live-streaming the launch event for the new open-source video model
Maybe there's just a lot of miscommunication going on. Perhaps there's still hope?
>>
File: 1733209331012221.png (984 KB, 1360x768)
984 KB
984 KB PNG
replace the subtitle text at the bottom with "time to shitpost on /LDG/!"

text edits are much better this version it seems, sometimes you'd get random characters if the string length was very different.
>>
>>106674107
likely decided to go api first for a year. it's over, just let it go
>>
File: 1751432461064515.png (980 KB, 1360x768)
980 KB
980 KB PNG
>>106674112
replace the subtitle text at the bottom with "make Miku's hair blonde!". Change the hair color of the anime girl to blonde.
>>
>>106674114
It would not be the first time a company has said they were going to release a model and then a suit decided to go all Bilbo Baggins on them and refuse to release it.
>>
https://wavespeed.ai/models/alibaba/wan-2.5/text-to-video
>Is literally out right now via api
Can't tell if this means it will be indeed open or not, does google license their models on other platforms as well?
>>
qwen edit
qwedit
heh
>>
>>106674158
Doesn't this website often intentionally mislabel models?
>>
I've been away for a while and now feeling utterly lost, could someone qrd me on the current models of note?
My impression so far is:
>qwen edit for editing and image gen
>chroma controversially for image gen but can do porn out of the box
>wan2.2 for videos
Am I missing anything?
>>
>>106674195
Seed dream diffused via API and sent to your UI locally.
Wan 2.5. Diffused via API and sent to your UI locally (We aren't allowed to critique them because they gave us free shit in the past)
>>
>>106674169
No.

Also we can already heavily inter from this and from rumors that WAN 2.5 was focused on tuning wan to be able to run at 1080p natively. If that affects general model intelligence in a good or bad way we have to see

If this is the main update to wan2 then I don't care (this is my cope feel free to use it)

Of course it could be a mislabelling which can immediately be checked if anyone actually spends a few shekels on a 1080p gen from wavespeed.
>>
>>106674158
Confirmed 1080p and 10secs.
Either they do not give us shit or we'll get a very cutdown version like 720p 5secs.
Grim.
>>
File: 1741607375959638.png (957 KB, 1360x768)
957 KB
957 KB PNG
replace the anime girl with rei ayanami.

qwen image v2 knows new characters! that's pretty cool.
>>
>>106674216
I specifically recall them doing something similar in the past and they just named another API as another TBA open source model.
Pretty sure wavespeed is scummy.
>>
>>106674232
That and it's $1 per run.
That is insanely expensive. My scam detector is going nuts.
>>
>>106674227
>10 seconds
He's right. Check out the examples tab on the wavespeed page

Damnit. There's no way this is getting released. If Ali releases this I will bow to the superiority of chinks because I absolutely would not release this if I had it lol

>>106674244
I don't remember anything about this but

$1 per run is the same they were asking for hunyuan (0.20 per second). If anything it's cheaper now
>>
>>106674158
Some dodgy sites like that and blogs tend to already have a "model" or article already up, probably placeholders for seo or someshit.

>>106674159
qwan 2.5
>>
>>106674249
You're being scammed bro.
>>
File: 1737258400117988.png (984 KB, 1360x768)
984 KB
984 KB PNG
replace the anime girl with a white gundam.
>>
>>106674255
It has examples already and wave speed isn't a dodgy site, I remember using them on replicate to generate free hunyuan videos of little girls in bikinis so much they had to put 3 levels of censorship in at the prompt and output level, and then they added a lora to slop the faces of young people lmao

>>106674260
I've already scammed them harder than they can ever scam me. See above
>>
Oh another thing, there is a a chance replicate will be serving the model completely for free for a few hours once it comes out. They did this for hunyanvideo and wan 2.1s release for sure. Since most people itt can't or won't do 1080p locally if this happens it may be your only chance to get some unwatermarked 1080p genning in wan 2.5 so pay attention to that
>>
>>106674297
If it's api only the can suck my balls and eat my shit. At that point just use fucking veo.
>>
>>106674304
Veo won't let you generate beautiful little girls in bikinis in the beach in 1080p while I promise you if replicate serves it they keep the filters loose for a few hours

Actually I don't promise that, maybe they learned their lesson but the odds are 1 million times higher with wan 2.5. If this isn't something you care about then ew stinky brown person but also carry on
>>
>>106674318
You should kill yourself, piece of garbage jew.
>>
>>106674304
Yeah if its a worse veo not sure why anyone would wanna use it. I wanna believe they know that but who knows, seen corpos make dummer decision.
>>
File: 1739979227867682.png (949 KB, 1360x768)
949 KB
949 KB PNG
The anime girl Miku Hatsune is holding a stack of boxes with the label "LDG gens" on them. keep her expression the same.
>>
There's no way the wavespeed examples aren't wan. They're all yapping so much

>>106674323
>No argument
Big words from a mutt that's circumsized and mutilated for literally no reason lmao. I accept you concession

You now remember that Tel Aviv University made a SOTA video model and never released it btw

>>106674327
I'm 100% certain it will be less censored than veo at least for the first week or so to gain userbase. There are legitimate use cases for this (like beautiful children)
>>
>>106674304
Basically this.
>>
I don't see how 1080p 10 second videos would even be viable on current consumer cards. 720/5secs already pushes 24GB cards to the limit
>>
>>106674327
>Yeah if its a worse veo
going by their examples it's def worse than veo
>>
>>106674391
We can use our local UIs to diffuse via API directly to our PC locally.
>>
Guys I'm feeling pretty black pilled right now.
>>
>>106674448
Nothing is confirmed until the live stream. I WILL keep coping until then
>>
>>106674448
The night is darkest before the dawn

>>106674391
5090 users exist, and resolutions between 720p and 1080p exist. Why are you niggers dooming so hard we have one consistent Chinese guy saying it's open source and then a literal who being like no its not
>>
>>106674465
Haven't seen a single tester or dev saying it is open source.
>>
>>106674476
Chinese guy who goes to conferences has only been alluding to it as an open source release

Also here's the link for the image 2 video. Cheaper to run than the t2v it seems??

https://wavespeed.ai/models/alibaba/wan-2.5/image-to-video
>>
>>106674490
Man stop linking that scam site. I'm not even fucking around here.
>>
>>106674490
yeah check the t2i ones they look quite slopped
audio sounds like shit in all the examples but i guess it's progress for them
>>
>>106674510
>>yeah check the t2i ones they look quite slopped
Funny because I've seen the completely opposite from other's peoples gens, it absolutely shits on flux and derivatives.
>>
>>106674500
Lmao kill yourself schizo if you want to see the beautiful children I make on replicate in a few hours you could have just asked (I was going to share them anyways for science)

>>106674510
After seeing their prompt guide I don't trust them to not prompt for slop anyways. I also don't even use their Chinese negs in my workflows because it also adds to the slop
>>
>>106674520
okay? they're still slop

>>106674524
yeah but if it isn't local it's not going to matter much for me since the alternatives would be better than what they're offering
>>
everything will be closed source from now on, its over for local
>>
>>106674551
After seeing how much was squeezed out of sdxl for years I'm genuinely not worried if wan 2.2 is the last open source video base model for the next two years (it won't be)
>>
>API JEW NODES ADDED https://github.com/comfyanonymous/ComfyUI/pull/9996
>NO LOCAL NODES
KEEEEEEEEK
It's so over
>>
>>106674570
>BFL Presents: Flux Video Mega Slop Edition
>>
>>106674524
>Lmao kill yourself schizo
>Hey guys, this website with no association with Alibaba just happens to have access to Wan 2.5 before it's even present. No, you can't look at the API.
Now give me a buck and maybe I'll spit out a video from some random API at a 200% markup.

Go fuck yourself scammer.
>>
>>106672879
way to ruin a beautiful woman and turn her into an ugly yellow beast
>>
>>106674622
Happens literally all the time in all industries. Happens in this industry very often. You must be over 18 to post on this website.
>>
>>106674653
You know what happens way more often? Intentionally mislabeling APIs for highly anticipated products and charging exorbitant prices.
Like do you work for wavespeed or something?
>>
>>106672879
>roundest asian ass
>>
>>106674666
No it actually doesn't lol
Yeah my dad works at wave speed his last name is Speed
>>
File: AnimateDiff_00001.mp4 (2.87 MB, 480x480)
2.87 MB
2.87 MB MP4
>turns out I've been in kijais workflow all along
>found some nodes that could have fixed the color shift
>it makes it worse
>it's now even showing up in the very simple images

I am about to shit myself in anger.
>>
>>106674695
>No it actually doesn't
Do you have any idea how many fake claudes there are out there right now? I cannot tell if you are genuinely stupid or this is some kind of next level vanryo shit.
>>
>106674718
How many of those fake claudes are official partners with replicate retard? No more actual replies for you.

https://replicate.com/wavespeedai


Anyways, assuming wavespeed isn't fucking around, I noticed in their i2v example prompt they type [Close-up] with square brackets. I know that they tried to do some stuff with prompt structure with 2.2 so I'm wondering if they're messing around again with 2.5
>>
File: Wanimate_00102.mp4 (2.01 MB, 738x1024)
2.01 MB
2.01 MB MP4
>>
>>106674737
>official partners with replicate
Is that supposed to hold some meaning to me? In don't care if they're official partners with my dick. They aren't alibaba and their model hasn't even been presented yet.
>>
why is anon shilling an api model?
>>
>106674762
HAHAHAHAHAHAHA
>>
>>106674763
Because anon likely gets a cut.

1 fucking dollar per gen. That is an insane markup.
>>
>>106674792
even sneedream wasn't that greedy
>>
Why is wanvideo node calling samplers schedulers?
>>
>>106674799
Yes but seedream was actually seedream.
>>
I thought they would wait until 3.0 before going saas
>Veo 3 exists
And I guess Wan's guys are planning to give a less good, but cheaper alternative to VEO 3's 250$ per month
>>
The pricing shows they’re serious about starting a business. This isn’t a hobby. They’ve started cashing out.
>>
>>106674602
fucking grim
>>
Before we all lose our shit. Do we actually have confirmation this is API only or is just extremely obvious from the context?
>>
>>106674832
As they feared, what’s unfolding looks like it can only end one way: picking a fight with the open-source community and having the livestream get raided.
>>
>>106674832
looks like 99% wan 2.5 is going to be api. Judging from that retard's tweet (he's an insider) and also api nodes already being PR'd in comfy.
It's unironically over. We'll have to cope with 2.2. Hopefully nunchaku makes it less painful
>>
>>106674820
>I guess Wan's guys are planning to give a less good, but cheaper alternative to VEO 3's 250$ per month
They are. From the wave speed description:
"Compared to Google Veo3, it offers a more affordable solution with faster video generation speeds, making it an excellent choice for creating audio-embedded videos efficiently."

I'm full doomed right now by the way, thanks guys. But we will only know for sure in about an hour or so

>>106674602
I am now willing to bet money on it being SaaS only as an initial release. Very sad

>>106674832
The comfy PR is the best knowledge we have so far.
https://replicate.com/wan-video/wan-2.5-t2i-preview
Hmm anyone else having this page take a long time instead of 404ing immediately?

>>106674846
>picking a fight with the open-source community and having the livestream get raided
Calling it now, we will see the first assassination during a model release in the next 5 years

One final wave speed link I found
https://wavespeed.ai/models/alibaba/wan-2.5/text-to-image
Also this claims they're getting it served from Alibaba's cloud. Since wave speed is a Chinese company I 100% think this is real
>>
File: were back its over.png (820 KB, 1192x900)
820 KB
820 KB PNG
Bros...
>>
>>106674863
we got memed by alibaba, it's unironically SUPER over, the west has fallen, millions must die
>>
>>106674863
this is the kind of propaganda they push when its over but they don't want you to realize it.
>>
>>106674860
Nevermind I'm schizo on the replicate link

>>106674863
This is literally exactly how the thread's reactions have gone kek
>>
bros what the fuck, sdxl is supported in nunchaku in the last release?
>>
>>106674906
https://huggingface.co/nunchaku-tech/nunchaku-sdxl
lol?
>>
>>106674906
Why would you ever need a speedup for sdxl
>>
>>106674923
yeah like sdxl takes... 3 secs? I'm gonna try it anyway because I live for hoarding
>>
>>106674743
me in the back
>>
>>106674928
just noticed they updated the nunchaku wheel, but not the comfyui wheel sooo only through diffusers for now?
>>
I don't understand why these retards don't publish the actual code for quanting, just for diffusing. Like who the fuck cares about base sdxl/turbo? release the quanting code you fucking chinkoids
>>
>>106674855
>It's unironically over. We'll have to cope with 2.2. Hopefully nunchaku makes it less painful
At least 2.2 isn't that painful of a cope, and bytedance has already shown you can do a lot with finetuning it

There's also that paper of the 2.2 tune that did well on benchmarks that said it will release weights soon

Yes this is the bargaining stage before depression and finally acceptance how could you tell
>>
>api
>abandoned by nunchaku

A new king must take the throne
>>
if wan2.5 remains api only, then it will become like sd3 and become irrelevant. wan2.2 will become the new sdxl of video gen and last many years.
>>
>>106674923
Chroma kekes on swicide line
>>
>>106674982
Yep. No one even knows the limits of 2.2 just yet because not enough experimentation has been done with promoting yet. It's also a top tier text to image model.

>>106674965
>must
Not sure why you think this. Getting wan 2.1 at all was an insane blessing for local
>>
File: 1731852789530640.jpg (49 KB, 1080x1016)
49 KB
49 KB JPG
>>106673860
>wan finally looks like veo 3
>it's a paid model
every time. as a gamer, i'm not surprised at all
>>
>>106674906
based
>>
>>106674906
its base sdxl, who cares
if i cant get a speedup using my favorite finetune then its useless
>>
https://xcancel.com/wavespeed_ai
Sounds like wavespeed will be rebroadcasting the conference on their Twitter and YouTube, this was 1 hour ago
>>
Why won't this context window node work with these nodes? It seems to fix a lot of the color shifting.
>>
File: G1h3QpQawAAm0Iy.jpg (955 KB, 643x4096)
955 KB
955 KB JPG
>>
File: 1733680414219958.png (27 KB, 1107x210)
27 KB
27 KB PNG
>>106673570
>all the fucking loras have been nuked on hf and civitai
wait what? I stepped away for a few months and everything is gone?

I am also in need of help.
Been using A1111 for SD and Forge for Flux.
I took a break and now A1111 just refuses to start now so I guess its time to move on.

What UI is the current recommended for SD (preferably with working controlnet for Pony/Illust models and something like regional prompter/latent couple built in or as a plugin)?
And what should I use for FLUX and FLUX derivatives?

Yeah I know comfy is the way but I really enjoy slapping prompts, generating 32 images and saving the ones I like right in the interface. Would be grateful for suggestions!
>>
>>106673860
>He trust'eded the chinks and thoughteted'ed that china would really save local
LMAO, I told you, the moment they got a good model they kept it for themselves, we'll never get a good model locally in our lifetime
https://www.youtube.com/watch?v=H47ow4_Cmk0
The day Qwen Image Edit will be as good as Nano Banana they'll keep it for themselves as well, be prepare for that btw
>>
>>106675096
Most have moved to comfyui, otherwise if you like forge you can try neoforge and/or one of its 99932553211 forks.
I'd personally stick with comfy for the time being, has a bit of a learning curve but its worth it imho
>>
>>106675096
Neo Forge
>>
>>106673966
>>106673972
are these done in comfy? did something local like nanobanana but not cringe drop?
>>
>>106675096
NeoForge branch o Forge classic, easy stable search in github Haoming02
>>
>>106675116
Wan 2.2 has made me orgasm over 50 times, that's pretty good for a price of 0 dollars and Apache 2 license, and we're still using the equivalent of base sdxl kind of

The new cope will be that fine-tune that impressed Gemini and benchmemes, and hoping they release weights for it soon
>>
>>106675143
I don't want to end up with a 2 models (high low) process, that's lame
>>
File: wav.jpg (15 KB, 715x186)
15 KB
15 KB JPG
kek
>>
when did illustrious go closed source, what the fuck? what do I use now instead?
>>
>>106673860
Holy shit it has sound like VEO3, no wonder they want to keep it for themselves
https://files.catbox.moe/orknbn.mp4
https://wavespeed.ai/models/alibaba/wan-2.5/text-to-video
>>
>>106675150
>that's lame
Explain why without getting upset or admitting to a skill issue. I agree that it's less convenient than a one sampler workflow but once you set it up it's not any different
>>
>>106675169
>Explain why
>it's less convenient than a one sampler workflow
you just did, loading/reloading the models take time, that's bullshit
>>
>>106675152
>>106675168
>$1.5 for a one 1080p 10 sec clip?
how big is this shit? I'm sure wan 2.5 is a 50+b model, even if it was local no one could've run that lol
>>
>>106675152
>>106675168
Shame on me for feeding the schizo when wan's own tweet had wave speed as an official partner listed like 4 times lol

I also had no idea tensor art was Chinese or that heavily associated with the industry since I never used it and thought they were a wild west for loras of real child models for the longest time

>>106675179
>loading/reloading the models take time, that's bullshit
Pretty sure this is either a DRAMlet cope or flat out not true or both. My workflows aren't much slower than they used to be for 2.1 idk
>>
>>106675163
Noob or Illust
Yeah, nothing changed for weebstuff since last year
>>
>>106675188
we could copequant it, as usual
>>
>>106675195
I see a new illust wai release, I guess I'll use that as a base.
>>
>>106675188
First of all you're not sure of that at all, second of all there's anons with 96GB cards in this general, third of all you need to be 18 to post on this website.
>>
>>106675200
>second of all there's anons with 96GB cards in this general
yeah sure, try to guess why step video (30b) got ignored everywhere? because it's too big to be run by people except 2 millionaires that have a A100, you need to be 18 to post on this website
>>
>>106675188
Man we get gargantuan LLMs for open source that people eventually end up running but constantly complain about this shit on image/video gen
>>
>>106675128
>>106675139
>>106675120
thanks! thats why I asked. theres a billion forks now, even in the op theres reForge, Forge Classic, SD.Next and I have no idea whats the difference.
I'll try the Forge Neo then. Are regional prompter and pony/illustrious controlnets available there? I will use it while I learn comfy probably.

Is there a way in Comfy to batch generate and select what to save like I do in WebUI? Something like an infinite image browser?
>>
>>106675205
Step video was worse than wan in every way and people did test it out. You need to be under 18 to turn me on
>>
>>106675210
you can batch gen in comfy, yes.
As for controlnets they're native but you'll get better mileage by using the controlnet aux plugin. Tbh I'm so lazy I just shove in the AUX AIO Preprocessor, and let it do its magic on its own..
You can have a preview output in comfy without saving, then you'll just check your queue and rightclick->save what you like.
Usually I just save everything, for me space is not a problem
>>
>>106675214
>Step video was worse than wan in every way and people did test it out.
Step video (19 feb 2025) was released before Wan 2.1 (25 feb 2025) you mongoloid, at that time of release it was the best model we had and no one cared because it was too big
>>
>>106675227
thanks, I meant controlnets in Forge Neo. I never managed to make them work on pony, no matter which controlnet models I tried.
For me the space isnt a problem either, cant thank my past self enough for investing time to build a personal NAS. The amount of failed slop piling up in the folders is the problem.
I guess its time to learn comfy after all, the quality I've seen out of those is pretty impressive.

BTW whats the mentions in the threads about LoRAs being nuked from Civitai?
>>
>>106675240
Sorry I meant it was worse than hunyuanvideo when it came out. Anything else or more ad hominems?
>>
>>106675240
you think people had enough time to test it in 5 days. It takes a few days just to have the shit implemented but people decided it was the best model? I never even saw one gen out of that model lol.
>>
>>106675259
>Sorry
kek
>>
>>106675168
Those eyes are deader than my future...
>>
>>106673671
Any smart anons plz
>>
>>106675261
>It takes a few days just to have the shit implemented
I went to verify that and surprisignly, Comfy has never implemented it on his official repo, maybe he hadn't a powerful enough GPU to test that out?
the company that made that model had a custom node though
https://github.com/stepfun-ai/ComfyUI-StepVideo
>>
>>106675134
yes comfy. that is qwen edit v2 (new), Q8. Even better for image/text edits/swaps.

https://huggingface.co/QuantStack/Qwen-Image-Edit-2509-GGUF/tree/main
>>
>>106675298
Every anon that actually trains loras has given the same advice which is "just start trying stuff".
>>
File: 1754135937379448.png (868 KB, 1770x843)
868 KB
868 KB PNG
>>106675227
Alright what is this trickery. it basically has everything covered. last time I tried comfy it was a barely working blank field on startup with no additional tools. Am I so much out of touch?
>>
>>106673966
>>106673972
can you stitch the 2 images instead? would be more convenient to see
https://github.com/BigStationW/Compare-pictures-and-videos
>>
>>106675258
not all loras, the one deemed unsafe by the 'AI SAFETY' faggots.
Basically loras about real people are verboten, same as undress loras or anything that can be used in a clear way to make deepfakes.
There's still tons of porn realism loras anyway sooo it's literally so these retards can say 'UGHH WE R FIGHTING THE BAD AI GUYS!!!'
Also of course loras for cunny, be it real or anime are also not allowed. BUT usually anime models do generate cunny without problems soooo lmao
>>
>>106675313
oh wow thanks! guess I'm learning comfy this weekend.
Sucks I only have a 16gb gpu, but oh well. We only have so many kidneys
>>
>>106675317
It's comfy, embrace the spaghetti bro
>>
>>106673307
Who the fuck would give you fundraising?
Holy fuck you actually believe your delusions and will post here full of more tall tales. You have been caught lying about this project for too long just go back to your shit tier thread vs begging us for attention
>>
>>106675323
>Sucks I only have a 16gb gpu, but oh well.
you can offload to the cpu, it's not that much of a big deal, he's the node to add some GB of the model to the ram
https://github.com/pollockjj/ComfyUI-MultiGPU
>>
File: c.png (1.38 MB, 832x1488)
1.38 MB
1.38 MB PNG
>>106674340
cute
>>
>>106675323
you dont need a lot. I am using 16gb as well (4080).

16 is plenty for wan 2.2, and despite the model being 20gb along with other stuff it works perfectly fine in comfy. no multigpu node needed.
>>
>>106675323
>16gb
>Q8 image edit 2509
each iteration/step will take roughly 20-15s~ depending if you're using sage/flash attention or nothing. Lightning loras will make it bearable, but I didn't try them with the new model desu
>>
>>106675313
Someone said you can't run those gguf on comfyui, is it still true?
>>
>>106675336
*I assume it's loading some or all of it into RAM, but gen times are fast (20-30s, with 8 step qwen lightx2v lora).
>>
File: 1728462367776889.png (28 KB, 705x246)
28 KB
28 KB PNG
>>106675347
nope i'm using them now just fine, all you have to do to make the previous qwen edit workflow work is to remove one node and replace it with this:

and add a second load image and connect it if you want to do stuff with 2 images (or 3). update comfy and that node should be there.
>>
>>106675356
thanks anon
>>
>>106675363
np, without that node I was getting random noise, then saw a comment about a new node, now it's fine.
>>
>>106675320
I'm really not interested in cunny but I am terminally ashamed that there are no guro capable base models/loras anywhere.
I really hate the safety faggots.
>>106675348
My full setup is 4080S, 7800X3D and 64GB ram. I doubt I will have money to upgrade in the near future since I am trying to save up for a mortgage and I lost a lot on surgery recently.
>>
File: ya like SAAS?.png (851 KB, 1923x1592)
851 KB
851 KB PNG
https://www.reddit.com/r/StableDiffusion/comments/1nog4kc/comfyui_business_wan25api/
kek, even the ledditors are making fun of CumfartAPI
>>
>>106675378
why upgrade? that's a great pc, you can do basically anything even wan video.
>>
Again that fucking fingolian retard, what an obnoxious piece of shit. Yeah, yeah we get it you are now also getting paid to shill comfy, calm down clown.
Just make your own fucking discord at this point.
>>
>>106675395
Ani is more of a cancer in this thread, I hope you have the same energy when he lies and shills a piece of software that is not even usable yet. Can you imagine the delusion it takes to behave like he does?
>>
>comfyui discussion is so inorganic guise!!!
shittledorf hands typed this
>>
>>106675395
>Yeah, yeah we get it you are now also getting paid to shill comfy
are you retarded or something? he was making fun of comfy
>>
>>106675411
>Lacks reading comprehension
Hallmark of Ani desu, he often misreads things and has melties over it
>>
>>106675411
The fuck you talking about, that was about Kijai not some schizo here.
>>
>>106675419
>>106675423
my filters are too good I'm not seeing any of this drama shit, you should do the same, filter "AniStudio" and your life will be better
>>
File: vu vill own nothing.png (97 KB, 1208x414)
97 KB
97 KB PNG
>2 oss
>3 apis
yep, we're fucked
>>
>>106675393
Sometimes I almost ran out of memory in WebUI with Illustrious and just a few LoRAs without controlnets. Before wan was released I tried hunyuanvideo thing in pinokio and it told me to eat my own face. Plus the LLMs I can run are just pathetic at this point.
I could go second GPU for the latter at least, but the MB I got doesnt have enough spacing for the second slot (the model I wanted was sold out in my area and I had to buy whatever was the closest alternative or risk being without a PC for months since the old one decided to an hero).
Overall its a very comfy machine.
>>
File: 1756309867017273.png (1.05 MB, 774x1024)
1.05 MB
1.05 MB PNG
I actually can't post on esim but I had a nice conversation with an ex gearbox employee. did you know randy pitchfor burned 2 million on diffusion models and has nothing to show for it? hilarious!
>>
>>106675451
>2 oss
>inb4 wan 2.4 HIGH/LOW
>>
>>106675465
>2 million
>not pocket change at that level
oh no!!!
>>
>>106675472
I doubt Alibaba is like that, either they release their best model, either they don't release anything, they're not giving subpar distilled shit like Tencent or BFL
>>
>i'm superior for not being able to engage in discussions properly
gotta admit, that's a new one from filterkeks

>>106675465
>randy pitchfor burned 2 million on diffusion models and has nothing to show for it? hilarious!
he could show that he spent 2 million dollars investigating "AI" to his stakeholders. that's not nothing to show for it. You'd understand this if you were jewish

>>106675451
>2 oss
My thoughts are a dedicated text to image model and one other something-to-video model. Voice and sound models will be kept behind API
>>
>>106675483
Above average bait or this anon doesn't know anything about Qwen, call it
>>
>>106675485
>My thoughts are a dedicated text to image model and one other something-to-video model.
probably two LLMs, Alibaba does that as well
>>
File: 1738539696737567.png (85 KB, 304x360)
85 KB
85 KB PNG
>>106675168
people spit on chroma creator. but the real monster was jewlibaba. fuck qwen and wan (even though i keep my shizo wan, kek)
>>
>>106675495
all their LLMs are already OSS. They already released Qwen 3 Omni. Unless this is just them re-celebrating that release in which case the chinks have started to KPI-maxx like the West and we are entering a global era of rent seeking and stagnation and I should learn how to grow vegetables
>>
>>106675505
>They already released Qwen 3 Omni.
didn't they tease a Qwen 3 vl moe or something too?
>>
wan2.5 looks ass lmao.
1080p doesn't mean fucking anything if it's essentially just 480p shitscaled up.

like really look at the examples. blurry, low detail kek
>>
where did it all go so wrong?
>>
save us debo
>>
the shills for wan2.5 are already out in full force lmao. they are astroturfing in that one thread that "api isn't so bad" in a local diffusion subreddit and that we should beg them like good little cucks to release it openly.

i forgot how truly retarded reddit is, holy fuck
>>
File: 00014-2243056077.jpg (362 KB, 2048x2480)
362 KB
362 KB JPG
After talking with lora gods, it's pretty clear you don't need to go over rank 16 for chroma loras but you will see a benefit going a higher resolution. I'm going to try a batch maxxed lora but last time I did it the results were lacking, I might need to do more steps also zone in on specific periods of the artist work. Chroma handles styles well even at 31 images and 4 repeats
>>106675465
Can we unsubscribe?
Like many anons have told you, this shit doesn't help your case. You don't discuss anything other than shill and you also keep trying to hijack thread OP.
>>
>>106675540
what is chroma? someone finally made something completely different to sdxl?
>>
>>106675540
>Can we unsubscribe
would have done it on your very first blogpost. why the fuck do we want to hear about you fucking up Lora training all the time? you haven't fucking figured it out while /h/ gooners that barely have a tech background run circles around you
>>
>>106675550
I'm learning how to make loras and like a retard I started with chroma which if you look online is full of bad misleading information.
>>
File: 00008-2812543434.png (1.53 MB, 1824x1248)
1.53 MB
1.53 MB PNG
>>106675451
why did i have to wake up to this news
>>
File: wavespeed.png (606 KB, 1586x1189)
606 KB
606 KB PNG
https://www.youtube.com/watch?v=1J2PiGYzy5I
this WaveSpeedAI video from 4 hours ago with only 5 views, this is the youtube channel in the Wan 2.5 preview announcement infographic

it's the same videos as TensorArt's 2.2 release stream video it seems. Why are the chinks grifting eachother

>>106675511
no idea but their VLs are SOTA at least :')
>>
>>106673951
>Beggars can't be choosers
we can, because they want us to beg lol >>106673860
>It is recommended that the community call for follow-up open source
>>
MOVE

>>106675569
>>106675569
>>106675569
>>106675569
>>106675569
>>
>>106675540
>You don't discuss anything other than shill and you also keep trying to hijack thread OP.
schizo logic
>>
>>106675557
NTA but /h/ uses chroma?
>>
>>106673860
>catches up to SaaS
>joins SaaS
anyone genuinely surprised by this is retarded



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.