Discussion of Free and Open Source Text-to-Image/Video Models and UIPrev: >>106669789https://rentry.org/ldg-lazy-getting-started-guide>UIComfyUI: https://github.com/comfyanonymous/ComfyUISwarmUI: https://github.com/mcmonkeyprojects/SwarmUIre/Forge/Classic/Neo: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassicneoSD.Next: https://github.com/vladmandic/sdnextWan2GP: https://github.com/deepbeepmeep/Wan2GP>Checkpoints, LoRAs, Upscalers, & Workflowshttps://civitai.comhttps://civitaiarchive.com/https://openmodeldb.infohttps://openart.ai/workflows>Tuninghttps://github.com/spacepxl/demystifying-sd-finetuninghttps://github.com/Nerogar/OneTrainerhttps://github.com/kohya-ss/sd-scripts/tree/sd3https://github.com/derrian-distro/LoRA_Easy_Training_Scriptshttps://github.com/tdrussell/diffusion-pipe>WanXhttps://comfyanonymous.github.io/ComfyUI_examples/wan22/https://github.com/Wan-Video>Chromahttps://huggingface.co/lodestones/Chroma1-BaseTraining: https://rentry.org/mvu52t46>Neta Luminahttps://huggingface.co/neta-art/Neta-Luminahttps://civitai.com/models/1790792?modelVersionId=2122326https://neta-lumina-style.tz03.xyz/>Illustrious1girl and Beyond: https://rentry.org/comfyui_guide_1girlTag Explorer: https://tagexplorer.github.io/>MiscLocal Model Meta: https://rentry.org/localmodelsmetaShare Metadata: https://catbox.moe | https://litterbox.catbox.moe/GPU Benchmarks: https://chimolog.co/bto-gpu-stable-diffusion-specs/Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-oneTxt2Img Plugin: https://github.com/Acly/krita-ai-diffusionArchive: https://rentry.org/sdg-linkBakery: https://rentry.org/ldgcollage>Neighbours>>>/aco/csdg>>>/b/degen>>>/b/realistic+parody>>>/gif/vdg>>>/d/ddg>>>/e/edg>>>/h/hdg>>>/trash/slop>>>/vt/vtai>>>/u/udg>Local Text>>>/g/lmg>Maintain Thread Qualityhttps://rentry.org/debo
>>106672246neat
>>106672283>cfg 1QIEP is distilled?
>>106672283cool, even. despite my stupid typo>>106672319i'm using the 4step lora, anon.
>>106672325>i'm using the 4step lora, anon.and it's working on the new version? nice
>>106672334qiep is just a simple finetune, there are no huge architectural changes. all loras still work with it.
>>106672283Can it make a more general shape depth map that can be used as a template or is it only doing 1:1?
>>106671842Here's a pic of my workflow, it's not pretty and a lot of it is unused but this is how the sausage is made when I gen Chroma.Also there's one error in there because iirc I use beta sampler with uni_pc_bh2, not ddim uniform. As I said in the last thread, in case you didn't see my brief reply there, I have 12gb of vram so if you're at 8 you might need to make further compromises on the model and t5 encoder
Any of these benefical or should I keep it default?
>>106672493also I'm a retard for calling them "groups of samplers" when I meant to say "groups of nodes corresponding to two different choices of sampler"... or something like that...
>>106672494Random flip should almost always be on.No idea what crop jitter is.Color stuff shouldn't be enabled for character and most style loras. Only in concept loras, maybe.
>>106672246>>106672283Sorry, I should have posted the prompt I got from another anon."Convert the character in the provided image into a fully 3D Nendoroid-style figure. The model should have a large head, small body, and chibi-style proportions while keeping recognizable features and outfit details. Include detailed textures and subtle shading to match the classic Nendoroid look. Pose the character dynamically on a small display base, with soft, even lighting and a neutral background. Render in high-quality 3D, focusing on making it look like a real Nendoroid collectible, not a flat illustration."
>>106672581Sorry again, I'm drunk.I used Comfy UI for the image and then Gemini/ChatGPT for the nendroid gen.
Is there anything like ic light but for the newer image models? Surely there must be something out there where I can take an existing image and modify the light to a high degree
>>106672590Fantastic composition
>>106672512>>106672493tf is wrong with me today. Beta scheduler*. Fuck me.I'm sure I wrote something else completely wrong too but I'm not posting any more corrections. Use my workflow at your own risk lol
>>106672608The edit models can do that
>>106672648Does it have a spline editor or those directional overlays where I can change the direction of the lighting? The ones I seen so far (kontext in this case) just changes the saturation, brightness and hue
>>106672677I know you can specify the direction of like (lit from above, lit from upper left side etc.) but not sure if you can take it a step further with adding light source. Maybe worth testing
>>106672616I thought so too.
How much loss is too much loss for lora training?
>>106672754Loss is a meme value for diffusion models.Noise injection fucks with it.Ignore it.
>>106672795what hath god wrought
>>106672795>on the leftwhat?
>>106672809The position of the images in the workflow does not correspond to the actual position of the images as the model sees them.
>>106672879It's a shame about her ass.
>>106672879Does this only work with pose CN? can it use depth or canny?
>>106672879I enjoy how her tits are the best part both of the gen and of her body
>>106672887Lookin' good boss
>>106672887doesn't look like him at all and it's ultra slopped, I think I'll pass on that one, I'll wait for the next month and see if they improved the model further
>>106672893Well I did use a slopped image from his a reference. It's likely double slopped.
>>106672893>doesn't look like him at allhe can afford much more cosmetics than a hair transplant with his pateron bux
>>106672899kek
>>106672899Do turks go to turkey for hair transplants or is there an even shittier equivalent of turkey that turks use to get hair transplants?
https://huggingface.co/QuantStack/Qwen-Image-Edit-2509-GGUF/blob/main/Qwen-Image-Edit-2509-Q8_0.ggufFINALLY, can we run it on ComfyUi though?
>>106672915Does it use the same clip as the previous edit model?
>>106672918yes
just how bad will Qwen-Image-Edit be at Q2_K?
>>106672915why not? qie+ was already merged in, it's not in stable yet thoughbeit
BROS???? NUNCHAKU GODS!?!?!
>>106672966The wan never meme is true.
>concept bleed>Ultraslopped>Barely looks like reference.Qwen edit bros?
I'm starting to wonder if the fact that the image source being AI made is fucking with the FFLF i2v getting so much colorshift.But the shift also only happens when it is FFLF, if just First frame there's no colorshift.With or without the Color Match node, the shift still happens.For these loops I make with old reactionimages, the shift doesn't happen. And If I remove the light loras all together, the results come out broken at 20steps, cfg4
>>106672979Did you set the noise argument too high? This isn't really an issue with kij's nodes.
>>106672879cool, catbox?
>>106672977is it worse than the original QIE?
>>106672979for example. Seems to work fine?
>>106672966>distracted yet again by another new release for the 7th timekek, holy shit, I give up
>>106673010There's no point when new Wan is around the corner anyway
>>106672971my understanding is that the nunchakuing process is hard only once per architecture, so they should be able to deliver qie2 in a speedy manner (it's still qwen_image arch).Hopefully for wan the first hurdle will be their 1st implementation of the arch on their quants, then we'll get speedy releases too.
>>106673003Worse? No and all things considered, Nano banana seems to do just as bad. I think it's just a hard prompt.
Using the depth map as the reference. It's super faithful to the depth map.
>>106672986>>106673006I'm afk atm, are the workflows in the custom nodes folder? If not, mind sharing?I used to be able to have no color shift, but then it stopped working. Found a new workflow that was twice as fast and no colorshift, but then it stopped working again.Goonrot, perhaps.
>>106672950Likely troll question but almost every single AI model of any kind is dogshit at Q2.Very limited, if any, use cases.
Seek the elden ring, become the elden lord...Probably needs to increase mask grow size, otherwise voluminous headed characters gets squished inside the bounding box.
>>106673056copequant bros...
I have seen the word troll used three times in these threads today. We have a new user amogus.
>>106673014>>106673016In regards to radial attention as well, here's what jumped the queue>fusionX lora>ligthx2v>flux kontext>flex krea>qwen (since its release)>pic rel apparentlyUntil people demand it, wanchaku is dead despite image models being easier to implement
>>106673118there's also qwen loras support, there's actually a PR open right now. I can't build the wheel so I can't try, but qwen bros are eating very good.
Interestingly for single images it does a really good job at preserving the identity. I think 3 images is just too much.
>>106673130Single image with pose controlnet preserves the output pretty well desu.
>>106673085
>>106673202Why was https://rentry.org/wan22ldgguide removed from the OP? Had to find it in the 'chivesFucked my comfyui install and it's basically the only retard proof way of getting triton+sage installed
>>106673216Meant for OP >>106672240
>>106673130
>>106673126Yeah, suppose image bros gotta eat too. Just hope this new wan will allow for 30 second gens
>>106673238I bet that really hurts desu.
Ok... everywhere I go, everyone is raving about how Illustrious is sooo much better than Pony. I have spent an ungodly amount of time trying to make it work for me, and every time I end up going back to my usual setup (Incursio's Meme Diffusion 2.7 + Incase Style + Vixon's Pony Styles - Gothic Neon + Vixon's Pony Styles - Emote v1.0 + Deep Negative Pony + Pony PDXL Negative Embeddings Photoreal)).Illustrious models invariably give me these washed out, pastel images, and all the random (not named) people they generate are so incredibly boring.I haven't been able to find anything that matches the skin texture, the sense of a real musculo-skeletal framework underneath, and just generally vibrant, interesting and pretty random characters my combo generates.And yet I've been using it for sooo long. Surely there must be something better somewhere after all these years. What am I doing wrong anon? Help me graduate my noobdom.
it helps if I post in the active thread lolI am in Japan now. most anistudio work while I'm here will just be cmake and splitting things off into shared libs. sorry I haven't been active on the repo recently but I'll be back at it. wish me luck with softbank fundraising!
>>106673307Be careful with getting money from softbank, they have the touch of death.
another fizzlejeet failure...
>>106673330if anything they spend tall on the stupidest shit like openai and we work then get burned
Spent last 8 hours gooning and ejaculated 4 times to AI slop.Good night anons, I am going to bed now.I will probably feel like shit tomorrow and my dick will hurt, but we will see.
>>106673307I've been working on the repo in the meanwhile.
didnt realize new thread. make sure you update to this node or you get random noise. qwen-image-edit-remove_clothes.safetensors still works if you want to do that.https://files.catbox.moe/y5y946.png
>>106673460and now the safe version:the japanese woman is holding a sign saying "LDG". keep her expression the same.kept her face the same, old edit would sometimes change the faces so this is better.
https://x.com/Ali_TongyiLab/status/1970401571470029070Some goober on reddit said 10 seconds for 1080 gens
>>106673464the japanese woman is holding a sign saying "LDG" with her right hand. keep her expression the same.
>>106673473mind you her right arm was behind her back in the original photo so it did a good job, it's definitely a step up over v1.
>>106673473Can you imagine if you had like a time machine to a year and half or so ago and showed people this? They'd flip their shit.We are truly spoiled.
https://x.com/RuihangChu/status/1970390357738430839Now the fucking amerimutt retard that always mention that the source for wan 2.5 was "JUST TWITTER BRO", gonna kill himself right?Or at least should give more of his tip to Israel.
>>106673488best edit model, best video model, noob/illustrious for anime, open source is eating well
added another image the two women are waving hello. keep their expressions the same. keep their expression the same.ZERO image stitching just connect a load image node to image2 of the new text node.
the two anime characters are sitting at a desk in a classroom in japan. keep their expression the same. the character with white hair has a black blindfold on.
>>106673490Exciting
did a neat upscale of OG miku too:
the problem with the new QIE is that it's actually much too obstinate about keeping things the same. it will exactly clone parts of the image instead of conforming them to the new setting.
seems to be a good upscaler, this was from a pixelized image:
man my comfy install is ancient... I wanted to upgrade to whats it called... pytorch 2.8 and the new triton.Is it worth the hassle?
>>106673460>qwen-image-edit-remove_clothes.safetensorsuhmm link? all the fucking loras have been nuked on hf and civitai
>>106673565At least make a backup.
>>106673556the blue hair anime girl and the red hair anime girl are having a hug.definitely better than before imo.>>106673570need time to upload this sites upload speed is kinda shit, 1s.
>>106673585the blue hair anime girl and the red hair anime girl are dancing and smiling.
>>106673465>10 seconds, 1080 genslet's hope the gen time is manageable
>>106673543Out of curiosity, are you using the lightning LoRA? I found it ruins a lot of its ability to change up styles. CFG also needs to be like 4.
>>106673465The post is gone :O
>>106673588rei is a doll, she cant smile
it thought rei was tanned cause of the shading kek, need to specify light skin
>>106673608>long hairI hate this with QIE, the fucking bleeds, I gotta gacha gens or change wording slightly
Try QIE with controlnets. They work shockingly well. Like they follow the controlnets perfectly without frying the image.
>>106673307*yawn*
>>106673570https://filebin.net/ex6rjtj8rot8vi7u/qwen-image-edit-remove_clothes.safetensors
>>106673623thanks bro
Can any one whos trained a wan 2.1 t2v lora confirm this guide is accurate? I fed AI information of what apparently worked for others who have 16gb vram and 32gb ram on a small video data set. The aim is to train a t2v model and hope it works on i2v in the end (this also apparently works). Never used musubi but its already installed and ready.>durrry why 2.1Because>>106673536Sweet
So now they deleted the announcement are we getting wan 2.5 tomorrow or not?
>>106673677wait nvm, it's backhttps://x.com/Alibaba_Wan/status/1970419930811265129
>>106673677>>106673685https://xcancel.com/Alibaba_Wan/status/1970419930811265129>please escape 5 second hell>please escape 5 second hell>please escape 5 second hell
>>106673608the blue hair anime girl and the red hair anime girl are eating ice cream at the beach in Japan.
>>106673707the problem is they look like temu versions of who they're supposed to be.
>>106673584had to reinstall nunchaku, now in the process of building the flash attention wheel (luckily I'm also a dev and have the cuda sdk + all the build tools already installed)
>>106673719you need to specify to keep the expression the same, I didn't.>>106673473for example is 1:1
>>106673729Try it without the lora.
the blue hair anime girl is relaxing in a hot spring in Japan. keep her expression the same.that's pretty good
>>106673761original:
>>106673765also what's neat is the model got the bottom half of the plugsuit right despite me not specifying rei, or evangelion.
>>106673765>>106673761ok but can you make her cum?
>>106673792not on a blue board, but you can do it anon
Wtf is chroma-2k-qc?
Wan 2.5?What about 2.3, 2.4?
We almost went a whole thread without mentioning Chroma.
>>106673761that was the fp8 one, this time I tried Q8 (was downloading)in general isnt fp8 close to q8? all I know is q8 is closer to fp16 in general.
>>106673839
Wansisters, I dont feel so good...https://xcancel.com/T8star_Aix/status/1970419314726707391
>>106673860I can't hate them because Wan 2.2 is already a great model but that is very disappointing
>>106673860I'm done.
the blue hair anime girl is at the beach in Japan and holding a sign saying "LDG" in stylish black text. Keep her expression the same.diff image, didnt change the face at all, success
>>106673860Who is this guy though. It just says "comfyui tutorials" on his profile. Is he someone of note?
>>106673860Why is comfyui listed as a partner here >>106673536They've got Comfyui people on the English panel too
>>106673860The fingolian is already going damage control mode in the discord, considering he's on comfy's payroll, had early access to some recent wan models and that he's literally listed in their hugginface dev group. I'm gonna say that sadly it does look like api only at least for the near future.
>>106673921Beggars can't be choosers
the anime girl is holding two silver pistols and is pointing them in opposite directions. keep her black blindfold the same.yep it's qwen edit v2 time. this is with the q8 version.
Explains the complete lack of hype I suppose, just a couple vague tweets and discord messages
>>106673966source image:
>>106673972
>>106673860its over
>>106673860No matter how much you pay me I am not using the API
change the location to a computer lab. the character on the left is typing on a laptop, keep his face and expression the same. the green cartoon frog is unchanged, with the same pose. A sign saying "all you can prompt" is on the wall in scribbled black text.pepe edit success
>>106673891Does a lot of comfyui workflows on civit I think and also see >>106673921Calling it now>free version: 5 second lock>api version: no lock, faster
>>106674025api first is DOA, 2.2 is fine anyway. why pay for 5 seconds more when you can stitch 2 clips.
>>106673860I WILL NOT use API nodes!I WILL NOT have my "barely legal teenage girl sucking dick" prompt sent directly to a company just so they can forward a tip to authorities!I WILL NEVER use it!
>>106674023Add Miku Hatsune behind the green cartoon frog, who is pointing and laughing at the character on the left.
>>106673307So a nothingburger like always
>>106673860>>106673921>>106674025Reminder that Comfy is also pushing their integrated "SAAS" and servers and they work hand and hand with WAN and Nvidia.
>>106674028Mostly yeah. True, if you're doing something that requires repetitive motion, the context nodes also work well for 10 - 15 sec gens, depending on hardware of course.
Was it Hunyuan that made their 720p model API and only open sourced the 480p? I really hope Wan don't go down that route
>>106674040background swaps also follow the style better in the new version, nice.
the character is pointing at the camera and smiling.
>>106674076They also went API only for their 3D models. We're fucked.This will just be just the first in a long line of API only shit models. Qwen Image Edit was a bone they threw to us because they knew we would be mad.
>>106674089the character is holding a sign that says "LDG" in black text with his right hand, and giving the thumbs up with his left hand.gj carlos
>I'll be live-streaming the launch event for the new open-source video modelMaybe there's just a lot of miscommunication going on. Perhaps there's still hope?
replace the subtitle text at the bottom with "time to shitpost on /LDG/!"text edits are much better this version it seems, sometimes you'd get random characters if the string length was very different.
>>106674107likely decided to go api first for a year. it's over, just let it go
>>106674112replace the subtitle text at the bottom with "make Miku's hair blonde!". Change the hair color of the anime girl to blonde.
>>106674114It would not be the first time a company has said they were going to release a model and then a suit decided to go all Bilbo Baggins on them and refuse to release it.
https://wavespeed.ai/models/alibaba/wan-2.5/text-to-video>Is literally out right now via apiCan't tell if this means it will be indeed open or not, does google license their models on other platforms as well?
qwen editqweditheh
>>106674158Doesn't this website often intentionally mislabel models?
I've been away for a while and now feeling utterly lost, could someone qrd me on the current models of note?My impression so far is:>qwen edit for editing and image gen>chroma controversially for image gen but can do porn out of the box>wan2.2 for videosAm I missing anything?
>>106674195Seed dream diffused via API and sent to your UI locally.Wan 2.5. Diffused via API and sent to your UI locally (We aren't allowed to critique them because they gave us free shit in the past)
>>106674169No.Also we can already heavily inter from this and from rumors that WAN 2.5 was focused on tuning wan to be able to run at 1080p natively. If that affects general model intelligence in a good or bad way we have to seeIf this is the main update to wan2 then I don't care (this is my cope feel free to use it)Of course it could be a mislabelling which can immediately be checked if anyone actually spends a few shekels on a 1080p gen from wavespeed.
>>106674158Confirmed 1080p and 10secs.Either they do not give us shit or we'll get a very cutdown version like 720p 5secs.Grim.
replace the anime girl with rei ayanami.qwen image v2 knows new characters! that's pretty cool.
>>106674216I specifically recall them doing something similar in the past and they just named another API as another TBA open source model.Pretty sure wavespeed is scummy.
>>106674232That and it's $1 per run.That is insanely expensive. My scam detector is going nuts.
>>106674227>10 secondsHe's right. Check out the examples tab on the wavespeed page Damnit. There's no way this is getting released. If Ali releases this I will bow to the superiority of chinks because I absolutely would not release this if I had it lol>>106674244I don't remember anything about this but $1 per run is the same they were asking for hunyuan (0.20 per second). If anything it's cheaper now
>>106674158Some dodgy sites like that and blogs tend to already have a "model" or article already up, probably placeholders for seo or someshit.>>106674159qwan 2.5
>>106674249You're being scammed bro.
replace the anime girl with a white gundam.
>>106674255It has examples already and wave speed isn't a dodgy site, I remember using them on replicate to generate free hunyuan videos of little girls in bikinis so much they had to put 3 levels of censorship in at the prompt and output level, and then they added a lora to slop the faces of young people lmao >>106674260I've already scammed them harder than they can ever scam me. See above
Oh another thing, there is a a chance replicate will be serving the model completely for free for a few hours once it comes out. They did this for hunyanvideo and wan 2.1s release for sure. Since most people itt can't or won't do 1080p locally if this happens it may be your only chance to get some unwatermarked 1080p genning in wan 2.5 so pay attention to that
>>106674297If it's api only the can suck my balls and eat my shit. At that point just use fucking veo.
>>106674304Veo won't let you generate beautiful little girls in bikinis in the beach in 1080p while I promise you if replicate serves it they keep the filters loose for a few hours Actually I don't promise that, maybe they learned their lesson but the odds are 1 million times higher with wan 2.5. If this isn't something you care about then ew stinky brown person but also carry on
>>106674318You should kill yourself, piece of garbage jew.
>>106674304Yeah if its a worse veo not sure why anyone would wanna use it. I wanna believe they know that but who knows, seen corpos make dummer decision.
The anime girl Miku Hatsune is holding a stack of boxes with the label "LDG gens" on them. keep her expression the same.
There's no way the wavespeed examples aren't wan. They're all yapping so much >>106674323>No argumentBig words from a mutt that's circumsized and mutilated for literally no reason lmao. I accept you concession You now remember that Tel Aviv University made a SOTA video model and never released it btw >>106674327I'm 100% certain it will be less censored than veo at least for the first week or so to gain userbase. There are legitimate use cases for this (like beautiful children)
>>106674304Basically this.
I don't see how 1080p 10 second videos would even be viable on current consumer cards. 720/5secs already pushes 24GB cards to the limit
>>106674327>Yeah if its a worse veogoing by their examples it's def worse than veo
>>106674391We can use our local UIs to diffuse via API directly to our PC locally.
Guys I'm feeling pretty black pilled right now.
>>106674448Nothing is confirmed until the live stream. I WILL keep coping until then
>>106674448The night is darkest before the dawn >>1066743915090 users exist, and resolutions between 720p and 1080p exist. Why are you niggers dooming so hard we have one consistent Chinese guy saying it's open source and then a literal who being like no its not
>>106674465Haven't seen a single tester or dev saying it is open source.
>>106674476Chinese guy who goes to conferences has only been alluding to it as an open source release Also here's the link for the image 2 video. Cheaper to run than the t2v it seems??https://wavespeed.ai/models/alibaba/wan-2.5/image-to-video
>>106674490Man stop linking that scam site. I'm not even fucking around here.
>>106674490yeah check the t2i ones they look quite sloppedaudio sounds like shit in all the examples but i guess it's progress for them
>>106674510>>yeah check the t2i ones they look quite sloppedFunny because I've seen the completely opposite from other's peoples gens, it absolutely shits on flux and derivatives.
>>106674500Lmao kill yourself schizo if you want to see the beautiful children I make on replicate in a few hours you could have just asked (I was going to share them anyways for science)>>106674510After seeing their prompt guide I don't trust them to not prompt for slop anyways. I also don't even use their Chinese negs in my workflows because it also adds to the slop
>>106674520okay? they're still slop>>106674524yeah but if it isn't local it's not going to matter much for me since the alternatives would be better than what they're offering
everything will be closed source from now on, its over for local
>>106674551After seeing how much was squeezed out of sdxl for years I'm genuinely not worried if wan 2.2 is the last open source video base model for the next two years (it won't be)
>API JEW NODES ADDED https://github.com/comfyanonymous/ComfyUI/pull/9996>NO LOCAL NODESKEEEEEEEEKIt's so over
>>106674570>BFL Presents: Flux Video Mega Slop Edition
>>106674524>Lmao kill yourself schizo>Hey guys, this website with no association with Alibaba just happens to have access to Wan 2.5 before it's even present. No, you can't look at the API.Now give me a buck and maybe I'll spit out a video from some random API at a 200% markup.Go fuck yourself scammer.
>>106672879way to ruin a beautiful woman and turn her into an ugly yellow beast
>>106674622Happens literally all the time in all industries. Happens in this industry very often. You must be over 18 to post on this website.
>>106674653You know what happens way more often? Intentionally mislabeling APIs for highly anticipated products and charging exorbitant prices.Like do you work for wavespeed or something?
>>106672879>roundest asian ass
>>106674666No it actually doesn't lolYeah my dad works at wave speed his last name is Speed
>turns out I've been in kijais workflow all along>found some nodes that could have fixed the color shift>it makes it worse>it's now even showing up in the very simple imagesI am about to shit myself in anger.
>>106674695>No it actually doesn'tDo you have any idea how many fake claudes there are out there right now? I cannot tell if you are genuinely stupid or this is some kind of next level vanryo shit.
>106674718How many of those fake claudes are official partners with replicate retard? No more actual replies for you. https://replicate.com/wavespeedaiAnyways, assuming wavespeed isn't fucking around, I noticed in their i2v example prompt they type [Close-up] with square brackets. I know that they tried to do some stuff with prompt structure with 2.2 so I'm wondering if they're messing around again with 2.5
>>106674737>official partners with replicateIs that supposed to hold some meaning to me? In don't care if they're official partners with my dick. They aren't alibaba and their model hasn't even been presented yet.
why is anon shilling an api model?
>106674762HAHAHAHAHAHAHA
>>106674763Because anon likely gets a cut.1 fucking dollar per gen. That is an insane markup.
>>106674792even sneedream wasn't that greedy
Why is wanvideo node calling samplers schedulers?
>>106674799Yes but seedream was actually seedream.
I thought they would wait until 3.0 before going saas>Veo 3 existsAnd I guess Wan's guys are planning to give a less good, but cheaper alternative to VEO 3's 250$ per month
The pricing shows they’re serious about starting a business. This isn’t a hobby. They’ve started cashing out.
>>106674602fucking grim
Before we all lose our shit. Do we actually have confirmation this is API only or is just extremely obvious from the context?
>>106674832As they feared, what’s unfolding looks like it can only end one way: picking a fight with the open-source community and having the livestream get raided.
>>106674832looks like 99% wan 2.5 is going to be api. Judging from that retard's tweet (he's an insider) and also api nodes already being PR'd in comfy.It's unironically over. We'll have to cope with 2.2. Hopefully nunchaku makes it less painful
>>106674820>I guess Wan's guys are planning to give a less good, but cheaper alternative to VEO 3's 250$ per monthThey are. From the wave speed description:"Compared to Google Veo3, it offers a more affordable solution with faster video generation speeds, making it an excellent choice for creating audio-embedded videos efficiently."I'm full doomed right now by the way, thanks guys. But we will only know for sure in about an hour or so>>106674602I am now willing to bet money on it being SaaS only as an initial release. Very sad>>106674832The comfy PR is the best knowledge we have so far. https://replicate.com/wan-video/wan-2.5-t2i-previewHmm anyone else having this page take a long time instead of 404ing immediately?>>106674846>picking a fight with the open-source community and having the livestream get raidedCalling it now, we will see the first assassination during a model release in the next 5 yearsOne final wave speed link I foundhttps://wavespeed.ai/models/alibaba/wan-2.5/text-to-imageAlso this claims they're getting it served from Alibaba's cloud. Since wave speed is a Chinese company I 100% think this is real
Bros...
>>106674863we got memed by alibaba, it's unironically SUPER over, the west has fallen, millions must die
>>106674863this is the kind of propaganda they push when its over but they don't want you to realize it.
>>106674860Nevermind I'm schizo on the replicate link >>106674863This is literally exactly how the thread's reactions have gone kek
bros what the fuck, sdxl is supported in nunchaku in the last release?
>>106674906https://huggingface.co/nunchaku-tech/nunchaku-sdxllol?
>>106674906Why would you ever need a speedup for sdxl
>>106674923yeah like sdxl takes... 3 secs? I'm gonna try it anyway because I live for hoarding
>>106674743me in the back
>>106674928just noticed they updated the nunchaku wheel, but not the comfyui wheel sooo only through diffusers for now?
I don't understand why these retards don't publish the actual code for quanting, just for diffusing. Like who the fuck cares about base sdxl/turbo? release the quanting code you fucking chinkoids
>>106674855>It's unironically over. We'll have to cope with 2.2. Hopefully nunchaku makes it less painfulAt least 2.2 isn't that painful of a cope, and bytedance has already shown you can do a lot with finetuning it There's also that paper of the 2.2 tune that did well on benchmarks that said it will release weights soon Yes this is the bargaining stage before depression and finally acceptance how could you tell
>api>abandoned by nunchakuA new king must take the throne
if wan2.5 remains api only, then it will become like sd3 and become irrelevant. wan2.2 will become the new sdxl of video gen and last many years.
>>106674923Chroma kekes on swicide line
>>106674982Yep. No one even knows the limits of 2.2 just yet because not enough experimentation has been done with promoting yet. It's also a top tier text to image model. >>106674965>mustNot sure why you think this. Getting wan 2.1 at all was an insane blessing for local
>>106673860>wan finally looks like veo 3>it's a paid modelevery time. as a gamer, i'm not surprised at all
>>106674906based
>>106674906its base sdxl, who caresif i cant get a speedup using my favorite finetune then its useless
https://xcancel.com/wavespeed_aiSounds like wavespeed will be rebroadcasting the conference on their Twitter and YouTube, this was 1 hour ago
Why won't this context window node work with these nodes? It seems to fix a lot of the color shifting.
>>106673570>all the fucking loras have been nuked on hf and civitaiwait what? I stepped away for a few months and everything is gone?I am also in need of help.Been using A1111 for SD and Forge for Flux.I took a break and now A1111 just refuses to start now so I guess its time to move on.What UI is the current recommended for SD (preferably with working controlnet for Pony/Illust models and something like regional prompter/latent couple built in or as a plugin)?And what should I use for FLUX and FLUX derivatives?Yeah I know comfy is the way but I really enjoy slapping prompts, generating 32 images and saving the ones I like right in the interface. Would be grateful for suggestions!
>>106673860>He trust'eded the chinks and thoughteted'ed that china would really save localLMAO, I told you, the moment they got a good model they kept it for themselves, we'll never get a good model locally in our lifetimehttps://www.youtube.com/watch?v=H47ow4_Cmk0The day Qwen Image Edit will be as good as Nano Banana they'll keep it for themselves as well, be prepare for that btw
>>106675096Most have moved to comfyui, otherwise if you like forge you can try neoforge and/or one of its 99932553211 forks.I'd personally stick with comfy for the time being, has a bit of a learning curve but its worth it imho
>>106675096Neo Forge
>>106673966>>106673972are these done in comfy? did something local like nanobanana but not cringe drop?
>>106675096NeoForge branch o Forge classic, easy stable search in github Haoming02
>>106675116Wan 2.2 has made me orgasm over 50 times, that's pretty good for a price of 0 dollars and Apache 2 license, and we're still using the equivalent of base sdxl kind ofThe new cope will be that fine-tune that impressed Gemini and benchmemes, and hoping they release weights for it soon
>>106675143I don't want to end up with a 2 models (high low) process, that's lame
kek
when did illustrious go closed source, what the fuck? what do I use now instead?
>>106673860Holy shit it has sound like VEO3, no wonder they want to keep it for themselveshttps://files.catbox.moe/orknbn.mp4https://wavespeed.ai/models/alibaba/wan-2.5/text-to-video
>>106675150>that's lameExplain why without getting upset or admitting to a skill issue. I agree that it's less convenient than a one sampler workflow but once you set it up it's not any different
>>106675169>Explain why>it's less convenient than a one sampler workflowyou just did, loading/reloading the models take time, that's bullshit
>>106675152>>106675168>$1.5 for a one 1080p 10 sec clip?how big is this shit? I'm sure wan 2.5 is a 50+b model, even if it was local no one could've run that lol
>>106675152>>106675168Shame on me for feeding the schizo when wan's own tweet had wave speed as an official partner listed like 4 times lolI also had no idea tensor art was Chinese or that heavily associated with the industry since I never used it and thought they were a wild west for loras of real child models for the longest time >>106675179>loading/reloading the models take time, that's bullshitPretty sure this is either a DRAMlet cope or flat out not true or both. My workflows aren't much slower than they used to be for 2.1 idk
>>106675163Noob or Illust Yeah, nothing changed for weebstuff since last year
>>106675188we could copequant it, as usual
>>106675195I see a new illust wai release, I guess I'll use that as a base.
>>106675188First of all you're not sure of that at all, second of all there's anons with 96GB cards in this general, third of all you need to be 18 to post on this website.
>>106675200>second of all there's anons with 96GB cards in this generalyeah sure, try to guess why step video (30b) got ignored everywhere? because it's too big to be run by people except 2 millionaires that have a A100, you need to be 18 to post on this website
>>106675188Man we get gargantuan LLMs for open source that people eventually end up running but constantly complain about this shit on image/video gen
>>106675128>>106675139>>106675120thanks! thats why I asked. theres a billion forks now, even in the op theres reForge, Forge Classic, SD.Next and I have no idea whats the difference.I'll try the Forge Neo then. Are regional prompter and pony/illustrious controlnets available there? I will use it while I learn comfy probably.Is there a way in Comfy to batch generate and select what to save like I do in WebUI? Something like an infinite image browser?
>>106675205Step video was worse than wan in every way and people did test it out. You need to be under 18 to turn me on
>>106675210you can batch gen in comfy, yes.As for controlnets they're native but you'll get better mileage by using the controlnet aux plugin. Tbh I'm so lazy I just shove in the AUX AIO Preprocessor, and let it do its magic on its own..You can have a preview output in comfy without saving, then you'll just check your queue and rightclick->save what you like.Usually I just save everything, for me space is not a problem
>>106675214>Step video was worse than wan in every way and people did test it out.Step video (19 feb 2025) was released before Wan 2.1 (25 feb 2025) you mongoloid, at that time of release it was the best model we had and no one cared because it was too big
>>106675227thanks, I meant controlnets in Forge Neo. I never managed to make them work on pony, no matter which controlnet models I tried.For me the space isnt a problem either, cant thank my past self enough for investing time to build a personal NAS. The amount of failed slop piling up in the folders is the problem. I guess its time to learn comfy after all, the quality I've seen out of those is pretty impressive.BTW whats the mentions in the threads about LoRAs being nuked from Civitai?
>>106675240Sorry I meant it was worse than hunyuanvideo when it came out. Anything else or more ad hominems?
>>106675240you think people had enough time to test it in 5 days. It takes a few days just to have the shit implemented but people decided it was the best model? I never even saw one gen out of that model lol.
>>106675259>Sorrykek
>>106675168Those eyes are deader than my future...
>>106673671Any smart anons plz
>>106675261>It takes a few days just to have the shit implementedI went to verify that and surprisignly, Comfy has never implemented it on his official repo, maybe he hadn't a powerful enough GPU to test that out?the company that made that model had a custom node thoughhttps://github.com/stepfun-ai/ComfyUI-StepVideo
>>106675134yes comfy. that is qwen edit v2 (new), Q8. Even better for image/text edits/swaps.https://huggingface.co/QuantStack/Qwen-Image-Edit-2509-GGUF/tree/main
>>106675298Every anon that actually trains loras has given the same advice which is "just start trying stuff".
>>106675227Alright what is this trickery. it basically has everything covered. last time I tried comfy it was a barely working blank field on startup with no additional tools. Am I so much out of touch?
>>106673966>>106673972can you stitch the 2 images instead? would be more convenient to seehttps://github.com/BigStationW/Compare-pictures-and-videos
>>106675258not all loras, the one deemed unsafe by the 'AI SAFETY' faggots.Basically loras about real people are verboten, same as undress loras or anything that can be used in a clear way to make deepfakes.There's still tons of porn realism loras anyway sooo it's literally so these retards can say 'UGHH WE R FIGHTING THE BAD AI GUYS!!!'Also of course loras for cunny, be it real or anime are also not allowed. BUT usually anime models do generate cunny without problems soooo lmao
>>106675313oh wow thanks! guess I'm learning comfy this weekend.Sucks I only have a 16gb gpu, but oh well. We only have so many kidneys
>>106675317It's comfy, embrace the spaghetti bro
>>106673307Who the fuck would give you fundraising?Holy fuck you actually believe your delusions and will post here full of more tall tales. You have been caught lying about this project for too long just go back to your shit tier thread vs begging us for attention
>>106675323>Sucks I only have a 16gb gpu, but oh well.you can offload to the cpu, it's not that much of a big deal, he's the node to add some GB of the model to the ram https://github.com/pollockjj/ComfyUI-MultiGPU
>>106674340cute
>>106675323you dont need a lot. I am using 16gb as well (4080).16 is plenty for wan 2.2, and despite the model being 20gb along with other stuff it works perfectly fine in comfy. no multigpu node needed.
>>106675323>16gb>Q8 image edit 2509each iteration/step will take roughly 20-15s~ depending if you're using sage/flash attention or nothing. Lightning loras will make it bearable, but I didn't try them with the new model desu
>>106675313Someone said you can't run those gguf on comfyui, is it still true?
>>106675336*I assume it's loading some or all of it into RAM, but gen times are fast (20-30s, with 8 step qwen lightx2v lora).
>>106675347nope i'm using them now just fine, all you have to do to make the previous qwen edit workflow work is to remove one node and replace it with this:and add a second load image and connect it if you want to do stuff with 2 images (or 3). update comfy and that node should be there.
>>106675356thanks anon
>>106675363np, without that node I was getting random noise, then saw a comment about a new node, now it's fine.
>>106675320I'm really not interested in cunny but I am terminally ashamed that there are no guro capable base models/loras anywhere.I really hate the safety faggots.>>106675348My full setup is 4080S, 7800X3D and 64GB ram. I doubt I will have money to upgrade in the near future since I am trying to save up for a mortgage and I lost a lot on surgery recently.
https://www.reddit.com/r/StableDiffusion/comments/1nog4kc/comfyui_business_wan25api/kek, even the ledditors are making fun of CumfartAPI
>>106675378why upgrade? that's a great pc, you can do basically anything even wan video.
Again that fucking fingolian retard, what an obnoxious piece of shit. Yeah, yeah we get it you are now also getting paid to shill comfy, calm down clown.Just make your own fucking discord at this point.
>>106675395Ani is more of a cancer in this thread, I hope you have the same energy when he lies and shills a piece of software that is not even usable yet. Can you imagine the delusion it takes to behave like he does?
>comfyui discussion is so inorganic guise!!!shittledorf hands typed this
>>106675395>Yeah, yeah we get it you are now also getting paid to shill comfyare you retarded or something? he was making fun of comfy
>>106675411>Lacks reading comprehensionHallmark of Ani desu, he often misreads things and has melties over it
>>106675411The fuck you talking about, that was about Kijai not some schizo here.
>>106675419>>106675423my filters are too good I'm not seeing any of this drama shit, you should do the same, filter "AniStudio" and your life will be better
>2 oss>3 apisyep, we're fucked
>>106675393Sometimes I almost ran out of memory in WebUI with Illustrious and just a few LoRAs without controlnets. Before wan was released I tried hunyuanvideo thing in pinokio and it told me to eat my own face. Plus the LLMs I can run are just pathetic at this point.I could go second GPU for the latter at least, but the MB I got doesnt have enough spacing for the second slot (the model I wanted was sold out in my area and I had to buy whatever was the closest alternative or risk being without a PC for months since the old one decided to an hero).Overall its a very comfy machine.
I actually can't post on esim but I had a nice conversation with an ex gearbox employee. did you know randy pitchfor burned 2 million on diffusion models and has nothing to show for it? hilarious!
>>106675451>2 oss>inb4 wan 2.4 HIGH/LOW
>>106675465>2 million>not pocket change at that leveloh no!!!
>>106675472I doubt Alibaba is like that, either they release their best model, either they don't release anything, they're not giving subpar distilled shit like Tencent or BFL
>i'm superior for not being able to engage in discussions properlygotta admit, that's a new one from filterkeks>>106675465>randy pitchfor burned 2 million on diffusion models and has nothing to show for it? hilarious!he could show that he spent 2 million dollars investigating "AI" to his stakeholders. that's not nothing to show for it. You'd understand this if you were jewish>>106675451>2 ossMy thoughts are a dedicated text to image model and one other something-to-video model. Voice and sound models will be kept behind API
>>106675483Above average bait or this anon doesn't know anything about Qwen, call it
>>106675485>My thoughts are a dedicated text to image model and one other something-to-video model.probably two LLMs, Alibaba does that as well
>>106675168people spit on chroma creator. but the real monster was jewlibaba. fuck qwen and wan (even though i keep my shizo wan, kek)
>>106675495all their LLMs are already OSS. They already released Qwen 3 Omni. Unless this is just them re-celebrating that release in which case the chinks have started to KPI-maxx like the West and we are entering a global era of rent seeking and stagnation and I should learn how to grow vegetables
>>106675505>They already released Qwen 3 Omni.didn't they tease a Qwen 3 vl moe or something too?
wan2.5 looks ass lmao.1080p doesn't mean fucking anything if it's essentially just 480p shitscaled up.like really look at the examples. blurry, low detail kek
where did it all go so wrong?
save us debo
the shills for wan2.5 are already out in full force lmao. they are astroturfing in that one thread that "api isn't so bad" in a local diffusion subreddit and that we should beg them like good little cucks to release it openly.i forgot how truly retarded reddit is, holy fuck
After talking with lora gods, it's pretty clear you don't need to go over rank 16 for chroma loras but you will see a benefit going a higher resolution. I'm going to try a batch maxxed lora but last time I did it the results were lacking, I might need to do more steps also zone in on specific periods of the artist work. Chroma handles styles well even at 31 images and 4 repeats>>106675465Can we unsubscribe?Like many anons have told you, this shit doesn't help your case. You don't discuss anything other than shill and you also keep trying to hijack thread OP.
>>106675540what is chroma? someone finally made something completely different to sdxl?
>>106675540>Can we unsubscribewould have done it on your very first blogpost. why the fuck do we want to hear about you fucking up Lora training all the time? you haven't fucking figured it out while /h/ gooners that barely have a tech background run circles around you
>>106675550I'm learning how to make loras and like a retard I started with chroma which if you look online is full of bad misleading information.
>>106675451why did i have to wake up to this news
https://www.youtube.com/watch?v=1J2PiGYzy5Ithis WaveSpeedAI video from 4 hours ago with only 5 views, this is the youtube channel in the Wan 2.5 preview announcement infographicit's the same videos as TensorArt's 2.2 release stream video it seems. Why are the chinks grifting eachother >>106675511no idea but their VLs are SOTA at least :')
>>106673951>Beggars can't be chooserswe can, because they want us to beg lol >>106673860>It is recommended that the community call for follow-up open source
MOVE>>106675569>>106675569>>106675569>>106675569>>106675569
>>106675540>You don't discuss anything other than shill and you also keep trying to hijack thread OP.schizo logic
>>106675557NTA but /h/ uses chroma?
>>106673860>catches up to SaaS>joins SaaSanyone genuinely surprised by this is retarded