Discussion of Free and Open-Source Diffusion models.Previous: >>103533408>UIMetastable: https://metastable.studioSwarmUI: https://github.com/mcmonkeyprojects/SwarmUIForge: https://github.com/lllyasviel/stable-diffusion-webui-forgereForge: https://github.com/Panchovix/stable-diffusion-webui-reForgeComfyUI: https://github.com/comfyanonymous/ComfyUIInvokeAI: https://github.com/invoke-ai/InvokeAI>Models, LoRAs, & Upscalershttps://civitai.comhttps://tensor.art/https://openmodeldb.info>Traininghttps://github.com/Nerogar/OneTrainerhttps://github.com/kohya-ss/sd-scripts/tree/sd3https://github.com/derrian-distro/LoRA_Easy_Training_Scripts>HunyuanVideoComfy: https://github.com/kijai/ComfyUI-HunyuanVideoWrapper/Windows: https://rentry.org/crhcqq54Training: https://github.com/tdrussell/diffusion-pipe>FluxForge Guide: https://github.com/lllyasviel/stable-diffusion-webui-forge/discussions/1050ComfyUI Guide: https://comfyanonymous.github.io/ComfyUI_examples/fluxDeDistilled Quants: https://huggingface.co/TheYuriLover/flux-dev-de-distill-GGUF/tree/main>MiscShare Metadata: https://catbox.moe | https://litterbox.catbox.moe/Generate Prompt from Image: https://huggingface.co/spaces/fancyfeast/joy-caption-alpha-twoArchived: https://rentry.org/sdg-linkSamplers: https://stable-diffusion-art.com/samplers/Open-Source Digital Art Software: https://krita.org/en/Txt2Img Plugin: https://kritaaidiffusion.com/Collagebaker: https://www.befunky.com/create/collage/Video Collagebaker: https://kdenlive.org/en/>Neighbo(u)rs>>>/aco/sdg>>>/aco/aivg>>>/b/degen>>>/c/kdg>>>/d/ddg>>>/e/edg>>>/h/hdg>>>/tg/slop>>>/trash/sdg>>>/u/udg>>>/vt/vtai>Texting Neighbo(u)r>>>/g/lmg
Can't say I love how these gens turned out on either model and I probably could come up with a better test prompt, but here is my baseline using the regular model. CFG 2, Guidance 8, Flow Shift 8, 30 steps. Will follow with FastVideo checkpoint in a moment.
>>103539859This is good.
>>103539913And here is the one using the Fast Video model. Same params including seed, but with 17 flow shift and 10 steps. I actually prefer this one, though the lighting isn't behaving.
Unless there's a different workflow for Fast Hyvid there seems to be weird flickering lighting artifacts and general crust
>>103539933what's with the light? it's flickering
>>103539956That's how candles do innit
>>103539956The prompt says it's lit by a single candle so I wouldn't read too much into that. Somebody give me a sfw prompt that isn't weird and I'll do one more test.
>>103539913>>103539933>CFG 2?? so are you using negative prompt as welli like your 1girl she is very cute can you post prompt/catbox for thiswhat is the speed difference on fastvideo vs normal hunyuan 10 step 17 flow shift? is it the same?
>>103539972Could you post your workflow, please sir
Blessed thread of frenship
>BTFOs Chinese modelshttps://x.com/GoogleDeepMind/status/1868703624714395907/video/1
does this fast shit mean I can run it on my njudea 3060 or
>>103540073>censored, online only, will bankrupt you to use
>>103540096It's quite janky and low quality so it likely needs to have some extra custom node work done to make it generate better outputs. For now you might as well just do 12 steps on the normal model and accept some blur.
>>103540073>no weightskek, worthless
>>103540073this is what GTA 8 will look like
>>103539859>>103539913Fantastic
>>103540119This is how most games will look in 10 years when they can get AI frame gen to work at 30fps.
>>103540125pfff, the ps5 pro an do it now
https://www.reddit.com/r/StableDiffusion/comments/1hfnnjg/a_comfyui_node_for_hunyuanvideo_that_lets_you/Does this work or is just a placebo?
>>103540172Schizo placebo
>>103540073>not localwhy should we care? at least pika 2.0 adds something new to the table that should be replicated >>103533581
>>103539976>>103539985Yes, I do use a negative prompt. Iterations per second are identical between models. Excuse the schizo negative; I'm experimenting. Here is the catbox.https://files.catbox.moe/8abeag.webm
>>103540073https://www.youtube.com/watch?v=nlIhROafItU
>>103540225Thanks
wtf I missed real Princess Slaying Hours
egg
>https://github.com/fszontagh/sd.cpp.gui.wxGui written with c++
>>103540506It's always princess slaying hours.
>>103540674Ani on suicide watch
>>103540674perfect time, SD has never been more popular
>>103540704top kek
>>103540674>GUI>no screenshotsMe no andastand.
>>103540766>no andastandhttps://youtu.be/RoSs9-NDP3E?t=9
>>103540694
>>103540704it does flux too, in a a1111 interface that justwurx
>>103540783Found a few on the linked site and it looks so uncomfy. Still, an actual program, cool.
what's the secret prompts to add to make HY output better quality videos?
>>103540847Why did you posting the png and not the mp4, sir?
>>103540870>what's the secret prompts to add to make HY output better quality videos?there's none, it's already a miracle it works this well with the wrong text encoder, to get improvements we need to get the official one, HunyuanMLLM
>>103540890vramlet
>>103540806can you take some screenshots of the gui
>>103540847Interesting pose and angle
>>103540674>>103540694HAHAHAHAHAHAHA
>>103540694>>103540792>Julien publicly announces some project>doesn't deliver>gets done by someone else>will act as if be deserves respect and credit How many times will this keep happening?
>>103540971at least trani is still the best animator of the univers- oh wait i forgot
>>103540985>TFW his only feat is blowing comfy for a job at a dying company
>>103540674okay share some screenshots or something I wanna know how good it is
>>103541039Even if it's shit that's more work put in than the fraud trani
Oh nononononoTrani sistershttps://private-user-images.githubusercontent.com/51741446/393540826-a0c02eff-11da-487e-a09b-9448d18006f8.gif
>>103541084https://github.com/fszontagh/sd.cpp.gui.wx/issues/23
>>103540225>I do use a negative prompt.Is it really useful?
>>103540225how faster is it?
>>103540073another day, another imgtovideo but not available locally
>>103540914Very nice, catbox?
>>103541084>>103541105looks like some bittorrent client
>>103541271sovl
>>103540674Janky as hell, needs polish.
>>103541105looks fine I guess, but why the hassle of doing that
>>103541105>Crated at
>>103541294Still more than anything trani has ever done
Has anyone attempted to make any img2video hacked comfy flow for hunyuan? Even if the output is not 100% identical to the original frame, but following the exact same compositionIn theory it should be possible to transform a static image into video, then doing something to it to simulate motion, then using video2videoor just using another local model that allows img2video and using the output from it on hunyuan vid2vid (would probably be garbage, but who knows)
>>103539859This is honestly incredible.
>>103541320kindahttps://github.com/kijai/ComfyUI-HunyuanVideoWrapper/tree/main/examples/ip2v
>>103540870Amazing video, cinematic... Stuff like that>>103540890
>>103541320It can kind of work that use generate the frames using one still image, but motion? Nope... It just isn't trained to make videos from one image, the frames need to be real video with motion.
>>103541444the example is pretty cool actually considering it was just one image? So erm, when can we have this?
>>103541505nvm, i just checked the githttps://github.com/kijai/ComfyUI-HunyuanVideoWrapper/tree/mainI'm guessing now, this pleases me.
>>103540847moody, I like
hello /ldg/, quick question, not sure to ask here or /sdg/ . I have a project idea in mind that would involve creating many simple 2d assets in a consistent style, like small images of cartoon furniture in an isometric perspective. I don't have a great graphics card to do anything locally and I don't mind spending money to get decent quality. what would my best options be? pic rel a misato plushie I made in imagen
>>103541572You could rent compute from a cloud service if you don't want to buy a new computer.
>>103541572I use vast.ai to rent computers with gpus, like 30-40 cents an hour for a 4090.
>>103541429having horns like this would be horrible. your skull could be split if some would pull on it because of the leverage effect.
>>103539836>Metastable>SwarmUI>Forge>reForge>InvokeAII'm assuming ComfyUI still isn't comfy so which one should I use? I used to use A1111 but it seems like it's not recommended anymore?
>>103541756easiest way to use Hunyuan at the moment
>>103541756>>103541766"easiest way to use Hunyuan at the moment is comfyUI, but for other models I think Forge is good enough"
>>103540674it's shit. looks like shit and needlessly abstracted>>103541766their gradio app demo technically is
>>103541707>>103541613Interesting, and a 4090 is sufficient to run the latest models? How does the workflow work like you have to install software I'm assuming, so then that takes time out of your reservation or you can load an image or something? Should I get experience creating images locally first to get a hang of it even though I just have a 2060 super rn
>>103541796now have it running through a street before devouring someone, make it realistic, i want to see their terror as this thing runs through the street eating people.
>
>>103541886can't wait for the ani UI that doesn't look like shit and is not needlessly abstracted
>>103541221I find it useful, but doubling gen time is definitely a hefty cost. Some prompts need it more than others.>>103541242Same speed per step, but fewer steps required. My impression is that 12 steps get you similar polish to 30 steps on the base model, so it's a pretty significant speedup. I don't have a strong enough grasp on either model's capabilities to evaluate how steep the loss in quality is, but it doesn't really feel worse.
>>103542003was the flickering light just a thing with onle one video?
>>103541105No preview for generation? I guess it works but man does that seem slow and not that much faster than using Pytorch.
>>103542003>12 steps get you similar polish to 30 stepsthanks anon, interesting
>>103542019Yes. My prompt called for candlelight so I think that's all it was.
>>103541913A 4090 can do pretty much any image stuff. Video stuff can get a bit tight.
>>103541994How can he work when his anus leaks stinky fluid that betrays him?
>>103542039I also want to stress that I'm just naïvely using the flow shift value I've seen recommended for 10 steps on the base model, but from what little I understand there isn't any particular reason why it should be the same. It does seem to work, but I'm sure it is worth experimenting with different stuff. The guy who created it seems to have just copy/pasted the model card from the base model and gives no suggestions for any of the parameters.
>>103542111i too hate it when my big fat juicy asshole betrays me
>>103542131More like blown out and loose after years of abuse, Imagine the sleeve of a wizard drenched in brown smelly aids fluid.
What's the difference between Forge and Reforge? Which should I use?
>>103542129>I also want to stress that I'm just naïvely using the flow shift valueI think that's pretty much everyone here.What I don't get is why wouldn't we let it be at high number (for example 20) regardless of the steps since at low steps it helps and in high steps it seems to not have any bad effect.
>>103542098>Video stuff can get a bit tight.I bet
>julien is seething
>>103541756ComfyUI is comfy in the way that my ex gf was not a lying whore
>ani is crying
>>103542044oh cool!
>>103541913>Should I get experience creating images locally first to get a hang of it even though I just have a 2060 super rnyou can do quite a bit with a 2060 - definitely see what you can do with that before thinking about renting
>>103540674But where all all my comfy custom nodes? :(
>>103541707How does that work? I can see the way we get the cards, but what about the os? Can I install whatever linux I want and comfy there ?
>>103542418The company you rent from spins up a virtual machine for you running what ever OS you tell them
>>103542431I see.Then what happens when I stop using the thing, should I basically reinstall everything every time I need compute?I'm coming from the VPS world so it's kind of confusing to me.
>>103542442pretty much, yeah. you can publish your own docker image if you wanna save time though since container startup time doesnt count
>>103542485OK thanks anon.Maybe worth my time testing what an A100 80GB can do.
>>103542442I wrote a couple sh scripts to install and download everything I'm using. Most of the machines will have gigabit up and down so after standing up the image I run the script and grab a drink and it'll be done when I'm back. You also will keep the hard drive space you used reserved unless you destroy it but it's kind of a pain in the ass because once you're done with the system you're likely to have it rented out from under you.
>>103542209im imagining it now what
>>103539913I'm in love.
>>103542539>keep the hard drive space you used reserved unless you destroy it but it's kind of a pain in the ass because once you're done with the system you're likely to have it rented out from under you.So basically only useful if you reuse the same system, so useless lol.What's the minimum renting time?
>>103542631I don't think there is one.
>>103542631Also there's bandwidth costs, keep an eye on that, each host chooses how much they bill per TB, some have it free, some have ludicrous amounts like 30 bucks a TB.
>>103539859Great work, Anon.
>>103540847how do you get it to be so dark?
>>103542707>>103542717ok thanks!
>>103542744vpred most likely
>>103542587I liked the other one better, even if she looks a bit bucc'd. The makeup and fish lips are too much for me. Usually my negative prompt takes care of that, but it seems to have slipped though.
>>103541756I'm with Swarm for a year at this point. I think it's the most noob friendly UI out there.
Tried the fast video hunyuan, doesn't feel like it follows the prompts as well,that's the main first impression. Like way worse. Also more likely to look AI like if that makes sense.
>>103543012very nice
Hunyuan wrapper repo has a sort of image prompting workflow now. It's not img2video, but supposedly will capture style and concepts using VLM. I guess I'll pull and give it a shot...
>>103543057Thanks!
>>103543154can i get a catbox for this one pls?
>>103543177It's just a generic 1girl. What produces this effect is (((high contrast, black palette, chiaroscuro, dark)))
>>103543196more interested in model/settings but ty anyway i assume some version of n00b
>>103543208It's just 1.1 epsilon
Working with hunyuang for anime is like working with an Alzheimer's patient. Most of the time it barely remembers your prompt but every once in a while is has a moment of lucidity.
>>103543275Vid not related, meant to post this one.
I shouldn't even try getting into video gen with 12 gb vram, right?
>>103543315Yeah just kill yourself instead
>>103543315it works just takes a bit longer at lower res
>>103543012>noobI hate this word so much it's unreal. Every time I read it my mental voice reads it in the thickest south east asian or indian accent.
The IP2V workflow actually does a pretty good job. Here is the image I used as input, with the following prompt:<image>. A Middle-Eastern young woman is sitting in shallow water surrounded by lilies. She she appears to be deep in thought.
>>103543441but where's the video
>>103543441And here is the output
>>103543441how much vram? It says minimum 20GB
>>103543458AI is still too retarded to understand hands
>>103543441>>103543458Butiful baby I luv u
>>103543441>>103543458it's just using a vision model to get the prompt out of your pictures, I can do that by asking GPT4V to get the prompts, and it'll be more accurate than that lol
>>103543040>Tried the fast video hunyuan, doesn't feel like it follows the prompts as well,that's the main first impression. Like way worse.yeah same, the quality is fine but if it doesn't want to listen to what I want I don't see the point, it's already hard enough to get it to understand your prompt with the vanilla model lol
>>103543489My total vram went up to 18.5 during encoding. I don't really know shit about shit but I would think you could use a quant for llava to reduce that.>>103543530According to the PR author it should be superior to using a VLM to describe the image because it is passing the embeddings directly to the model. Maybe I'll try it sometime. How do you talk to GPT4 for free these days? I didn't buy a 3090 to use OAI shit but if they have the best image captions I can hold my nose.
>>103543583>How do you talk to GPT4 for free these days?the usual? on chatgpt.com
https://civitai.com/models/1038199/nsfw-hunyuan-lora?modelVersionId=1164548wtf? they removed the hunyuan nfsw lora? why?
>>103543628Feds know how powerful this tech is and they will do anything to stop it from getting in "the wrong hands"
>>103543643they haven't banned the hunyuan model on civitai though, that's the "too much powerful" tech already
>>103543648Soon, they're still working out the logistics
>>103543628wont civitai put a little notice on it if they take it down? the page 404ing makes me think the author did it desu
>>103543628>>103543643I wonder if that's why we haven't seen any celebrity jeetslop loras yet
>>103543724there are https://civitai.com/models/1035770/hunyuan-video-bogged-lora?modelVersionId=1166218https://civitai.com/models/1034630/hunyuan-video-arnold-schwarzenegger-lora?modelVersionId=1160430
>>103543730i want hot FEMALE celeb jeetslop
>>103543730Bogs are a meme and Arnold is kino. I'm talking, reedem buzz 4 emma watson lora saaar
>>103543628author removed it
>>103543628https://civitai.com/user/CubeyAIlooks like the author is still on civitai>>103543747how do you know that?
>>103543746>emma watsonThis perpetual ugly smug face being the crush of a generation is a disaster for lewd making.
What flow shift is recommended for 50 steps+ on hunyuan ?1?
>>103543599>the usualDon't sass your betters. A localchad has little reason to know which model the trash over at OAI is hosting as chatgpt these days. Anyway, here is what I got using a description written by chatgpt. I won't say it didn't do a decent job, but it is pretty badly slopped. Which pretty much sums up anything OAI.
>>103543776>Don't sass your betters.you are too retarded to know where chatgpt could be and you call yourself a better? lmaooooooo
>>103543788I think you mean "too wealthy and intelligent to be concerned about where and what slop the poorfags are consuming." Anyway I appreciate your idea but it looks like llava works better after all.
>>103543829>too wealthy and intelligent to be concerned about where and what slop the poorfags are consuming.>>>103543583>How do you talk to GPT4 for free these days? I didn't buy a 3090 to use OAI shit but if they have the best image captions I can hold my nose.bruh... anyways, it was interesting to see the comparison though, nothing beats the true i2v process that's for sure, I hope we'll get that quickly
>>103543583>18.5 during encodingprobably not gonna work on my 12GB card then, yeah it does mention on the git that something isn't supported yet and well i'm sure they work on getting it to work for lower end cards in the next few days. Fingers crossed because I'd really like to use some sort of image2vid for this model.
why no one tried to change the code so that it forces the model to get the image on the first frame during the denoising? I know it's not a model specialized for i2v, but we could try and see what it can do?
>>103543583I tried the pull yesterday and it's just meh. Like does it work? Yeah, but it's not performing magic or anything. it's just putting the LLM's best description of the image into the prompt.
>>103543891Nice try, Furk
>>103543530its actually doing something similar to ipadapter to influence the style, probably the reason for high vram. Anyway things are looking cool for us if this is what they are achieving. >>103543628lol, that lora barely did anything imo. The output always looks cartoon.
>>103543910>lol, that lora barely did anything imo. The output always looks cartoon.not at all, I managed to get some actual porn with it, you had to remove "hentai" and all those "camera lens, f1/8" tokens nonsense though
>>103543902>See problem on github>Look at the thread>See furk>He's just asking random users for random things and contributing nothing to the solution.It's like he exists to consume.
>>103543776try gemini, honestly i was impressed with its ability to give a decent prompt that results in a decent video. Just feed it the template and ask it to make whatever you want.
>>103543891Make it so
>>103543929>try geminithis, and you can go for their latest 2.0 flash version for free herehttps://aistudio.google.com/prompts/new_chatand it's completly uncensored, it's an API you can change the cucking if you want but by default it's on "based" mode
I do not understand this part about the image as a prompt thing in hyvid.There is no option on any of the notice for vision language nor is there anything in config.json.What does this step even mean?
>>103543898I don't know, I'm pretty impressed. My expectations were pretty low, though. It isn't going to replicate your image, but it does a decent job of steering and is able to convey certain concepts that I haven't had a lot of luck getting it to understand with text alone. It doesn't replace a prompt but it definitely augments it.
>>103543984>It doesn't replace a prompt but it definitely augments it.For sure. It has its uses and it's a powerful too to get what might be lost in just describing a subject, but it's not an ip adapter like they want you to think.
>>103543981I don't understand what he's saying there either. Use the workflow json in custom_nodes\ComfyUI-HunyuanVideoWrapper\examples\ip2v
>>103544020idk, it seems to work fine without touching this step. I just wish I knew what they meant by it.
So fast hunvid works on six steps?
>>103544111yeah I went for 10 steps and the quality seems ok, the issue is with the prompt understanding, it doesn't listen to your prompts really well compared to before, which is a deal breaker for me
>>103543917It's fine to ask questions but it annoys me he just pops in and start asking questions.People should learn ignoring him.
>>103543949what's with the red warning sign
>>103544125Hyvid at its best kind of sucks at prompts too desu. I'll fiddle around a bit and see how big a deal breaker it is.
>>103544172it just says that it's an offensive answer, but they won't prevent it from happening so it's a nothingburger >>103544173>Hyvid at its best kind of sucks at prompts too desu.by "at its best" you mean the API version? because if you say local it's far from its best, it's using the wrong encoder for the moment
>>103544071catbox? thats amazing
>>103544184>it just says that it's an offensive answer, but they won't prevent it from happening so it's a nothingburgeroh ok
>>103544195when you go to the filter section you'll get a disclaimer that you're the only guy responsible of anything happening with the llm, which should be the normal thing, I wanted something like that for so long, we're fucking adults we're responsible and the consequences are for us only
>>103544185>catbox? thats amazinghttps://files.catbox.moe/4jfph4.webmironically hunyuan is *too* realistic, where genning stuff like that is difficult just because the girl isn't hot enough and only slightly pretty or just average
>>103544218oh that's nice, though won't stop google from brainwashing the base model, at least no moderation backed forced on like chatgpt
>>103544237what's with the awful looking teeth
>>103544248prompted for braces lol
>>103544169I made a simple feature request one time and after the dev implemented it Furukan appeared and asked me to give an example of how to use the feature and when it might be useful so he could make a tutorial. Like he couldn't even bother to read the issue where I explained exactly why it would be useful. I just ignored him.>>103544125I'm basically spamming variations of 1girl and I can't say I've noticed a major loss in prompt adherence. But again, my prompts are nothing too complex. It's very hit-or-miss, but so was the base model. I'm going to fuck around and try it with the IP2vid workflow.
>>103544184I've yet to see a direct comparison of API vs local on how the text encoder is affecting output. I understand the quality will probably be better on the MMLM that hyvid made, but I don't have much reason to believe it will be the black and white different we've seemingly hyped ourselves into expecting.
>>103544267>I've yet to see a direct comparison of API vs local on how the text encoder is affecting output.there was one a few threads ago, he made the comparisons with the exact same settings (seed, resolution, steps, number of framesn prompt... ) the quality was superior on the API, like it could actually render text well and shit
Can Hunyuan do bouncing boobs yet?
>>103544290>yet?It always could?
>>103544258>Furukan appearedHow the fuck does he manage to appear everywhere like that, which means to a degree he follows all the developments, and yet still be this awfully bad and rude.>I just ignored him.That's the best policy.
>>103544290maybe you're thinking of mochi
>>103544267no one has hyped anything into anything, it was like two anons circlejerking. if llava wasn't 95% good enough to use as an MLLM they wouldn't have released hunyuan at all>>103544290yeah its good at that
>>103544290>Can Hunyuan do bouncing boobs yet?they literally made a bouncing boobs demo on their site kekhttps://aivideo.hunyuan.tencent.com/
>>103544317>if llava wasn't 95% good enough to use as an MLLM they wouldn't have released hunyuan at allwhy they released the ducktape and not the real one though, if it's """95%"""" as good, why would it hurt them to release the one that's 100% as good
>>103544237>the girl isn't hot enough and only slightly pretty or just averageIsn't that what a LoRA is for
>>103544332Their website has mochi running with open pose controlnets, i2v and other amazing stuff. Why haven't they released that too?
>>103544062>>103544071lol the fed is back... I don't even have to click to know, the thumbnail is so obvious.
>>103544343>i2voh really? didn't know they had the i2v version on the API, desu I don't care about mochi anymore, it's worse than hunyuan and more censored
>>103544358>i2v version on the APIThey don't. I'm talking about their demo reels on their promotional site.
>>103544358>>103544365Also I don't know why I said Mochi, I meant hyvid. I don't give a shit about mochi. All my homies hate mochi.
>>103544237its just too young maybe, if she was older looking it could pass like the typical boutine bikini model
Testing my bog LoRA with the new fast model and I'm not sure it's applying.
>>103544347Drown him out with your vids m8
Yeah can anyone else confirm? These guys aren't bogged, they're just Chinese.
>>103544389I don't think it is but that result still looks pretty good. Would you mind sharing your catbox? My tests with the fast model ended up looking incoherent.
>>103544305He really is omnipresent. I told my wife about the whole situation and she didn't really get it but her reaction was hilarious. She thought it was so damned cool that I was talking directly with a dev and an internet "celebrity".
>>103544240That's because AI Studio is Google's "developer playground". Their equivalent of ChatGPT, Gemini Chat (former Bard), has about the same level of cuckening.While you can get a free API key there with some daily prompts, probably enough for videogen, I'd suggest to try Mistral. Their chatbot allows to create "agents" i.e. custom prompts for free and their free API has very generous limits so you could hook it up to Comfy with both image input and text input-output. And for Google you gotta have a burner because they log and screen their llms since this month and promise to revoke access (haven't heard of this happening yet). Mistral doesn't care.Btw is there a node that explicitly supports MistralAI API?See >>103525707 (on le chat).
>>103544343im assuming its because they want to make money off of their model for the chinese audiences (like how flux serves their dev model on API too)>>103544380if you're talking about, the braces, no its just a training data thing. when the model is more confident placing braces in a mouth it can do them wellif you're talking about the attractiveness of the girl, i think its more to do with the captioning of the data especially for young girls. "hot" definitely is the word that has the most association with the "hot highschooler" look, "beautiful" is almost worthless in my testing
>>103544390Nah i'm done with this place, its time to start working on serious projects and fix my life. I've learn enough, now its time to apply.
Okay anon
>>103544418I see. Anything front for normies/journalists is basically maximally SAFE AND CONSENSUAL or something like that.I don't mind that.And thanks, I'll try Mistral, hopefully it's not too dumb.
>>103544468D-do you happen to live in Turkey?
How do I apply both depth and canny controlnet to flux at the same time?
>>103544389>I'm not sure it's applying.it's a completly different model because it's distilled in another way, for example loras of flux dev don't work on flux schnell
>Look into thread after his last drunken fit>See the newsI now understand why he drowns himself in booze and yells at me saying I'm nothing unprompted. it's pure projection on his part.The difference between me and him is that I don't place my self worth on this hobby. I have fun for a bit and keep it moving because I have other hobbies and passions.I guess it hurts him because he's not in this for fun.
>>103544565wtf? who are you?
>>103544588Nobody but I have a full suite in a drunken loser that dreams of being a lactating transsexual cat girl
pov: you are a taco>its time to start working on serious projectsthats the natural progression of this stuff unless you just burn out forever. im excited to go back to mochi with a purpose in a couple of weeks during the holiday break and not feel like i'm wasting time and money if i'm not innovating with every prompt/gen with hunyuan>and fix my lifeoh well this explains why you project about me so much, hope things get better for you anon <3
>>103544602why do you think anyone gives a shit about your unrelated drama? if you want to argue with mentally ill people, /sdg/ was made for that
>>103544620I'm a thread founder so cope.
>>103544626me in the back giving you the finger
>complaining about mentally ill peopleyou seem to be forgetting where you are
>>103543284This is pretty nice, actually
>>103544651I'm not above rubbing the cigarette in said sperg's face
>>103544665what is your opinion on python?
>>103544565post something new
>>103544696post something first>>103544689meh
>>103539933Can they distill it one more time for the sake of science. I want to see what it'll look like.
drama queen bunk. whatever.
https://xcancel.com/AI_Jasad/status/1868853363150143666#mDamn, Google is cooking, they are starting to slowly catch up, far are the days they were laughed at when they were releasing Bard
>>103544742at some point it should become crab
>>103544760I'm surprised sora allowed that, looks a bit lewd
>>103544760I don't care about SaaS
>>103544785neither does the general public. that's what makes it hilarious
>>103544760meaningless test, but SORA is bad, and hunyuan fine
>>103544790desu I like that, the less the normies are knowledgable about the video model stuff, the better
>>103544780I'm so glad there are so many competitors to Sora, I fully expected them to be the only one in town for years.
>>103544801>I fully expected them to be the only one in town for years.same, I have a question though, why is sora the only model that's not allowed in the EU? Even Pika labs is making fun of them for that lolhttps://xcancel.com/pika_labs/status/1867641187898995179#m
>>103544813>OpenAI: "Please you have to control and censor every AI model makers, for the safety of the world, and not because we want a monopoly or something...">EU: "All right, we'll do that, only for your Sora model though"dare I say based?
>>103544813>why is sora the only model that's not allowed in the EUIt's not banned, it's just not released, probably gdpr issues as always.
>>103544831yeah I know there's a reason, but so far only Sora got this treatment, even though it's the most cucked and boring video model ever, which is quite ironic
>>103544813it has nothing to do with being allowed or not, it's like saying wallmarts aren't allowed in the EUthere is zero ruling about that, they're just not there by the decision of the company itself
>>103544847>they're just not there by the decision of the company itselfabsolutely nothttps://www.euronews.com/next/2024/12/10/openai-releases-ai-video-creator-sora-but-it-wont-be-coming-to-europe-yet>OpenAI said the product would not be available in the European Economic Area, Switzerland, or the United Kingdom yet, possibly due to regulations, which has been the case for OpenAI’s other products, which reached the European market at a later date.>“We're going to try our hardest to be able to launch there,” OpenAI CEO Sam Altman said on Monday.
>>103544839Someone in the legal team at OAI freaked out and told internally that they need more time to release in the EU (and the UK, and Switzerland).It's often the case.
>>103544300>>103544317>>103544321I feel like Hunyuan sucks ass at bouncing boobs. All of my attempts to generate bouncing boobs have gone pretty miserably.Show me your best bouncing boobs gens.Hard mode: No treadmills. It's not "good at bouncing boobs" if it can only make them bounce in one hyper-specific scenario.
>>103544865so the video on the right is hunyuan right? why is it so stuttery? did you put 24 fps on the node?
>>103544858yes, the company decided to not release in the EU/UK/CHit's their own decision, there is no court in Switzerland who told them not to release
>>103544865try same aspect ratio, and select a good result
>>103544885there has to be a reason on why they aren't releasing their products on certain countries and not on others, it has to do with the EU being more into AI regulations than the US for example, if the EU was as chill as the US it would've been released here already
>>103541796If you look closely you can see the demolition charges.
>>103544488help
>>103544894yes, it's the legal freaking out like >>103544864 wrote, many such cases, but note that didn't stop others from releasing thereit's temporary anyway
>>103543427Bloody nooby bastard fuck.
ComfyUI is finally starting to implement support for Hunyuan
>>103540674how fast is stable-diffusion.cpp compared to normal comfyui and python shit?
>>103544935???
>>103544916>the legal freaking outto be fair, every single eyes are onto OpenAI, they are the most scrutanized AI company in the world, and desu they deserve it, they wanted this authoritarian AI regulations in the first place, at least OpenAI is the big tree that hides the forest, when Pika labs is making a release, there hasn't any waves, even though they managed to make a process that'll make deepfake as easy as it gets kek
https://files.catbox.moe/xw6l5n.mp4Someone asked for my workflow for the hyvid fast model. Here you go. Still think it looks like deep fried shit so I'll be switching back to slow soon
>>103544945he looks like the guy who slap people on Squid Game https://www.youtube.com/watch?v=X34pVRkbni8
>>103544942https://github.com/comfyanonymous/ComfyUI/commit/bda1482a272e36315d39ef84663073f88d0a0367
>>103544935>ComfyUI is finally starting to implement support for Hunyuanwhat's the point? kijai's node works well, and when I was trying Comfy's implementation on Mochi, his vae decoder sucked, always had OOM, even on automatic tilted mode
>>103544944>they wanted this authoritarian AI regulations in the first placethe most schizophrenic industry ever>when Pika labs is making a releaseman their videos are fucking smooth
>>103544944you are wise in your ways, chang
>>103544911you can't.you gen sdxl from controlnets first then refine with flux.
>>103544892>select a good resultThat's the problem. I can't make any good results, because Hunyuan sucks at making boobs bounce.
>>103544936slower if you have an nvidia card
I’m starting to worry good anime gens are going to be fundamentally impossible.
Why haven't you fags invented looping gifs yet?That's such an obvious step needed for porn.
>>103545039give me your prompt I'll try it once I ended my current thing
>>103545052>Why haven't you fags invented looping gifs yet?you need the i2v model for that, so that you can put the same image on the first and last frame
>>103545061having guide frames like what ani was doing seems key
>>103545052desu its annoying
>>103545039Why do the gens on the left move like ass?
>>103545060>An 18-year-old alluring beautiful cute enticing fit gorgeous hot pretty seductive sexy young Korean gravure idol with light-pink hair and an ample bust is running on a beach while wearing a bikini with a plunging neckline.>She has a narrow waist. She is skinny. She is slender. She is slim. She has a thin waist.>Each step is causing her enormous bust to bounce, jiggle, and wobble gratuitously.>She is making eye contact with the camera. She is running towards the camera.>Filmed at normal speed, played at normal speed.>black bikini, micro bikini, mini bikini, shiny bikini, string bikini, tiny bikini
>>103545096Check out this coomer.
>>103545096holy shit... I don't know if I should be impressed or disgusted kek
>>103545096I guess you tried, but this doesn't explain why it looks like it runs at low fpsI'll try it somewhere tomorrow when all my current gens are done
>>103545096
>>103545165Shoulda used the bog lora.
>>103545165would
>>103545165Lookin good grandma
>>103545113My "Pink haired asian woman running on beach in bikini" folder has 474 items in it, but it's not even the biggest folder.
>>103545096your prompt is bad. simply writing "breasts bouncing" would get you breasts bouncing so you overengineered your prompt into garbageyou're probably coming from hailuo (i recognize you) so you're used to having to censor yourself. just write what you want aka "large cleavage bouncing"start from first principles. get your close up of cleavage bouncing then get a medium shot of the asian gyaru then finally make her running
so this...is the power...of hunyuan...whoa...
This this enough jiggle?
>>103545210Post example.
>>103545222I like her jig, she just oozes confidence.
>>103545212its real power is unlocked at its true resolution (1280x720) but we don't have the gpu for that, I hope the 5090 will be enough
>>103545236https://www.youtube.com/watch?v=2oHV-ZKocEk
>>103545047what the flying fuck is the point then
>>103545283>what the flying fuck is the point thenthere's none, autists are gonna autism
>>103545298A better native UI that's not built on gradio would be pretty nice. Something that actually uses the file system and organizes your work with a undo/redo buffer. Some simple drawing tools with layers, etc. You could still use a python backend for all I care.
>>103540674>[EXTPROCESS] Can not load shared library: F:\StableDiff\StableDiffusionGUI 0.2.3\stable-diffusion_cuda.dllSend help plox
I think there's something wrong with the latest hyvid wrapper pull. These are supposed to be bogs.
And this is supposed to be furk. I think something got fucked. The LoRA weights either aren't applying or are too weak to notice.
>>103545344>>103545350oh, maybe that's also why the loras weren't working on FastHunyuan?
>>103545344>>103545350thats the bad thing about Kijais wrappers. hes always pushing commits with little testing
>>103544935>ComfyUI is finally starting to implement support for HunyuanOk? What would be the reason to go for that one and not stick to kijai's node?
>>103545364Well that's what was bugging me. If the LoRAs straight up didn't work with fast hyvid, it would give a key error or something. This just looks like weight s aren't applying or are extremely weak.After looking at this gen, I think they are applying, but something is off.
>>103545344>>103545350>>103545381>the master of Bogs pulledEven the bests make this mistake kek
I think the most impressive part of HunyuanVideo is its VAE, you can go for a slow ass resolution of 540x320, and the humans look great even at a far away distance, not even Flux comes even close to that, I feel like image models are cheaping out on the VAE, I'd much rather prefer a high quality VAE that uses a lot of vram (we can use the tilted method to get away with this) rather than something cheaper but destroys the details
>>103545556you're getting better at this anon, cool
Okay I just tested an old seed and it seems like the LoRAs are not broken. I guess my prompt either really sucked or the subject matter of scientists strongly overpowered bogs and furk.
>>103545664>Okay I just tested an old seed and it seems like the LoRAs are not broken.that's why I'm always keeping some old kino renders, becauses sometimes you can get such a bad seed you start to believe there's a bug in the code
>>103545676We may never know. I do like the offering LoRAs have for Hyvid, but from what I've seen you can't be too ambitious with them. They work better at inserting characters than going to extremes. LoRAs based on 3D work I've done looked pretty good though and transferred the style while still extrapolating animation. Animation of the 2D variety seems to be a bit of a hit and miss affair though. I haven't seen any good aggressive 2D LoRAs for hyvid yet.
>>103545567Different anon just testing that prompt. Simplified it quite a bit>A beautiful young Korean gravure idol with light-pink hair and an ample bust is running on a beach while wearing a bikini with a plunging neckline. Her large breasts are bouncing and wobbling with each step
>>103545772But I was told that there was no jiggle.
>>103545664Domp eet
Hibernation mode
>>103544760Sorry for doubting you Google, you invented the transformers architecture after all
noobAI vpred models (using a cyberfix merge) are something else, the lineart/shading/color is so nice.
>>103546082>tfw google can make better anime slop than pony/noob finetuners :(
>>103546131>ask google gemini to make a japanese samurai>get african man
>>103546147Yasuke was a samurai you bigot!!1!1!1!
>>103546164he was literally a circus act that nobunaga paraded around
>>103544760https://xcancel.com/bilawalsidhu/status/1868873130791649584#m>it can do pepeIs it me or since Trump got president again, all companies are based now?
>>103546186even elon's grok generator does pepes, we're entering a new age of rare pepe media.
>>103546212>even elon's grok generator does pepesI mean, that's obvious for elon's model, he's a right winger, we're talking about Google there, one of the most woke companies ever
>>103546223they might just be saving face after the gemini humiliation, showing black emperors of Japan.
>>103546186>Early-accessStill time to neuter it
>>103546186https://xcancel.com/1littlecoder/status/1868847655927116231#mwtf, this model is mogging the competition so hard, how did Google become so good??
the ship crashed but she's okay.
>>103546243wtf how did they train it on Sekiro, just game footage?
>>103545556Not bad. I actually want to see the catbox, or even just the prompt.
>>103546240>Still time to neuter ityeah, this shit can render Star Wars out of the box, they probably went Hunyuan mode and trained their model with *cough* *public available data* of course
>>103546164I must have missed this. Was there actually an internal leak that said this?
This fucking sucks. Why does everything of value sit in some crusty datacenter out of the hands of the public?
>>103546287I'm gonna miss the monstruosities the old models did because at some point those models will all be so consistently good you won't have those funny glitches anymore
>>103546331because it's too powerful and dangerous for you goy
>>103546331To be fair, it's already a miracle we got locally a model as good as Hunyuan, I expected something like that in 5 years, not now
>>103546345Cool
>>103546335I don't get why they went for 8 sec, why not 10? Would be the perfect length to display an idea
>>103546268>just game footageyeah, something like that
>>103546372ah damn that one was kinda rad
>>103546372>Can't render Will SmithDOA
>>103546387whats wild is the video actually follows game rules/physics, the character is actually running around a 3d map. idk how they do it, I know Nvidia's dlss3 framegen stuff can predict frames with info to make "fake" frames that look real, and boost framerates without artifacting.
>>103546404
what python/comfyui versions do i need to get hunyuanvideo wrapper set up? it fails to import into cui and it blames lacking diffusers custom node as a dependency, which then i think i found, but then it requires omegaconf to import which i have no clue what to do about
>>103546459bruh
>>1035464633.11 and 3.12 work, you can use this guidehttps://rentry.org/crhcqq54
>>103544760>Google is cookingI mean, google owns Youtube, they have infinite video data at their disposal, of course they can make their model good with what they own
If it doesn't run on my consumer GPU I don't want to fucking hear about it! Get it out!
>>103546509this
>>103546372>yayyy, now that we got Hunyuan, we got closer to the SOTA mode-ACKthis is depressing, everytime we are catching up to them, they go further away
https://videocardz.com/newz/zotac-confirms-geforce-rtx-5090-with-32gb-gddr7-memory-5080-and-5070-series-listed-as-well>the 5060 will have 6gb of vramin the year 2024 of our lord? is this a joke? lol
>>103546587>600W...
>>103544976>>103545371This is what he does. Next he will make "improvements" that just happen to be breaking changes for Kijai's node.
>>103546587Please intel, save us from this shit.
There is a new cogmodelxfun that has been released for a week and no one posted about it here
>>103546587That says 8gb which is fair enough for a shit cheap budget card. Still not having a reasonable 24gb card is a joke though
>>103546459>>103546404>>103546372>>103546287This is a local diffusion thread for local diffusion people, there is nothing for you here!
>>103546478tysm anon
>>103546641why should we care about it? hunyuan exists now
>>103546752>>103546752new>>103546752>>103546752
>>103546587>600wunironically need a 1200+ PSU unless you are using an efficient 7800x3D or something