[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip / qa] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/g/ - Technology


Thread archived.
You cannot reply anymore.


[Advertise on 4chan]


File: HunyuanVideo_00350.webm (472 KB, 960x544)
472 KB
472 KB WEBM
Discussion of Free and Open-Source Diffusion models.

Previous: >>103533408

>UI
Metastable: https://metastable.studio
SwarmUI: https://github.com/mcmonkeyprojects/SwarmUI
Forge: https://github.com/lllyasviel/stable-diffusion-webui-forge
reForge: https://github.com/Panchovix/stable-diffusion-webui-reForge
ComfyUI: https://github.com/comfyanonymous/ComfyUI
InvokeAI: https://github.com/invoke-ai/InvokeAI

>Models, LoRAs, & Upscalers
https://civitai.com
https://tensor.art/
https://openmodeldb.info

>Training
https://github.com/Nerogar/OneTrainer
https://github.com/kohya-ss/sd-scripts/tree/sd3
https://github.com/derrian-distro/LoRA_Easy_Training_Scripts

>HunyuanVideo
Comfy: https://github.com/kijai/ComfyUI-HunyuanVideoWrapper/
Windows: https://rentry.org/crhcqq54
Training: https://github.com/tdrussell/diffusion-pipe

>Flux
Forge Guide: https://github.com/lllyasviel/stable-diffusion-webui-forge/discussions/1050
ComfyUI Guide: https://comfyanonymous.github.io/ComfyUI_examples/flux
DeDistilled Quants: https://huggingface.co/TheYuriLover/flux-dev-de-distill-GGUF/tree/main

>Misc
Share Metadata: https://catbox.moe | https://litterbox.catbox.moe/
Generate Prompt from Image: https://huggingface.co/spaces/fancyfeast/joy-caption-alpha-two
Archived: https://rentry.org/sdg-link
Samplers: https://stable-diffusion-art.com/samplers/
Open-Source Digital Art Software: https://krita.org/en/
Txt2Img Plugin: https://kritaaidiffusion.com/
Collagebaker: https://www.befunky.com/create/collage/
Video Collagebaker: https://kdenlive.org/en/

>Neighbo(u)rs
>>>/aco/sdg
>>>/aco/aivg
>>>/b/degen
>>>/c/kdg
>>>/d/ddg
>>>/e/edg
>>>/h/hdg
>>>/tg/slop
>>>/trash/sdg
>>>/u/udg
>>>/vt/vtai

>Texting Neighbo(u)r
>>>/g/lmg
>>
File: 00037-4112077858.jpg (1.16 MB, 1600x1280)
1.16 MB
1.16 MB JPG
>>
File: HunyuanVideo_00013.webm (737 KB, 400x720)
737 KB
737 KB WEBM
Can't say I love how these gens turned out on either model and I probably could come up with a better test prompt, but here is my baseline using the regular model. CFG 2, Guidance 8, Flow Shift 8, 30 steps. Will follow with FastVideo checkpoint in a moment.
>>
>>103539859
This is good.
>>
File: HunyuanVideo_00012.webm (374 KB, 400x720)
374 KB
374 KB WEBM
>>103539913
And here is the one using the Fast Video model. Same params including seed, but with 17 flow shift and 10 steps. I actually prefer this one, though the lighting isn't behaving.
>>
Unless there's a different workflow for Fast Hyvid there seems to be weird flickering lighting artifacts and general crust
>>
>>103539933
what's with the light? it's flickering
>>
>>103539956
That's how candles do innit
>>
>>103539956
The prompt says it's lit by a single candle so I wouldn't read too much into that. Somebody give me a sfw prompt that isn't weird and I'll do one more test.
>>
>>103539913
>>103539933
>CFG 2
?? so are you using negative prompt as well
i like your 1girl she is very cute can you post prompt/catbox for this

what is the speed difference on fastvideo vs normal hunyuan 10 step 17 flow shift? is it the same?
>>
>>103539972
Could you post your workflow, please sir
>>
Blessed thread of frenship
>>
File: Google Veo 2.mp4 (1.41 MB, 1280x720)
1.41 MB
1.41 MB MP4
>BTFOs Chinese models
https://x.com/GoogleDeepMind/status/1868703624714395907/video/1
>>
does this fast shit mean I can run it on my njudea 3060 or
>>
>>103540073
>censored, online only, will bankrupt you to use
>>
>>103540096
It's quite janky and low quality so it likely needs to have some extra custom node work done to make it generate better outputs. For now you might as well just do 12 steps on the normal model and accept some blur.
>>
>>103540073
>no weights
kek, worthless
>>
>>103540073
this is what GTA 8 will look like
>>
>>103539859
>>103539913
Fantastic
>>
>>103540119
This is how most games will look in 10 years when they can get AI frame gen to work at 30fps.
>>
>>103540125
pfff, the ps5 pro an do it now
>>
File: LongClip.jpg (1.02 MB, 1920x1632)
1.02 MB
1.02 MB JPG
https://www.reddit.com/r/StableDiffusion/comments/1hfnnjg/a_comfyui_node_for_hunyuanvideo_that_lets_you/

Does this work or is just a placebo?
>>
>>103540172
Schizo placebo
>>
>>103540073
>not local
why should we care? at least pika 2.0 adds something new to the table that should be replicated >>103533581
>>
>>103539976
>>103539985
Yes, I do use a negative prompt. Iterations per second are identical between models. Excuse the schizo negative; I'm experimenting. Here is the catbox.
https://files.catbox.moe/8abeag.webm
>>
>>103540073
https://www.youtube.com/watch?v=nlIhROafItU
>>
>>103540225
Thanks
>>
wtf I missed real Princess Slaying Hours
>>
File: HunyuanVideo_00043.mp4 (497 KB, 640x480)
497 KB
497 KB MP4
egg
>>
>https://github.com/fszontagh/sd.cpp.gui.wx
Gui written with c++
>>
File: HunyuanVideo_00034.mp4 (478 KB, 960x544)
478 KB
478 KB MP4
>>103540506
It's always princess slaying hours.
>>
>>103540674
Ani on suicide watch
>>
>>103540674
perfect time, SD has never been more popular
>>
>>103540704
top kek
>>
>>103540674
>GUI
>no screenshots
Me no andastand.
>>
>>103540766
>no andastand
https://youtu.be/RoSs9-NDP3E?t=9
>>
>>103540694
>>
>>103540704
it does flux too, in a a1111 interface that justwurx
>>
>>103540783
Found a few on the linked site and it looks so uncomfy. Still, an actual program, cool.
>>
File: tmpdcvd35e0.png (603 KB, 1248x896)
603 KB
603 KB PNG
>>
what's the secret prompts to add to make HY output better quality videos?
>>
>>103540847
Why did you posting the png and not the mp4, sir?
>>
>>103540870
>what's the secret prompts to add to make HY output better quality videos?
there's none, it's already a miracle it works this well with the wrong text encoder, to get improvements we need to get the official one, HunyuanMLLM
>>
>>103540890
vramlet
>>
>>103540806
can you take some screenshots of the gui
>>
>>103540847
Interesting pose and angle
>>
File: 00015-420696969.png (1.45 MB, 999x776)
1.45 MB
1.45 MB PNG
>>
>>103540674
>>103540694

HAHAHAHAHAHAHA
>>
>>103540694
>>103540792
>Julien publicly announces some project
>doesn't deliver
>gets done by someone else
>will act as if be deserves respect and credit
How many times will this keep happening?
>>
>>103540971
at least trani is still the best animator of the univers- oh wait i forgot
>>
>>103540985
>TFW his only feat is blowing comfy for a job at a dying company
>>
>>103540674
okay share some screenshots or something I wanna know how good it is
>>
>>103541039
Even if it's shit that's more work put in than the fraud trani
>>
Oh nonononono
Trani sisters
https://private-user-images.githubusercontent.com/51741446/393540826-a0c02eff-11da-487e-a09b-9448d18006f8.gif
>>
>>103541084
https://github.com/fszontagh/sd.cpp.gui.wx/issues/23
>>
>>103540225
>I do use a negative prompt.
Is it really useful?
>>
>>103540225
how faster is it?
>>
>>103540073
another day, another imgtovideo but not available locally
>>
>>103540914
Very nice, catbox?
>>
>>103541084
>>103541105

looks like some bittorrent client
>>
>>103541271
sovl
>>
>>103540674
Janky as hell, needs polish.
>>
>>103541105
looks fine I guess, but why the hassle of doing that
>>
>>103541105
>Crated at
>>
>>103541294
Still more than anything trani has ever done
>>
Has anyone attempted to make any img2video hacked comfy flow for hunyuan? Even if the output is not 100% identical to the original frame, but following the exact same composition
In theory it should be possible to transform a static image into video, then doing something to it to simulate motion, then using video2video
or just using another local model that allows img2video and using the output from it on hunyuan vid2vid (would probably be garbage, but who knows)
>>
>>103539859
This is honestly incredible.
>>
File: tmp2v5_ya62.png (1.04 MB, 1248x768)
1.04 MB
1.04 MB PNG
>>
>>103541320
kinda
https://github.com/kijai/ComfyUI-HunyuanVideoWrapper/tree/main/examples/ip2v
>>
>>103540870
Amazing video, cinematic... Stuff like that>>103540890
>>
>>103541320
It can kind of work that use generate the frames using one still image, but motion? Nope... It just isn't trained to make videos from one image, the frames need to be real video with motion.
>>
>>103541444
the example is pretty cool actually considering it was just one image? So erm, when can we have this?
>>
>>103541505
nvm, i just checked the git
https://github.com/kijai/ComfyUI-HunyuanVideoWrapper/tree/main
I'm guessing now, this pleases me.
>>
>>103540847
moody, I like
>>
hello /ldg/, quick question, not sure to ask here or /sdg/ . I have a project idea in mind that would involve creating many simple 2d assets in a consistent style, like small images of cartoon furniture in an isometric perspective. I don't have a great graphics card to do anything locally and I don't mind spending money to get decent quality. what would my best options be?

pic rel a misato plushie I made in imagen
>>
>>103541572
You could rent compute from a cloud service if you don't want to buy a new computer.
>>
File: 00021-1205026481.jpg (287 KB, 1440x810)
287 KB
287 KB JPG
>>
>>103541572
I use vast.ai to rent computers with gpus, like 30-40 cents an hour for a 4090.
>>
File: tmp7vve6gjo.png (861 KB, 768x1280)
861 KB
861 KB PNG
>>
>>103541429
having horns like this would be horrible. your skull could be split if some would pull on it because of the leverage effect.
>>
File: 1732552800411861.jpg (274 KB, 1145x592)
274 KB
274 KB JPG
>>103539836
>Metastable
>SwarmUI
>Forge
>reForge
>InvokeAI
I'm assuming ComfyUI still isn't comfy so which one should I use? I used to use A1111 but it seems like it's not recommended anymore?
>>
>>103541756
easiest way to use Hunyuan at the moment
>>
File: HunyuanVideo_00055.mp4 (1.44 MB, 640x480)
1.44 MB
1.44 MB MP4
>>
>>103541756
>>103541766
"easiest way to use Hunyuan at the moment is comfyUI, but for other models I think Forge is good enough"
>>
>>103540674
it's shit. looks like shit and needlessly abstracted

>>103541766
their gradio app demo technically is
>>
>>103541707
>>103541613
Interesting, and a 4090 is sufficient to run the latest models? How does the workflow work like you have to install software I'm assuming, so then that takes time out of your reservation or you can load an image or something? Should I get experience creating images locally first to get a hang of it even though I just have a 2060 super rn
>>
>>103541796
now have it running through a street before devouring someone, make it realistic, i want to see their terror as this thing runs through the street eating people.
>>
>
>>
>>103541886
can't wait for the ani UI that doesn't look like shit and is not needlessly abstracted
>>
>>103541221
I find it useful, but doubling gen time is definitely a hefty cost. Some prompts need it more than others.
>>103541242
Same speed per step, but fewer steps required. My impression is that 12 steps get you similar polish to 30 steps on the base model, so it's a pretty significant speedup. I don't have a strong enough grasp on either model's capabilities to evaluate how steep the loss in quality is, but it doesn't really feel worse.
>>
>>103542003
was the flickering light just a thing with onle one video?
>>
>>103541105
No preview for generation? I guess it works but man does that seem slow and not that much faster than using Pytorch.
>>
>>103542003
>12 steps get you similar polish to 30 steps
thanks anon, interesting
>>
>>103542019
Yes. My prompt called for candlelight so I think that's all it was.
>>
>>103541913
A 4090 can do pretty much any image stuff. Video stuff can get a bit tight.
>>
>>103541994
How can he work when his anus leaks stinky fluid that betrays him?
>>
>>103542039
I also want to stress that I'm just naïvely using the flow shift value I've seen recommended for 10 steps on the base model, but from what little I understand there isn't any particular reason why it should be the same. It does seem to work, but I'm sure it is worth experimenting with different stuff. The guy who created it seems to have just copy/pasted the model card from the base model and gives no suggestions for any of the parameters.
>>
>>103542111
i too hate it when my big fat juicy asshole betrays me
>>
>>103542131
More like blown out and loose after years of abuse, Imagine the sleeve of a wizard drenched in brown smelly aids fluid.
>>
What's the difference between Forge and Reforge? Which should I use?
>>
>>103542129
>I also want to stress that I'm just naïvely using the flow shift value
I think that's pretty much everyone here.
What I don't get is why wouldn't we let it be at high number (for example 20) regardless of the steps since at low steps it helps and in high steps it seems to not have any bad effect.
>>
>>103542098
>Video stuff can get a bit tight.
I bet
>>
>julien is seething
>>
>>103541756
ComfyUI is comfy in the way that my ex gf was not a lying whore
>>
>ani is crying
>>
>>103542044
oh cool!
>>
>>103541913
>Should I get experience creating images locally first to get a hang of it even though I just have a 2060 super rn
you can do quite a bit with a 2060 - definitely see what you can do with that before thinking about renting
>>
>>103540674
But where all all my comfy custom nodes? :(
>>
>>103541707
How does that work?
I can see the way we get the cards, but what about the os? Can I install whatever linux I want and comfy there ?
>>
>>103542418
The company you rent from spins up a virtual machine for you running what ever OS you tell them
>>
>>103542431
I see.
Then what happens when I stop using the thing, should I basically reinstall everything every time I need compute?
I'm coming from the VPS world so it's kind of confusing to me.
>>
>>103542442
pretty much, yeah. you can publish your own docker image if you wanna save time though since container startup time doesnt count
>>
>>103542485
OK thanks anon.
Maybe worth my time testing what an A100 80GB can do.
>>
>>103542442
I wrote a couple sh scripts to install and download everything I'm using. Most of the machines will have gigabit up and down so after standing up the image I run the script and grab a drink and it'll be done when I'm back. You also will keep the hard drive space you used reserved unless you destroy it but it's kind of a pain in the ass because once you're done with the system you're likely to have it rented out from under you.
>>
>>103542209
im imagining it now what
>>
>>103539913
I'm in love.
>>
>>103542539
>keep the hard drive space you used reserved unless you destroy it but it's kind of a pain in the ass because once you're done with the system you're likely to have it rented out from under you.
So basically only useful if you reuse the same system, so useless lol.
What's the minimum renting time?
>>
>>103542631
I don't think there is one.
>>
>>103542631
Also there's bandwidth costs, keep an eye on that, each host chooses how much they bill per TB, some have it free, some have ludicrous amounts like 30 bucks a TB.
>>
>>103539859
Great work, Anon.
>>
>>103540847
how do you get it to be so dark?
>>
>>103542707
>>103542717
ok thanks!
>>
>>103542744
vpred most likely
>>
>>103542587
I liked the other one better, even if she looks a bit bucc'd. The makeup and fish lips are too much for me. Usually my negative prompt takes care of that, but it seems to have slipped though.
>>
File: 1705711260949419.jpg (1.61 MB, 1824x1248)
1.61 MB
1.61 MB JPG
>>103541756
I'm with Swarm for a year at this point. I think it's the most noob friendly UI out there.
>>
Tried the fast video hunyuan, doesn't feel like it follows the prompts as well,that's the main first impression. Like way worse. Also more likely to look AI like if that makes sense.
>>
>>103543012
very nice
>>
Hunyuan wrapper repo has a sort of image prompting workflow now. It's not img2video, but supposedly will capture style and concepts using VLM. I guess I'll pull and give it a shot...
>>
File: 1723845242964820.jpg (826 KB, 1024x1024)
826 KB
826 KB JPG
>>103543057
Thanks!
>>
File: 1731531532344599.jpg (1.14 MB, 1824x1248)
1.14 MB
1.14 MB JPG
>>
>>103543154
can i get a catbox for this one pls?
>>
File: 1734062088123684.jpg (937 KB, 1824x1248)
937 KB
937 KB JPG
>>103543177
It's just a generic 1girl. What produces this effect is (((high contrast, black palette, chiaroscuro, dark)))
>>
>>103543196
more interested in model/settings but ty anyway i assume some version of n00b
>>
File: 1727932127228102.jpg (1.57 MB, 1824x1248)
1.57 MB
1.57 MB JPG
>>103543208
It's just 1.1 epsilon
>>
File: HunyuanVideo_00059.mp4 (517 KB, 960x544)
517 KB
517 KB MP4
Working with hunyuang for anime is like working with an Alzheimer's patient. Most of the time it barely remembers your prompt but every once in a while is has a moment of lucidity.
>>
File: HunyuanVideo_00060.mp4 (703 KB, 960x544)
703 KB
703 KB MP4
>>103543275
Vid not related, meant to post this one.
>>
I shouldn't even try getting into video gen with 12 gb vram, right?
>>
>>103543315
Yeah just kill yourself instead
>>
>>103543315
it works just takes a bit longer at lower res
>>
>>103543012
>noob
I hate this word so much it's unreal. Every time I read it my mental voice reads it in the thickest south east asian or indian accent.
>>
File: 2218-flux1-dev-2116063110.png (1.85 MB, 1024x1024)
1.85 MB
1.85 MB PNG
The IP2V workflow actually does a pretty good job. Here is the image I used as input, with the following prompt:
<image>. A Middle-Eastern young woman is sitting in shallow water surrounded by lilies. She she appears to be deep in thought.
>>
>>103543441
but where's the video
>>
File: HunyuanVideo_00209.mp4 (473 KB, 720x480)
473 KB
473 KB MP4
>>103543441
And here is the output
>>
>>103543441
how much vram? It says minimum 20GB
>>
>>103543458
AI is still too retarded to understand hands
>>
>>103543441
>>103543458
Butiful baby I luv u
>>
>>103543441
>>103543458
it's just using a vision model to get the prompt out of your pictures, I can do that by asking GPT4V to get the prompts, and it'll be more accurate than that lol
>>
>>103543040
>Tried the fast video hunyuan, doesn't feel like it follows the prompts as well,that's the main first impression. Like way worse.
yeah same, the quality is fine but if it doesn't want to listen to what I want I don't see the point, it's already hard enough to get it to understand your prompt with the vanilla model lol
>>
>>103543489
My total vram went up to 18.5 during encoding. I don't really know shit about shit but I would think you could use a quant for llava to reduce that.
>>103543530
According to the PR author it should be superior to using a VLM to describe the image because it is passing the embeddings directly to the model. Maybe I'll try it sometime. How do you talk to GPT4 for free these days? I didn't buy a 3090 to use OAI shit but if they have the best image captions I can hold my nose.
>>
>>103543583
>How do you talk to GPT4 for free these days?
the usual? on chatgpt.com
>>
https://civitai.com/models/1038199/nsfw-hunyuan-lora?modelVersionId=1164548
wtf? they removed the hunyuan nfsw lora? why?
>>
>>103543628
Feds know how powerful this tech is and they will do anything to stop it from getting in "the wrong hands"
>>
>>103543643
they haven't banned the hunyuan model on civitai though, that's the "too much powerful" tech already
>>
>>103543648
Soon, they're still working out the logistics
>>
>>103543628
wont civitai put a little notice on it if they take it down? the page 404ing makes me think the author did it desu
>>
>>103543628
>>103543643
I wonder if that's why we haven't seen any celebrity jeetslop loras yet
>>
>>103543724
there are
https://civitai.com/models/1035770/hunyuan-video-bogged-lora?modelVersionId=1166218
https://civitai.com/models/1034630/hunyuan-video-arnold-schwarzenegger-lora?modelVersionId=1160430
>>
>>103543730
i want hot FEMALE celeb jeetslop
>>
>>103543730
Bogs are a meme and Arnold is kino.
I'm talking, reedem buzz 4 emma watson lora saaar
>>
>>103543628
author removed it
>>
>>103543628
https://civitai.com/user/CubeyAI
looks like the author is still on civitai
>>103543747
how do you know that?
>>
>>103543746
>emma watson
This perpetual ugly smug face being the crush of a generation is a disaster for lewd making.
>>
What flow shift is recommended for 50 steps+ on hunyuan ?
1?
>>
File: HunyuanVideo_000014.webm (593 KB, 720x480)
593 KB
593 KB WEBM
>>103543599
>the usual
Don't sass your betters. A localchad has little reason to know which model the trash over at OAI is hosting as chatgpt these days. Anyway, here is what I got using a description written by chatgpt. I won't say it didn't do a decent job, but it is pretty badly slopped. Which pretty much sums up anything OAI.
>>
>>103543776
>Don't sass your betters.
you are too retarded to know where chatgpt could be and you call yourself a better? lmaooooooo
>>
>>103543788
I think you mean "too wealthy and intelligent to be concerned about where and what slop the poorfags are consuming." Anyway I appreciate your idea but it looks like llava works better after all.
>>
>>103543829
>too wealthy and intelligent to be concerned about where and what slop the poorfags are consuming.
>>>103543583
>How do you talk to GPT4 for free these days? I didn't buy a 3090 to use OAI shit but if they have the best image captions I can hold my nose.
bruh... anyways, it was interesting to see the comparison though, nothing beats the true i2v process that's for sure, I hope we'll get that quickly
>>
>>103543583
>18.5 during encoding
probably not gonna work on my 12GB card then, yeah it does mention on the git that something isn't supported yet and well i'm sure they work on getting it to work for lower end cards in the next few days. Fingers crossed because I'd really like to use some sort of image2vid for this model.
>>
why no one tried to change the code so that it forces the model to get the image on the first frame during the denoising? I know it's not a model specialized for i2v, but we could try and see what it can do?
>>
>>103543583
I tried the pull yesterday and it's just meh. Like does it work? Yeah, but it's not performing magic or anything. it's just putting the LLM's best description of the image into the prompt.
>>
>>103543891
Nice try, Furk
>>
>>103543530
its actually doing something similar to ipadapter to influence the style, probably the reason for high vram. Anyway things are looking cool for us if this is what they are achieving.

>>103543628
lol, that lora barely did anything imo. The output always looks cartoon.
>>
>>103543910
>lol, that lora barely did anything imo. The output always looks cartoon.
not at all, I managed to get some actual porn with it, you had to remove "hentai" and all those "camera lens, f1/8" tokens nonsense though
>>
>>103543902
>See problem on github
>Look at the thread
>See furk
>He's just asking random users for random things and contributing nothing to the solution.
It's like he exists to consume.
>>
>>103543776
try gemini, honestly i was impressed with its ability to give a decent prompt that results in a decent video. Just feed it the template and ask it to make whatever you want.
>>
>>103543891
Make it so
>>
File: 1705538525234934.png (142 KB, 2475x1057)
142 KB
142 KB PNG
>>103543929
>try gemini
this, and you can go for their latest 2.0 flash version for free here
https://aistudio.google.com/prompts/new_chat
and it's completly uncensored, it's an API you can change the cucking if you want but by default it's on "based" mode
>>
File: Untitled.png (9 KB, 710x308)
9 KB
9 KB PNG
I do not understand this part about the image as a prompt thing in hyvid.
There is no option on any of the notice for vision language nor is there anything in config.json.
What does this step even mean?
>>
>>103543898
I don't know, I'm pretty impressed. My expectations were pretty low, though. It isn't going to replicate your image, but it does a decent job of steering and is able to convey certain concepts that I haven't had a lot of luck getting it to understand with text alone. It doesn't replace a prompt but it definitely augments it.
>>
>>103543984
>It doesn't replace a prompt but it definitely augments it.
For sure. It has its uses and it's a powerful too to get what might be lost in just describing a subject, but it's not an ip adapter like they want you to think.
>>
>>103543981
I don't understand what he's saying there either. Use the workflow json in custom_nodes\ComfyUI-HunyuanVideoWrapper\examples\ip2v
>>
File: HunyuanVideo_00601.webm (934 KB, 960x544)
934 KB
934 KB WEBM
>>
File: HunyuanVideo_00600.webm (1.15 MB, 960x544)
1.15 MB
1.15 MB WEBM
>>
>>103544020
idk, it seems to work fine without touching this step. I just wish I knew what they meant by it.
>>
So fast hunvid works on six steps?
>>
>>103544111
yeah I went for 10 steps and the quality seems ok, the issue is with the prompt understanding, it doesn't listen to your prompts really well compared to before, which is a deal breaker for me
>>
>>103543917
It's fine to ask questions but it annoys me he just pops in and start asking questions.
People should learn ignoring him.
>>
>>103543949
what's with the red warning sign
>>
>>103544125
Hyvid at its best kind of sucks at prompts too desu. I'll fiddle around a bit and see how big a deal breaker it is.
>>
>>103544172
it just says that it's an offensive answer, but they won't prevent it from happening so it's a nothingburger
>>103544173
>Hyvid at its best kind of sucks at prompts too desu.
by "at its best" you mean the API version? because if you say local it's far from its best, it's using the wrong encoder for the moment
>>
>>103544071
catbox? thats amazing
>>
>>103544184
>it just says that it's an offensive answer, but they won't prevent it from happening so it's a nothingburger
oh ok
>>
File: 1703334449199706.png (113 KB, 1086x1404)
113 KB
113 KB PNG
>>103544195
when you go to the filter section you'll get a disclaimer that you're the only guy responsible of anything happening with the llm, which should be the normal thing, I wanted something like that for so long, we're fucking adults we're responsible and the consequences are for us only
>>
File: HunyuanVideo_00602.webm (744 KB, 960x544)
744 KB
744 KB WEBM
>>103544185
>catbox? thats amazing
https://files.catbox.moe/4jfph4.webm
ironically hunyuan is *too* realistic, where genning stuff like that is difficult just because the girl isn't hot enough and only slightly pretty or just average
>>
>>103544218
oh that's nice, though won't stop google from brainwashing the base model, at least no moderation backed forced on like chatgpt
>>
>>103544237
what's with the awful looking teeth
>>
>>103544248
prompted for braces lol
>>
>>103544169
I made a simple feature request one time and after the dev implemented it Furukan appeared and asked me to give an example of how to use the feature and when it might be useful so he could make a tutorial. Like he couldn't even bother to read the issue where I explained exactly why it would be useful. I just ignored him.

>>103544125
I'm basically spamming variations of 1girl and I can't say I've noticed a major loss in prompt adherence. But again, my prompts are nothing too complex. It's very hit-or-miss, but so was the base model. I'm going to fuck around and try it with the IP2vid workflow.
>>
>>103544184
I've yet to see a direct comparison of API vs local on how the text encoder is affecting output. I understand the quality will probably be better on the MMLM that hyvid made, but I don't have much reason to believe it will be the black and white different we've seemingly hyped ourselves into expecting.
>>
>>103544267
>I've yet to see a direct comparison of API vs local on how the text encoder is affecting output.
there was one a few threads ago, he made the comparisons with the exact same settings (seed, resolution, steps, number of framesn prompt... ) the quality was superior on the API, like it could actually render text well and shit
>>
Can Hunyuan do bouncing boobs yet?
>>
>>103544290
>yet?
It always could?
>>
>>103544258
>Furukan appeared
How the fuck does he manage to appear everywhere like that, which means to a degree he follows all the developments, and yet still be this awfully bad and rude.

>I just ignored him.
That's the best policy.
>>
>>103544290
maybe you're thinking of mochi
>>
>>103544267
no one has hyped anything into anything, it was like two anons circlejerking. if llava wasn't 95% good enough to use as an MLLM they wouldn't have released hunyuan at all

>>103544290
yeah its good at that
>>
File: 1721272103151622.mp4 (650 KB, 1248x704)
650 KB
650 KB MP4
>>103544290
>Can Hunyuan do bouncing boobs yet?
they literally made a bouncing boobs demo on their site kek
https://aivideo.hunyuan.tencent.com/
>>
>>103544317
>if llava wasn't 95% good enough to use as an MLLM they wouldn't have released hunyuan at all
why they released the ducktape and not the real one though, if it's """95%"""" as good, why would it hurt them to release the one that's 100% as good
>>
>>103544237
>the girl isn't hot enough and only slightly pretty or just average
Isn't that what a LoRA is for
>>
>>103544332
Their website has mochi running with open pose controlnets, i2v and other amazing stuff. Why haven't they released that too?
>>
>>103544062
>>103544071
lol the fed is back... I don't even have to click to know, the thumbnail is so obvious.
>>
File: 1716543101493600.png (2.45 MB, 1484x1005)
2.45 MB
2.45 MB PNG
>>103544343
>i2v
oh really? didn't know they had the i2v version on the API, desu I don't care about mochi anymore, it's worse than hunyuan and more censored
>>
>>103544358
>i2v version on the API
They don't. I'm talking about their demo reels on their promotional site.
>>
>>103544358
>>103544365
Also I don't know why I said Mochi, I meant hyvid. I don't give a shit about mochi. All my homies hate mochi.
>>
>>103544237
its just too young maybe, if she was older looking it could pass like the typical boutine bikini model
>>
File: HunyuanVideo_00439.mp4 (699 KB, 960x544)
699 KB
699 KB MP4
Testing my bog LoRA with the new fast model and I'm not sure it's applying.
>>
>>103544347
Drown him out with your vids m8
>>
File: HunyuanVideo_00440.mp4 (491 KB, 960x544)
491 KB
491 KB MP4
Yeah can anyone else confirm? These guys aren't bogged, they're just Chinese.
>>
>>103544389
I don't think it is but that result still looks pretty good. Would you mind sharing your catbox? My tests with the fast model ended up looking incoherent.
>>
>>103544305
He really is omnipresent. I told my wife about the whole situation and she didn't really get it but her reaction was hilarious. She thought it was so damned cool that I was talking directly with a dev and an internet "celebrity".
>>
>>103544240
That's because AI Studio is Google's "developer playground". Their equivalent of ChatGPT, Gemini Chat (former Bard), has about the same level of cuckening.
While you can get a free API key there with some daily prompts, probably enough for videogen, I'd suggest to try Mistral. Their chatbot allows to create "agents" i.e. custom prompts for free and their free API has very generous limits so you could hook it up to Comfy with both image input and text input-output. And for Google you gotta have a burner because they log and screen their llms since this month and promise to revoke access (haven't heard of this happening yet). Mistral doesn't care.
Btw is there a node that explicitly supports MistralAI API?
See >>103525707 (on le chat).
>>
File: co0wt0.webm (776 KB, 960x544)
776 KB
776 KB WEBM
>>103544343
im assuming its because they want to make money off of their model for the chinese audiences (like how flux serves their dev model on API too)

>>103544380
if you're talking about, the braces, no its just a training data thing. when the model is more confident placing braces in a mouth it can do them well
if you're talking about the attractiveness of the girl, i think its more to do with the captioning of the data especially for young girls. "hot" definitely is the word that has the most association with the "hot highschooler" look, "beautiful" is almost worthless in my testing
>>
>>103544390
Nah i'm done with this place, its time to start working on serious projects and fix my life. I've learn enough, now its time to apply.
>>
Okay anon
>>
>>103544418
I see. Anything front for normies/journalists is basically maximally SAFE AND CONSENSUAL or something like that.
I don't mind that.
And thanks, I'll try Mistral, hopefully it's not too dumb.
>>
>>103544468
D-do you happen to live in Turkey?
>>
How do I apply both depth and canny controlnet to flux at the same time?
>>
>>103544389
>I'm not sure it's applying.
it's a completly different model because it's distilled in another way, for example loras of flux dev don't work on flux schnell
>>
File: CatJak_00430.png (2.37 MB, 832x1280)
2.37 MB
2.37 MB PNG
>Look into thread after his last drunken fit
>See the news
I now understand why he drowns himself in booze and yells at me saying I'm nothing unprompted. it's pure projection on his part.
The difference between me and him is that I don't place my self worth on this hobby. I have fun for a bit and keep it moving because I have other hobbies and passions.
I guess it hurts him because he's not in this for fun.
>>
>>103544565
wtf? who are you?
>>
File: CatJak_00403.png (967 KB, 1024x832)
967 KB
967 KB PNG
>>103544588
Nobody but I have a full suite in a drunken loser that dreams of being a lactating transsexual cat girl
>>
File: HunyuanVideo_00612.webm (639 KB, 960x544)
639 KB
639 KB WEBM
pov: you are a taco

>its time to start working on serious projects
thats the natural progression of this stuff unless you just burn out forever. im excited to go back to mochi with a purpose in a couple of weeks during the holiday break and not feel like i'm wasting time and money if i'm not innovating with every prompt/gen with hunyuan

>and fix my life
oh well this explains why you project about me so much, hope things get better for you anon <3
>>
>>103544602
why do you think anyone gives a shit about your unrelated drama? if you want to argue with mentally ill people, /sdg/ was made for that
>>
File: CatJak_00402.png (963 KB, 1024x832)
963 KB
963 KB PNG
>>103544620
I'm a thread founder so cope.
>>
>>103544626
me in the back giving you the finger
>>
File: HunyuanVideo_00613.webm (773 KB, 960x544)
773 KB
773 KB WEBM
>complaining about mentally ill people
you seem to be forgetting where you are
>>
>>103543284
This is pretty nice, actually
>>
File: 0.jpg (166 KB, 896x1152)
166 KB
166 KB JPG
>>
File: CatJak_00404.png (950 KB, 1024x832)
950 KB
950 KB PNG
>>103544651
I'm not above rubbing the cigarette in said sperg's face
>>
>>103544665
what is your opinion on python?
>>
>>103544565
post something new
>>
>>103544696
post something first
>>103544689
meh
>>
>>103539933
Can they distill it one more time for the sake of science. I want to see what it'll look like.
>>
drama queen bunk. whatever.
>>
https://xcancel.com/AI_Jasad/status/1868853363150143666#m
Damn, Google is cooking, they are starting to slowly catch up, far are the days they were laughed at when they were releasing Bard
>>
>>103544742
at some point it should become crab
>>
>>103544760
I'm surprised sora allowed that, looks a bit lewd
>>
>>103544760
I don't care about SaaS
>>
>>103544785
neither does the general public. that's what makes it hilarious
>>
>>103544760
meaningless test, but SORA is bad, and hunyuan fine
>>
>>103544790
desu I like that, the less the normies are knowledgable about the video model stuff, the better
>>
>>103544780
I'm so glad there are so many competitors to Sora, I fully expected them to be the only one in town for years.
>>
>>103544801
>I fully expected them to be the only one in town for years.
same, I have a question though, why is sora the only model that's not allowed in the EU? Even Pika labs is making fun of them for that lol
https://xcancel.com/pika_labs/status/1867641187898995179#m
>>
>>103544813
>OpenAI: "Please you have to control and censor every AI model makers, for the safety of the world, and not because we want a monopoly or something..."
>EU: "All right, we'll do that, only for your Sora model though"
dare I say based?
>>
>>103544813
>why is sora the only model that's not allowed in the EU
It's not banned, it's just not released, probably gdpr issues as always.
>>
>>103544831
yeah I know there's a reason, but so far only Sora got this treatment, even though it's the most cucked and boring video model ever, which is quite ironic
>>
>>
>>103544813
it has nothing to do with being allowed or not, it's like saying wallmarts aren't allowed in the EU
there is zero ruling about that, they're just not there by the decision of the company itself
>>
>>103544847
>they're just not there by the decision of the company itself
absolutely not
https://www.euronews.com/next/2024/12/10/openai-releases-ai-video-creator-sora-but-it-wont-be-coming-to-europe-yet
>OpenAI said the product would not be available in the European Economic Area, Switzerland, or the United Kingdom yet, possibly due to regulations, which has been the case for OpenAI’s other products, which reached the European market at a later date.
>“We're going to try our hardest to be able to launch there,” OpenAI CEO Sam Altman said on Monday.
>>
>>103544839
Someone in the legal team at OAI freaked out and told internally that they need more time to release in the EU (and the UK, and Switzerland).
It's often the case.
>>
>>103544300
>>103544317
>>103544321
I feel like Hunyuan sucks ass at bouncing boobs. All of my attempts to generate bouncing boobs have gone pretty miserably.
Show me your best bouncing boobs gens.
Hard mode: No treadmills. It's not "good at bouncing boobs" if it can only make them bounce in one hyper-specific scenario.
>>
>>103544865
so the video on the right is hunyuan right? why is it so stuttery? did you put 24 fps on the node?
>>
>>103544858
yes, the company decided to not release in the EU/UK/CH
it's their own decision, there is no court in Switzerland who told them not to release
>>
>>103544865
try same aspect ratio, and select a good result
>>
>>103544885
there has to be a reason on why they aren't releasing their products on certain countries and not on others, it has to do with the EU being more into AI regulations than the US for example, if the EU was as chill as the US it would've been released here already
>>
File: 007675.png (3.55 MB, 1224x2144)
3.55 MB
3.55 MB PNG
>>
>>103541796
If you look closely you can see the demolition charges.
>>
>>103544488
help
>>
>>103544894
yes, it's the legal freaking out like >>103544864 wrote, many such cases, but note that didn't stop others from releasing there
it's temporary anyway
>>
>>103543427
Bloody nooby bastard fuck.
>>
ComfyUI is finally starting to implement support for Hunyuan
>>
>>103540674
how fast is stable-diffusion.cpp compared to normal comfyui and python shit?
>>
>>103544935
???
>>
File: 1704465981793840.mp4 (1.75 MB, 1080x720)
1.75 MB
1.75 MB MP4
>>103544916
>the legal freaking out
to be fair, every single eyes are onto OpenAI, they are the most scrutanized AI company in the world, and desu they deserve it, they wanted this authoritarian AI regulations in the first place, at least OpenAI is the big tree that hides the forest, when Pika labs is making a release, there hasn't any waves, even though they managed to make a process that'll make deepfake as easy as it gets kek
>>
File: HunyuanVideo_00443.mp4 (560 KB, 960x544)
560 KB
560 KB MP4
https://files.catbox.moe/xw6l5n.mp4

Someone asked for my workflow for the hyvid fast model. Here you go. Still think it looks like deep fried shit so I'll be switching back to slow soon
>>
>>103544945
he looks like the guy who slap people on Squid Game
https://www.youtube.com/watch?v=X34pVRkbni8
>>
>>103544942
https://github.com/comfyanonymous/ComfyUI/commit/bda1482a272e36315d39ef84663073f88d0a0367
>>
>>103544935
>ComfyUI is finally starting to implement support for Hunyuan
what's the point? kijai's node works well, and when I was trying Comfy's implementation on Mochi, his vae decoder sucked, always had OOM, even on automatic tilted mode
>>
>>103544944
>they wanted this authoritarian AI regulations in the first place
the most schizophrenic industry ever

>when Pika labs is making a release
man their videos are fucking smooth
>>
>>103544944
you are wise in your ways, chang
>>
>>103544911
you can't.
you gen sdxl from controlnets first then refine with flux.
>>
>>103544892
>select a good result
That's the problem. I can't make any good results, because Hunyuan sucks at making boobs bounce.
>>
>>103544936
slower if you have an nvidia card
>>
I’m starting to worry good anime gens are going to be fundamentally impossible.
>>
Why haven't you fags invented looping gifs yet?
That's such an obvious step needed for porn.
>>
>>103545039
give me your prompt I'll try it once I ended my current thing
>>
>>103545052
>Why haven't you fags invented looping gifs yet?
you need the i2v model for that, so that you can put the same image on the first and last frame
>>
>>103545061
having guide frames like what ani was doing seems key
>>
>>103545052
desu its annoying
>>
>>103545039
Why do the gens on the left move like ass?
>>
File: image.png (2.37 MB, 2215x1192)
2.37 MB
2.37 MB PNG
>>103545060
>An 18-year-old alluring beautiful cute enticing fit gorgeous hot pretty seductive sexy young Korean gravure idol with light-pink hair and an ample bust is running on a beach while wearing a bikini with a plunging neckline.
>She has a narrow waist. She is skinny. She is slender. She is slim. She has a thin waist.
>Each step is causing her enormous bust to bounce, jiggle, and wobble gratuitously.
>She is making eye contact with the camera. She is running towards the camera.
>Filmed at normal speed, played at normal speed.
>black bikini, micro bikini, mini bikini, shiny bikini, string bikini, tiny bikini
>>
>>103545096
Check out this coomer.
>>
>>103545096
holy shit... I don't know if I should be impressed or disgusted kek
>>
>>103545096
I guess you tried, but this doesn't explain why it looks like it runs at low fps
I'll try it somewhere tomorrow when all my current gens are done
>>
File: HunyuanVideo_00062.mp4 (538 KB, 960x544)
538 KB
538 KB MP4
>>103545096
>>
>>103545165
Shoulda used the bog lora.
>>
>>103545165
would
>>
>>103545165
Lookin good grandma
>>
File deleted.
>>
>>103545113
My "Pink haired asian woman running on beach in bikini" folder has 474 items in it, but it's not even the biggest folder.
>>
File: HunyuanVideo_00630.webm (634 KB, 960x544)
634 KB
634 KB WEBM
>>103545096
your prompt is bad. simply writing "breasts bouncing" would get you breasts bouncing so you overengineered your prompt into garbage
you're probably coming from hailuo (i recognize you) so you're used to having to censor yourself. just write what you want aka "large cleavage bouncing"
start from first principles. get your close up of cleavage bouncing then get a medium shot of the asian gyaru then finally make her running
>>
File: HunyuanVideo_00039.webm (553 KB, 640x352)
553 KB
553 KB WEBM
so this...is the power...of hunyuan...whoa...
>>
File: HunyuanVideo_00063.mp4 (640 KB, 960x544)
640 KB
640 KB MP4
This this enough jiggle?
>>
>>103545210
Post example.
>>
>>103545222
I like her jig, she just oozes confidence.
>>
>>103545212
its real power is unlocked at its true resolution (1280x720) but we don't have the gpu for that, I hope the 5090 will be enough
>>
>>103545236
https://www.youtube.com/watch?v=2oHV-ZKocEk
>>
>>103545047
what the flying fuck is the point then
>>
>>103545283
>what the flying fuck is the point then
there's none, autists are gonna autism
>>
>>103545298
A better native UI that's not built on gradio would be pretty nice. Something that actually uses the file system and organizes your work with a undo/redo buffer. Some simple drawing tools with layers, etc. You could still use a python backend for all I care.
>>
>>103540674
>[EXTPROCESS] Can not load shared library: F:\StableDiff\StableDiffusionGUI 0.2.3\stable-diffusion_cuda.dll
Send help plox
>>
File: HunyuanVideo_00449.mp4 (153 KB, 640x480)
153 KB
153 KB MP4
I think there's something wrong with the latest hyvid wrapper pull. These are supposed to be bogs.
>>
File: HunyuanVideo_00446.mp4 (264 KB, 640x480)
264 KB
264 KB MP4
And this is supposed to be furk.
I think something got fucked. The LoRA weights either aren't applying or are too weak to notice.
>>
>>103545344
>>103545350
oh, maybe that's also why the loras weren't working on FastHunyuan?
>>
>>103545344
>>103545350
thats the bad thing about Kijais wrappers. hes always pushing commits with little testing
>>
>>103544935
>ComfyUI is finally starting to implement support for Hunyuan
Ok? What would be the reason to go for that one and not stick to kijai's node?
>>
File: HunyuanVideo_00450.mp4 (302 KB, 640x480)
302 KB
302 KB MP4
>>103545364
Well that's what was bugging me. If the LoRAs straight up didn't work with fast hyvid, it would give a key error or something. This just looks like weight s aren't applying or are extremely weak.
After looking at this gen, I think they are applying, but something is off.
>>
File: 1723071248558258.png (603 KB, 686x386)
603 KB
603 KB PNG
>>103545344
>>103545350
>>103545381
>the master of Bogs pulled
Even the bests make this mistake kek
>>
File: 007703.png (2.67 MB, 1944x1328)
2.67 MB
2.67 MB PNG
>>
I think the most impressive part of HunyuanVideo is its VAE, you can go for a slow ass resolution of 540x320, and the humans look great even at a far away distance, not even Flux comes even close to that, I feel like image models are cheaping out on the VAE, I'd much rather prefer a high quality VAE that uses a lot of vram (we can use the tilted method to get away with this) rather than something cheaper but destroys the details
>>
File: HunyuanVideo_00003.mp4 (1.15 MB, 960x544)
1.15 MB
1.15 MB MP4
>>
>>103545556
you're getting better at this anon, cool
>>
File: HunyuanVideo_00453.mp4 (139 KB, 640x480)
139 KB
139 KB MP4
Okay I just tested an old seed and it seems like the LoRAs are not broken. I guess my prompt either really sucked or the subject matter of scientists strongly overpowered bogs and furk.
>>
>>103545664
>Okay I just tested an old seed and it seems like the LoRAs are not broken.
that's why I'm always keeping some old kino renders, becauses sometimes you can get such a bad seed you start to believe there's a bug in the code
>>
File: HunyuanVideo_00452.mp4 (399 KB, 640x480)
399 KB
399 KB MP4
>>103545676
We may never know. I do like the offering LoRAs have for Hyvid, but from what I've seen you can't be too ambitious with them. They work better at inserting characters than going to extremes. LoRAs based on 3D work I've done looked pretty good though and transferred the style while still extrapolating animation. Animation of the 2D variety seems to be a bit of a hit and miss affair though. I haven't seen any good aggressive 2D LoRAs for hyvid yet.
>>
File: HunyuanVideo_00005.mp4 (1.35 MB, 960x544)
1.35 MB
1.35 MB MP4
>>103545567
Different anon just testing that prompt. Simplified it quite a bit
>A beautiful young Korean gravure idol with light-pink hair and an ample bust is running on a beach while wearing a bikini with a plunging neckline. Her large breasts are bouncing and wobbling with each step
>>
>>103545772
But I was told that there was no jiggle.
>>
File: 1712350992637095.mp4 (263 KB, 640x400)
263 KB
263 KB MP4
>>103545664
Domp eet
>>
File: HunyuanVideo_00454.mp4 (278 KB, 640x480)
278 KB
278 KB MP4
>>
Hibernation mode
>>
File: 1729671873386190.png (507 KB, 1559x1376)
507 KB
507 KB PNG
>>103544760
Sorry for doubting you Google, you invented the transformers architecture after all
>>
File: 1725896797081781.png (860 KB, 1024x1024)
860 KB
860 KB PNG
noobAI vpred models (using a cyberfix merge) are something else, the lineart/shading/color is so nice.
>>
File: 1727658793556198.png (2.26 MB, 1024x1024)
2.26 MB
2.26 MB PNG
>>103546082
>tfw google can make better anime slop than pony/noob finetuners :(
>>
>>103546131
>ask google gemini to make a japanese samurai
>get african man
>>
File: 1730016846934576.png (241 KB, 460x484)
241 KB
241 KB PNG
>>103546147
Yasuke was a samurai you bigot!!1!1!1!
>>
File: 1724111605375430.webm (1.32 MB, 1280x720)
1.32 MB
1.32 MB WEBM
>>103546164
he was literally a circus act that nobunaga paraded around
>>
File: 1733135862821476.mp4 (2.05 MB, 1280x720)
2.05 MB
2.05 MB MP4
>>103544760
https://xcancel.com/bilawalsidhu/status/1868873130791649584#m
>it can do pepe
Is it me or since Trump got president again, all companies are based now?
>>
>>103546186
even elon's grok generator does pepes, we're entering a new age of rare pepe media.
>>
>>103546212
>even elon's grok generator does pepes
I mean, that's obvious for elon's model, he's a right winger, we're talking about Google there, one of the most woke companies ever
>>
>>103546223
they might just be saving face after the gemini humiliation, showing black emperors of Japan.
>>
>>103546186
>Early-access
Still time to neuter it
>>
File: 1729650599587528.mp4 (1.87 MB, 1280x720)
1.87 MB
1.87 MB MP4
>>103546186
https://xcancel.com/1littlecoder/status/1868847655927116231#m
wtf, this model is mogging the competition so hard, how did Google become so good??
>>
File: 1724586126040604.png (1.37 MB, 1024x1024)
1.37 MB
1.37 MB PNG
the ship crashed but she's okay.
>>
>>103546243
wtf how did they train it on Sekiro, just game footage?
>>
>>103545556
Not bad. I actually want to see the catbox, or even just the prompt.
>>
File: 1721177976814047.mp4 (3.41 MB, 1920x1080)
3.41 MB
3.41 MB MP4
>>103546240
>Still time to neuter it
yeah, this shit can render Star Wars out of the box, they probably went Hunyuan mode and trained their model with *cough* *public available data* of course
>>
File: HunyuanVideo_00458.mp4 (597 KB, 640x480)
597 KB
597 KB MP4
>>
>>103546164
I must have missed this. Was there actually an internal leak that said this?
>>
This fucking sucks. Why does everything of value sit in some crusty datacenter out of the hands of the public?
>>
File: 1719854918214349.mp4 (1.73 MB, 1280x720)
1.73 MB
1.73 MB MP4
>>103546287
I'm gonna miss the monstruosities the old models did because at some point those models will all be so consistently good you won't have those funny glitches anymore
>>
File: 1705794320358936.png (57 KB, 130x174)
57 KB
57 KB PNG
>>103546331
because it's too powerful and dangerous for you goy
>>
File: HunyuanVideo_00459.mp4 (1.35 MB, 640x480)
1.35 MB
1.35 MB MP4
>>
>>103546331
To be fair, it's already a miracle we got locally a model as good as Hunyuan, I expected something like that in 5 years, not now
>>
>>103546345
Cool
>>
File: 1704954041812434.mp4 (1.13 MB, 1280x720)
1.13 MB
1.13 MB MP4
>>103546335
I don't get why they went for 8 sec, why not 10? Would be the perfect length to display an idea
>>
>>103546268
>just game footage
yeah, something like that
>>
>>103546372
ah damn that one was kinda rad
>>
File: 1714566207948580.mp4 (531 KB, 1280x720)
531 KB
531 KB MP4
>>103546372
>Can't render Will Smith
DOA
>>
>>103546387
whats wild is the video actually follows game rules/physics, the character is actually running around a 3d map. idk how they do it, I know Nvidia's dlss3 framegen stuff can predict frames with info to make "fake" frames that look real, and boost framerates without artifacting.
>>
File: spider12.webm (2.43 MB, 1280x720)
2.43 MB
2.43 MB WEBM
>>103546404
>>
what python/comfyui versions do i need to get hunyuanvideo wrapper set up? it fails to import into cui and it blames lacking diffusers custom node as a dependency, which then i think i found, but then it requires omegaconf to import which i have no clue what to do about
>>
File: 1719002465556481.png (724 KB, 851x598)
724 KB
724 KB PNG
>>103546459
bruh
>>
>>103546463
3.11 and 3.12 work, you can use this guide
https://rentry.org/crhcqq54
>>
>>103544760
>Google is cooking
I mean, google owns Youtube, they have infinite video data at their disposal, of course they can make their model good with what they own
>>
If it doesn't run on my consumer GPU I don't want to fucking hear about it! Get it out!
>>
>>103546509
this
>>
>>103546372
>yayyy, now that we got Hunyuan, we got closer to the SOTA mode-ACK
this is depressing, everytime we are catching up to them, they go further away
>>
File: 1726437084862975.png (310 KB, 2649x811)
310 KB
310 KB PNG
https://videocardz.com/newz/zotac-confirms-geforce-rtx-5090-with-32gb-gddr7-memory-5080-and-5070-series-listed-as-well
>the 5060 will have 6gb of vram
in the year 2024 of our lord? is this a joke? lol
>>
>>103546587
>600W
...
>>
>>103544976
>>103545371
This is what he does. Next he will make "improvements" that just happen to be breaking changes for Kijai's node.
>>
>>103546587
Please intel, save us from this shit.
>>
There is a new cogmodelxfun that has been released for a week and no one posted about it here
>>
>>103546587
That says 8gb which is fair enough for a shit cheap budget card.
Still not having a reasonable 24gb card is a joke though
>>
>>103546459
>>103546404
>>103546372
>>103546287
This is a local diffusion thread for local diffusion people, there is nothing for you here!
>>
>>103546478
tysm anon
>>
>>103546641
why should we care about it? hunyuan exists now
>>
>>103546752
>>103546752
new
>>103546752
>>103546752
>>
>>103546587
>600w
unironically need a 1200+ PSU unless you are using an efficient 7800x3D or something



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.