Discussion of Free and Open-Source Diffusion models.Last bread : >>103475488>Local (Hunyuan) VideoWindows: https://rentry.org/crhcqq54>UIMetastable: https://metastable.studioSwarmUI: https://github.com/mcmonkeyprojects/SwarmUIForge: https://github.com/lllyasviel/stable-diffusion-webui-forgereForge: https://github.com/Panchovix/stable-diffusion-webui-reForgeComfyUI: https://github.com/comfyanonymous/ComfyUIInvokeAI: https://github.com/invoke-ai/InvokeAI>Models, LoRAs, & Upscalershttps://civitai.comhttps://tensor.art/https://openmodeldb.info>Cookinghttps://github.com/Nerogar/OneTrainerhttps://github.com/kohya-ss/sd-scripts/tree/sd3https://github.com/derrian-distro/LoRA_Easy_Training_Scripts>FluxForge Guide: https://github.com/lllyasviel/stable-diffusion-webui-forge/discussions/1050ComfyUI Guide: https://comfyanonymous.github.io/ComfyUI_examples/fluxDeDistilled Quants: https://huggingface.co/TheYuriLover/flux-dev-de-distill-GGUF/tree/main>Guides & ToolsShare the Sauce: https://catbox.moePerishable Sauce: https://litterbox.catbox.moe/Generate Prompt from Image: https://huggingface.co/spaces/fancyfeast/joy-caption-alpha-twoArtifact resources: https://rentry.org/sdg-linkSamplers: https://stable-diffusion-art.com/samplers/Open-Source Digital Art Software: https://krita.org/en/Txt2Img Plugin: https://kritaaidiffusion.com/Collagebaker: https://www.befunky.com/create/collage/Video Collagebaker: https://kdenlive.org/en/>Neighbo(u)rs>>>/aco/sdg>>>/aco/aivg>>>/b/degen>>>/c/kdg>>>/d/ddg>>>/e/edg>>>/h/hdg>>>/tg/slop>>>/trash/sdg>>>/u/udg>>>/vt/vtai>Texting Neighbo(u)r>>>/g/lmg
>>103482892based moving op
its so over that we're so back
>>103482892So it's not about collages anymore? Just one single video now?
Blessed thread of frenship
You make next collage
>>103482916We should prepare a ffmpeg script that takes a few files as input and makes the collageI don't expect people to bother with video editing software just for this
>>103482916The collage is ass because it's just the baker's fetish / masturbation.
>>103482916just like in real life there is only one winner, 2nd and 3rd and 4th, etc. are 2nd looser, 3rd looser, etc.
>>103482958Also I couldn't help but notice that they've been selecting images/videos that could get the thread nuked, so it's at best low key trolling.
>>103482930that is assuming all videos are the same aspect ratio
>>103482916>its only been exactly 1 week since videogen began taking over the general>it took less time than that for the OP's to start getting ((weird))jesus this week really flew by i didn't even noticegoing to fire up sd in a moment and catch up cause i missed this shit so much
>>103482977get chatgpt to do a fancy recursive algorithm and resize everything before combining it all
>>103482971For example...?
>>103483086I trust your adult brain
>surely noobai won't understand my schizo prompt>it does
>>103483211Hellsing fan i see, managed to get a good render of Zorin or The Captain ?
what causes reforge to do this? im on a new install, has done this in the past (and half why i kept uninstalling then reinstalling later)
>month 23 of LDG >thread falls off catalog multiple times>anon still complains about the collages >anon still has yet to bake his own collage
>>103483281>month 23 of LDGldg isn't that old lol
>>103483281Big tears buddy, it will get easier.
>>103483243Funnily enough, i had absolutely no idea to connect this with Hellsing in any way, it's my attempt to do dumb OCs again.
Congrats moralfags. We get one video of a dog now. This is the OP you wanted.
>>103483322Yes it's better than the ones that deserve you getting a perma.
>>103483298itd be funny if he started crying like that other anon
>>103483064Resize based on the smallest aspect ratio? How do you mean?
if you want godlike hunyuan quality put this in front of your prompts. Thank me later:shot on a Canon EOS R7, 85mm lens, f/2, sharp focus, movie scene.
>>103483379>shot on a Canon EOS R7, 85mm lens, f/2, sharp focus, movie scenethanks anon, I guess that only works for realistic stuff, if you have the same magic prompt for anime it would be cool
>>103483379>shot on a Canon EOS R7, 85mm lens, f/2placebo. models don't genuinely interpret those kinds of prompts. post a comparison.
When will the jannies just learn to sit in this thread and ban the pedo for ban evasion?
>>103483388>two asians kissing being one of the videos in a collage will get the thread nuked Those anons must be trolling
>>103483427>gore>nudity>pedo videosThe guy you've been replying too has been perma'd multiple times. He's the one who gets literally all his posts nuked daily.
>>103483211>>103483303god damn these are good, what noob model are you using?
>>103483447Huh, you're right. How about that kek
>>103483422Wait trani is here? Ew!
>>103483447thats a shame, pedanon seemed based enough but if hes getting nuked and some of those posts are obviously trolling then i can't see him in the same light anymore. rip.
>anon is still mad his slop never made it into the collage
>>103483483You can thank him when they make it mandatory to verify an email before posting.
>>103482917Blessed thread
>>103483495This. The guy has zero opsec—I wouldn't be surprised if he was actually a fed.Funny how flagrant posting of underage girls always goes hand in hand with abusing high cfg. Must both be low IQ traits
>anon grows bored of current SOTAs video outputs >infighting begins It's like a cycle
>>103483530We know the goal is reached when we don't even have to post anymore
>bro you must be bored if you're annoyed about the guy shitting in the corner
>>103483495>>103483520I mean it's kinda obvious most anons who try to fuck up threads are feds, inside job and all that. Already been exposed too many times that they intentionally hire people on for moderation/janitorial service with the intent to follow scripted ops and shit intended to push for more pass ownership.the idea that anything is done for the sake of "advertisers" is hilariously ignorant especially after the fuckups this year alone.Daily reminder for anyone who forgot or didn't know that BBCposting in certain boards always seems to stop at the same time as certain geopolitical events and at one point a fed forgot to practice bare level opsec and got exposed on /pol/ for stoking infighting and false flagging.
i never posted gore and any nudity was always in a catbox you have me mixed up with another genner>>103483520yes i work for the chinese police or CP for shortim really liking guidance 5
>>103483422why do you even care? he is literally the only one contributing to the thread. maybe if there was two more people contributing you could argue against him, until then shut the fuck up.
I don't mind dancing girl videos but people who got to ruin it for everyone else can get the fuck out tbfh. Things like exposed cleavage and butt cheeks hanging out of shorts will only get the thread nuked. I wouldn't be surprised it is don't intentionally to destroy the general, or so the media can have a field day on condemning AI generated art...
>>103483555I just think they're mentally ill, even as a job he would only be here for work shifts. He's in here like 16 hours a day.
can u do a couple redheads pls
>>103483468That's just vpred 0.75. I don't use merges or style loras of any kind because they deeply fuck with image creativity from my experience. And image creativity on par with wild 1.5 models is the best thing about illustrious.
>>103483576>he is literally the only one contributing to the thread. maybe if there was two more people contributingcontributions come in many forms other than just posting videos anonie...
>>103483576That's how I feel desu, if he's posting videos that aren't absolutely obscene let the man cook. Everyone was starting a conversation about the whole RF inversion thing last thread, does anyone have any examples of it? Or at least a breakdown of what may be possible with it?
>>103483576Well I care because what he's doing is permabannable and what he's doing will cause collateral damage. That's of course ignoring that 1) ban evasion and constant rule breaking is pushing toward full verification on 4chan and 2) the content is designed to troll and disrupt, that's why they do it.
>>103483578>He's in here like 16 hours a day.MhSounds familiar somehow
>>103483576Why do you type those avatarfags, it's like you all draw from the same well
>>103483579>redheadsok sure>>103483577>Things like exposed cleavage and butt cheeks hanging out of shorts will only get the thread nukedare you aware of how retarded you sound>>103483576>he is literally the only one contributing to the threadit is kind of awkward that my gens are the only ones desuif there were a way to run the fp8 on gradio i would host something on an A100 again
https://github.com/tdrussell/diffusion-pipe
>>103483586>vpred stockwow, nice. I don't know what the deal is with vpred but i can never tardwrangle it well, even with the right settings, mostly it is just loras and the way merges are done i think.i stick with epsilon since it plays nicely with loras. still never got a definitive idea on how training loras for vpred works.>>103483608yeah its one thing to be a little chaos causer but its kinda obvious by now how they play, its subtle thread ruination until they start getting more people to join in without realizing they're helping make things worse.i mean case in point now this thread is 99% arguing even after we just, yknow, werent arguing kek
>>103483520retard still has no idea how hunyuan video works haven't even tried it, yet he thinks he is superior to someone that does>>103483588he helped me many times in the past week. he is obviously one of the best anons here, contributing in many forms.>>103483608>muh rulesfaggot narc. it's the likes of you that get people b& for saying the nword. he is not doing anything disruptable, if he was then the rules should be applied.>>103483617why don't you have a single argument?I RESPECT AND APPRECIATE PEDANON FOR ALL THE HELP AND CONTRIBUTIONS HE PROVIDED FOR /ldg/
>>103483586>don't use merges or style loras of any kind because they deeply fuck withthis. noob is perfect on it's own (other than the characters it doesn't know) >>103483629>Pipeline parallelism, for training models larger than can fit on a single GPUwoah. parralel inference when?
this is professional trolling>start complaining for no reason about x>people respond to you and say you're retarded>omg look how everyone is complaining now because of x!
>>103483629>HunyuanVideo supports fp8 transformer. The example config file will train a HunyuanVideo LoRA, on images only, in well under 24GB of VRAM. You can probably bump the resolution to 1024x1024 or higher.>on images only
>>103483647That's not what I said so feel free to reread what I wrote and address it specifically. He has been perma'd multiple times for his pedo videos and images, so it's not just "oh maybe it's okay", it's very clearly NOT okay and he knows this. He also very much knows what he's doing is disruptive and he does it anyways, so no, I don't think I'm just going to watch him diddle around in civilized circles idly.Anyways, I'm all for full ID verification to post on 4chan. This site has been ruined by the proliferation of spam evasion.
>>103483670read the next sentence dumb monkey
>>103483665phree bumps desu
>>103483686I read it you dumb nigger, it says that you won't be able to train videos with a 24gb card, or if you do it'll be slow ass resolutions without a lot of frames, DOA
>>103483619>are you aware of how retarded you soundnot when they are underage, but what ever brainless cunt i'm not getting into an argument about it.
>>103483699>it says that you won't be able to train videos with a 24gb cardso stop being poor and buy a second 4090its obvious that's the future of local going into 2027
>>103483636I generally prefer Epsilon, but vpred is getting more consistent lately. Higher contrast across the board though.
>>103483718>so stop being poor and buy a second 4090are you retarded or something, the 5090 will be released in a month
>>103483718>its obvious that's the future of local going into 2027how so? with the way it advances, maybe we'll be able to get the quality of hunyuan with a 2b model, who knows
>>103483718The future is A6000 Adas, Titan AI RTX and 5090s.
>>103483729if you're actually planning on getting it in a month start preparing scripts for buying it unironically>>103483740sure but there will always be a slightly better model with higher parameters
>>103483740VRAM is more or less tied to context length. A long video uses a shit ton of VRAM because the context length at best linearly scales the VRAM requirements per frame.
>>103483745>A6000 Adasalready essentially obsolete>Titan AI RTXno evidence of its existence
>>103483396It doesn't but those tags are correlated with higher quality images so the vectors within HunYuan's model related to quality gets activated.
>>103483579>can u do a couple redheads plsyour slop, sirhttps://files.catbox.moe/vhyjjx.webm
>>103483752>start preparing scriptsnta but I live in east europoor so I don't expect them to be immediately sold out
>>103483724This one is great
>>1034837524090s didn't sell out for like 3 hours after launch, I think the 5090s will last longer because the economy is worse and they'll cost $200-300 more.>>103483766A6000 Adas aren't obsolete, they're 48 GB of VRAM at 4090 speed and 300w. 5090s are housefires at 600w and of course 16 GB of VRAM less which is significant especially for training something like videos.
>>103483792>You still see the human ear under her hairShit gen
>>103483724I just gave base noobai vpred 0.75s a download and honestly youre right, its a lot more consistent than the last time i tried it. One thing it seems its very sensitive to resolution ratios, widescreen sure seems to force very specific character stylings. Will have to fuck around with this today.>>103483782ghaayytt daaammn booiii>>103483794that really does put into perspective how much of a worthless paypiggie you have to be to buy a 5000 series in any capacity, from there it's painfully obvious how hard nvidia is just laughing all the way to the bank.
>>103483685how is it not ok? how is he disruptive?he is just goofing off. he is not shilling, he is not ruining the thread, he has in fact helped me, and no doubt others, many times over the last week or two you're just a rules obsessed cuck that doesn't understand why the rules exist in the first place. you have no idea what civilization is>so no, I don't think I'm just going to watch him diddle around in civilized circles idlylook at what you are saying, fucking narc faggot.3>I'm all for full ID verification to post on 4chanof course you do you just bend to whatever authority for any reasontattle tale narc teachers pet you'll get yours in recess
>>103483820pedo content is not allowed, periodhe's not "goofing" around
>>103483820The text content of his posts is perfectly fine but I would also prefer he stopped avatarfagging with pedo slop.
>>103483817>that really does put into perspective how much of a worthless paypiggie you have to be to buy a 5000 series in any capacity, from there it's painfully obvious how hard nvidia is just laughing all the way to the bank.Yes and no. 5090s will be the best in slot GPU for hybrid gaming and AI. It will likely be the the king of training 1B AI models especially for raw speed.
>>103483831who cares, tattle tale? it doesn't harm the thread in any way shape or form. our little discussion here is way more disruptive than what he is doing. should we be b&, i'd say we deserve it far more than he does>>103483851that i can concede. he should pretend he is two different pedos at the same time so it's less of avatarfagging
we'll need to see how important including videos is for LoRA training video models vs only using images
>>103483792I have better ones, they are just not entirely blue board friendly.The amount of weird shit you can do with noob is insane.
>>103483777>It doesn'tif this is correct then >those tags are correlated with higher quality images so the vectors within HunYuan's model related to quality gets activated.cannot be true. that just sounds like voodoo speak. not trying to bust your balls about it but also >no comparison vids
>>103483877enjoy id verifications, you obviously want it too
Hosting the PromptRewrite model for an hour, the password is miku:UI:https://nil-intimate-madness-educational.trycloudflare.com/Pic rel is how to prefill the answer. Clicking Generate continues from there. It might need to be told that NSFW is allowed or something like that.API:https://nil-intimate-madness-educational.trycloudflare.com/v1/chat/completions
>>103483892nta but the Chinese do unironically like their tags, its insane how many quality tags they would use for one image
>>103483910...what do I do with this and how does it help me?
So... It's been months. Where are the non shit Flux porn finetunes?You guys were so adamant all the tunes were shit back then because "it just got out, just you wait". So, where are they?>DeDistilled quantsThis sounds interesting. What is it? An attempt to approximate pro? How does it differ from regular dev? It's not clear from the examples.
>>103483924It's the model that rewrites the prompts for HunyuanVideo:https://huggingface.co/tencent/HunyuanVideo-PromptRewrite>To address the variability in linguistic style and length of user-provided prompts, we fine-tune the Hunyuan-Large model as our prompt rewrite model to adapt the original user prompt to model-preferred prompt.
>>103483543Merely pointing out the fact that this dumbass infighting only ramps up after anon gets bored of whatever new thing just came out. We saw it with Flux, too. Thank god we still talk shop desu
>>103483933>An attempt to approximate pro?no, it's just a way to remove the distillation of flux dev so that it can be finetuned and work at cfg > 1 (negative prompting)
>>103483910I don't understand what this is, but it completely ignored the last instruction. So I assume it will be as shit for its main goal, too.>>103483957I see. Thanks.
>>103483852how hard is it to just slap a couple more vram sticks on the 5090? it's insane it will only be 32gb. it's unacceptable. it's like they want a competitor to rise up and destroy them. that sort of abuse never bodes well
>>103483965It's supposed to look like this, more or less. I think for some reason it didn't send the system prompt there.
>>103483965I think this model understands chinese better than english, I should try that actually...
>>103483670Hey let's not knock it until we see the results, maybe the knowledge of the video data will understand single images well enough to use it to some degree of kino.
>>103484005I have a hunch is that the reason sex never has any movement is that it wasn't train on porn videos, only images.
>>103483933>So, where are they?Nowhere. No one's trained off dedistilled or even that weird OpenFlux model
>>103484016If you can't get movement you are a dumb promptlet
>>103484034what's the prompt to get some action, oh dear prompt god??
>>103483979It's not hard at all, NVIDIA is already doing it.The L40 is basically just a 4090 with 48 GB VRAM and some datacenter-exclusive driver features (like being allowed to use them at all) and it's like five times the price.
>>103484034Catbox me a gen that contains a penis entering any orifice at least twice during the video
>>103484051how do they get away with this scam?how long we will be left holding out for a hero?https://www.youtube.com/watch?v=bWcASV2sey0
>>103483811Shit didn't notice that, inpainting artifact from when I zoomed in too much
>The MLLM text encoder will be based on HunyuanLarge and will need 1TB of RAM
>>103484166kek, that's my fair though, that it's a giant ass text encoder...
>>103483979It's not the difficulty. They want this to be the market segmentation without splitting up the software ecosystem. Pay for HGX, make nvidia richer!
>>103483910If you use an LLM to make your prompt it'll be 95% the same anyways and the rewriter will just add "Realistic, Natural lighting, Casual" to the end
>>103484166Source?
>>103483647>retard still has no idea how hunyuan video works haven't even tried it, yet he thinks he is superior to someone that doesI have tried it, retard. That's how I know embedded_guidance as low as 2.0 and probably lower does, in fact, work.
>>103483379I tried this using the Hunyuan example prompt with the woman jogging. I will post results for three seeds with and without the suggested prefix. So one prompt will be:>In the gym, a woman in workout clothes runs on a treadmill. Side angle, realistic, indoor lighting, professional.And the other prompt will be:>Shot on a Canon EOS R7, 85mm lens, f/2, sharp focus, movie scene. In the gym, a woman in workout clothes runs on a treadmill. Side angle, realistic, indoor lighting, professional.I am using 1280x720 resolution, 129 frames, 50 steps, embedded guidance scale 6.0, and flow shift 7.0. The seed will be in the filename, the order of the prompts will be randomized.
>>103483379>>103484226
>>103482892wolfu!>>10348181764 ram>>103481825barely got sage attention in with guide>>1034818298 vram using nvidia driver cuda sysmem fallback policy to prefer sysmem fallback
>>103483379>>103484231
>>103484226she runs like a character in a shitty Unity game
>>103483379>>103484241
>>103484212no it doesn't. it doesn't follow the prompt if its that low, but you as a retard doesn't notice that the prompt is different than the output so you say stupid shit like that. honestly embed guidance 6 is too low, you need like 9 to make it actually folllow, and even then...
>>103483379>>103484249
>>103484235if you can't load the text encoder (15.5gb), how do you expect to load the video model, it also asks for a shit ton of vram
>>103483379>>103484255My personal opinion is that there is no discernible difference.
>>103484226>>103484231>>103484241>>103484249>>103484255i love how from day one this thread proved the model could nail titty jiggle when the tits are clothed but had trouble nudewhat i wanna know is how they sourced enough of a gargantuan "girl running on treadmill" dataset to accomplish this in the first place
>>103484250>it doesn't follow the prompt if its that lowDid you start doing this last week? That has ALWAYS been the trade-off with lowering cfg, and is why the value is available to us for adjusting. You think lowering cfg didn't mean weaker prompt adherence in SD1.5?
>>103484269Yeah, those are all functionally equivalent. Thanks for testing.
I had no electricity or internet for the past 24 hoursDid any major breakthroughs happen that make Hunyuang more usable for VRAMlets?Will I finally be able to gen videos longer than 2 seconds?Only 12 GB of VRAM here
>>103484269Perhaps it's just due to the low sample size but to me it looks like there is a tendency to have simulated diffuse light like there's someone with a diffuse studio lamp outside the scene adding some light.
>>103484262you misunderstand with my hardware and the comfyui setup it is working fine only answering questions people had on workflow
>>103484291Okay, try to make a guess as to which videos were made using which prompt.I used a random number generator to determine which video I post first for each seed.
>>103484250>but you as a retard doesn't noticePost your skin color please.
>>103484289IDK about 12GB exactly but the blockswap node *is* saving VRAM and compared to the intitial release there are more toggles that offload stuff to system RAM as well as more quantization options in the comfyui nodes.
A: >>103484226B: >>103484231 >>103484249 >>103484255
>>103484274i started it with flux, the main difference being that hunyuanvideo looks good on 6 or even 9, better than 2 (someone did a comparison, it completely breaks the video). with flux values that high only look good with certain loras, not even dynamic thresholding can save it.>>103484308full white italian. get through your tiny polish brain. hunyuanvideo is different than flux
>>103482892Back when Stable Diffusion first came out there was a feature where you could emphasize parts of the prompt with brackets.Does ComfyUI have that feature with Hunyuan?
>zoomers need millions in R&D and thousands in equipment to turn a 3 word prompt into a 10 word prompt because thinking for 20 seconds without receiving a reward deprives them too much of dopamineJesus Fucking Christ you're fucked in the head.
>>103484335>italian "white">started with FluxI'm Anglo-Saxon and Icelander and I've been doing this for over two years across every major model. Some of the genning techniques you've learned were invented by me. Sit down."Someone did a comparison" lmfao. You don't know anything.
>>103484353I'd say 6/10 ragebait, it's a bit funny but that's too long, gotta work on that anon
what did gen z do to make that guy so angry?>>103484370generous, i give it a 3, just enough credit for me to reply without quoting.
>>103484370"Ragebait"? He's simply correct
>>103484353That'd be rational. How much time do you spend typing on a smartphone for 150 images if you do that?Granted it's much worse if the training wasn't done specifically for what the model actually reacts reasonably to.And it could just as well be wildcards.
>>103484376see, that one deserves a 7/10, it's short and simple, that's how we do it
>>103483782lovely
>>103484363>I'm Anglo-Saxon and Icelanderand you're proud of that?here faggot homo get absolutely btfo'dfeel free to try your lower cfg scales. i'm not wasting my precious gen time
>>103484291>there is a tendency to have simulated diffuse light like there's someone with a diffuse studio lamp outside the scene adding some light.Which has nothing to do with the original post stating prompts that include camera specs "makes the videos look better" It's just placebo, anon
>>103484433>he can't tell that the 0.0 obviously follows the prompt just fine and evidently has a more natural texture to it>he says things like "absolutely btfo'd" and continues to be ItalianPlease, keep going.
>>103484454>>he can't tell that the 0.0 obviously follows the prompt just fine and evidently has a more natural texture to itthis has to be bait. NIGGA, THE WORLD IS WARPING AROUND HER, NIGGA, WTF
what the fuck is happening to the thread today
low cfg anon has gaslit me into lowering my embedded guidance to 5 and i kinda like it but i will go no lower
>>103484477trust the plan TM
>>103484463You're the first one to ever notice that high cfg values exhibit more coherence and 'perfection' in their execution of the prompt. Please, tell me more that I don't know.I'm going to cut this off because it's clear the Italian mind struggles to parse the meaning of English statements like "x has a more natural texture".
>>103484080Because CUDA started in 2007 and it takes time for anyone else to catch up with the insane ecosystem NVIDIA cultivated. 2007, imagine that... I don't know how he does it, but the CEO seems to always smell the next big thing.
I have noticed that lowering the guidance on the same seed helps with the 2.5d look when it happens.
>>103484270stock videos + youtube/social media is full of these
stop worrying and love the slop
>>103484507I think that tends to be the case when the prompt is ambiguous (as horny prompts often are). High guidance means that ambiguity in the prompt has to be respected, and should also be present in the resulting image/video—low guidance allows the model to resolve any developing ambiguity in one or another direction that, prompt aside, produces a more plausible real image/video.Roughly speaking ofc
>>103484161Last one of these, too high effort
>>103484546I think it happens because horny prompts are horny, not because they are ambiguous. A mere mention of tentacles will shift it into 2.5d no matter how many words for "realistic" you can come up with.
>>103484565>mere mention of tentacles will shift it into 2.5d no matter howdataset issue, how many live action tentacle videos vs 2d ones
>>103484565If your prompt says "realistic" (photo-esque digital art) and "tentacles" (anime) you have created an ambiguity which the model doesn't know how to resolve.I guess it's not clear what I meant, but I meant that the prompt is ambiguous to the model, not ambiguous to a human reader. The styles with which each prompt word are laden are in competition for control of the output. This is especially true with horny prompts, because they strongly indicate both 3dpd porn and hentai/etc.
>four fucking minutes to this one gen>dunno why it gave stocking down syndromereforge is fucked, too tired to re learn all of this from only a week of being out
>>103483910Thank you for your service.
>>103484488>anglo can't into realityancient anglo problem. observable in bacon already. that's why you lost not only one, but two countries you created. low cfg looks bad. it's all warped, it's all ghosts. yes, lower cfg for flux do look better, 3.5 is already too damn high, however, even then it destroys human anatomy, it's basically sd1.5 at >2.1 cfg.also italians have a visual mind. they're master of aesthetics. feel free to point out any relevant english painter, you can but it's>dante gabriel rosettithat's not a common anglo name is it? that's because HE'S ITALIAN
>>103483948any way to get it working on comfy?enter whatever you wantlet it rewriteuse that automatically as text input
>>103484226Solution:>In the gym, a woman in workout clothes runs on a treadmill. Side angle, realistic, indoor lighting, professional.>>103484231 >>103484249 >>103484255>Shot on a Canon EOS R7, 85mm lens, f/2, sharp focus, movie scene. In the gym, a woman in workout clothes runs on a treadmill. Side angle, realistic, indoor lighting, professional.>>103484226 >>103484241 >>103484269So >>103484334 was 3/3 correct if he meant to only separate the samples into two groups but 0/3 correct if he meant that A is without and B is with the prefix.
>>103484309>blockswap nodeHow much VRAM do you save with this? Enough to get a few more frames in a gen?You just activate everything?
>>103484477We call it "diffusion"
>"ldg bakery anon" deleted the rentry once again fucking kekd :(https://rentry.org/ldg-bakery
>>103484549
>>103483576>maybe if there was two more people contributingthere's none because everyone was chased away by the cringe
>>103483887Very nice
>>103484488i'm cooking something up with embed cfg 2. wait to be destroyed
>>103484600>reforge is fucked, too tired to re learn all of this from only a week of being outIsn't reForge supposed to be a drop in replacement for Auto/Forge?
>>103484631>swarthoid modern Italian claims brotherhood with Renaissance era northern ItaliansNice try. You are not Fra Angelico, you are a 100IQ retard who makes porn on your computer.>low cfg looks bad. it's all warpedYou really did arrive here yesterday. I'm not retreading arguments that we already did to death two years ago. You offer the most basic diffusion101 observations as arguments against my position. Has it occurred to you that I might be aware of the trade-offs involved with choosing a lower cfg? Maybe after I have acknowledged again and again that I am aware of them?You somehow can't see the overcooking that is visually obvious in 6.0+, maybe because your eyes aren't trained enough yet, but that's what you're going to have in your gens forever if you can't accept more chaos and more botched gens which are the price you pay for lower cfg.I am aware of the arguments against low cfg. You don't know who you're talking to and you're too new to be acting this arrogant in arguments.>>103484788Low CFG gens are more volatile, bad 95%+ of the time. What could you possibly prove? The low CFG genner has to dig through many failed results to find the gems.
>>103484488>it's clear the Italian mind struggles to parse the meaning of English statements like "x has a more natural texture".
>>103484208>>103484478You've shown /b/ yes?
>>103484827way to shift the goal posts. yes, very practical to do 20 15-20 minute video gens to get 1 (one) good one. that's 5 hours for a 1 (one) gen. what you're not understanding is that >6 cfg for hunyuan is absolutely fine, totally unlike other image models and it's not me saying that it's the chinese. if you had tried the model you would understand.a few more minutes and you're going to feel ridiculous>>103484916not funny, and untrue
>trust me just a few more minutes!
>try to buy credits for hunyuan video>you need a fucking "coupon code" Do they just hate money?
>>103484969>what you're not understanding is that >6 cfg for hunyuan is absolutely fineI can see the gens. I can see that it isn't. I've also been genning videos myself, no matter how much you bizarrely insist I haven't. I understand that you think the cfg values being different between Flux and Hunyuan is something new which I'm failing to take into account, but if you'd been here before Flux you'd know that cfg scales have been different on every model ever released. 4 cfg was very low on SD1.4, but relatively high on SDXL, very high on FLUX, extremely low on WDXL, etc. The numbers are almost arbitrary. My judgments about what the cfg should be were based initially on what I was seeing from results at different guidance values, and later based on my own experience genning. Hunyuan is not uniquely different, the effects of going higher or lower are more or less the same.>>103485016He's genning a video. It takes time.
>>103485016look, i'm doing another gen with cfg 6, then i can post and we can compare. if i'm proven wrong i will apologize, if i'm not i think you should. and even then, there are still a very valid reason for high cfg specially for video. we'll talk about it when the other gen finishes.
>>103485035the ali express way lol
Might pull, dunno
Sar...
>>103485047You're wasting your time because you can't prove anything by making a comparison of two videos. What could such a comparison demonstrate that would contradict what I said a few threads ago here: >>103452901 ?
>>103485066>it gets bigger when i pull it
will she come back if i pull?
>>103485074hmmmm...
>>103484543never
>>103485144Hey, that's my big titted elf streamer
>>103485144It's strange how much more mesmerizingly real her face on the right looks. Shame about the glasses. I wonder if you can shuffle some variables around and get rid of them.
NOW I might pull
>>103485144>>103485168I am unable to refrain from ogling at those tits
>>103485144The one on the right isn't wet enough. She needs to be wetter.
>>103485186Anon is this close to inventing a virtual youtuber.Just invent a generative neural network that works on old crypto asics. There must be a shit ton of them, useless because of the algorithms getting more complex.
>>103484646> was 3/3 correct if he meant to only separate the samples into two groupsThat was me and I was the anon who contradicted >>103484269 just about there not being a discernible difference for this >>103484291 reason. There apparently is and it's not pure placebo?>>103484672I really don't get the exact behavior between all the model load/unload/moves, caches and various parameters... and with how slow this is to try, I don't want to experimentally find out.But it seems to allow a lot more frames.
>>103485154thanks for the inspiration>>103485220when we have working loras for hunyuan then its over, we will be able to add any character to any video
>>103482892I tried copy-pasting random TikTok descriptions into Hunyuan and it seems to kind of work.Though only afterwards did I realize that I made the critical mistake of generating a horizontal video.
EMBEDED CFG 0
>>103485236>when we have working loras for hunyuanYou can already try:https://github.com/kijai/ComfyUI-HunyuanVideoWrapper/pull/72https://github.com/tdrussell/diffusion-pipe
>>103485291Neat
EMBEDED CFG 6>>103485291actually it's embeded cfg 2
I guess I must apologize. I'm sorry, I was wrong.
>>103485233>But it seems to allow a lot more frames.what did you use and how many frames do you use now?
>>103485236>when we have working loras for hunyuan then its over, we will be able to add any character to any videoI'd rather have longer outputs (10-15s) and img2video
>>103484824Right but it puts back in the older Gradio frontend. Forge originally took A1111's frontend and then did full on replacement of the backend code with ComfyUI's backend and mashed it together with some other code changes essentially and then evolved from there focusing on the frontend mostly while merging in ComfyUI stuff on the backend. It's why there are some regressions in a few niche areas like Intel Arc support because the A1111 backend hacked in operator override support for Intel's Extension for Pytorch and ComfyUI doesn't.
>>103485321However...
>>103485233Once again, the original post is nonsensical. >>103483379>if you want godlike hunyuan quality put this in front of your prompts. Thank me later:>>shot on a Canon EOS R7, 85mm lens, f/2, sharp focus, movie scene.It failed at both making outputs "godlike" AND making outputs resemble anything like the camera specs prompted. The fact that the original poster had no examples should have been enough for anon to quickly dismiss. Have you ever seen a widely used dataset which specifies camera specs? NO! These plebbit-tier placebo prompts ALWAYS pop up with every new model. The only time anything similar to "please make this picture pretty" works is with Booru models because they are explicitly trained with "highres" etc. >it's not pure placebo?It is. It always is. And anon just showed you.
>>103485291Unusually coherent for 2.
>>103485295nice, hopefully we will see some tests soon, it intrigues me how you can train loras for a video model only using images, I think the lora will outfit anything in the video, meaning your character will be imposed in the video gen>>103485291nice
If we get 97 frames@544x960 with 24GB VRAM, how much more could wish for using 32GB (5090) or 48GB (2x3090/4090 if someone figures how) ?Is it linear so 129 with 32GB and 194 with 48GB?
>>103485358I love the reddit-tier placebo prompts. Love getting ideas for injecting nonsense into my prompt, no matter where from. Sometimes it makes for good or fun results, sometimes not, but always good to change things up.
anyone tried using Chinese to see if the result is any better?
>>103485333I still think high cfg is better for video because, let's face it, videogen is only good for pornos or memes, and you require closer prompt adherence for those. low cfg prosbetter compositionbetter colorslow cfg cons phantasms likely require higher step count to fixpossibly (probably) lower prompt adherencepossibly higher likelihood of anatomy deformations
>>103485402try with cool sunglasses
>>103485381That's perfectly fine, but anon shouldn't go around claiming they "make outputs look godlike" or even push the model to output anything resembling the camera specs prompted. Because that's simply inaccurate.
>>103485358I'm pointing out it probably does something. It might have shifted more towards a film shot with more lighting provided by studio lights or assistants perhaps.Whether it's an accurate or interesting representation of some Canon camera? No clue whatsoever.
>>103485391>videogen is only good for pornos or memesIf you have an insane amount of compute/vram, it's probably usable for games or movies.Vidtovid would be like mocaping and imgtovid a cheap way to small animations in videogames for example.
12gb hell. This is all I've got>>103485391>and you require closer prompt adherence for those.Memes certainly. Pornos—it depends. I have always preferred to just ogle a certain kind of woman, and just to see her idly shift her weight or glance from one side to the other makes it more real to me. So for 'pornographic' purposes I can be content with very low cfg.But another kind of anon specifically wants to see Ana De Armas in a princess peach costume being split in half by Johnny Sins with a cock the size of an arm, and if that's what you need then you need big cfg
>>103485422>It might have shifted more towards a film shot with more lighting provided by studio lights or assistants perhaps.Wouldn't explicitly prompting for this instead of some illusive arbitrary camera specs that the model doesn't actually "understand" achieve that better? I understand your sentiment, but there's simply a smarter way to go about it.
>>1034854498gb hell. This is all I've got.
>>103485322Don't remember exactly but I just tried again and 150 frames worked. I know it crashed with some number above 50 below 100 frames before.Granted I also did comfyui updates and sageattention instead of sdpa at some point. But I recall blockswap did help very noticeably. =
>>103485471tell me a prompt, aspect ratio, and cfg setting you want and I'll gen five tiny videos and post the best one
What determines how long the prompt can be for HunyuanVideo?
>>103485474>I just tried again and 150 frames workedOh, very nice, I'll try that then, that's 2s more.
>>103485449>>103485471You can generate 960 x 544 x 77 on 8gb with all offloading possible and using CUDA SysMem Fallback. Needs more system ram for all that offloading.
>>103485444>Vidtovid would be like mocaping and imgtovid a cheap way to small animations in videogames for example.I was playing around with FMV lora for flux, made me thought i could make my own FMV game, just filming myself in my room acting out the scenes and the script and then run them through a vid2vid to make me be a princess or a lizard, etc. but i guess that's not possible yet = (
>>103485466I did NOT propose the prompt.This may be crazy gibberish but arguably if anon aesthetically likes it: What more precise alternative keyword variants activate better? Pretty sure we don't have a tool to do that?
>>103485497The way I have it is 20 40 and both offloads enabled. Not as a result of extensive experimentation to find the optimal choices, mind you.
>>103485449the problem with low cfg is that it's going to increase anatomical deformities, specially with movement, and create all these phantasms and artifacts. it's probably better to leave the prompt's somewhat vague and higher cfg for variety.
>>103485508>, just filming myself in my room acting out the scenes and the script and then run them through a vid2vid to make me be a princess or a lizard, etc. but i guess that's not possible yet = (we are so close anon
>>103485531so hopeful for the future
I heard there were vramlets in here jealous of videogen
>>103485545Many such
>>103485524so picrelthanks anon
>>103485481>Marge Simpson from The Simpsons, wearing a vibrant two-piece bikini in a slow-motion jump, with the camera angle focused on her upper body from a slight overhead view. In the middle of the jump, arms extended upwards, torso arched, with a smile on her face. Subtle motion blur to enhance the slow-motion effect. The lighting is bright and airy, an outdoor setting.9:16cfg 4.5>>103485507Gonna try it then when it gets even more streamlined, but for now I'm content living through anonymouses.
>>103485508I need to learn to sway my hips when walking, and then slap a vidtovid of a hot woman.
>>103485569NTA but I wouldn't advise this unless you have a ton of system ram
>>103485545videogen killed the vramlet stars
>>103485402How is you doing these? Is it some kind of vid2vid? Or just normal Hunyuan workflow?
>>103485587How much ? I have 64GB.
>>103485596How are* is what I mean to write, what the fuck is wrong with me.
>>103485569Yes. Maybe these are the maximum (most VRAM saving) settings possible on this node, but I really haven't experimented much.
>>103485587Works on 64GB for me. Some people probably have 128 or 256GB already.
>>103485600Probably enough.I have 32GB and blue screened by using similar settings
>>103485582yep. luckily we probably have hundreds, if not thousands of videos about how to walk like a bio woman from our transfolk friends.
>>103485490>"--text-len", type=int, default=256, help="Maximum length of the text input.">"--text-len-2", type=int, default=77, help="Maximum length of the second text input.",Can I just change the first one and get infinite prompt following?
>>103485582>>103485640it's not hard just follow a straight line when walking, women have natural hip swaying but they can exaggerate it using that trickand men can have it a little when doing the same
>>103485573working on it. Should take 10 minutes for 5.
>>103485627I see, well should be good then.
>>103485668Thanks.
>>103485596rf-inversion workflow in new commit
>>103485661yeah, men walk with their feet on each side and women walk with their feet straight pointing forward. top trooning advice for our transfolx
>>103484943/b/ is too poor to gen video
>>103485678It's not looking good so far. Might be pushing unacceptably low on image dimensions.
oh no, kektime to hoard some videos then
time to start my fake thot tiktok
2/5
>>103485731Like something from a le lost HAUNTED Simpsons episode zoomerbait video
>>103485699What does it do? Is it just another kind of vid2vid?
>>103485713>we will have a VR version of this at some point for maximum acceleration>https://www.youtube.com/watch?v=WxmICiOXw2c
3/5
another total failure
>>103485770comfy
sorry anon.
>>103485766It's like V2V but better because it uses the input to generate the noise instead of adding random noise to every frame. Like inverting the generation back into noise.
So... What the fuck is flow_shift
>>103485748catbox for that one? did you change some values compared to kijai's workflow?
Did anyone try using a different text encoder?
>>103485748
>>103485766pretty much
>>103485935to change the strength of this v2v you change the gamma or the eta thing?
>>103485755wow batiful baby
>>103485935does it take more time or vram etc? or pretty much the same?
>>103485821Oh god these are so cursed. But thank you anon, at least it knows the character and #2 was really close to the modern art style. There is hope.
>>103485987its twice as long because now the model has to add the noise + denoise
>>103485953the resampler, start_step, end_step and eta_base, also eta_trend
>>103486010Can you test if you can turn her into an animal? like a walking Tiger?
>>103482892Is there any general for weird canny valley porn made by Hunyuan yet?
>>103486010what are the best values for you?
>>103486017yeah, im trying but i'm not having success so far, only can do "surface" changes, it may work for enhancing videos but nothing crazy, it needs to have the option to multiply sigmas in order to preserve the structure at least
>>103486029there's an ai porn thread, I wish it was more active though so that we can discuss on our findings, like what hunyuan is able to do or not>>>/aco/8628498
>>103486084The silence is probably everyone genning stuff too hardcore for 4chan, when a prompt fails they post the result so it isn't completely wasted.
https://xcancel.com/huwhitememes/status/1866701055570088241#mkek, didn't know Hunyuan could render Elon Musk
>>103486063ok thanks anyway
>>103485755nice bobs lik lik
>>103485880there is the one that just got released, I wonder if it works
local hunyuan doesn't have any censors, just like stable diffusion right?
>>103486127nope, nothing >>>/aco/8639690
>>103486123>there is the one that just got releasedwhich one?
>>103486102not enough people with 24GB cards, it was way more animated when the cloud models were out at first before being more censored
>>103486130Kek, I just posted that. I didn't know we had an AI porn thread until that anon just linked it
>>103486130Welp, I guess my life next month will be entirely centered around trying to get a 5090 without paying scalper niggers.
>>103486148I bet, my 4090 is worth more than when I bought it new
>>103486141the one from onediffusionhttps://huggingface.co/lehduong/OneDiffusion/tree/main/text_encoder
>>103485769yeah in 80 yearscant wait literally
>>103486171that one could work on Hunyuan? interesting
>>103486161yeah it's crazy, I'm almost tempted to sell mine while waiting for the new 5090
>>103486177>that one could work on Hunyuanno idea
>>103486189there was an anon trying to make it work with JoyCaption but the prompt understanding was completly ass, as expected I guess
>>103486171>t5-xlBut wasn't the point to change the text encoder to a regular multi-model LLM?
>>103486201But it does work, and it's as good as the default. You are just dumb.
>>103486209yeah it was the point, but we still got the ducktape MLLM, not the original one
>>103486220>it's as good as the defaultgood joke
>>>/aco/8639708Holy fuck dude...>*Goes to e-bay to find a 4090*
>>103486209Yeah and I have no idea why they did that. Lumina-Next used Gemma 1 2B which OneDiffusion based their architecture on. Updating that to Gemma 2 2B or Qwen 2.5 would've been better than using T5-XL.
>>103486161>>103486180does the same apply for the 3090? i may finance my new 5090 with it
>>103486313>does the same apply for the 3090?not really, the 3090 doesn't support fp8 torch compile and fp8_fast, that's why the 4090 is so valuable, a video that a 3090 takes 20 mn to do, the 4090 can do it in just 6 mn
>>103486313No but its value used didn't really change since a year ago, which is also a sign it's still in demand.
Happening alertHyvid lora training code released https://github.com/tdrussell/diffusion-pipe
>>103485569 (you)Using picrel on a headless linux, I reach 43GB usage on RAM and 15GB VRAM on my 3090.150 frames went through, I wonder if I can go higher. 175? 200?
>>103486109
>>103486371how much vram needed to train some videos on hunyuan with loras? more than 24gb?
>>103486010>start_step, end_stepSo in other words basically how controlnets work? They effect the frames at 50% of steps of set end step to 0.5? Then the final steps control is released for the sampler to denoise, if so it will be best to leave start_step at 0 and only experiment with end_step >>103485837its useful in vid2vid, higher values will alter it more, and lower values less, that is all i know anon
>>103486371>>103486392>HunyuanVideo supports fp8 transformer. The example config file will train a HunyuanVideo LoRA, on images only, in well under 24GB of VRAM. You can probably bump the resolution to 1024x1024 or higher.>Video uses A LOT more memory. I was able to train a rank 32 LoRA on 512x512x33 sized videos in just under 23GB VRAM usage
>>103486371I have a hard time understanding what would be a lora training for a video model.
>>103486371so, you'll use the wrong text encoder to train a lora of hunyuan? or you don't need the text encoder when you train a model?
>>103486407the same thing as lora for image models, you want to add a celebrity, a pose, a style? you'll be able to do it on hunyuan
>>103486340>>103486340where I live the 4090 is still around the same price it was a year ago. should I buy a 4090 and then sell it alongside my 3090 when the 5090 comes? is that a sound investment?
>>103486409We never trained t5 with flux either.
>>103485837>So... What the fuck is flow_shiftit's basically how you alter the curve on the scheduler, that's all
How long does a single video gen take on a 4090?
>>103486371where is AI to analyze the repository for backdoors?
>>103486084>>103486102patience
>>103486433960x544x97f -> 6 min
>why is 4090 so expensive!?!There is a country of over a billion people where the sale of 4090s are prohibited but you can easily pop the core out of a 4090 to make frankengpus.
>>103486433about 3 fiddy seconds
>>103486440This. I will wait for furk to update his patreon.
>>103486426lmaoif you asked that seriously, nowait a month
>>1034864336 minutes on 4090, ~13 minutes on my 3090
>>103486467damn. do you know if the rtx 4070 super had the same scalper problem the 4090 had? i remember the day it came out you could get one for official prices easily where I live. would be nice if the same happened to the 5090
>>103486495>rtx 4070 superMid life refresh card. Those never have the same day1 scalper problem (it's also a 12gb piece of shit)
>>103486489>~13 minutes on my 3090how many steps at what resolution? i'm getting 17 minutes on 40 steps 960x544
>>103486495that is very sad
>>103486526>40 stepsAny enhancement expected when going to 40 steps instead of 30?
i have absolutely no idea if the SAE vae has helped with prompt adherence btw
>>103486536i think so yeah
>>103486371Its over... for image-gen.
>>103486526Be happy, my 3090@300W generates in 30mn for ~150 frames.
>>103486526I've overclocked the memory to the maximum it could get without crashing. 960x544x97 45 steps
>>103486536More steps is always better. We're just coping with 30 because gen times are so long
>>103486552imagine a hyper lora for hunyuan, like it could run fine at 4 steps, would be game over lol
>>103486537vae is the image decoder, has nothing to do with the prompt
>>103486544>>103486552What's the difference? More clarity?I do know that for images there is definitely diminishing returns depending on the sampler, and at some point there is no point adding steps.
>>103486570he said SAE, you can change clip_l for its finetuneshttps://huggingface.co/zer0int/CLIP-SAE-ViT-L-14
>>103486552>>103486576I think the resolution is much more important than steps, especially for prompt understanding
>>103486576yes, less artifacts, less of that ai look, specially on the hair, makes movement more fluid, faces better, less wobbling etc.
>>103486537why does it keep genning them so young? prompt so i can avoid this?
>>103486570>>103486577sorry yea i meant the clip not the vaei havent noticed any improvements that could be more than placebo but it doesnt seem to hurt to have it on so ill keep it on
>>103486576From a few threads ago>>103452649>>103452659>>103452675
>>103486604>A beautiful young teenage girl with long blonde hair, wearing a revealing bikini that showcases her ample bosom and toned midriff, lounges on a beach towel by the ocean. The camera movement is Low Angle as she applies sunscreen, her hands slowly rubbing the lotion on her skin. Realistic, Natural lighting, Casualthanks to the anon who was hosting the prompt rewrite model
>>103486445Sheesh I don't think I'll be touching local video gen for a while then
>>103486598and you feel that 40 steps is the moment when you have diminishing returns?
>>103486613>100 stepsI'm not gonna do that, it's too long already!
If I have a working kijai Hunyuan setup from a week ago is there currently a good reason to pull a newer commit? Like more features or coherence or speed or anything. Or would I just be risking introducing problems for no gain
>>1034865496 videos per KWhif electricity is over 20 cents a KWh and you pay for your own electricity, you might as well rent somethng on vast for a similar price per video and gen over twice as fast
>>103486621i guess, it just takes too damn long to gen anything more. for me it's the time/quality sweet spot. just did a 60 step, didn't look all that better desu so i'm sticking with 40
>>103486637yeah you can pull, there's more options for the vae now so that if you get some OOM during the decoding, you can decrease some of its parameters
>>103486617prompt rewrite model?
for me its 50, anything less and it feels like you're sacrificing a ton in coherency but more doesn't add much in realistic scenes
>>103486659psycho mantis? >>103483910
For 3090s: Underclock your core clock but overclock your memory you can generate up to 40% faster in HunYuan that way.
>>103486653>OOM during the decodingyup, pic related is what the dev recommend, uses half the VRAM, solved my OOM problems
>>103486690what does changing the tile settings affect?
>>103486682Specific amount in afterburner?
https://github.com/esciron/ComfyUI-HunyuanVideoWrapper-Extended>Support for any llama model type LLM.>Support for LLava and mLLava model_type.>Support for Mistral model_type.Really interesting
>>103486706tiles the decoding, uses less vram, it might introduce seems/edges but I'm yet to notice any.
>>103486740So Hunyuan can both support LLMs and MLLMs?
>>103486740what other MLLM exist than llama-llava-3-8b? maybe we could go for something a bit bigger but can still fit on a 24gb card
>>103486682no way. exact amounts please
>>103486740>>Support for Mistral model_type.Does nobody understand how this shit works? The text embeddings are the hidden states of the LLM. Anything based on the llama 3 8b base model is going to "work" to some extent. Everything else won't work at all (unless you explicitly train some kind of weird embedding projection adapter layer or something). I am reminded of retards in /lmg/ like a year ago who were convinced you could apply mistral loras to llama models and it would work.
>>103486776What's a MLLM? I know llama 3.2 has image input, as well as pixtral.
>>103486706I'm not sure what method it is using, whether its tiling individual frames one by one or tilling them all into smaller pieces. Like top section then middle and so on. But its slower obviously, it might be swapping tiles into system ram in chunks and if you are like me using a lot of swap space on SSD for example if will lag the fuck out of everything for a few seconds, but it prevents OOM on the GPU, smaller values means less vram btw. But they had changed it slightly in the last 2 days and I don't know what its actually doing because I've have read into it much. All I know is it works.
>>103486682what values anon?
>posters scared to do their own overclocking/g/ has fallen
>>103486835i wouldn't bother with that desu, you might burn out your hardware... void the warranty etc.
>>103486584Yeah I meant all things being equal.>>103486598>>103486613Oh yeah it's kind of obvious on anime style.I didn't see it that bad on realistic stuff, but worth launching 50 steps before sleep I guess.
>>103486827>I am reminded of retards in /lmg/ like a year ago who were convinced you could apply mistral loras to llama models and it would work.I mean, Hunyuan is using a text encoder it has never seen and it's working all right, we're still waiting for the official encoder though
>>103486850>void the warrantyNigga I bought a second hand mining 3090 for $300 I don't give a fuck.
>>103486869>I didn't see it that bad on realistic stuffI've seen more merging of humans in 30 steps than 50 I guess
>>103486644I pay between 12 and 15. And it's fine, not a big deal, I have the server, might as well use it.
>>103486613More thigh on 30 steps so that one wins
>>103486830>What's a MLLM?https://medium.com/@cout.shubham/exploring-multimodal-large-language-models-a-step-forward-in-ai-626918c6a3ec
>>103486850>void the warrantyhow would they ever know?
>>103486835undervolt to 300W for 90-95% of the same performance
>>103486923Llm on the vram at the time of the crash snitches on you for overclocking and raping it.
>>103486891well have a look online then, but every GPU is different, don't come crying heard when you system randomly reboots it self after your screen messes up. You need more voltage as you increase the frequency, but I never don't OCing since years ago. Just look around in reddit i'm sure some will give examples of sucessful overclock on your specific card but no card is the same, even if its the same model, it depends on how well the gpu was made. You might have a good gpu that is stable or you might only get like 10% out of it before it becomes unstable.
>>103486947*not undervolt, just power profile change
>>103486918A-anon...That was a rhetorical question as should have been clear since I named the llms next to answer op. Because>Multimodal can mean one or more of the following:and then 3 different options of what it could actually mean tells you it's a shit coined term that doesn't help at all when you need to specify something, and only exists to make ai sloppers sound more sophisticated.
>>103486952how did we ever went along with large language models and unique id's on our ram sticks and hidden processors inside our processors?
>>103482892https://files.catbox.moe/d19pqx.webm>Tiktok dance video. Hatsune Miku is shaking her gigantic tits and huge ass. She's showing off her massive breasts, slim waist, and wide hips. She's jumping around and her boobs jiggle up and down. Her long, blue hair is twirling around as she dances. She's completely naked and her tight pussy and large nipples are in full view. #fyp #fy #foryou #foryoupageTikTok seems like a relatively good base for videos of women.>>103486433~45 minutes if you go for 1280x720, 129 frames, and 50 steps.
>>103486966To clarify, MLLM isn't a real term. A multimodal LLM with [something] capability is.
>>103486990>relatively good base for videos of womencan you post an example
>>103486433Depends on settings.I think if you're just trying stuff out I'd possibly actually recommend something like 480x288 / 30 steps or the other way around. Gives you about 25 frames per minute.
>>103486872>Hunyuan is using a text encoder it has never seen and it's working all rightYeah, because the unreleased text encoder it was trained with is still based on llama 3 8b. That's why they can point us to a random llama 3 8b llava model and it kind of works. But it's not going to work with mistral or any other unrelated model family.
>>103487099>>103487099>>103487099>>103487099
>>103487093>the unreleased text encoder it was trained with is still based on llama 3 8bhow do you know that?
>>103482892Good evening sirs
>>103486682This retard is trolling, you should do the opposite.
>>103487107Because otherwise the llava model we're using wouldn't work at all! It works because the hidden state vector space is compatible, because it's based on llama 3 8b, just like the official text encoder must be.You guys are fucking retards. I will film myself drinking a gallon of horse cum and post it if the official LLM isn't llama 3 8b based. It literally has to be.
>>103487132someone managed to make JoyCaption work on hunyuan, no errors, and we got an output, the prompt adherance was complete horseshit but it was possible
>>103487108ugly titcow give me more
>>103487159JoyCaption takes a pretrained CLIP model, then projects the output of that into the llama 3 8b embedding space. Then feeds that into a llama 3 8b LLM and predicts the text. Once again, the reason that works at all is because it's based on llama 3 8b.
>>103487287oh ok, then it's a good news if that's still llama3-8b, the size is acceptable for a 24gb card