I'm Getting Litty EditionDiscussion of Free and Open Source Text-to-Image/Video Models and UIPrev: >>106585705https://rentry.org/ldg-lazy-getting-started-guide>UIComfyUI: https://github.com/comfyanonymous/ComfyUISwarmUI: https://github.com/mcmonkeyprojects/SwarmUIre/Forge/Classic/Neo: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassicneoSD.Next: https://github.com/vladmandic/sdnextWan2GP: https://github.com/deepbeepmeep/Wan2GP>Checkpoints, LoRAs, Upscalers, & Workflowshttps://civitai.comhttps://civitaiarchive.com/https://openmodeldb.infohttps://openart.ai/workflows>Tuninghttps://github.com/spacepxl/demystifying-sd-finetuninghttps://github.com/Nerogar/OneTrainerhttps://github.com/kohya-ss/sd-scripts/tree/sd3https://github.com/derrian-distro/LoRA_Easy_Training_Scriptshttps://github.com/tdrussell/diffusion-pipe>WanXhttps://comfyanonymous.github.io/ComfyUI_examples/wan22/https://github.com/Wan-Video>Chromahttps://huggingface.co/lodestones/Chroma1-BaseTraining: https://rentry.org/mvu52t46>Neta Luminahttps://huggingface.co/neta-art/Neta-Luminahttps://civitai.com/models/1790792?modelVersionId=2122326https://neta-lumina-style.tz03.xyz/>Illustrious1girl and Beyond: https://rentry.org/comfyui_guide_1girlTag Explorer: https://tagexplorer.github.io/>MiscLocal Model Meta: https://rentry.org/localmodelsmetaShare Metadata: https://catbox.moe | https://litterbox.catbox.moe/GPU Benchmarks: https://chimolog.co/bto-gpu-stable-diffusion-specs/Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-oneTxt2Img Plugin: https://github.com/Acly/krita-ai-diffusionArchive: https://rentry.org/sdg-linkBakery: https://rentry.org/ldgcollage>Neighbours>>>/aco/csdg>>>/b/degen>>>/b/realistic+parody>>>/gif/vdg>>>/d/ddg>>>/e/edg>>>/h/hdg>>>/trash/slop>>>/vt/vtai>>>/u/udg>Local Text>>>/g/lmg>Maintain Thread Qualityhttps://rentry.org/debo
>>106589837you didn't add anistudio to the OP
>>106589848correct
>>106589837faggot OP forgot thishttps://github.com/FizzleDorf/AniStudio/releases/tag/pre-release>>106589855>t. faggot
Blessed thread of frenship
>nunchaku qwen image models released>deepcompressor last updated 6 months ago>no quantization code in deepcompressor so you can't quantize your own merged modelsthey are going closed source
SaaS is here to stay.>>106589823The only difference between you and me is how far my GPU sits from my setup. I'm running local just like you, images save to my local drive in the same folder as my API nodes and local UI workflow.Or are you telling me you're somehow generating with autonomous electricity and ibterbet access while staying local, retard?
>>106589911ani is saving local. saas lost
>>106589911>The only difference between you and mekek, you are using the fisher price version of AI generation, you are like someone who thinks Facebook is the internetgo back to /saasdg/ and wallow in your misery
Doe lora files contain metadata? And if yes, is there a way to strip it?
>>106589911>The only difference between you and me is how far my GPU sits from my setup.
>>106589866>jumpscares you
>>106589944Doubt it, since that would have to be placed there by the trainer program, for what purpose ?
>>106589911>how far my GPU sits from my setup>my GPULOL
>>106589911>The only difference between you and me iskek, you litterally have a jew standing behind you saying 'no goy, you can't generate that, you can only generate this'(((SAAS))) shill begone
>>106589837What is the alternative to buying a GPU to run AI models? Surely all those server farms aren't built off GPUs right?
>>106589978I think they call them accelerator cards in the enterprise market, but yes they're graphics cards
>>106589978nah you're right. they use pixie dust and unicorn farts
/ldg/ bross welcome SaaS users too! Share your gens, don't be shy about your outputs here!We're all working with generative models, local or cloud based doesn't matter. /ldg/ accepts everyone, post your creations and let's see what you've been making!!Pic related generated in my local pc with local nodes with my local billing adress, same as you.
*yawn*
>>106589978>Surely all those server farms aren't built off GPUs right?NVidia is swimming in endless money, surely those server farms aren't using GPU's, right ?
Qwen SRPO waiting room:https://github.com/Tencent-Hunyuan/SRPOWaiting for a hero to try this on Qwen
>>106589978They're not built off gaming gpus, no. That's why nvidia can charge an arm and a leg for tripling your power bill, they don't care about you because gaymers don't financially matter (anymore)
>>106590010It was tried on Flux and results weren't that good. Flux is more realistic than Qwen by default. It is over.
>>106589987big if true
>>106589837>previous thread >ctrl + f "litty">zero results what did OP mean by this
>>106590019>Flux is more realistic than Qwen by defaultlel
>>106590027This is a zoomer thread now, he's bussin on god no cap fr fr on ohio skibidi rizzler's gyatt
>>106589911KYS shizo
>>106589987>the true state of local
>>106590040fr fr ong no cappin' bruhski we uh, fuck i don't know more zoomer lingo
>>106590040O.o
>>106590019Even if it doesn't completely unslop the model, lowering the bias towards 4o garbage would already be a victory, don't you agree?
>>106590038It's the truth. Krea was no accident. It is based off an non-distilled Flux model. Chroma was no accident, similar. Qwen is not even distilled and it's already slopped.
>>106589958gib catbox for godzilla tits miku
>>106589995Imagine yawning at objective truth. Imagine yawning when your "local" models were literally trained on LAION scraped by German academics using cloud compute clusters, then distributed by Stability AI (a company). Imagine yawning when your loras were trained on datasets someone else curated and uploaded to HuggingFace (owned by investors). Imagone yawning when your "local" inference is running on hardware designed by NVIDIA/AMD and drivers maintained by corpos. Sure, keep yawning about how clicking download on civitai makes you some kind of digital freedom fighter while SaaS chads just cut out the middleman. At least we are honest about using corpo services instead of larping as tech libertarians.
>>106590063bro dumped like a dozen catboxes recently FAGGOT GO LOOK FOR THEM IN THR PREVIOUS THREADS
>106590071holy cloudkek cope
>>106590058Krea attempted to specifically remove Flux slop, and even that failed.Flux is THE slop model, nothing else comes close.
>>106590065i should call her...
Decided to try out flux and apparently my 4090 is not enough lol
>>106590052https://en.wikipedia.org/wiki/Glossary_of_2020s_slangsome i remember being used when i was a kid and probably before thatquite a few are p funny
>>106590065Why are you posting pics of my waifu
>>106590090that's one way to out yourself as a total retard
>>106590063that just made my dih leak a fluid, sure budhttps://files.catbox.moe/w7lnm4.pngthough the background is a bit fried, i think i busted the cfg scale in this one though i forgor>>106590073i appreciate the enthusiasm but its fiiine>>106590095>99% of this shit is just black/urban simplifications of already simple wordsshits the way the cookie crumbles i guess, personal favorite has to go to "ligma" and its variations though honestly.
>>106590052Model/catbox? I'm trying to improve my proompting when it comes to realistic gens
>>106589862>Commercial License>If you prefer to use this project under a commercial license, please contact us at [your-email@example.com]kek
>>106590110>https://files.catbox.moe/w7lnm4.pngno i wanted the catbox of miku, unless that wasn't you
>>106590075Aother yawn from "I downloaded someone else's model" anon? Let me guess, you're running it on Windows too? Microsoft thanks you for your "local" rebellion. While you yawn your GPU drivers getting updates from NVIDIA servers?While you yawn your Python packages from PyPI? Your model checkpoints from Google Drive links posted by randoms? Please, tell me about your sovereignty while your entire stack depends on corpo infrastructure you have no control over. Yawn harder, SaaS brothers welcome! We are /ldg/!
>>106590113this catbox >>106590110same model, same prompt, just cfg i'd float around 3.5-4.5 depending on the lora,though honestly if you have a gpu with more than 8gb of vram, i'd look into doing a refiner pass with an already realistic sdxl modelthe second i pick up the 16gb card i'm looking at, i'm dedicating entire weekends to trying that out. nova animal is good but, it has its weaknesses.
>>106590135>no i wanted the catbox of miku, unless that wasn't youoh i thought by the way you worded it, you wanted to TRADE for the catbox of my image.periods and commas are important my m8. pretty sure that guy just used wan and asked for some titty jiggles, it's not difficult to pull off.
>>106588315sexo with jenny
>>106590128Qwen does an okay job with the prompt. The model is superior to Seedream 4.
>>106590129where does it say that?
>>106590139Thanks>more than 8gb vramI do, but aren't there options to unload the base model if you're using a refiner? After all, the base model is not needed during the refiner pass
>>106590183reforge as far as i can tell, fully unloads the first model then loads the second, then when you do a new model it does the process over again.very slow on my near 10 year old card, i'm sure its near instantaneous anything rtx 3000 and newer.with comfyui there's specific nodes you're supposed to use for better memory management, like unloading models and clearing memory. some wan workflows use that for vid2vid passes.>sorry i can't be of more help i haven't touched anything like this in months again due to aging gpu
>>106590175sounds like a prompt skill issue
out of curiosity, there's no way to monetize genning an animation from an image right? since i assume there's a lot of legal trouble even if its extremely privatelike theres i2v websites but if you were a named individual on like patreon or something, you'd probably get taken down and sued fast.
>>106590190The prompt is as straightforward as you can get. Works on Chroma. Works on actual decent SaaS models. It doesn't get any better than that.
>>106590146noooo i'm the one who animated it
>see insane outputs on civitai model page>"dude, why do my outputs always look like ass but everyone else's are fire?">download image>import into comfyui>at least 30 nodes with multiple groups for upscaling, skin enhancers, hand fixers, face detailers and so onDo... do you guys actually do all that? Am I just not autistic enough?
javascript:quote('106589609')Escape from Tokyo with Snakette Pliskenawa
>>106590233holy fucking booba animation dude>>106590248you're not autistic enough for the autism club
>>106590248>Am I just not autistic enough?correct
>106590249
>>106590263that's how chatgpt said I should do it man
>>106590255>>106590257Okay then, what do you guys recommend? Are upscalers needed when the default resolution range of sdxl models (so like 768x1280 for example) is fine for me? What about the other groups I mentioned?
>>106590289If you have to ask, you'll never make it.
>>106589609Escape from Tokyo with Snakette Pliskenawa
>>106590294How dare I ask for advice instead of wasting hours experimenting with what works and what's needed, shame on me
>>106590206Where there's a demand, there's a way, but discussion falls outside of the scope of this site.
>>106590289>(so like 768x1280 for example) is fine for me?90% of the wow factor for a gen is how high the res is desu
>>106590289>>106590359also its, im pretty sure, 1216x832, not 1280x768 for some reason.
>>106590359I don't need 8k images when I'm trying to generate some decent-looking smut, especially since I'm on a 1080p monitor>>106590371As long as the total amount of pixels is the same and the dimensions are a power of 64
>>106590405Upscaling+refiner second pass is good for ironing out minor errors and artifacts. And just 1.25x is fine really.
>>106590289>Are upscalers neededyou can upscale if you want, adds more detail and is a easy way to fix shit. usually the way people usually inpaint is by cropping a square around their masked area -> upscaling or downscaling it to their desired resolution (either 1024x or higher) -> denoising the masked area. -> stitching the masked area to the original image (this is done automatically using the 'masked only' option in a1111-like guis or the crop and stitch node in comfyui). upscaling the original image gives you more pixels to work with and is less lossy. optional though if you know what you are doing>hand fixers, face detailersi recommend you skip this stuff and stick to doing it yourself manually
>>106590405>I don't need 8k imagesIt's not about being 8k. Doing at least a single second pass amplifies how "good" an image looks. >especially since I'm on a 1080p monitorThat's the thing about "highresfix", it adds details beyond what we think of as being just a higher resolution. I'm on a 480p screen and I gen images that are at least double if not triple the base res.
>>106590429>I'm on a 480p screencrt monitor? or just those TFT dell monitors from like 2006?
>>106590436It's a MAG 321UP, I just like 480P.
>>106590448Why would you pretend to be me though.
>>106590448>4k ultrawide monitor>uses it in 480p
>>106590248>do i have to do all those things that make a gen good to make my gens look good? lots of workflows can be trimmed due to bypassed or disconnected nodes but in general yes
>>106590063back from the gym, here at catbox, i used wan2gphttps://files.catbox.moe/5s6c7t.mp4https://files.catbox.moe/4yulk1.png
>>106590452i like to add to the narrative
>>106590289look into tiled diffusion
>>106590547I think people underestimate Wan for being able to deslop poses and composition. Like you can use wan to have the character move to a place with a more appealing background or take a position or look in a direction that makes the piece unique and different from other things you might see from SDXL outputs.
But can I use booru tags in Wan? Huh? Yeah, thought so.
>2.2 vace model is over 32gbHow the hell am I supposed to use this?
>>106590419>>106590429Fair enough, I'll give it a shot then. Do you use latent upscaling or a dedicated upscaler model?>>106590428Nice write-up, thank you
>>106590588I use latent upscale. Also if you modify the prompt in the conditioning hook going to the second sampler you can even do minor edits to the original pic with high enough denoise.
>>106590577get a job, hombre
>>106590595How low denoise do you typically set for second pass if you're just looking to upscale / fix small defects ?
>>106590577Not worth it
>>106590577did they release the 2.2 vace model or are you talking about "vace fun". i thought that just plugs into the existing t2v model like a lora
>>106590595Alright, cool, time to hopefully elevate my 1girl experience to the next level
>>106590577you mean the vace-fun version? real vace 2.2 is not out yet
>>106590640Idk if all models share this but on chroma 0.65-0.70 for plain upscale and 0.75+ it starts reiterating things and chaning stuff
>>106590464thank the Lord in Heavenbtw 4 days late lol>>106556008>how do I prompt for blow kiss but without the hand movement?>the woman stretches her arms out in front of her as if to give a hug. the camera zooms in on her face. her lips fill the entire screen and she kisses the camera with her lipsthat at least keeps them from blowing a kiss with their hands
>>106590723Thanks man
>>106590749The cigarettes are better
>>106590749the surprise dick in the mouth are better
>>106590803Please Furkan, enough with that
>>106590071>larping as tech libertarians.Keep crying pajeet, the only reason you are seething is cause you don't have the skill and the hardware, go back to your general /aids/ and /dale/, here is not for brown skin poorfags like you
>>106590784Fuck off this is the zoomer thread
>>106590825inb4 lefties claiming local ai generation is white supremacy
>>106590784>>106590803Impossible to please I swear.
>>106590888Now that's more like it
can chroma be prompted to do anime or cartoon illustration?
>>106591089yes of course but the style won't be consistent between seeds
>>106591035>>106591072>>106591111model? It's not cloudshit / seedream, isn't it?
>>106591089yes, pretty extensively.if you want to prompt specific characters with their standard outfits tho you'll usually find that easier on illustrious or noobai
>>106591123worse than that it's qwen but least it's not chroma
>>106591123mindbrokenwhat a great troll that was
>>106591123It's just qwen with an awful whinnie the pooh lora I cam across while looking for porn and thought "why not?"
Still suffering from horrible colorshifts.Makes the whole video genning useless really.
>>106591171Based on everything else here, this seems to be an entirely you issue.
I heard sd3 launched and came to see if it's any good
>>1065911712.2 fast loras caused this last time I tried
>>106591177Yeah and Trump became president AFTER like two assassination attempts, bitcoin passed 100k, Thailand and Burma had a brief conflict. Putin Visited the US. Biden has colon cancer. And SD3 is shit.
go to bed debo
>>106591171how are you specifically struggling so much with video gen. it's the easiest shit in the world
>>106591189welp, back to cave I guess
>>106590738>kisses the cameraincluding this has a 50/50 chance of wan spawning a camera and having her kiss that in my experience
>>106591172>>106591188>>106591196The lightx2v loras introduce motion to the gen, it seems impossible to not use them.The others aren't using first frame last frame loops.
>>106591212I really don't understand how they come out so bad on your end. Maybe switch to Kijai's workflows? They are retard proof.
>>106591171some people can't be helped
>>106591212catbox your workflow and i will tell you how you are fucking up
>>106591211has not happened to me once
>>106591229happens to me almost every time i prompt something involving a camera. most recently i tried prompting for handheld camera movement and it just spawned a camera in
>>106591235It only ever happened to me once on a throwaway gen and it's because I specifically prompted ronald to "bat the camera away with his hand" which it interpreted as a camera appearing to be pushed away.
>>106591251i love that gen. still cracks me up
>>106591220>>106591222Guess I'm retarded, works fine with kijai.>>106591227https://files.catbox.moe/2u1ndd.mp4
>>106591283I just use kij these days desu. Something about comfy workflows that always come out wrong.Which kind of makes sense since Kij is literally just the script wan gives you to run the model and all kij is doing is making nodes in comfy that can interact with that script.
ugh bros my 90s candid amateur out of focus grainy gens are so SOVLFUL!!! chromabros we're so back!!!!!!!!
>>106589872bro just merge the attention layers manually DUH
>>106591283okay yeah nvm aside from the questionable prompting im not sure there's any glaring issues lol. comfy implementation simply might be worse
>>106591235i don't know what you're doing wrong, using a retarded allinone model, or using a <8 cope quant but it just does not happen to me
It seems that when i use more than 3 regions in regional prompter (forge UI), it breaks and only takes the first region into account. any idea what's the problem? I'm sure I used at least 4 regions in the past.
>>106591328nope im just using Q8 wan + 2.1 lightx2v.
>>106591341I can't help myself, the problem is you aren't using comfy UI. I think my head would have exploded if I didn't at least type that. Sorry.
>>106591295>>106591310I should have listened to my own notes. Now that I've swapped back and forth between two different 'Save Video' nodes, the color shift is gone. There has to be some shit going on that glitches it out.
>>106591371Flawless loops now, with lightx2v, REEEE
https://files.catbox.moe/2mxzj8.pnghttps://files.catbox.moe/mfy6zj.mp4
>>106591442The consistency is impressive, some small artifacts in the bodypaint but overall it looks like in game footage
>>106591442ummm metadata?
>>106591111If only the bears weren't such bearlets.
>>106591474Whinnie the pooh is a canonical bearlet. It can't be helped.
Is it possible to save each new gen into it's own new folder? I realized with png, I can get rid of even more distortion in a loop.
>>106591458this shit to me a whole 45 minutes to get it decently right. very frustrating trying to do convenient censorship just right enough to avoid triggering jannies with the ban hammer.this is the original lora if curious. https://civitai.com/models/1714926/tomb-raider-lara-croft-survivor>>106591470drag the image to png info on forge anon. video metadata is in its comments. drag it to wan2gp.
>>106591496filename_prefix can do https://blenderneko.github.io/ComfyUI-docs/Interface/SaveFileFormatting/ and also you can do folder/subfolder/[...]/filename
>>106591442*steals your workflow.*
>>106591550Thumbs up!
>>106591341>>106591352ok it works in a1111 so it's a problem with forge. using comfy is out of the question so i guess i'll try reforge
What samplers do you all use for wan? I've been using res_multistep and while good it makes gens look sorta plasticy
>>106591647>it makes gens look sorta plasticyThat would be the 4step LoRAs
>>106591652nunchaku wan never ever
https://huggingface.co/bytedance-research/UMObytedance has released the full model of UMO (not just the lora)
>>106591722@grok what is this
https://s2guidance.github.io/Babe wake up, it's time for Alibaba to go for a new cope, "The next replacement of CFG(TM)1!1!!!1!"
>>106591722I highly doubt ByteDance is going to give local a model like Seedream. It's either going to be Wan or Qwen that does it. Probably Wan since they don't go for slop evals.
>>106591731>THING is all you needWhenever I see this I assume it's trash.
>>106591740idk alibaba has started to keep some things behind closed doors lately
>>106591759>lately
>>106591759>alibaba has started to keep some things behind closed doors latelywhat happened?
>>106591770Qwen has started keeping certain models api-only
>>106591771>it's beginingit was bound to happen, they're starting to get models that can be serious rivals to the best API ones, no way they're gonna release SOTA models, no one is doing that, ever
>>106591731interesting paper, but id like all these bold claims to be backed up by actual code.
https://xcancel.com/bdsqlsz/status/1967431792992129065#mHis English is rough, but if I understand correctly, in a week we will have a new editing model and another video model.
>>106591795that's why europe got rid of most of the bears
>>106591731that's funny, all the other replacements cope of CFG were actually making it worse than CFG itself
>>106591845Did they get nostalgic and decide to mass reintroduce wild creatures to assault the local population?
>>106591731I want to say "nothingburger" but since it's from Alibaba I want to believe, so far they showed that they are a serious company.
>>106591558It's saving to a folder, but each subsequent gen goes into the same folder. For png renders that's going to be a mess, hence a new folder for each time you gen.I'm a complete beginner so I can't make sense of that page.
>>106591731>We use a De-distilled version of Flux Labs (2024) in our experiments.excuse me? how did they get that? I doubt they collaborated with BFL so I guess they used this model? https://huggingface.co/nyanko7/flux-dev-de-distill
>>106591855italy and romania and probably others had a number of bear situations where people died without having done anything particularly stupid, yes
>>106591873Change your "filename_prefix" field from AnimateDiff to %date:yyMMdd-hhmmss%/AnimateDiff
>>106591896>how did they get that?Seems they give it out to anyone so long as you're not a filthy local shitter.
>>106591911>as you're not a filthy local shitter.that's why you don't have it either :(
>>106591873the page explains that you can use %node_name.widget_name% or %date:FORMAT% to define the foldername it goes to.use that for a foldername with the date and seed for example
>>106591910Hell yeah, that worked. I bring you raunchy frieren and orc wip as thanks: https://files.catbox.moe/ze7qky.mp4>>106591922Ah so the date format decides each new folder, because it's counted in seconds. If I were to keep it just to the days, everything I gen goes into that one folder for the day?
>>106591731>https://s2guidance.github.io/The outputs look a little deep fried to me desu.
>>106591974it does, everytime a "replacement" of CFG comes in, it's always some ultra slopped, ultra fried shit (but it follows the prompt better though !!!!)
>>106591948>Ah so the date format decides each new folder, because it's counted in seconds. If I were to keep it just to the days, everything I gen goes into that one folder for the day?yes. and as it says with %node_name.widget_name% you could also use any other information from any other node, such as a seed or other random number from a random number generating node
>>106591982>Makes the image better by destroying it.Can't wait to play with it for a few hours and never use it again.
>>106591442>>106591567that's a good idea
>>106591989The widget name would be the parameter inside of that named node, for example %Ksampler:noise_seed%?
>>106591997>webmThat's quite the glowup of her.
>>106591997Why is she so captivating?
>>106591997She could have a future in porn
>>106592046Wasn't her father like very high up at Goldman Sachs?She never ever has to worry about money. Ever.
>>106591969
>>106592064>Bomb expert, dual bomb expert to be exact
Spent some time with SRPOTerrible prompt comprehension and medium knowledge compared to Chroma or even Qwen. Half the time it just ignores a chunk of my prompt. Still slops hands regularly. Can only decently do 3d, but is completely incapable of good nsfw. Beyond faster generation, i honestly don't understand what the fuck is even the point of that finetune is or why it has been shilled around lately.Pic related is how SRPO understands OIL PAINTING.
>>106592064>>106591997>>106591567you guys are getting it all wrong, i trying tooth and nails to get the camera to do a 360 degree orbit around the subject and not the subject doing a 360 body spinning in from of a static camera. This shit is pissing me off to no end.
>>106592086SRPO seems to be a good method to unslop renders, but doing it on Flux was a retarded move, you can't save Flux it's obvious at this point, can't wait to see them try on qwen image though
>>106592107Wasn't the SRPO guys from Alibaba, if so why didn't they do it on Qwen to begin with ?
>>106592093damn I forgot about her game, should I buy some lube to play it?
>>106592093what's your prompt for this? it does seem to ignore "camera orbiting around character" prompt a lot
>>106592114no, SRPO was invented by Tencent
>>106592003i believe it is %KSampler.noise_seed% but you got the concept right.i'm using another node than the vanilla KSampler so I can't definitely check it
>>106592118Ahh, that makes more sense then
Is it possible to pass an existing video into an img2vid generation with Wan2.2? I try doing it, but it fucks with the colors in a way that regular img2vid generations don't do.
>>106592179What are you trying to achieve exactly?
>>106592039Because you're white
>>106591810Why so much focus on video and imagen now? Not complaining of course.
>>106591997Can you post the starting pic?
>>106592249They already have a niche cut out in the llm space. Now they are trying to cut out one in the image gen space. They throw us the failed attempts along the way.
>>106592249Whatever tech is easiest to improve on for the return will be the focus. LLMs are now in an incremental codemaxxing era while there's a lot to improve on in video gen
>>106592253
>>106592262>LLMs are now in an incremental codemaxxing eraI hate this. I just want one where its reward training is extracting semen from balls.
>>106592210I've got a video with glitchy eyes, and I'd like to pass it through the refiner to see if it gets better. It does get better, but it fucks with the colors.
>>106592267You can't just pass it through T2V at a very low denoise?
>>106592116prompt "The camera is orbiting 360 degrees around the woman's showing the viewer her side, back, other side of body before completing the full rotation back to her starting position. The 360 degree camera orbit around the woman's body is fast and smooth. The lighting is cinematic and dramatic, with soft shadows and realistic detail." good luck getting this work, frankly after 6 hours straight with multiple failures, i give up. here a catbox for a spicy gen attempt.https://files.catbox.moe/764uam.mp4
After trying video gen for a couple days. My respect for blacked content has increased. It's impossible to get it working right without a shit ton of loras and different strengths etc etc. for regular sex it's so easy just plug in and go. I wish black girls were more attractive then because I just want contrasting content during sex.
>>106592256>They throw us the failed attempts along the way.feelsbad being a local fag, because this is true, look at bytedance, they failed with UMO so they gave it to us, and once they struck gold with seedream 4.0 they kept it for themselves >>106591722
>>106592272I've tried that, but it changes the rest of the image too much, even at extremely low deneoise, ie. 0.05.
>>106592288Okay, and you tried i2v but with the first frame of the video as the input and vae encoded the broken video into the low noise sampler? Did you add noise to the samples?
>>106592284Genuinely curious as to what makes seed dream so good. I'm not sure Ive even seen an output from it.
Image models from 2021 (sovlMaxxing) vs image models from 2025 (slomMaxxing)
>>106592323>Genuinely curious as to what makes seed dream so good.they released the paperhttps://xcancel.com/bdsqlsz/status/1966034419183124527#mhttps://arxiv.org/abs/2509.08826
>>106592314Yeah, and that fucks with the colors, unfortunately.
/adt/ got deleted
>>106592333Yeah, but I've never seen an output from it.
>>106592323>I'm not sure Ive even seen an output from it.>>106577845>>106578184>>106576677>>106576615
>>106592368Hmm well, they are pretty sharp.
>>106592267read this, you can use a segment model to mask a face -> crop around it and upscale before sending it to the low model for a 1 step i2i and paste it back in>https://www.notion.so/bedovyy/WanFaceDetailer-261ce80b3952805f8aaefb1cdb90ec04
>>106592390I'll look into it. Thanks for sticking around and giving me a way forward, whether it pans out or not.
Can someone help me figure out what to try next?https://files.catbox.moe/fal8ln.mp4Just want the dude to be black. and unfortunately neg prompts don't work for me and this setup.
>>106592390That is interesting.
>>106592408>Just want the dude to be black
>>106592408pajeet you gotta master the english language before you can master prompting, so things like negro, basketball american, breathing simulator 9000 all evoke images of 'black skinned' people in the ai's mind
>>106592306amazing consistency
>>106592390I wonder how this actually works for non-anime stuff.
>>106592327>Image models from 2021there was none
>>106592408>Just want the dude to be black.a BBC enjoyer I see https://www.youtube.com/watch?v=oNsNjMuevXw
>>106592340That flame, else it's really good
>>106592327What's the actual difference here, beyond the tan and him getting older?
>>106592466A man oblivious to the concept of plastic surgery.
>>106592472>>106592466he was in a motorcycle accident and needed facial reconstruction.
>>106592466You can't see his cheeks, lips and chin have been bogged to hell and back?Is this that face blindness austists talk about?
>>1065924592021 was the year this paper predicted that the future of image models would be diffusion models.https://youtu.be/W-O7AZNzbzQ?t=3235https://arxiv.org/abs/2105.05233
>>106592408Wan BLACKED lora when?
>>106592466>>106592480>You can't see his cheeks, lips and chin have been bogged to hell and back?>Is this that face blindness austists talk about?those are the same "people" who see no problem with Chroma btw
>106592491obsessed
>>106592480I can see there being lip filler in 2025, but he already looks bogged in 2021. I don't see much difference in the other features that couldn't be explained by aging, a wider smile, and weight gain.
>>106592486tried one and it made weird 4 legged, cock and pussy monsters fucking.
>>106591810Surely they aren't releasing inferior model to what we have. If they ain't on par with at least Qwen edit and Wan 2.2, nobody is gonna use them, kinda like omnigen 2, Hunyuan i2v and so on
>>106592476>he was in a motorcycle accident and needed facial reconstruction.it was in 2013 though, he looked fine before the 2020's, he started looking like bog way after that
>>106592264Thankshttps://files.catbox.moe/a9w5zo.mp4
>>106592503they don't care that they don't compete with local SOTA, they just want some good boi points and be treated like the "nice guy company", optics are important, especially for investors
>image2image>staring images is a white man on the button of the woman.>10 lora setupperhaps you sittings and start image is bad to begin with.
>>106592466>What's the actual difference here
>>106592446can't believe basketball player worked...almost there but the penis is all wrong https://files.catbox.moe/mus9ye.mp4
>>106592466Left is normal, right is someone using img2img on his face
>>106592466left is training a model with real data, right is training the same model with synthetic data
>>106592571try using "bestiality" next time in the prompt to get that skin
>>106592509>in an AI thread>can't tell when something is obviously using AI and/or photoshopgrim
>>106588114Can I get a box for this?
>>106592390I can't be assed to download all of the segmentation models for this right now, but as far as I can tell, it just takes an input video. Segments and boxes the faces, upscales them and denoises them at a higher detail the pastes them back over the video, right?I think the most interesting takeaway is that he uses causevid for the lowpass.
>>106592613unfortunately, that's the real face of Zac nowdays, what a waste...
>>106592624Did you train the lora on Mickey Rourke's face
>>106592605K i'll try it next. Hell if dog knots is easier than black guys I'll go that route instead
>>106592282you have to be autistic with prompting with wanmention all four angles step by step, mention how the background moves step by step
>>106590560That's how I keep consistency for characters in my project. I started with one good image and then do gens with her changing positions then use that as a starting image, it's not perfect but better than the gacha of trying to gen the exact same outfit/appearance again. Sometimes I'll run the final image through img2img in an image model or to refine it. The main drawback is the resolution is a bit low so if the character isn't close up close you can lose some details.
what lora can i use in wan2.2 to pull their top down and expose breasts?
>>106592886>The woman pulls down her shirt, exposing her breastsNo lora needed.
>>106592886FLF
>>106592889i had >she exposes her large breasts at the startand nothing. I'll try this
https://files.catbox.moe/2tyce3.png
>>106592811Just turned off my pc due too hearing weird noises with my 5090 and i gotta sleep. Please Post the entire prompt positive and negative prompts. I suffered seven hours straight with multiple failured gens and feel very burnt out. Help an anon out please :'). Please do a 360 orbit of pic related.
>>106592971prompt>The camera is orbiting 360 degrees around the girl showing the viewer her left side, the background moves showing the left side of the room, wall. Then the camera continues to rotate showing the viewer her back, the background moves showing the back side of the room, audiences. Then the camera continues to rotate showing the viewer her right side, the background moves showing the right side of the room, wall. Then the camera continues to rotate showing the viewer her front, the background moves showing the front side of the room.
https://files.catbox.moe/gzd7ba.png
>>106592971>>106592985also use first and last frame and set the two frames to the same image
https://files.catbox.moe/v5c1s5.png
lmk if such type of gore is allowed here, idk how to spoiler images https://files.catbox.moe/n9fkqr.png
>>106592985>>106592997Will test this out later, thank you very much anon.
https://files.catbox.moe/kpdgky.pngCatpcha:YGANG
>Chroma-DC-2K-T2-SL4These niggas will train anything except the qwen text encoder for it. Someone stop them
>>106590405>As long as the total amount of pixels is the same and the dimensions are a power of 64I'm pretty sure SDXL isn't trained on 1280x768. It means a lot what specific resolutions it's trained on. SDXL is trained on multiple different resolutions. within the 2048 pixel plane meaning 1024x1024 and 1216x832 and some others i believe; you can look it up. You're going to get the best results if you stick to the exact dimensions it was trained on.It's the same way where you get very bad results with SD1.5 if you make it any other dimension than 512x512.The training dimensions are kind of hardcoded into the model, and when you don't follow them, you kind of warp the vector space and it's associations with the pixel space. You can generate at the native resolutions, and then upscale to a high resolution and then you can downsize and crop the images later. That's the way you'll get the best qualitative outcome with the current models.
guys which joycaption nodes to use?
>>106592889>>106592605Thanks! it works way better than what I had originally. still need to work on the legs, either the guy has legs coming out of his hips to the side, or the girl is missing the bottom half of her legshttps://files.catbox.moe/ke3kv9.mp4
>heh I haven't updated comfy in weeks>update>now all my gens come out a blurry messI'm fed up
>>106593201turn off fast optimizations retard
>>106593211There are no fast optimizations, retard.
>>106593226Turn off your eyeballshttps://files.catbox.moe/ct8jyn.png
If i want to continue a video, do i use a different prompt? I used the same prompt as the initial and it just. slowed down and barely moved.
I enjoy Chroma.
i enjoi haveng extra chromasomee..
>>106592825do you think ani is on the right track for making a game engine with diffusion mechanics built in? you are the only anon I know that's making a game
>>106593279ymmv, generally the same prompt shouldn't do LESS on averagemaybe you have too many speedup things eabled that interfere with motion, or not enough steps, or maybe you want to try the hps/mps reward or movement lora, or other things
>>106593149>joycaption>nodeslol ur gay
>>106589978>Surely all those server farms aren't built off GPUs right?Yes they are, but it's not the same GPU's you can buy to put in your PC. If you watch some of the nVidia presentations, you'll see that they're super huge, and they're making them bigger and bigger each time.nVidias market is no longer really focused on consumer GPU's. They're more so in the business of designing custom systems for big businesses that need datacenters and software solutions for training and analyzing all kinds of stuff with AI.
>>106593149Just get taggui
Specs: 32GB RAM, 12GB GPU, i12900k Running ComfyUI with SDXL realisticslopSwitched from Forge to Comfy recently. Two questions:Is Comfy actually faster for gen/checkpoint loading than Forge or just me?Anyone else notice Comfy outputs seem slightly softer/less sharp/quality?
>>106593343this anon just mentioned it >>106593201
>>106593349I have to move back to Forge?
>>106593149probably decide between some of the most recently updated
>>106593201>he pulled
>>106590090You know you can use quantized versions right?
>>106593334I just love noodles, what can I say>>106593338id rather not have YET another conda enviro PLEASE>>106593362I wanted to go to a 'mostly' generic route using a generic LLAVA wrapper which uses llama-cpp-python, but the generic vision nodes were not updated recently, the other nodes have descriptions all written and chink and id rather xi not see what im captioning
so different seeds can give you a completely fucked unusable gen? is there any way to know how bad the gens will be? I did 2 gens and one had a random mystery guy added behind the girl and the other was perfect.
>>106593426xi has a "few" chinese that can write english - but the chinese sometimes write in their language.just ignore it as long as you can use it.
>>106593510depending on model and prompt sureyou can't know in advance unless you use a very special model type but you can generate preiews with like TAESD on each step as it's crunching the tensors... on most models at least
>>106593529yeah i have previews on but by the time i can tell if it's fucked or not it's too late to abort,
Qwen SRPO when.
>>106590560>my boy discovering that wan is the best edit model
new>>106593668>>106593668>>106593668>>106593668
>>106593319I don't know the specifics of what he's doing, is it just allowing the devs to create a prompt along with parameters to generate images in game? I think using image/video gen at runtime will be pretty common eventually so starting something like that for games to come out a few years from now is probably a good idea. I don't think it's super viable right now because it's too slow or just not possible on the average consumer's hardware. Players won't like waiting a minute for a scene to generate. There's also the issues with discontinuity but maybe it will be solved in future models or players just won't care.
>>106593537i don't see how there could be a way that lets you see it even earlier
>>106593319i think trani (read, you) should kill himself immediately