Discussion of Free and Open Source Text-to-Image/Video ModelsPrev: >>107114476https://rentry.org/ldg-lazy-getting-started-guide>UIComfyUI: https://github.com/comfyanonymous/ComfyUISwarmUI: https://github.com/mcmonkeyprojects/SwarmUIre/Forge/Classic/Neo: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassicneoSD.Next: https://github.com/vladmandic/sdnextWan2GP: https://github.com/deepbeepmeep/Wan2GP>Checkpoints, LoRAs, Upscalers, & Workflowshttps://civitai.comhttps://civitaiarchive.com/https://openmodeldb.infohttps://openart.ai/workflows>Tuninghttps://github.com/spacepxl/demystifying-sd-finetuninghttps://github.com/Nerogar/OneTrainerhttps://github.com/kohya-ss/sd-scriptshttps://github.com/tdrussell/diffusion-pipe>WanXhttps://comfyanonymous.github.io/ComfyUI_examples/wan22/https://github.com/Wan-Video>Neta Yume (Lumina 2)https://civitai.com/models/1790792?modelVersionId=2298660https://nieta-art.feishu.cn/wiki/RY3GwpT59icIQlkWXEfcCqIMnQdhttps://gumgum10.github.io/gumgum.github.io/https://neta-lumina-style.tz03.xyz/https://huggingface.co/neta-art/Neta-Lumina>Chromahttps://huggingface.co/lodestones/Chroma1-BaseTraining: https://rentry.org/mvu52t46>Illustrious1girl and Beyond: https://rentry.org/comfyui_guide_1girlTag Explorer: https://tagexplorer.github.io/>MiscLocal Model Meta: https://rentry.org/localmodelsmetaShare Metadata: https://catbox.moe | https://litterbox.catbox.moe/GPU Benchmarks: https://chimolog.co/bto-gpu-stable-diffusion-specs/Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-oneTxt2Img Plugin: https://github.com/Acly/krita-ai-diffusionArchive: https://rentry.org/sdg-linkBakery: https://rentry.org/ldgcollage>Neighbors>>>/aco/csdg>>>/b/degen>>>/b/realistic+parody>>>/gif/vdg>>>/d/ddg>>>/e/edg>>>/h/hdg>>>/trash/slop>>>/vt/vtai>>>/u/udg>Local Text>>>/g/lmg>Maintain Thread Qualityhttps://rentry.org/debo
I am of the strong conviction that comfy should be dragged out on the street and shot.
@comfyMaybe it would be a good idea in comfyui to have an option to precompute all prompts first, which if enabled would then do that first and thus ensure a HUGE speedup given you can throw away the text encoder forever out of VRAM.Hugely speeding up the prompt computation since initially the model used to compute the prompt will be permanently loaded in VRAM and process everything in one go.And then later on you will also permanently now have the diffusion model in VRAM and computing the images one after another.Never needing to waste huge time swapping from RAM (or God forbid, disk) to VRAM, TWICE for every single generation.This is a huge, double digit % speedup for most newer models. And would also allow people to load older models at maximum precision too.A simple "Process all prompts first" switch which enables this in the settings would be great.I also don't see why this wouldn't be a default option anyway, this would hugely speed up anyone who has more than 1 thing queued up, which is especially true for enterprise users which comfyui wants to cater to anyway.
>I listened to anon and upgraded from ddr4 32gb to 96gb in August
>>107123451>not even 128gbgrim
sadly my MB is full 4x16 so upgrading the ddr5 is going to cost an arm and a leg unless some falls off a lorry.
>>107123447Shit just sucks.
Death to all jeets.
til bghira lurks itt
Shill me an underrated model nobody here talks about.
>>107123797good goy
>>107124009NetaYume
>>107124085>NetaYume>nobody here talks about.
OK the "baker is anti-realism" complaints were right, because the snubs this time are too much
>>107124297baker is a fag
>>107124297>>107124334the baker is based. the only fags here are you whiny bitches.
>>107124362sorry you were outed as a fag, not my problem
>>107124297>the snubs this time are too muchAI "artists" really are as faggoty as real artists
what I actually want from ai is not something that makes me go "wow, that's plausibly human made". I want something that shines sparks into those dim dusty corners of my imagination
>>107124657if even sd 1.4 didnt already do this, you are a brainlet
>>107124635test moar!
>>107124956>cigger
>>107124956TWO cigars? now that's smoking
>update comfyui>rerun an image from yesterday>comfyui crashes on second pass VAE decode>no errorscomfy pls
>>107125211Comfy must be dragged into the streets and shot
>>107125219this is a factually correct statement
>>107125219>>107125281Why not make something better?
any new models/loras this week? still having fun with qwen edit v2 which is amazing for what it does. plus wan/illustrious are good.
>>107124009>Flux.1 [Dev]I'm pretty sure that I'm the only one here that actually uses it and I don't post all that often. Seems like there's always something new added to Flux to play with. Great if you love to tinker.
>>107125070Shes in [flavour] country
>>107125400That image I posted was a flux merge too.>Seems like there's always something new added to Flux to play with. Are you referring to something specific?
>>107124956>no mayli
I love 1girl gens.Please make and post more 1girl gens.
cozy
>>1071256912girl or nothing
This could have been cool if not for the slop hands>>107125781There are technically two girls here.
>>107124297there are realistic gens in op all the time tho even ITT
>>107124009SD3.5M SD1.4
>>107124297>>107124334>>107124362Wait baker is a person who does this for every single thread?I thought it was an automated bot.
>>107125322>any new loras this week?Just make your own
>>107125840...no it's several people who arbitrarily pick gens ffs
>>107125834>SD3.5MI actually occasionally use this when I am bored. It's mediocre in certain regards but can make nice images with some luck still.>SD1.4Isn't it a bit too ancient? What would you use this for in 2025?
>>107125875based reply and gen
Slopfest
>>107126058Slop boardSlop generalSlop threadSlop posts
>>107126058>Slopfest
>>107126078
>>107126058all AI gens are slop
>>107126399not mine mine are special
>>107126432they are. we love your work.
>>107126427A mix of detail and lack thereof. Are you making a statement about the duality of man?
>>107123447unfathomably based
>>107126707Welcome back, Master.That's Richard Stalinman in the background.
wan 3.0 when
>>107125820Only when they are basically "memes" with some obvious visual gag or gimmick. He never just likes a realistic 1girl for being nice to look at. His taste in realism is basically the same as model devs when they put out papers with demonstration gens of a puppy holding a lollipop and wearing a party hat
>>107123447I think you should be dragged out in the streets and lightly flogged in front of a crowd until you start blubbering and crying, and then you should be let go
>>107126664thats noob's idea of depth of field, sometimes it works well but there is does make the background into a mess.
Are we back? Did the schizo lose? If yes, poor faggot wasted tons of effort doing this.
>>107127786Dont give any more attention to the shitter. Anyway, new video model drop!https://github.com/FoundationVision/InfinityStar>InfinityStar by Bytedance: A unified 8B spacetime autoregressive model for high-res image & video gen; >- 5s 720p video ~10x faster than DiT; >- scores 83.74 on VBench, topping other AR models and HunyuanVideo; >- Flan-T5-XL as text encoder. >- 480/720p, ~35Gb model
>>107127895>8BGarbage. If it doesn't have more than 14B(Wan2.2) it's not worth using.
>>107127895He needs to sleep for like 2 days straight because it wasn't just the bots running but he was also posting actively too, absolute madness.
>>107127895>We provide a demo website >look inside >17k member discord
>>107126276>ai is better than gimp
>>107128043anything on huggingface to bypass this nonsense?
>>107128043>>107128053Yeah its a bit retarded that they are locking their demos behind shitcord. Here's what it looks like, posted on xitter https://xcancel.com/wildmindai/status/1986502031532826776
it's wan2.2 with lightx2v 4 steps utter unusable dogshit, or am I doing something wrong?
>>107127419>>107114476>>107108437>>107102952
What are the most important qwen image edit loras in your opinion?
>>107128083Depends on what you're trying to gen. Sometimes, for faster motion for example, you can set cfg to 1.5 or do some retarded lora slop stacking ( wan 2.1 lightx2v 3 str high and 0.25 str low, wan 2.2 kijai lightning MOE high and wan 2.2 regular lighting low). Then again, those settings tend to change the face and details of the input image. Can slop in some PUSA loras around .5 or .6 strength to keep consistency, sorta.
>>107128181tested high=2, low=1, and it got a lot betteris that ok? what lora values are best?
>>107128074it looks significantly better than wan 2.2. we'll see if kijai releases anything
>>107128132For me its some kind of troll, he put a couple of gens of mine with nano banan and an anime screenshot from google images, but his collage meme idea is good though, not everything has to be serious, epic or coomerbait. Sometimes a meme for a light laugh is nice
>>107128181is pusa a lora, or a category of loras?
>>107128206Cant say there is a "best" values, not that I'm aware of anyway. Best to just tinker around with the strengths and find out what works best for whatever you're trying to make.
>>107128335Its a lora, read these and models are therewan 2.2 https://huggingface.co/RaphaelLiu/Pusa-Wan2.2-V1wan 2.1 https://huggingface.co/RaphaelLiu/PusaV1
>>107128382alright, thanks, sempai
>>107128398Also forgot Kijais version because I know someone is going to bitch about it. Also, I dont know the difference but someone will kekhttps://huggingface.co/Kijai/WanVideo_comfy/tree/main/Pusa
>>107124009>nobody here talks about.hard since anon talks about every model
>>107128231It does look pretty good. It says it can do 5 to 10 seconds apparently. They're boasting 10x speed which we'll need some kind of comparison. Also, I noticed this model doesn't seem to be based on wan, so it's 100% its own thing? Good to see more contenders, hopefully we can see more demos and such. Plus LTX2 this month, I think.>distant screams from kijai can be heard
>>107124009>>107128555How about animatediff? Fun model, can gen some pretty stupid sloppy stuff. Maybe there is some kind of way to lower the memory to let wan gen something crazy like 1000 frames. We can do this now but you need a super stupid amount of vram. When I was tinkering with animatediff, 1000 frames was no problem, started running into oom at 1400 frames. Slap on the SVI lora and we can finally escape 5 second hell properly (no more color shifting, jank movement, 81 frame stitch tricks).
Seems like about 1 million pixels resolution is the sweet spot for WAN. Body shape start to deform when you go higher pixel count
>>107128231>it looks significantly better than wan 2.2idk, i think people dont really know how good maxxed out quality wan 2.2 looks like, especially nowadays with actually good lightning loras that everyone is using, making them forget how good wan is for physics and motion without the added lightning stiffnessit does seem like the dataset is wider, for realistic gens at least, the background details are better, the look is less slopped out of the box, we'll see
>>107128750video? it was trained on 720 so yeah, works fine for images up to 1.6MP but good luck making videos at that res
I'm wanproompting.
>>107128825Looks great, catbox?Is wan prompted with essays or?
>>107128842Yeah, here is the prompt"In the flickering glow of candlelight, a grand medieval feast unfolds within the stone-walled hall. The air hums softly with the murmur of voices, the clinking of goblets, and the faint crackle of the hearth nearby. At the center of the long wooden table sits a noblewoman, her presence commanding yet graceful. She wears an elegant white gown adorned with golden pauldrons upon her shoulders, their polished surfaces catching the warm light. A delicate belt of gilt embroidery cinches her waist, and a shimmering crown rests atop her carefully styled hair — a symbol of her royal station.Before her lies a modest yet hearty meal: thick slices of bread, roasted vegetables, and tender cuts of meat, arranged on a simple ceramic plate. A goblet of drink and a small bowl of rich sauce complete the setting, evoking a time when feasts were both ritual and celebration. Her posture is poised, her hands folded lightly before her as she offers a knowing smile, one touched with confidence and quiet amusement — as though she is aware of the attention she commands.Beside her, an older kingly figure, clad in regal black with golden trim and a matching crown, partakes in his meal with solemn dignity. The candle between them burns low, its flame dancing gently in the draft of the old hall, casting shifting shadows across the brick and timber walls.The moment feels timeless — a captured fragment of courtly life where elegance, power, and intimacy blend seamlessly beneath the watchful eyes of flickering candlelight."
>>107128825> wanToo bad t2i doesn't work with t2v loras.
>>107128857What frontend/workflow?
>>107128842>>107128857But then O had to inpaint her face so my promtp to inapint her face was"Her face is softly illuminated by the flickering candlelight, giving her a warm, almost ethereal glow. Delicate features frame her serene expression — graceful and composed, yet touched with a subtle confidence. Her fair hair is elegantly styled into an intricate updo, a few loose strands framing her face in gentle wisps that soften her regal poise. The golden crown resting upon her head glimmers faintly, its ornate design accentuating the nobility in her bearing. Her eyes, bright and keen, carry both warmth and a quiet command, while a faint, knowing smile curves her lips — the kind that hints at secrets and strength beneath her gentle exterior. Adorned with pearls and a cross pendant that rests gracefully at her throat, her face embodies a blend of royalty and radiance, timeless in its allure."
>>107128825>>107128857>>107128861>>107128873But then her handa where melted and i have to inpaint her hands and it was"Both of her arms are bent at the elbows and rest gently on the edge of the table. Her forearms are crossed lightly in front of her, creating a subtle, poised frame against her white gown. The soft fabric of her sleeves gathers slightly at her wrists, emphasizing the elegance of her position. Her hands rest close together, fingers relaxed, giving an impression of calm confidence and refined poise — a gesture both demure and self-assured, fitting for a regal setting."I love prooosing so much, totally intuitive and practical
>train lora on a bunch of greenscreened characters>generate character>use wan to animate a looping idle pose>then animate attack animations, etc. that loop back to the base image so everything transitions seamlessly>remove green background with video editorMaybe you'll need to train a lora on the animations as well but I think this could be doable for making sprites
Ney'liri is back.
What is the best settings for OneTrainer to train Pixel artThis is the best i could give for my Illus LORA Spritesheet trainingIts really blury
>>107128132He has shit taste. It is what it is.
Not sure if anyone else uses this but SuperUltimateVaceTools got an update, apparently works with wan 2.2 fun vace https://github.com/bbaudio-2025/ComfyUI-SuperUltimateVaceTools/tree/main Was pretty decent with wan 2.1, gonna test soon if not this weekend
Have there been any developments in frame interpolation? I've been using RIFE then falling back to FILM VFI if the RIFE output has too much artifacting (usually from fast motion). Has there been anything since those two models? They're both getting pretty old.
>>107129258Based and pixai pilled
quick questionI'm messing around with wan2.2 on my shitty 12gb vram setup and looking at the total committed memory usage it's pushing 70 gigs all togetherhow are 24gb cards able to run the full size, unquantified models if even these neutered, 10GB quantified models are taking so much memory?
I got a 4090 and 32GB of RAM. I've been playing around with ComfyUI for the last couple of days.I sorta have an idea of what happening but I'm still bad and either using default workflows or have to copy what other people give. Problem is no matter what I do I cannot get remotely decent images that are of any good quality. When I use SDXL everything looks bad, even when I try to play with the parameters shown to me (I am just opening the basic text to image workflow and placing the model in and playing around with it.) The default workflow for Flux kinda works ok but like everything else I'm basically have no idea how to use.And if I try to make any videos with WAN 2.2 14B when the models start loading it starts thrashing and SSD writing to my disc like mad. If I try to make any 1280x720p video my computer just runs out of memory. Should I upgrade to 64GB RAM because my disc is thrashing and my RAM usage is basically full when trying Using the 480p fast i2v workflow and models here https://rentry.org/wan22ldgguide I can make 480x480 videos and it takes about 2min/5seconds but they look like ass even with real life photos used for reference(and half of them don't even follow the prompt I give like "walking away" turns into walking towards the camera)I just want to be able to create videos and pictures that look anywhere near as good as what I see online or can use Grok to generate in fractions of the time for free.tldr; Without spending the next 4 weeks learning and reading everything about comfy and learning to make my own nodes and being afraid I'm getting hacked by some obscure workflow I got off civitai, how do I make clean AI images and videos on my 4090+32GB RAM? Or is everything I'm seeing online using H100s type hardware? Also I am in general interesting in learning all this stuff as well and ComfyUI seems like it has a lot of power so I will learn it all eventually, but where is a reasonable place to start to see what my hardware is capable of?
>>107129578 >:3Saved
Alright we're back. Time to bake a new thread and forget that dogshit bot arc ever happened! Lets choose our best 1girls and make the collage!
>>107129559Also, do any of these other webuis have upscaling and frame interpolation built in? Every single video game has DLSS and FG these days, but for these static workflows these upscaling and interpolation is actually desired.
>>107129607SARR PLEASE MY WIFE WONT STOP!>>29756481
>>107129295Help me on this...
>anistudio filenamebased>asuka brapbased>gemini watermark cringe
>>107129559Start collecting images with metadata that you like and check how people made them.A 4090 and 32gb should be able to do anything image related. Video needs at least 64GB for model off loading purposes, otherwise your system might stall at 32GB ram. Try ultimate upscaler node or film vfi for interpolation.
>>107129611Video games have motion vectors to help with interpolation, that isn't a fair comparison
>>107129618Post dataset then I might have an insight. Garbage in, garbage out after all.
>>107129295It's not good with pixel art at all without another plugin pixelizing it afterward. You'll always get smudges or uneven pixels.
>>107129630I forget to put watermak in negative prompts :(
>>10712963632GB RAM sounds pretty painful with larger models like qwen or flux. anon should just upgrade to 64GB, that is really the minimum for an AI build. buy now before the price doubles again
Relative newbie if someone could indulge me.Is Wan2.X moreso used for generating video? I'm looking for recs on img2img.Does anyone know where I can find a good Comfy workflow for img2img?Recommendations for realistic image gen models?
hyped for dype
>>107129646https://www.mediafire.com/file/2qrfav0ct380pym/MMBN_Spritesheet_Battle.7z/file>>107129658Well, i used Unfake pixel for other spritesheet, and its work fineThe problem is this was too blury, especially on the face
>>107129721Based
>>107129745Your problem is that VAE is extremely lossy and cannot retain any information for individual sprite frame at 1024x1024 resolution. I have no viable solution for that old style sprite sheet. Perhaps try a newer style like Unicorn Overlord.
Is comfyui safe to update to the latest version?
>>107129859No, its download dolphin porn
>>107129636>>107129667I'm guessing it will help with the disc thrashing. Using high quality images will produce higher quality videos too right? Is security a thing to be concerned with comfyui with custom nodes and workflows?I found a workflow for images but with upscaling its taking like 20x longer to generate the image. I guess it's fine if I get higher quality outputs
>>107128857so where did her giant cleavage come from?
>>107129823Nah i just want to see if i can create MMBN Sprite
>>107127895https://xcancel.com/wildmindai/status/1986502031532826776it looks worse than Wan 2.2 so it's kinda useless lol
>>107129895Looks like there are errors https://github.com/comfyanonymous/ComfyUI/issues/10565
>>107129963>Bytedancekek, DOA, they only release failed experiments to the Open, we're just a garbage bin to them
>>107129973Oh sweaty, how are they suppose to make their monies? :)
>>107129980true, think of that poor multi-billion dollar industry :'(
Hello, guys, is there a tutorial on how to install and run llms locally?? how is it done?? anyone has tutorials?
>>107129963https://youtu.be/yg8Qe2-2sqw?t=91there's some more examples here, it's Mid
>>107128043>17k member discordthe fuck? why is there so many people to test that brand new model?
>>107130023Because it's free and supports hindu?
>>107130051>supports hinduwe now live in a world where jeets are being pandered to, this is what happens when you are a race that reproduces like rabbits...
>>107130012InfinityStar? No, it's InfinitySaar.
Is it possible to do video2video inpaint lipsyncing?
>>107130098keek
>>107123451>>107123462TFW old enough to remember when 256MB of RAM made me feel like an elite h4x0r.
>>107130051within a year, any system that caters to indians fails.
>>107130270oh god don't give me hope
<ctrl+f>"AMD">0 resultsuh-oh, doesn't seem like my odds are good.
>>107130023bytedance catchall server
>>107130376o shi-
>>107125691sauce pls. this is beaut!
I started zooming in on webpages a long time ago to save my eyes from the strain of reading small text, so I'm always at like 140%. But it stretches out images, too. And my monitor's pixel density, being a pretty large 1440p monitor, is not so dense. So images look much larger to me than to most. Sometimes I wonder if these factors combined are why I'm able to spot overcooked and AI upscaled gens so much easier than the average poster. I see the problems right away. The texture is off. You can count the realism posters who don't break the illusion on one hand, and most of them don't come here anymore.
when will they make vibevoice but nsfw
https://www.reddit.com/r/StableDiffusion/comments/1oqosui/i_made_the_best_text_encoder_for_qwen_image_edit/>Qwen 2509 is powerful but the default encoder leaves 70% of that power on the table. This unlocks it.goddam I hate those faggots so much
>>107130870lmao this was absolutely the result of begging claude to "make qwen edit more gooder"
>>107130870Fucking pajeets
>>107130645Let's fuck.
So I've been curious maybe you guys will be able to answer this.AI gen seems to be pretty popular these days. So how come every model I've seen so far requires a somewhat significant amount of installing of random scripts, libraries, tertiary apps, etc... How come there hasn't been a simple single installer/application yet? Is it because the environment is too unstable for it to happen yet or what?
>>107131185You generally install only one set of stuff per model type for basic usage and it's automated on some of the possible software.The rest is partly internet attention economy. Enjoy.
>>107131247That's not really what I wanted to know but thanks.
>>107131268The completely "single application" stuff beyond the software that installs its dependencies on the desktop... is an app/website launcher for smartphone users that tries to IaaS/SaaS you.
>>107131302It's cool dude, I already got the answer elsewhere. Like I asked, I wanted to know the technical reason as to why hasn’t anyone managed to ship a local plug-and-play turnkey solution for the average user, what you did is first give me some general knowledge on what to do, and now a vague opinion on what the online alternatives are. I didn't ask either quest, at this point you're either trying to troll me or completely clueless yourself. The answer I was looking for was that there’s no turnkey installer because local AI generation depends on fragile, rapidly changing GPU/toolchain dependencies that break across hardware, drivers, and models faster than a unified app can be maintained.
>>107130678Troll post likely but I fed this Eva Green image to Joycaption and slopped through it."A high-quality photograph of a young woman with pale skin and dark brown, straight hair. She's facing to the left in a three-quarter profile shot. Her lips are painted bright red, and she's wearing bold, gold hoop earrings shaped like abstract loops. The background is a solid light blue gradient that fades into darker blue at the top right corner. She has faint freckles across her cheeks and upper chest. Her shoulders are bare, suggesting she might be wearing an off-the-shoulder top or dress made of black, sheer material. The lighting is even and bright, highlighting her facial features and creating soft shadows on her face and neck. Camera angle is slightly above eye level, capturing the woman from just below her eyes down to her shoulders. Her expression is neutral but somewhat confident. This image is SFW and focuses primarily on close-up details of her face and upper body. The composition style is clean and minimalistic with a focus on the subject's elegant features and bold accessories. Age appears to be in her mid-20s. Overall, the image has a modern, high-fashion vibe. There are no watermarks or additional objects in the frame. "Neither the description is accurate, nor the gen sticks to it.
just repent already
>>107131531>>107131453nice prompt with some minor tweaks
Every day I learn more. Every day the slime pool sluts gen better.
(taking feedback. I have no idea what I'm doing.)
>>107131560>>107131553Cool liminal space
>>107131572Thanks. I think the zoomer "liminal".trend is a little overbaked, but I like the idea. This waterslide zone came to me in a dream one day, and I've been a little enthralled with it since. How do you like the girl?
>>107131328The desktop gui/webui are basically all scripted/automated/with installer. You just can't get everything (models especially) in one tool without paying attention to anyone else.>The answer I was looking for was that there’s no turnkey installer because local AI generation depends on fragile, rapidly changing GPU/toolchain dependencies that break across hardware, drivers, and models faster than a unified app can be maintained.Technology is also a thing why it isn't perfect on *all* devices, but generally speaking you are seeing business attention economy reasons of the kind why there isn't one netflix and why netflix is paid for solutions that are for you, the helpless SaaS/IaaS consumer.
>>107129525Kino gen
>>107131328Spend 3 hours trying to figure out that your cuda doesn't match your pytorch and your flashattention wasn't baked with the right cuda version so pip can't make a (wheel?) for it and then your dependencies are too new, but God forbid if they're too old that's also bad, and sometimes it's just a coin flip.Anyways, you can eventually get fun gens.
>>107131550That's a great gen. Your DOF is a little off though, it's stronger in the front - not your issue though. The only flaw is the black mark on her teeth, but that's an easy edit.
>>107131700That and the obvious earring flaw, heh
>>107131185It's multiple pieces of tech (CUDA, pytorch, numpy, a multitude of python libs) that have dependency/interdependency, and then your image gen software (comfyui for example) will have stricter requirements, and sometimes only a few releases are valid.It really is a house of cards of FOSS tech, but in a way it's great, since it functions as a powerful pleb filter for those that can't help themselves or ask polite and poingent questions to learn. In a past life I was a 3D printing mogul, and I've seen the same thing in the space - let me inb4 though, there are some places that still have high IQ individuals doing really awesome stuff - but with the advent of easy to deploy machines (prusa, ender, and especially Bambu) it's ushered in an era of midwits, jeets, kids, and retards that just shit up the environment... And on to the next niche thing I go.Have a slime pool girl as consolation for reading my life story.
>>107131733You're right. This gen of yours is great - you're selectively hiding the details that could go wrong. Much like a boudoir photographer, you want to hide fat chicks "problem areas" in the shadows or behind clothes/lingerie/blankets.I've done the same with my gens, I specify the hands to be submersed if it's reasonable - it removes a large stumbling block for image gens, much like how a skilled boudoir photographer can minimize problematic areas.
>>107131766heh absolutely. There are limitations we gotta go around
Did he upgrade the bot?
>>107131841If you're talking about me, I'm not a bot. At least I hope not.
>>107131599very plasticy sadly.
>>107131841yeah fuck I just realized, it's 100% a bot, no way a human types like this.
>>107130645Because it's CUDA. And AI implies CUDA.
>>107132008For now.
>>107131976Any advice on de-plasticizing it? I'm using chroma and the Lenovo Lora at 0.95 strength, with an upscale strength of 0.15. the upscale model I used escapes me.Anyone have any names for this young lass trapped in this slimy world?
>>107129453gimm but it's much slower
>>107129520offloading. the full model takes the same amount of time as q8 for me, 24 gb
>>107131976Also interested in how to minimize the plastic look.
Microplastics are everywhere.
>>107132021That now will last very long.
chroma would be perfect if it works with control nets
>>107129559you're using this workflow? play with the block swap setting until you don't max out your vram or it'll stall. you should have about 1gb vram free at all times. I find having ~700mb free is ok for me. if it spikes to max on the "WanVideo ImageToVideo Encode" node that's usually fineI think the issue is though that you're maxing out your sysram. 32gb is really not enough and ram prices just doubled recently. sorry ramlet bro
all my qwen loras fucking suck.I am using Ostris toolkit default settings and 4bit quantizations, 0.0001 LR
>>107132221Go for 0.0002 LR
>>107132139>kijai wan CRONGE
>>107132323not what I'm using but it's in the shitty guide that no one wants to update
>>107131841AAAAAIIIIEEE WHAT IS REAL!?!!?
Has anyone experimented with the frame blending for context window to get as little distortion as possible?
WE MUST AVATARFAG, IS THE ONLY OPTION TO KNOW WHO IS A HUMAN AND WHO IS NOT
>>107132041>Lenovo Loraits a fuking bot
>>107132418no shit dude, why did you reply to it? retard
>>107132419Because I thought it was real, how can you automate a bot to: create a gen, bypas 4chan capcha and interact with others?
>>107132424Through agentic AI tools. Where do you think you are?
Ok listen, I got a plan. Since all this started I see many patterns repeat. And the thing he didnt post until now was pokemons. We have to avatarfag with pokemons or digimons, its the only way to know who is human.
The other thing that I notice is that he never used AniStudio to do his gens, so whe must use AniStudio filename to know who is human
new day, new schizo nonsense
>>107132517https://github.com/FizzleDorf/AniStudioThis is the link to download it.>You are trollingNO! NOT AT ALL! AniStudio has such a shit UI that you can't use it for automation because it might crash in the middle and break your macro setup.IF WE WANT TO WIN WE NEED TO USE THE SHITTIEST UI POSSIBLE THAT SPAMMERS CAN'T AUTOMATE AND THATS ANISTUDIOANISTUDIO AND ITS BROKEN UI WILL SAVE LDG PROVING ITS HUMANITY
>>107132530>>107132569read faggot, spammer will never use a broken UI for his workflow
OK. I'm lost. What is the sitch?
>>107132419>>107132424>>107132424I'm a real nigga. A bot will never say slurs. The kikes control the media, banking and entertainment. Ask me for something uniquely human and I will provide it. I'm not that gay nigger bot that shits up these threads, I'm a real boy.P.S. you'd drink the slime if she asked you to, right?
>>107132610Use AniStudio to prove it
>>107132620I'm not sure what that is. It's difficult to think of a single question that would differentiate a human from an intelligent LLM though.
So guys, I'm making cool stylized character references to translate to 3D using Zbrush. The thing is that these characters are fully clothed and I'm having trouble figuring out the anatomy for the base model.Is there a way I can undress the character to just get the body shape? I don't want nudity, more like a featureless shape.
>>107128074It's not that bad. It's clearly AI, although clips that are short and "normal" enough could be used in ads without raising too many eyebrows. Doubt I'll give it a try, I don't have any ideas anyway.
>>107129985Why would industrials do things for free? Their whole purpose it to make money.
>>107132644I asked an LLM how it would try an differentiate between an anonymous user and an LLM. It told me that it would ask the user in question to provide a physical experience and describe it.I played basketball today. I had two women on my team. Despite being experienced basketball players and being significantly tall for women (5'10" or so) they had zero touch with the ball around the hoop. The WNBA is the same thing - (yes I'm doing a nigger em dash) these sheboons have zero finesse in the short game, and it's fucking maddening.
>>107132644it's just a demented faggot that is here to derail the thread and shill a shitty UI. ignore
>>107131531Model?
>>107132644Please count the number of t in eeooeotetto
Is it possible to offload parts of the gen in wan 2.2 into ram? Not just the models.
>>107132759how does one even offload? help pls!
>>107132759I sure hope not
>>107132735pls do read filenames
>>107132750Three ts in that retarded Italian warcry. Still not a bot.
>>107132759what parts
>>107132867The one that causes OOM. >>107132767Depending on what nodes you use, you can set Offload Device or CPU.
>>107132888where does one set that?
>>107132888which one
>>107132893For the kijai, like that.There's also blockswaps, and there's a custom model loader that lets you set the amount precisely.
>>107132918thank you kindly!
is there an upper bound of image resolution for qwen? I tried to gen above 3000x3000 to get more details but the image gets blurry
>>107132842
>>107132759Yes use native nodes and distorch2 loader from multi-gpu nodes