[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/g/ - Technology

Name
Options
Comment
Verification
4chan Pass users can bypass this verification. [Learn More] [Login]
File
  • Please read the Rules and FAQ before posting.
  • You may highlight syntax and preserve whitespace by using [code] tags.

08/21/20New boards added: /vrpg/, /vmg/, /vst/ and /vm/
05/04/17New trial board added: /bant/ - International/Random
10/04/16New board for 4chan Pass users: /vip/ - Very Important Posts
[Hide] [Show All]


Janitor application acceptance emails are being sent out. Please remember to check your spam box!


[Advertise on 4chan]


Discussion of Free and Open Source Text-to-Image/Video Models

Prev: >>107114476

https://rentry.org/ldg-lazy-getting-started-guide

>UI
ComfyUI: https://github.com/comfyanonymous/ComfyUI
SwarmUI: https://github.com/mcmonkeyprojects/SwarmUI
re/Forge/Classic/Neo: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassicneo
SD.Next: https://github.com/vladmandic/sdnext
Wan2GP: https://github.com/deepbeepmeep/Wan2GP

>Checkpoints, LoRAs, Upscalers, & Workflows
https://civitai.com
https://civitaiarchive.com/
https://openmodeldb.info
https://openart.ai/workflows

>Tuning
https://github.com/spacepxl/demystifying-sd-finetuning
https://github.com/Nerogar/OneTrainer
https://github.com/kohya-ss/sd-scripts
https://github.com/tdrussell/diffusion-pipe

>WanX
https://comfyanonymous.github.io/ComfyUI_examples/wan22/
https://github.com/Wan-Video

>Neta Yume (Lumina 2)
https://civitai.com/models/1790792?modelVersionId=2298660
https://nieta-art.feishu.cn/wiki/RY3GwpT59icIQlkWXEfcCqIMnQd
https://gumgum10.github.io/gumgum.github.io/
https://neta-lumina-style.tz03.xyz/
https://huggingface.co/neta-art/Neta-Lumina

>Chroma
https://huggingface.co/lodestones/Chroma1-Base
Training: https://rentry.org/mvu52t46

>Illustrious
1girl and Beyond: https://rentry.org/comfyui_guide_1girl
Tag Explorer: https://tagexplorer.github.io/

>Misc
Local Model Meta: https://rentry.org/localmodelsmeta
Share Metadata: https://catbox.moe | https://litterbox.catbox.moe/
GPU Benchmarks: https://chimolog.co/bto-gpu-stable-diffusion-specs/
Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-one
Txt2Img Plugin: https://github.com/Acly/krita-ai-diffusion
Archive: https://rentry.org/sdg-link
Bakery: https://rentry.org/ldgcollage

>Neighbors
>>>/aco/csdg
>>>/b/degen
>>>/b/realistic+parody
>>>/gif/vdg
>>>/d/ddg
>>>/e/edg
>>>/h/hdg
>>>/trash/slop
>>>/vt/vtai
>>>/u/udg

>Local Text
>>>/g/lmg

>Maintain Thread Quality
https://rentry.org/debo
>>
I am of the strong conviction that comfy should be dragged out on the street and shot.
>>
@comfy
Maybe it would be a good idea in comfyui to have an option to precompute all prompts first, which if enabled would then do that first and thus ensure a HUGE speedup given you can throw away the text encoder forever out of VRAM.

Hugely speeding up the prompt computation since initially the model used to compute the prompt will be permanently loaded in VRAM and process everything in one go.
And then later on you will also permanently now have the diffusion model in VRAM and computing the images one after another.
Never needing to waste huge time swapping from RAM (or God forbid, disk) to VRAM, TWICE for every single generation.
This is a huge, double digit % speedup for most newer models. And would also allow people to load older models at maximum precision too.

A simple "Process all prompts first" switch which enables this in the settings would be great.
I also don't see why this wouldn't be a default option anyway, this would hugely speed up anyone who has more than 1 thing queued up, which is especially true for enterprise users which comfyui wants to cater to anyway.
>>
>I listened to anon and upgraded from ddr4 32gb to 96gb in August
>>
>>107123451
>not even 128gb
grim
>>
sadly my MB is full 4x16 so upgrading the ddr5 is going to cost an arm and a leg unless some falls off a lorry.
>>
>>107123447
Shit just sucks.
>>
Death to all jeets.
>>
til bghira lurks itt
>>
File: ComfyUI_04220_.png (1.29 MB, 832x1216)
1.29 MB
1.29 MB PNG
Shill me an underrated model nobody here talks about.
>>
>>107123797
good goy
>>
>>107124009
NetaYume
>>
>>107124085
>NetaYume
>nobody here talks about.
>>
OK the "baker is anti-realism" complaints were right, because the snubs this time are too much
>>
>>107124297
baker is a fag
>>
File: 00004-2000405452.jpg (2.15 MB, 2560x2048)
2.15 MB
2.15 MB JPG
>>
>>107124297
>>107124334
the baker is based. the only fags here are you whiny bitches.
>>
>>107124362
sorry you were outed as a fag, not my problem
>>
>>107124297
>the snubs this time are too much
AI "artists" really are as faggoty as real artists
>>
File: test_00025_.png (1.82 MB, 1216x832)
1.82 MB
1.82 MB PNG
>>
File: 00005-3909016455.jpg (1.82 MB, 2560x2048)
1.82 MB
1.82 MB JPG
>>
File: 1758182464121016.jpg (1.87 MB, 1248x1824)
1.87 MB
1.87 MB JPG
>>
what I actually want from ai is not something that makes me go "wow, that's plausibly human made". I want something that shines sparks into those dim dusty corners of my imagination
>>
>>107124657
if even sd 1.4 didnt already do this, you are a brainlet
>>
File: 00006-486300998.jpg (1.05 MB, 2560x2048)
1.05 MB
1.05 MB JPG
>>
>>107124635
test moar!
>>
File: 1740315536289401.png (2.56 MB, 2048x3072)
2.56 MB
2.56 MB PNG
>>
>>107124956
>cigger
>>
>>107124956
TWO cigars? now that's smoking
>>
>update comfyui
>rerun an image from yesterday
>comfyui crashes on second pass VAE decode
>no errors

comfy pls
>>
>>107125211
Comfy must be dragged into the streets and shot
>>
>>107125219
this is a factually correct statement
>>
>>107125219
>>107125281
Why not make something better?
>>
any new models/loras this week? still having fun with qwen edit v2 which is amazing for what it does. plus wan/illustrious are good.
>>
File: ComfyUI_21183.png (2.53 MB, 1200x1800)
2.53 MB
2.53 MB PNG
>>107124009
>Flux.1 [Dev]
I'm pretty sure that I'm the only one here that actually uses it and I don't post all that often. Seems like there's always something new added to Flux to play with. Great if you love to tinker.
>>
>>107125070
Shes in [flavour] country
>>
>>107125400
That image I posted was a flux merge too.
>Seems like there's always something new added to Flux to play with.
Are you referring to something specific?
>>
>>107124956
>no mayli
>>
File: more 1girl.png (1.19 MB, 832x1216)
1.19 MB
1.19 MB PNG
I love 1girl gens.
Please make and post more 1girl gens.
>>
cozy
>>
>>107125691
2girl or nothing
>>
File: ComfyUI_04236_.png (1.65 MB, 832x1216)
1.65 MB
1.65 MB PNG
This could have been cool if not for the slop hands
>>107125781
There are technically two girls here.
>>
File: symmetry.png (1.69 MB, 832x1216)
1.69 MB
1.69 MB PNG
>>
>>107124297
there are realistic gens in op all the time tho even ITT
>>
>>107124009
SD3.5M
SD1.4
>>
>>107124297
>>107124334
>>107124362
Wait baker is a person who does this for every single thread?
I thought it was an automated bot.
>>
File: 1756228151641560.png (3.8 MB, 1328x2048)
3.8 MB
3.8 MB PNG
>>107125322
>any new loras this week?
Just make your own
>>
>>107125840
...no it's several people who arbitrarily pick gens ffs
>>
File: double exposure slop 2.png (1.34 MB, 832x1216)
1.34 MB
1.34 MB PNG
>>107125834
>SD3.5M
I actually occasionally use this when I am bored. It's mediocre in certain regards but can make nice images with some luck still.
>SD1.4
Isn't it a bit too ancient? What would you use this for in 2025?
>>
>>107125875
based reply and gen
>>
File: wings.png (1.64 MB, 832x1216)
1.64 MB
1.64 MB PNG
>>
File: beautiful face.png (1.82 MB, 832x1216)
1.82 MB
1.82 MB PNG
>>
Slopfest
>>
File: 00008-3274024266.jpg (2.52 MB, 2560x2048)
2.52 MB
2.52 MB JPG
>>
File: slop woman in space.png (1.12 MB, 832x1216)
1.12 MB
1.12 MB PNG
>>107126058
Slop board
Slop general
Slop thread
Slop posts
>>
File: 1750014303999610.png (2.87 MB, 2048x1328)
2.87 MB
2.87 MB PNG
>>107126058
>Slopfest
>>
File: slop.png (366 KB, 1144x672)
366 KB
366 KB PNG
>>107126078
>>
>>107126058
all AI gens are slop
>>
File: 00251-2581452911.png (2.92 MB, 1248x1824)
2.92 MB
2.92 MB PNG
>>
>>107126399
not mine mine are special
>>
>>107126432
they are. we love your work.
>>
>>107126427
A mix of detail and lack thereof. Are you making a statement about the duality of man?
>>
File: 00010-445967462.jpg (1.23 MB, 2560x2048)
1.23 MB
1.23 MB JPG
>>
>>
>>107123447
unfathomably based
>>
>>107126707
Welcome back, Master.
That's Richard Stalinman in the background.
>>
File: 00311-966916063.png (1.35 MB, 1144x672)
1.35 MB
1.35 MB PNG
>>
wan 3.0 when
>>
>>107125820
Only when they are basically "memes" with some obvious visual gag or gimmick. He never just likes a realistic 1girl for being nice to look at. His taste in realism is basically the same as model devs when they put out papers with demonstration gens of a puppy holding a lollipop and wearing a party hat
>>
>>107123447
I think you should be dragged out in the streets and lightly flogged in front of a crowd until you start blubbering and crying, and then you should be let go
>>
>>
>>107126664
thats noob's idea of depth of field, sometimes it works well but there is does make the background into a mess.
>>
Are we back? Did the schizo lose? If yes, poor faggot wasted tons of effort doing this.
>>
>>107127786
Dont give any more attention to the shitter. Anyway, new video model drop!

https://github.com/FoundationVision/InfinityStar

>InfinityStar by Bytedance: A unified 8B spacetime autoregressive model for high-res image & video gen;
>- 5s 720p video ~10x faster than DiT;
>- scores 83.74 on VBench, topping other AR models and HunyuanVideo;
>- Flan-T5-XL as text encoder.
>- 480/720p, ~35Gb model
>>
>>107127895
>8B
Garbage. If it doesn't have more than 14B(Wan2.2) it's not worth using.
>>
>>107127895
He needs to sleep for like 2 days straight because it wasn't just the bots running but he was also posting actively too, absolute madness.
>>
>>107127895
>We provide a demo website
>look inside
>17k member discord
>>
>>107126276
>ai is better than gimp
>>
>>107128043
anything on huggingface to bypass this nonsense?
>>
>>107128043
>>107128053
Yeah its a bit retarded that they are locking their demos behind shitcord. Here's what it looks like, posted on xitter https://xcancel.com/wildmindai/status/1986502031532826776
>>
it's wan2.2 with lightx2v 4 steps utter unusable dogshit, or am I doing something wrong?
>>
>>107127419
>>107114476
>>107108437
>>107102952
>>
What are the most important qwen image edit loras in your opinion?
>>
>>107128083
Depends on what you're trying to gen. Sometimes, for faster motion for example, you can set cfg to 1.5 or do some retarded lora slop stacking ( wan 2.1 lightx2v 3 str high and 0.25 str low, wan 2.2 kijai lightning MOE high and wan 2.2 regular lighting low). Then again, those settings tend to change the face and details of the input image. Can slop in some PUSA loras around .5 or .6 strength to keep consistency, sorta.
>>
>>107128181
tested high=2, low=1, and it got a lot better
is that ok? what lora values are best?
>>
>>107128074
it looks significantly better than wan 2.2. we'll see if kijai releases anything
>>
>>107128132
For me its some kind of troll, he put a couple of gens of mine with nano banan and an anime screenshot from google images, but his collage meme idea is good though, not everything has to be serious, epic or coomerbait. Sometimes a meme for a light laugh is nice
>>
>>107128181
is pusa a lora, or a category of loras?
>>
>>107128206
Cant say there is a "best" values, not that I'm aware of anyway. Best to just tinker around with the strengths and find out what works best for whatever you're trying to make.
>>
>>107128335

Its a lora, read these and models are there

wan 2.2 https://huggingface.co/RaphaelLiu/Pusa-Wan2.2-V1
wan 2.1 https://huggingface.co/RaphaelLiu/PusaV1
>>
>>107128382
alright, thanks, sempai
>>
>>107128398
Also forgot Kijais version because I know someone is going to bitch about it. Also, I dont know the difference but someone will kek

https://huggingface.co/Kijai/WanVideo_comfy/tree/main/Pusa
>>
>>107124009
>nobody here talks about.
hard since anon talks about every model
>>
>>107128231
It does look pretty good. It says it can do 5 to 10 seconds apparently. They're boasting 10x speed which we'll need some kind of comparison. Also, I noticed this model doesn't seem to be based on wan, so it's 100% its own thing? Good to see more contenders, hopefully we can see more demos and such. Plus LTX2 this month, I think.

>distant screams from kijai can be heard
>>
>>107124009
>>107128555
How about animatediff? Fun model, can gen some pretty stupid sloppy stuff. Maybe there is some kind of way to lower the memory to let wan gen something crazy like 1000 frames. We can do this now but you need a super stupid amount of vram. When I was tinkering with animatediff, 1000 frames was no problem, started running into oom at 1400 frames. Slap on the SVI lora and we can finally escape 5 second hell properly (no more color shifting, jank movement, 81 frame stitch tricks).
>>
Seems like about 1 million pixels resolution is the sweet spot for WAN. Body shape start to deform when you go higher pixel count
>>
>>107128231
>it looks significantly better than wan 2.2
idk, i think people dont really know how good maxxed out quality wan 2.2 looks like, especially nowadays with actually good lightning loras that everyone is using, making them forget how good wan is for physics and motion without the added lightning stiffness

it does seem like the dataset is wider, for realistic gens at least, the background details are better, the look is less slopped out of the box, we'll see
>>
>>107128750
video? it was trained on 720 so yeah, works fine for images up to 1.6MP but good luck making videos at that res
>>
File: 1749716275563416.png (3.81 MB, 1664x2432)
3.81 MB
3.81 MB PNG
I'm wanproompting.
>>
>>107128825
Looks great, catbox?
Is wan prompted with essays or?
>>
>>107128842
Yeah, here is the prompt
"In the flickering glow of candlelight, a grand medieval feast unfolds within the stone-walled hall. The air hums softly with the murmur of voices, the clinking of goblets, and the faint crackle of the hearth nearby. At the center of the long wooden table sits a noblewoman, her presence commanding yet graceful. She wears an elegant white gown adorned with golden pauldrons upon her shoulders, their polished surfaces catching the warm light. A delicate belt of gilt embroidery cinches her waist, and a shimmering crown rests atop her carefully styled hair — a symbol of her royal station.

Before her lies a modest yet hearty meal: thick slices of bread, roasted vegetables, and tender cuts of meat, arranged on a simple ceramic plate. A goblet of drink and a small bowl of rich sauce complete the setting, evoking a time when feasts were both ritual and celebration. Her posture is poised, her hands folded lightly before her as she offers a knowing smile, one touched with confidence and quiet amusement — as though she is aware of the attention she commands.

Beside her, an older kingly figure, clad in regal black with golden trim and a matching crown, partakes in his meal with solemn dignity. The candle between them burns low, its flame dancing gently in the draft of the old hall, casting shifting shadows across the brick and timber walls.

The moment feels timeless — a captured fragment of courtly life where elegance, power, and intimacy blend seamlessly beneath the watchful eyes of flickering candlelight."
>>
>>107128825
> wan
Too bad t2i doesn't work with t2v loras.
>>
>>107128857
What frontend/workflow?
>>
>>107128842
>>107128857
But then O had to inpaint her face so my promtp to inapint her face was
"Her face is softly illuminated by the flickering candlelight, giving her a warm, almost ethereal glow. Delicate features frame her serene expression — graceful and composed, yet touched with a subtle confidence. Her fair hair is elegantly styled into an intricate updo, a few loose strands framing her face in gentle wisps that soften her regal poise. The golden crown resting upon her head glimmers faintly, its ornate design accentuating the nobility in her bearing. Her eyes, bright and keen, carry both warmth and a quiet command, while a faint, knowing smile curves her lips — the kind that hints at secrets and strength beneath her gentle exterior. Adorned with pearls and a cross pendant that rests gracefully at her throat, her face embodies a blend of royalty and radiance, timeless in its allure."
>>
>>107128825
>>107128857
>>107128861
>>107128873
But then her handa where melted and i have to inpaint her hands and it was
"Both of her arms are bent at the elbows and rest gently on the edge of the table. Her forearms are crossed lightly in front of her, creating a subtle, poised frame against her white gown. The soft fabric of her sleeves gathers slightly at her wrists, emphasizing the elegance of her position. Her hands rest close together, fingers relaxed, giving an impression of calm confidence and refined poise — a gesture both demure and self-assured, fitting for a regal setting."

I love prooosing so much, totally intuitive and practical
>>
File: 00010-926861646.png (2.29 MB, 1280x1920)
2.29 MB
2.29 MB PNG
>>
>>
>train lora on a bunch of greenscreened characters
>generate character
>use wan to animate a looping idle pose
>then animate attack animations, etc. that loop back to the base image so everything transitions seamlessly
>remove green background with video editor
Maybe you'll need to train a lora on the animations as well but I think this could be doable for making sprites
>>
File: media_1762457388.png (1.24 MB, 768x1280)
1.24 MB
1.24 MB PNG
Ney'liri is back.
>>
What is the best settings for OneTrainer to train Pixel art
This is the best i could give for my Illus LORA Spritesheet training
Its really blury
>>
>>107128132
He has shit taste. It is what it is.
>>
Not sure if anyone else uses this but SuperUltimateVaceTools got an update, apparently works with wan 2.2 fun vace https://github.com/bbaudio-2025/ComfyUI-SuperUltimateVaceTools/tree/main Was pretty decent with wan 2.1, gonna test soon if not this weekend
>>
>>
Have there been any developments in frame interpolation? I've been using RIFE then falling back to FILM VFI if the RIFE output has too much artifacting (usually from fast motion). Has there been anything since those two models? They're both getting pretty old.
>>
>>
>>107129258
Based and pixai pilled
>>
quick question
I'm messing around with wan2.2 on my shitty 12gb vram setup and looking at the total committed memory usage it's pushing 70 gigs all together
how are 24gb cards able to run the full size, unquantified models if even these neutered, 10GB quantified models are taking so much memory?
>>
>>
I got a 4090 and 32GB of RAM. I've been playing around with ComfyUI for the last couple of days.

I sorta have an idea of what happening but I'm still bad and either using default workflows or have to copy what other people give. Problem is no matter what I do I cannot get remotely decent images that are of any good quality. When I use SDXL everything looks bad, even when I try to play with the parameters shown to me (I am just opening the basic text to image workflow and placing the model in and playing around with it.) The default workflow for Flux kinda works ok but like everything else I'm basically have no idea how to use.

And if I try to make any videos with WAN 2.2 14B when the models start loading it starts thrashing and SSD writing to my disc like mad. If I try to make any 1280x720p video my computer just runs out of memory. Should I upgrade to 64GB RAM because my disc is thrashing and my RAM usage is basically full when trying

Using the 480p fast i2v workflow and models here https://rentry.org/wan22ldgguide I can make 480x480 videos and it takes about 2min/5seconds but they look like ass even with real life photos used for reference(and half of them don't even follow the prompt I give like "walking away" turns into walking towards the camera)

I just want to be able to create videos and pictures that look anywhere near as good as what I see online or can use Grok to generate in fractions of the time for free.

tldr; Without spending the next 4 weeks learning and reading everything about comfy and learning to make my own nodes and being afraid I'm getting hacked by some obscure workflow I got off civitai, how do I make clean AI images and videos on my 4090+32GB RAM? Or is everything I'm seeing online using H100s type hardware?

Also I am in general interesting in learning all this stuff as well and ComfyUI seems like it has a lot of power so I will learn it all eventually, but where is a reasonable place to start to see what my hardware is capable of?
>>
File: 1746782705501959.png (2.36 MB, 1344x1728)
2.36 MB
2.36 MB PNG
>>
>>107129578
>:3
Saved
>>
File: AniStudio-00395.png (1.78 MB, 1120x928)
1.78 MB
1.78 MB PNG
Alright we're back. Time to bake a new thread and forget that dogshit bot arc ever happened! Lets choose our best 1girls and make the collage!
>>
>>107129559
Also, do any of these other webuis have upscaling and frame interpolation built in? Every single video game has DLSS and FG these days, but for these static workflows these upscaling and interpolation is actually desired.
>>
>>107129607
SARR PLEASE MY WIFE WONT STOP!>>29756481
>>
>>107129295
Help me on this...
>>
>anistudio filename
based
>asuka brap
based
>gemini watermark
cringe
>>
File: film vfi.png (281 KB, 1753x971)
281 KB
281 KB PNG
>>107129559

Start collecting images with metadata that you like and check how people made them.

A 4090 and 32gb should be able to do anything image related. Video needs at least 64GB for model off loading purposes, otherwise your system might stall at 32GB ram. Try ultimate upscaler node or film vfi for interpolation.
>>
>>107129611
Video games have motion vectors to help with interpolation, that isn't a fair comparison
>>
>>107129618

Post dataset then I might have an insight. Garbage in, garbage out after all.
>>
>>107129295
It's not good with pixel art at all without another plugin pixelizing it afterward. You'll always get smudges or uneven pixels.
>>
>>107129630
I forget to put watermak in negative prompts :(
>>
>>107129636
32GB RAM sounds pretty painful with larger models like qwen or flux. anon should just upgrade to 64GB, that is really the minimum for an AI build. buy now before the price doubles again
>>
Relative newbie if someone could indulge me.

Is Wan2.X moreso used for generating video? I'm looking for recs on img2img.
Does anyone know where I can find a good Comfy workflow for img2img?
Recommendations for realistic image gen models?
>>
File: media_1762490058.png (1.31 MB, 768x1280)
1.31 MB
1.31 MB PNG
>>
hyped for dype
>>
>>107129646
https://www.mediafire.com/file/2qrfav0ct380pym/MMBN_Spritesheet_Battle.7z/file
>>107129658
Well, i used Unfake pixel for other spritesheet, and its work fine
The problem is this was too blury, especially on the face
>>
>>107129721
Based
>>
>>107129745

Your problem is that VAE is extremely lossy and cannot retain any information for individual sprite frame at 1024x1024 resolution. I have no viable solution for that old style sprite sheet. Perhaps try a newer style like Unicorn Overlord.
>>
Is comfyui safe to update to the latest version?
>>
>>107129859
No, its download dolphin porn
>>
>>107129636
>>107129667

I'm guessing it will help with the disc thrashing. Using high quality images will produce higher quality videos too right?

Is security a thing to be concerned with comfyui with custom nodes and workflows?

I found a workflow for images but with upscaling its taking like 20x longer to generate the image. I guess it's fine if I get higher quality outputs
>>
>>107128857
so where did her giant cleavage come from?
>>
File: 52551994460_4536bb79df_h.jpg (202 KB, 1600x1067)
202 KB
202 KB JPG
>>107129823
Nah i just want to see if i can create MMBN Sprite
>>
File: 1746457487775210.mp4 (3.74 MB, 1280x720)
3.74 MB
3.74 MB MP4
>>107127895
https://xcancel.com/wildmindai/status/1986502031532826776
it looks worse than Wan 2.2 so it's kinda useless lol
>>
>>107129895
Looks like there are errors https://github.com/comfyanonymous/ComfyUI/issues/10565
>>
>>107129963
>Bytedance
kek, DOA, they only release failed experiments to the Open, we're just a garbage bin to them
>>
>>107129973
Oh sweaty, how are they suppose to make their monies? :)
>>
>>
>>107129980
true, think of that poor multi-billion dollar industry :'(
>>
Hello, guys, is there a tutorial on how to install and run llms locally?? how is it done?? anyone has tutorials?
>>
>>107129963
https://youtu.be/yg8Qe2-2sqw?t=91
there's some more examples here, it's Mid
>>
>>107128043
>17k member discord
the fuck? why is there so many people to test that brand new model?
>>
>>107130023
Because it's free and supports hindu?
>>
>>107130051
>supports hindu
we now live in a world where jeets are being pandered to, this is what happens when you are a race that reproduces like rabbits...
>>
>>107130012

InfinityStar? No, it's InfinitySaar.
>>
Is it possible to do video2video inpaint lipsyncing?
>>
>>107130098
keek
>>
>>107123451
>>107123462
TFW old enough to remember when 256MB of RAM made me feel like an elite h4x0r.
>>
>>107130051
within a year, any system that caters to indians fails.
>>
File: 1754176345812913.png (180 KB, 640x360)
180 KB
180 KB PNG
>>107130270
oh god don't give me hope
>>
<ctrl+f
>"AMD"
>0 results
uh-oh, doesn't seem like my odds are good.
>>
>>107130023
bytedance catchall server
>>
>>107130376
o shi-
>>
>>107125691
sauce pls. this is beaut!
>>
I started zooming in on webpages a long time ago to save my eyes from the strain of reading small text, so I'm always at like 140%. But it stretches out images, too. And my monitor's pixel density, being a pretty large 1440p monitor, is not so dense. So images look much larger to me than to most.

Sometimes I wonder if these factors combined are why I'm able to spot overcooked and AI upscaled gens so much easier than the average poster. I see the problems right away. The texture is off. You can count the realism posters who don't break the illusion on one hand, and most of them don't come here anymore.
>>
when will they make vibevoice but nsfw
>>
File: 1737623272970734.png (95 KB, 1324x653)
95 KB
95 KB PNG
https://www.reddit.com/r/StableDiffusion/comments/1oqosui/i_made_the_best_text_encoder_for_qwen_image_edit/
>Qwen 2509 is powerful but the default encoder leaves 70% of that power on the table. This unlocks it.
goddam I hate those faggots so much
>>
>>107130870
lmao this was absolutely the result of begging claude to "make qwen edit more gooder"
>>
>>107130870
Fucking pajeets
>>
>>107130645
Let's fuck.
>>
So I've been curious maybe you guys will be able to answer this.
AI gen seems to be pretty popular these days. So how come every model I've seen so far requires a somewhat significant amount of installing of random scripts, libraries, tertiary apps, etc... How come there hasn't been a simple single installer/application yet? Is it because the environment is too unstable for it to happen yet or what?
>>
>>107131185
You generally install only one set of stuff per model type for basic usage and it's automated on some of the possible software.

The rest is partly internet attention economy. Enjoy.
>>
>>107131247
That's not really what I wanted to know but thanks.
>>
>>107131268
The completely "single application" stuff beyond the software that installs its dependencies on the desktop... is an app/website launcher for smartphone users that tries to IaaS/SaaS you.
>>
>>107131302
It's cool dude, I already got the answer elsewhere. Like I asked, I wanted to know the technical reason as to why hasn’t anyone managed to ship a local plug-and-play turnkey solution for the average user, what you did is first give me some general knowledge on what to do, and now a vague opinion on what the online alternatives are. I didn't ask either quest, at this point you're either trying to troll me or completely clueless yourself. The answer I was looking for was that there’s no turnkey installer because local AI generation depends on fragile, rapidly changing GPU/toolchain dependencies that break across hardware, drivers, and models faster than a unified app can be maintained.
>>
File: 1744431872094635.jpg (976 KB, 2016x1152)
976 KB
976 KB JPG
>>
File: eva green 11.jpg (634 KB, 2398x3260)
634 KB
634 KB JPG
>>107130678
Troll post likely but I fed this Eva Green image to Joycaption and slopped through it.
"A high-quality photograph of a young woman with pale skin and dark brown, straight hair. She's facing to the left in a three-quarter profile shot. Her lips are painted bright red, and she's wearing bold, gold hoop earrings shaped like abstract loops. The background is a solid light blue gradient that fades into darker blue at the top right corner. She has faint freckles across her cheeks and upper chest. Her shoulders are bare, suggesting she might be wearing an off-the-shoulder top or dress made of black, sheer material. The lighting is even and bright, highlighting her facial features and creating soft shadows on her face and neck. Camera angle is slightly above eye level, capturing the woman from just below her eyes down to her shoulders. Her expression is neutral but somewhat confident. This image is SFW and focuses primarily on close-up details of her face and upper body. The composition style is clean and minimalistic with a focus on the subject's elegant features and bold accessories. Age appears to be in her mid-20s. Overall, the image has a modern, high-fashion vibe. There are no watermarks or additional objects in the frame. "
Neither the description is accurate, nor the gen sticks to it.
>>
File: CHR_09.png (2.08 MB, 1024x1536)
2.08 MB
2.08 MB PNG
just repent already
>>
File: CHR_11.png (1.73 MB, 1024x1536)
1.73 MB
1.73 MB PNG
>>107131531
>>107131453
nice prompt with some minor tweaks
>>
File: ComfyUI_00109_.png (3.02 MB, 1536x1536)
3.02 MB
3.02 MB PNG
Every day I learn more. Every day the slime pool sluts gen better.
>>
File: ComfyUI_00107_.png (3 MB, 1536x1536)
3 MB
3 MB PNG
(taking feedback. I have no idea what I'm doing.)
>>
>>107131560
>>107131553
Cool liminal space
>>
File: ComfyUI_00095_.png (3.13 MB, 1536x1536)
3.13 MB
3.13 MB PNG
>>107131572
Thanks. I think the zoomer "liminal".trend is a little overbaked, but I like the idea. This waterslide zone came to me in a dream one day, and I've been a little enthralled with it since. How do you like the girl?
>>
>>107131328
The desktop gui/webui are basically all scripted/automated/with installer.

You just can't get everything (models especially) in one tool without paying attention to anyone else.

>The answer I was looking for was that there’s no turnkey installer because local AI generation depends on fragile, rapidly changing GPU/toolchain dependencies that break across hardware, drivers, and models faster than a unified app can be maintained.
Technology is also a thing why it isn't perfect on *all* devices, but generally speaking you are seeing business attention economy reasons of the kind why there isn't one netflix and why netflix is paid for solutions that are for you, the helpless SaaS/IaaS consumer.
>>
>>107129525
Kino gen
>>
File: ComfyUI_00051_.png (3.23 MB, 1536x1536)
3.23 MB
3.23 MB PNG
>>107131328
Spend 3 hours trying to figure out that your cuda doesn't match your pytorch and your flashattention wasn't baked with the right cuda version so pip can't make a (wheel?) for it and then your dependencies are too new, but God forbid if they're too old that's also bad, and sometimes it's just a coin flip.

Anyways, you can eventually get fun gens.
>>
File: ComfyUI_00037_.png (3.29 MB, 1536x1536)
3.29 MB
3.29 MB PNG
>>107131550
That's a great gen. Your DOF is a little off though, it's stronger in the front - not your issue though. The only flaw is the black mark on her teeth, but that's an easy edit.
>>
File: CHR_19.png (1.82 MB, 1088x1344)
1.82 MB
1.82 MB PNG
>>107131700
That and the obvious earring flaw, heh
>>
File: ComfyUI_00035_.png (3.21 MB, 1536x1536)
3.21 MB
3.21 MB PNG
>>107131185
It's multiple pieces of tech (CUDA, pytorch, numpy, a multitude of python libs) that have dependency/interdependency, and then your image gen software (comfyui for example) will have stricter requirements, and sometimes only a few releases are valid.

It really is a house of cards of FOSS tech, but in a way it's great, since it functions as a powerful pleb filter for those that can't help themselves or ask polite and poingent questions to learn. In a past life I was a 3D printing mogul, and I've seen the same thing in the space - let me inb4 though, there are some places that still have high IQ individuals doing really awesome stuff - but with the advent of easy to deploy machines (prusa, ender, and especially Bambu) it's ushered in an era of midwits, jeets, kids, and retards that just shit up the environment... And on to the next niche thing I go.

Have a slime pool girl as consolation for reading my life story.
>>
File: ComfyUI_00023_.png (3.06 MB, 1536x1536)
3.06 MB
3.06 MB PNG
>>107131733
You're right. This gen of yours is great - you're selectively hiding the details that could go wrong. Much like a boudoir photographer, you want to hide fat chicks "problem areas" in the shadows or behind clothes/lingerie/blankets.

I've done the same with my gens, I specify the hands to be submersed if it's reasonable - it removes a large stumbling block for image gens, much like how a skilled boudoir photographer can minimize problematic areas.
>>
File: CHR_29.png (2.39 MB, 1088x1536)
2.39 MB
2.39 MB PNG
>>107131766
heh absolutely. There are limitations we gotta go around
>>
Did he upgrade the bot?
>>
File: ComfyUI_00131_.png (3.1 MB, 1536x1536)
3.1 MB
3.1 MB PNG
>>107131841
If you're talking about me, I'm not a bot. At least I hope not.
>>
>>107131599
very plasticy sadly.
>>
>>107131841
yeah fuck I just realized, it's 100% a bot, no way a human types like this.
>>
>>107130645
Because it's CUDA. And AI implies CUDA.
>>
>>107132008
For now.
>>
File: ComfyUI_00117_.png (2.96 MB, 1536x1536)
2.96 MB
2.96 MB PNG
>>107131976
Any advice on de-plasticizing it? I'm using chroma and the Lenovo Lora at 0.95 strength, with an upscale strength of 0.15. the upscale model I used escapes me.

Anyone have any names for this young lass trapped in this slimy world?
>>
>>107129453
gimm but it's much slower
>>
>>107129520
offloading. the full model takes the same amount of time as q8 for me, 24 gb
>>
>>107131976
Also interested in how to minimize the plastic look.
>>
Microplastics are everywhere.
>>
>>107132021
That now will last very long.
>>
chroma would be perfect if it works with control nets
>>
File: 1741869676865235.png (540 KB, 2287x1011)
540 KB
540 KB PNG
>>107129559
you're using this workflow? play with the block swap setting until you don't max out your vram or it'll stall. you should have about 1gb vram free at all times. I find having ~700mb free is ok for me. if it spikes to max on the "WanVideo ImageToVideo Encode" node that's usually fine

I think the issue is though that you're maxing out your sysram. 32gb is really not enough and ram prices just doubled recently. sorry ramlet bro
>>
all my qwen loras fucking suck.
I am using Ostris toolkit default settings and 4bit quantizations, 0.0001 LR
>>
>>107132221
Go for 0.0002 LR
>>
>>107132139
>kijai wan
CRONGE
>>
>>107132323
not what I'm using but it's in the shitty guide that no one wants to update
>>
>>107131841
AAAAAIIIIEEE WHAT IS REAL!?!!?
>>
File: file.png (2.27 MB, 1472x1136)
2.27 MB
2.27 MB PNG
>>
Has anyone experimented with the frame blending for context window to get as little distortion as possible?
>>
WE MUST AVATARFAG, IS THE ONLY OPTION TO KNOW WHO IS A HUMAN AND WHO IS NOT
>>
>>107132041
>Lenovo Lora
its a fuking bot
>>
>>107132418
no shit dude, why did you reply to it? retard
>>
>>107132419
Because I thought it was real, how can you automate a bot to: create a gen, bypas 4chan capcha and interact with others?
>>
>>107132424
Through agentic AI tools. Where do you think you are?
>>
File: 151.jpg (32 KB, 475x475)
32 KB
32 KB JPG
Ok listen, I got a plan. Since all this started I see many patterns repeat. And the thing he didnt post until now was pokemons. We have to avatarfag with pokemons or digimons, its the only way to know who is human.
>>
File: AniStudio-00032.jpg (67 KB, 478x606)
67 KB
67 KB JPG
The other thing that I notice is that he never used AniStudio to do his gens, so whe must use AniStudio filename to know who is human
>>
new day, new schizo nonsense
>>
File: AniStudio-00033.jpg (43 KB, 475x475)
43 KB
43 KB JPG
>>107132517
https://github.com/FizzleDorf/AniStudio
This is the link to download it.
>You are trolling
NO! NOT AT ALL!
AniStudio has such a shit UI that you can't use it for automation because it might crash in the middle and break your macro setup.
IF WE WANT TO WIN WE NEED TO USE THE SHITTIEST UI POSSIBLE THAT SPAMMERS CAN'T AUTOMATE AND THATS ANISTUDIO

ANISTUDIO AND ITS BROKEN UI WILL SAVE LDG PROVING ITS HUMANITY
>>
>>107132530
>>107132569
read faggot, spammer will never use a broken UI for his workflow
>>
OK. I'm lost. What is the sitch?
>>
File: ComfyUI_00251_.png (3.08 MB, 1536x1536)
3.08 MB
3.08 MB PNG
>>107132419
>>107132424
>>107132424
I'm a real nigga. A bot will never say slurs. The kikes control the media, banking and entertainment. Ask me for something uniquely human and I will provide it. I'm not that gay nigger bot that shits up these threads, I'm a real boy.

P.S. you'd drink the slime if she asked you to, right?
>>
>>107132610
Use AniStudio to prove it
>>
File: ComfyUI_00257_.png (3.18 MB, 1536x1536)
3.18 MB
3.18 MB PNG
>>107132620
I'm not sure what that is. It's difficult to think of a single question that would differentiate a human from an intelligent LLM though.
>>
So guys, I'm making cool stylized character references to translate to 3D using Zbrush. The thing is that these characters are fully clothed and I'm having trouble figuring out the anatomy for the base model.
Is there a way I can undress the character to just get the body shape? I don't want nudity, more like a featureless shape.
>>
>>107128074
It's not that bad. It's clearly AI, although clips that are short and "normal" enough could be used in ads without raising too many eyebrows. Doubt I'll give it a try, I don't have any ideas anyway.
>>
>>107129985
Why would industrials do things for free? Their whole purpose it to make money.
>>
File: ComfyUI_00265_.png (3.11 MB, 1536x1536)
3.11 MB
3.11 MB PNG
>>107132644
I asked an LLM how it would try an differentiate between an anonymous user and an LLM. It told me that it would ask the user in question to provide a physical experience and describe it.

I played basketball today. I had two women on my team. Despite being experienced basketball players and being significantly tall for women (5'10" or so) they had zero touch with the ball around the hoop. The WNBA is the same thing - (yes I'm doing a nigger em dash) these sheboons have zero finesse in the short game, and it's fucking maddening.
>>
>>107132644
it's just a demented faggot that is here to derail the thread and shill a shitty UI. ignore
>>
>>107131531
Model?
>>
>>107132644
Please count the number of t in eeooeotetto
>>
Is it possible to offload parts of the gen in wan 2.2 into ram? Not just the models.
>>
>>107132759
how does one even offload? help pls!
>>
>>107132759
I sure hope not
>>
File: media_1762457398.png (1.26 MB, 768x1280)
1.26 MB
1.26 MB PNG
>>
File: 00140-997524247.png (1.55 MB, 1144x672)
1.55 MB
1.55 MB PNG
>>
File: CHR_36.png (2.64 MB, 1088x1536)
2.64 MB
2.64 MB PNG
>>107132735
pls do read filenames
>>
>>107132750
Three ts in that retarded Italian warcry. Still not a bot.
>>
>>107132759
what parts
>>
>>107132867
The one that causes OOM.

>>107132767
Depending on what nodes you use, you can set Offload Device or CPU.
>>
>>107132888
where does one set that?
>>
>>107132888
which one
>>
>>107132893
For the kijai, like that.
There's also blockswaps, and there's a custom model loader that lets you set the amount precisely.
>>
>>107132918
thank you kindly!
>>
File: 00051-2142675812.png (1.49 MB, 1144x672)
1.49 MB
1.49 MB PNG
>>
is there an upper bound of image resolution for qwen? I tried to gen above 3000x3000 to get more details but the image gets blurry
>>
File: CHR_53.png (2.67 MB, 1088x1536)
2.67 MB
2.67 MB PNG
>>107132842
>>
>>107132759
Yes use native nodes and distorch2 loader from multi-gpu nodes
>>
>>107125211
>update comfyui and GGUF nodes again
>now the same workflow runs to the finish and produces the expected image
>run it for a second time
>pure noise image

maybe tomorrow
>>
I've tried various models with so many combinations of settings and prompts and can't get a high quality realistic girl gen base image (to be upscaled later on).
Am I limited by being a 16GB vramlet?
What models that fit are best for this purpose?
>>
>>107133527
just look at this thread man. so many good 1grills
>>
>>107133527
chroma
https://huggingface.co/rocca/chroma-nunchaku-test/tree/main
mystery meat: https://huggingface.co/Disty0/Chroma1-HD-SDNQ-uint4-svd-r32
>>
People hate ran because he is even worse than ****. All of his gens are identical.
>>
>>107133558
>copechaku
never began
>>
>>107133649
best 4.0 bit quant doe
25% of space
90% of quality
needs as little as 1.5gb vram due to genius memory allocation
>>
>>107129559
>>107129636
>>107132139

Even the default qwen text to image workflow is complete BS. Using everything out of the box and my memory is at 100% and thrashing. The note here says my max vram should be 86%(it isn't)

That and in the manager under models, it doesn't even list the size of the files correctly. The diffusion model is almost 20GB, it says 4GB, the LORA is 1.6B, it says 10GB
>>
>>107133665
>90% of quality
oh no no no
>>
File: 1752306272178646.png (38 KB, 978x256)
38 KB
38 KB PNG
>>107130870
>link to linkedin
wtf
>>
File: file.png (811 KB, 1815x379)
811 KB
811 KB PNG
>>107133723
>>
>>107133745
>the same simple as shit prompt official image
try regenning any of the good copechaku images you have with bf16, no fp16 accumulation etc, and see if it wont be better
whats the point of gimping your output quality at all when at worst it will take less than a minute to gen the image anyway
>>
>>107133787
the point is so i can run it on a 12gb vram card? you cant run BF16 on a 12gb vram card, it's 4x slower
could you post an image genned with BF16, and I'll use the workflow to gen a nunchaku image
>>
whuddabout SVD quant doe
>>
>>107133804
>12gb vram card
rip
>>
>>107133815
yea
>>
File: 00000-1784164830.jpg (1.26 MB, 2560x2048)
1.26 MB
1.26 MB JPG
>>
>>
File: 00001-1934210134.jpg (1.96 MB, 2560x2048)
1.96 MB
1.96 MB JPG
>>
Damn this is taking so long. Sure, it's 4x the frames, but shouldn't this be cpu focused task?
>>
>>107134064
what are you doing
>>
>>107133978
Keep on rock-coon!
>>
>>
use case of all that shit?
>>
>>107134116
Ah not much, sitting here watching videos, genning, waiting to go to bed.
>>
File: 111992037188192.jpg (60 KB, 640x360)
60 KB
60 KB JPG
>gen 125 frames
>161 comes out
>>
I feel fortunate the most dedicated idiots that have issues with me are alcoholic low functioning losers that have been exposed a lowcows that can't even match the average /ldg/ poster in skill. They also like to ERP pretending to be girls with other men and can't even read filenames to realize who's posting what
>>
>>107133464
Powerful.
>>
>years later
>Gens are still dog shit
>Exiled to a containment thread
>My quality is still top notch
>More respected than you ever will be
>Everyone makes fun of you on the discord you tried so hard to be accepted in
Was it worth it?
>>
LOOK AT IT GO



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.