Discussion of free and open source text-to-image modelsPrevious /ldg/ bread : >>101639278>Beginner UIEasyDiffusion: https://easydiffusion.github.ioFooocus: https://github.com/lllyasviel/fooocusMetastable: https://metastable.studio>Advanced UIAutomatic1111: https://github.com/automatic1111/stable-diffusion-webuiComfyUI: https://github.com/comfyanonymous/ComfyUIInvokeAI: https://github.com/invoke-ai/InvokeAISD.Next: https://github.com/vladmandic/automaticSwarmUI: https://github.com/mcmonkeyprojects/SwarmUI >Use a VAE if your images look washed outhttps://rentry.org/sdvae>Model Rankinghttps://imgsys.org/rankings>Models, LoRAs & traininghttps://civitai.comhttps://huggingface.cohttps://aitracker.arthttps://www.modelscope.cn/homehttps://github.com/Nerogar/OneTrainerhttps://github.com/derrian-distro/LoRA_Easy_Training_Scripts>Pixart Sigma & Hunyuan DIThttps://huggingface.co/spaces/PixArt-alpha/PixArt-Sigmahttps://huggingface.co/spaces/Tencent-Hunyuan/HunyuanDiThttps://huggingface.co/comfyanonymous/hunyuan_dit_comfyuiNodes: https://github.com/city96/ComfyUI_ExtraModels>Kolorshttps://gokaygokay-kolors.hf.spaceNodes: https://github.com/kijai/ComfyUI-KwaiKolorsWrapper>AuraFlowhttps://fal.ai/models/fal-ai/aura-flowhttps://huggingface.co/fal/AuraFlows>Index of guides and other toolshttps://rentry.org/sdg-linkhttps://rentry.org/rentrysd>View and submit GPU performance datahttps://vladmandic.github.io/sd-extension-system-info/pages/benchmark.htmlhttps://docs.getgrist.com/3mjouqRSdkBY/sdperformance>Try online without registrationtxt2img: https://www.mage.spaceimg2img: https://huggingface.co/spaces/huggingface/diffuse-the-restsd3: https://huggingface.co/spaces/stabilityai/stable-diffusion-3-medium>Related boards>>>/h/hdg>>>/e/edg>>>/d/ddg>>>/b/degen>>>/vt/vtai>>>/aco/sdg>>>/trash/sdg
blessed thread of frenship
SAAAAAAAAAAAAAAAAAAS!!!!!!!!!!!!!!!!!!!!!!!
THERE'S A FUCKING SAAS IN THE COLLAGE!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!
>>101655632>>101655654Baker has been compromised!
https://files.catbox.moe/2cbnsx.png
>>101655488got a message for the top right corner image, *hyuckptui*
>>101655785why so smug
>https://github.com/jhc13/taggui/releases/tag/v1.30.0>Phi-3-vision-128k-instructI wonder if it's crap
>>101656014probably censored to hell and back. remember seeing people mention how some models would only use gender neutral pronouns and refuse to mention skin color, not sure if phi is like that but i wouldn't be surprised.
>>101655818When you're the savior of the Six Faced World, you tend to be that way.
>>101656062I love when the caption is "the person is holding their breasts"
>>101656062>probably censored to hell and backI wouldn't be surprised. For now I just need something that captions objects and colors really well. Previous MS models have been decent and lightweight>>101656085I like "person is getting her tongue and her tongue pushed up"
>https://huggingface.co/SmilingWolf/wd-vit-large-tagger-v3have to test this too
>>101656269>Trained on Danbooru images You have my attention.
which one should i use
>https://huggingface.co/docs/peft/main/en/package_reference/boftand what the fuck is this>Diag-OFTwaht>>101656515seems pretty good, tried on photos
official bigma
>>101655488No Sigma the past two collages. SaaS in latest collage. >Is this an anime betrayal?>>101641968>>101642309>>101642323>>101646306>>101646584>>101646714>>101648131
>>101655488Can anyone replicate this style, it is sharp and soft at the same time
>>101656616>boftInteresting, will need to read a paper on it. The description left me with more questions than answers
Maybe?>A realistic full color detailed drawing of a beautiful woman with tribal tattoos and clothed in a fur bikini looking at the camera
>>101657543>https://github.com/derrian-distro/LoRA_Easy_Training_ScriptsDev branch has it implemented, but I cannot test it yet. I'll have time tomorrow
>>101657622 is for >>101657498Low sleep, many mistakes>>101657628ty, will reference that against what I read
>>101657622Its sharp but not soft, and definitely not beautiful
>>101657756Oh well
>IPNDM samplerworth using?
>>101657423two imgs in the last collage were sigma, and before that, three! :P i dont think ive even seen a completely non sigma collage desu
>>101657970I thought not initially, but it might be sharper than deis when testing again>>101657984Previous /ldg/ bread : >>101639278Which ones? I see SD and Cascade. There were a bunch of OP's w/o Sigma, but I never cared until SaaS creeped in. Tired and tilted so maybe don't mind me
>>101656576you must test every single option
Anyone have tips for prompting for a character holding weapons? I find it's really rare to get a gen where it doesn't break everything.
>>101659255Any decent model will respond well to prompts like "wielding" or "holding"
>>101655745nice
>>101659255I often see gens with the weapon to the side >>101660179ty
I summon the latent underworld to correct what was done.We must flush this thread from our memories
>>101655488oni SEX
first time using comfyui
We must unburden ourselves from what has been
>>101661333Good first go!
>>101661556i like this one
>>101661802Mid Century Art of a cat
>>101661895
>>101661083>>101661333>>101661802>>101662136VERY nice
>>101662519ty
absurd quality ITT
>>101656616>what the fuck is thisApparently there's a guide and it's really good imo https://huggingface.co/docs/peft/main/en/conceptual_guides/oft>>101662662Nice
>>101662687my feeling is that there isn't much "abstract photography" in the dataset but maybe im just failing to elicit it from the model
regardless, SD is outdated at this point. pixart looks better and better everyday and constantly surprises me
>>101662740100% boring normie IRL stuff
>>101662817>constantly surprises meMy same experience. It doesn't stop. How is 2k the same param count as 1k? Controlnet, etc. are definitely missing for Sigma though
Pixart IPAdapter pls
>>101662923Seems simple to train.. https://github.com/tencent-ailab/IP-Adapter/blob/main/tutorial_train.pyOn the project list!
Good night
>>101661333What model?
>>101663059>Seems simple to train..That's what they said about pixart itself yet there are only a few people doing it :d
>>101663139kolors, it's surprisingly good for not being a finetune
>>101663076>>101663097Very cool
>>101663558Thanks, I notice it now
>>101664006How do you prompt for this?
>>101664101lucky seed https://files.catbox.moe/9bj8li.png
>>101662662>>101662740>>101662776>>101662817>>101663097>>101664006>>101664169>>101664450>>101664501S P A MPAM
>>101664439
>>101657675Cool
I'm unable to refrain from exploring latent space
>>101662687>Apparently there's a guide and it's really good imoDid you try it? I wonder if it's upgrade for loras
>>101664551what model is that? gnarly nails aside it looks quite nice, good gen.
official pixart bigma and lumina 2 and that hunyuan finetune waiting room
>>101665627Pony realism
Are there any models or local apps that can do video inpainting like pika with 8gb of vram? I want to use ai to make women look pregnant
>>101666646ty, mind if i ask for catbox? really digging the style.
>>101663299There's no one training anything, it's a dumb metric.
>>101666880>its my comfy workflow and i get to choose the fetishes keke
https://blackforestlabs.ai/announcing-black-forest-labs/
>>101663299Hey anon, I train at night so there's fun results to check in the morning. Maybe speak for yourself>>101663690ty>>101664710ty>>101664958>I'm unable to refrain from exploring latent spaceThis ^>>101665134Seems like a more intelligent way of preserving "Hyperspherical Energy" without adding another linear layer like DoRA.. tldr; probably better and smaller but need to test still>>101667606>no one training anythingNo u
>>101667993>12bw-what a formidable fatty. any vram rich wanna give this a spin? seems like it's already supported by comfyui according to their huggingface repo.
>>101667993>trainingNo matches found
>>101668071i would but i don't see any comfyui workflow and i really don't want to fuck around with settings only to wind up disappointed
>>101668015>I train at night What did you train last night?
>>101668015My point is saying "Sigma isn't being trained" is stupid because there's basically no movement anywhere, it's just a couple people, as always, doing 99% of the work.
A new 12B parameter model just got open sourced, the examples are looking pretty good too.https://blackforestlabs.ai/announcing-black-forest-labs/
>>101668456It's not open source, there is no training code. It's open weights and not even that, they're delivering only the cucked distilled version. Means no training.
>>101668468There's two versions, if you're talking about schnell: https://huggingface.co/black-forest-labs/FLUX.1-dev
>>101668495>FLUX.1 [dev] is an open-weight, guidance-distilled model for non-commercial applications.
comfy thread
>>101668427you're mom
>>101668427Sigma 2k>>101668438It's okay anon, nobody expects you to know everything. Sigma _is_ being trained. There have been a lot of Sigma fine tunes released recently. And recently, I train it nightly.>>101668456Impressive but so huge. The gens coming out are absurdly good>>101668946kek
>>101668015>>101669184why do you think there arent MORE people training pixart? there are people doing it, this is true. i just expected more adoption at this point.
>>101669354no one is training anything, if you stand back for a second you'd realize that Pony is the only real training SDXL ever got.
>>101669364>no one is training anythingfair >Pony is the only real training SDXL ever got.perhaps so why arent furfags jumping to sigma?
>>101669364>pixart>hunyuan>kolors>lumina>auraflow>fluxall of this junk and not a single good finetune for any of them. this is how it's going to be for the next 5 years. endless pumping out 'almost good enough' base models that get forgotten in a week thanks to boring datasets and local finetuners lacking the compute needed to make anything with them.
>>101668094>comfyui workflowhttps://comfyanonymous.github.io/ComfyUI_examples/flux/
>>101669411Because people are dumb and need to be lead to water. Also base Sigma simply doesn't have enough parameters for something like Pony so you need someone to do something like 1.3B.>>101669427Auraflow is still in training, why would anyone fine tune something that is still in the oven? Pixart Next will be coming and that's got Nvidia money and their team actually gives a fuck about local training. Kolors is DOA because it's Unet. Hunyuan and Lumina require 40GB+ VRAM computers to train.
bigma status?
>>101669475>white dogIt seems to be working on finer details now but things are a lot more exploded than usual.
>>101669488this made me smile :)
>>101669488I love him.
>>101668456>12Bwill that work on my 24gb gpu though?
>>101668468>It's not open source, there is no training code. It's open weights and not even that, they're delivering only the cucked distilled version. Means no training.that's pretty easy to make the training code, stop bitching we got the weights lol
>>101669559I expect to see your training code soon then.
>>101669567anon, the worst part of imagegen model is to spent millions of dollars to train a good model, making some code with chatgpt is easy as fuck in comparaison, why are you crying?
>>101668456https://huggingface.co/black-forest-labs/FLUX.1-dev/discussions/1#66ab9dc4fd4ae9a7c49be855>I have a 3090 with 24gb vram. But 12b parameters in float16 format are still ~24GB and this does not include the two text encoders nor the internal state of the model.lmaoooo, what's the point then if no one can run it?
>>101669636Don't have to worry about competition if you make your model too big to run. It's what I would've done if I was SAI unironically.
>>101669657How do you convince your investors to spend millions of dollars on a model no regular user can use though? Sounds like suicide
>>101669706wtf? it's the FLUX.1-dev model?
>>101669732under the false premise that you can make money selling access via an api
>>101669706what's this picture? can you provide the source? that's interesting
>>101669354A two months ago there were 0 fine tunes. Momentum starts slow. How long was SDXL out before Pony?>>101669458>Because people are dumb and need to be lead to water.It's worse. They need to follow someone who says it's drinkable that they trust already.>>101669488HYPE!!! Mostly white background too
here are some 12b gens by an /lmg/ anon >>101668789>>101668964>>101669042>>101669149
>>101655488
>>101669873lmg sama i love you
>>101669873I'll see if I can get it running since bigma crashed
>>101669912>bigma crashedNOOOOOOOOOOOOOOOOOOOOOOOO!!!!!!!!!!!!!!!!!!!
>>101669925No, training sometimes crashes the GPUs and I have to restart the computer and it takes 15 minutes to load all the images so might as well dick around for a bit.
>>101669873that's insane, are we back? I waited so long for that day to happen!
>>101669873Wow looks really good, ty for linking!>>>/g/ldg eating good
>>101669939if you get it running could you try this prompt? >>101669706
>>101669912Bigma anon doesn't stop winning even during a crash. I've never had it crash during training btw. Are you getting random OOM's or something else?
>>101669982It's probably because I use my computer while it's training, I assume there's some sort of memory leak at the Nvidia driver level as it completely kills the video drivers where the screen just starts to stutter then freezes.
it doesnt fuck up anatomy of crouching/sitting subjects like SD3 holy shit thats nice
bruh it uses 999gb of vram swapping all that shit takes 2 min+ to make a image, barely uses the GPU (4090)
you can try out the model for free on replicate>https://replicate.com/black-forest-labs/flux-dev
tried this >>101669706oh wow
>>101670100https://replicate.com/black-forest-labs/flux-proThis one also works without an account, crank the safety tolerance to 5 so it doesn't stall on you.
>>101670081fellow 4090 user, same here i'm trying it out on /h/. the swapping is fucking brutal which is a shame because the gen speeds aren't that bad if it could actually stay loaded
>>101670081Meant RAM*, it hits the disk swap like crazy on 32gb>>101670155yeah, hopefully the model can be trimmed a bit to fit
>>101670148>without an accountim using it fine without an account>flux-proi believe flux-pro is their api only version, the flux-dev model is they released
>touhou, crino, 1girl, she has gigantic ass tits wooow
>an image of hatsune miku holding out both her hands, on her right hand is a red pill, on her left hand is a blue pill
>an image of hatsune miku, a large number of blue and red pills are coming out of her nostrilsi'll try "nose" instead next
>>101670277>an image of hatsune miku, a large number of blue and red pills are coming out of her nose
>>101670246We're so back.
>a picture of new york city, there is line of giant blue and red pill shaped buses on the road. the blue pill buses all have the face of hatsune miku on them
>>101670326wrong Pitbull
>a deranged serial killer using a crude cutout of hatsune miku's face as a mask, it is taped onto his face. he is very muscular with big nipples that look like sharp spearsnot what i asked for but alright. maybe it will do better with non esl prompting
>>101669427>>101669636Maybe this is a good time for me to complain for a bit.How come no existing training scripts can make efficient use of multiple consumer GPUs? I made an LLM training script (qlora-pipe on Github) that does pipeline parallelism. With that + full bf16 training + Kahan summation in the optimizer, I can match the performance of mixed precision while full finetuning something like llama 3 8b on 4x4090. But with SDXL, despite being a mere 2.6B parameters, I can't FFT it (not without compromises) using any training script.OneTrainer doesn't even support multi-GPU (lol, lmao even). With kohya, FSDP doesn't work. Deepspeed only got support recently, but DS Zero forces you into mixed precision training (where weights, grads, and optimizer state are all kept in fp32). Plus Zero has high inter-GPU bandwidth requirements and a decent amount of VRAM overhead it seems like. Basically I can't do a proper FFT of SDXL even on a fucking 4x4090 machine.Full bf16 training + adam with kahan summation uses 10 bytes per parameter. SDXL should easily be able to be FFT'd on just 2 3090s, which is a common setup for AI enthusiasts (at least in LLM land). No training script can even get close to this. And for the new flux 12b for example, a pipeline parallel training script ought to be able to do a decent rank lora on 2x3090 as well.At this point I just need to make a pipeline parallel training script for diffusion models I guess.
fluxsisters... I need more ram
If Flux can't do female feet, it's not worth using.
>>101670013Whoa.. I can even game while training on arch/kde. Base KDE/X uses 1GB VRAM and no more>>101670080Look at that hand!>>101670222Why so blurry?>>101670246Amazing
since it's a transformer i hope it can be quanted like llms>a holy painting of jesus christ caressing his big pregnant belly. there is a speech bubble above him saying "I shall name him PixArt Bigma"
>>101670414Because I asked for it. It's unfortunate that low-quality stuff like that is unreliable at best with these models, they filter out the low-quality images for training and that's all I want to generate.Maybe in the future there's gonna be image upload/customization options like Midjourney, that would really be something.
>>101670441kekd
>>101670395Basically no effort was done on this because they know that true democratization of training kills them. There's a reason why they keep the requirements >24 GB.
>>101670184I don't see much difference in image quality between pro and dev, that's cool
>an image of a cat just sitting there looking at the viewer. a speech bubble above him says "Near a tree by a riverThere's a hole in the Ground Where an old man of AranGoes around and aroundAnd his mind is a beaconIn the veil of the NightFor a Strange kind of FashionThere's a wrong and a Right"
>>101670413>female feet
>an image of "the last supper" but everybody has been replaced with drag queens and shemales, there is a child in a yellow coat for some reason
>ai generated
>>101670556I'm counting 5 toes on each foot, it's worth using.
>an image of hatsune miku holding out both her hands, on her right hand is a red pill, on her left hand is a blue pill, there is a speech bubble above her saying "the right makes you constipated, the left gives you diarrhea. choose wisely.", 1badassit gave me a 1cute instead
KEK
>ERROR>You have reached the free time limit.Death to non-local
>>101670629hunyuan feet anon is going to love this
>>101670632>Death to non-local>ERROR>You are out of memory.
>woman bowing forward, seen from behind, bikiniwill not make her bow at all
this had to be trained on synthslop, i can just tell. betting the paper will confirm another journeyDB masterpiece
>>101670725yeah it looks like it desu
>>101670725yes look at those feet
>an image of a cute little chibi anime girl smiling at the viewer, a speech bubble above her says "i'm going to say the nigger word"
https://replicate.com/black-forest-labs/flux-pro/examples>Hardware: CPU>Total duration 20.7swtf? how can it be so fast with cpu?
>>101670767AMD MI300
i'm impressed by stable-fast-3d with how it handles non character imageshttps://huggingface.co/stabilityai/stable-fast-3d
>>101670303KEK
>>101670767https://replicate.com/pricingAlso says the CPU is 4X. Is it regular 4 cores? or are they using some server grade 64 core x 4 = 256 cores?
>>101670725>synthslopyep
>>101670148I'm not feeling this model at all. I am not happy with the results from the pro version and that is the one you don't get to download. Dev is supposedly even worse. 12B model too.I ran bunch of tests and I prefer Dall-E 3 outputs over this pro model.This does not seem like a local model competitor, but more like a DE3, SD3 Large and Midjorney competitor.
>>101670643It's pretty bad at making feet actually. Worse than the Chinese models.Also I think you can easily finetune the china models for feet, but this not so much.
>>101670733something about it looks like Midjourney v4 outputs aesthetically. doesn't look authentic. ai trained on ai vibe. cool comprehension and a fun model but doesn't seem like the improvements line up with increased resource requirements. feels like an 8b model at max
>>101670794I really don't give a shit about 3D modeling
>>101670836hunyuan feet anon is going to hate this...>Patchouli Knowledge from touhou, her hair and eyes are purple and has many ribbons tied to her hair and other parts of her clothing. She wears pink pajama-like clothing and a night-cap with a gold crescent moon on it. Her dress has stripes of purple and violet, she is eating a cigarette
>>101670811you can test out the dev version herehttps://replicate.com/black-forest-labs/flux-devit's not that different to pro imo, and it's easily the best local model we ever had, that's a great day for me, fuck SAI
>>101670148>>101670424>Great image quality>Ok prompt understanding>Can do NFSW>Nice anatomy>Apache 2.0 LicenceThat's insane, I never thought we would get that day, WE ARE SO BACK
https://huggingface.co/black-forest-labs/FLUX.1-schnell/tree/main24GB model. Need a distilled version under 8GB
>>101670932nah i dont think this is it
>a little anime girl sitting on a dirty couch looking wasted, dark circles under her eyes. she has a cigarette in her mouth, hand and nose. next to her is a beer can with a cigarette in it instead of a straw. her room is dusty and decrepit, there is a thought bubble forming above her head, inside it is an image of a pack of cigarettes
>>101670081why you use the schnell one? it's the worst version
>>101670942what about the dev version? it's the better one no?
https://comfyanonymous.github.io/ComfyUI_examples/flux/>If you don’t have t5xxl_fp16.safetensors or clip_l.safetensors already in your ComfyUI/models/clip/ directory you can find them on: this link. You can use t5xxl_fp8_e4m3fn.safetensors instead for lower memory usage but the fp16 one is recommended if you have more than 32GB ram.How does that work? you can run the model on the GPU and the text encoder on the CPU?
>>101670954Dev version is bit better and is trained on the PRO versions
>A tense diplomatic negotiation in a grand hall, featuring representatives from 20 different countries, each wearing traditional attire. The scene should include interpreters, aides whispering to their leaders, and visible emotional reactions ranging from frustration to hope.Duonald trump
>>101670947idk, I picked the one that was running on the HF space
>>101670303where is he off to?
>>101670932yes anon, we're at insane level of back
>>101670932>Can do NFSWcan not, try to do do woman from behind bowing
Dallefags in shambles. We actually won
>>101670980it can do nude, and has great anatomy, finetunes will help for the poses
>>101670978/sdg/
>>101670932>requires RTX 4090>most likely bitch to train and make loras for>anatomy still lacking compared to Kolors for exampleI doubt that this is it, but maybe. Apache 2.0 license is the biggest thing that sets it apart from all the others.
>>101670856wasn't this supposed to be a safe space for alternative things?
>>101670725Can't copyright AI output so likely the investor-safe route. Still Sigma 0.6B btfo by this 20x larger model though got dayum
>>101671010>Apache 2.0 license is the biggest thing that sets it apart from all the others.not just that, the image quality is insane, and it has great prompt understanding and is perfect at text, this one is truly at API levels >>101670979
>>101670979
>>101671030it's also great at hands, holy fuck I never expected to get such a great local model in my lifetimehttps://reddit.com/r/StableDiffusion/comments/1ehknmh/new_ai_model_flux_fixes_hands/
https://comfyanonymous.github.io/ComfyUI_examples/flux/Can someone provide the links for the flux nodes?
>>101671016that's pretty cool ngl kek
https://huggingface.co/camenduru/FLUX.1-dev/tree/main>flux1-dev.sftwhat's a .sft? that's the model? why isn't it a safetensor model like the others?
>>101671016It's not a safe space but he doesn't even have a gen. Fuck that guy and keep posting your local gens like the OP suggests
>>101671091>SFT>SaFeTensors
>>101671065comfy has native support for this type of model. doesn't need extra nodes.
>>101671091.sft is safetensors according to comfy
>>101671065flux came with day 1 native comfy support
>ginger woman squatting, she is wearing round glasses and a stripped top with overalls a small frilled skirt and knee high pink boots,
>>101671030I believe when I can see that people can easily make finetunes and loras for this beast. I want to see good paper with no ClosedAI bullshit where all the sauce is hidden. I want training examples, training code etc.Just dumping the weights on the internet is not "open source" enough for me.
>a profound image of an anime girl in deep meditation, a white glow emanating from her head as she attains nirvana, the mighty glow from her eyes cause the entire image to tremble and warp. there is a speech bubble above her head saying "what if pixart 12b"
>>101671054But terrible at feet. What a price to pay. Still better than SD3.
>>101671104>>101671100oh ok my b I'm a retard, thanks kek>>101671146it's way harder to get millions of dollars to train a 12b model than making a training code, don't worry about it, the model is so good everyone will make the training work
>>101671169Also, keep in mind file extensions are just cosmetic, they merely inform programs of what to expect, you can rename a model to .jpg and still load it just file as long as the program recognizes it should try to load .jpg files as tensors
shrek lying on a recliner next to a pool, hes is drinking margarita and saying on a speech bubble "life is good"
>>101670478>they keep the requirements >24 GBI mean it's not some big conspiracy. Larger models are better, and will require high VRAM GPUs or multiple smaller GPUs to even train a lora. That's fine. I'm just surprised that with how popular imagegen is, none of the training script creators have put much effort into efficiently splitting models across 2+ GPUs so they can be trained with consumer hardware. It's possible and not even particularly difficult, it's just nobody seems to care. Like I said, at this point I'm seriously considering making my own pipeline parallel training script (will be open source if I do it), especially if this flux model or the new larger pixart model are any good.
we're so back
>>101671222wtf pikachu
>>101671216>it's just nobody seems to carenot enough people care, images are useless, meanwhile LLMs can actually do things, hence most people who know about ml are working on that instead
>>101671103>comfy has native support for this type of model. doesn't need extra nodes.I always thought I would never use this Spaggheti shit but here we go... the model is too god to be avoided at this point
>>101671216I'm just saying that as a business decision targeting >24 GB is a smart choice as it gives you brownie points and publicity having a "local" model while forcing most people to use your API. Honestly the best license would be something like:"Commercial use except for on-demand image generation via an API"
>>101671222This image made me realise I definitively have some weird cloth fetish.
>image of an heavenly immortal anime girl seated in deep mediation, her cultivation breaking through to 12b pixart biggerma realm, heaven and earth shatter as a speech bubble appear above her head saying "what if pixart 12b"
>>101671216>I mean it's not some big conspiracy. Larger models are betterthis, don't blame the model creators, blame Nvdia for nerfing the VRAM, it's still at 24gb since FUCKING 2018 (Rtx-Titan)
>>101671241That is changing now as all the modalities seem to be converging. Audio, text and image all on a single multimodal LLM.
>>101671286a colossal model that no one will be able to run too
>photo electric effect
The way it works with texts, even in cursive, is fucking amazing
>>101671366LMAOOOOOO
>>101671366so you made in run in comfy ui anon? how much VRAM does it ask? (image model + text encoder)
>>101671241For productive use, LLMs are more impactful than imagegen, yes. But for hobbyist use for "fun" (porn), I think imagegen is way more popular than anything people do with LLMs. Look how many loras and models are on civit compared to community LLM finetunes on huggingface.You see all these anons in this thread complaining that larger models are way too hard to finetune. This is only because existing training scripts are shit and can't do it, theoretically it's easily achievable. With a 2x3090 machine you should be able to FFT SDXL or train a lora on Hunyuan or the new flux 12b model.Fuck it, this weekend I'll make an attempt at a pipeline parallel training script for diffusion models, at least just to try to judge how much work it would be. Probably it shouldn't even be that much work if I reuse all the dataset loading code from kohya and make it based on HF Diffusers.
>>101671401Even in offload mode it maxes out my 4090
>A colossal anime woman towers above a plain field, her gigantic form stretching across the sky. The left side is bathed in a brilliant blue sky, while the right side is shrouded in a deep, velvety night sky, which she wears like a cape. Stars twinkle like diamonds across her gown, and the moon casts a silver glow on her majestic form.not what i asked but looks quite nice
https://huggingface.co/camenduru/FLUX.1->ae.sft>clip_l.safetensors>flux1-dev.sft>t5xxl_fp16.safetensors>t5xxl_fp8_e4m3fn.safetensorscan someone help a retard that will use Comfy for the first time of his life? do I have to download everything? what does those files mean?
Fresh bread is ready to eat...>>101671236>>101671236>>101671236
>>101671446>A colossal anime woman towers above a plain field, her gigantic form stretching across the sky. But it's what she's wearing that's truly striking: the night sky itself, draped across her shoulders like a majestic cape. Stars twinkle like diamonds, and the moon casts a silver glow on the folds of her celestial garment, as if the very fabric of the universe has come to life to adorn her.
>>101671426>Even in offload mode it maxes out my 4090what do you offload? the encoder text?
>>101671453ty baker
>>101655488I haven't come to /g/ in ages. What's the difference between this general and the stable diffusion one? It seems like anon are posting the exact same type of content in both.
>>101671426what if you load in 8bit?>--fp8_e5m2-text-enc --fp8_e5m2-unet
>>101671544/sdg/ allows saas gens, /ldg/ is local only
>>101671544the amount of free mental healthcare available in the country of the frequenters
>>101671544>exact same type of content in both.lurk long enough and you'll realize sdg is just a discord chatroom kek
>>101671617>lurk long enough and you'll realize 4chan is just a discord chatroom kek
>>101671630>he doesn't know most of sdg is avatarfags saying gm and gn to eachother and sharing suno songs
>>101671544There's none. This thread has no right to exists.
Ok. I came here because of Flux. I noticed /ldg/. and /sdg/ for some time now but I don't know the exact difference. I guess /ldg/ is more about tech than sharing gens, where /sdg/ is just about sharing gens? Or is there some drama that I missed that explains the split? If you could explain it as a veteran /sdg/ fag I'd be grateful
>>101672532/sdg/ allows dalle and gemma gens, /ldg/ is strictly local
>>101672553hmm. That doesn't sound right. Why is it called /sdg/ then? Also... that's it? No tripcode drama war or spamming autists fighting each other?
>>101672586it's a avatarfag and blogpost central as well>That doesn't sound right. Why is it called /sdg/ then?no clue, they just stopped caring for whatever reason
>>101672614fwiw, I like the "local [ai topic] general" naming better (like /lmg/) although iirc it didn't exist when /sdg/ came out. (or maybe it didn? I forget which came first) I know /lmg/ split from /aicg/ which is actually cancerous.
>>101672685anon frequently compares sdg to aicg and ldg to lmg, this is true
>>101672685>I know /lmg/ split from /aicg/ which is actually cancerous./ldg/ spilt from /sdg/ for similar reasons, it happened a while before the sd3 launch
nice
>>101675016sure...
>posting in the previous previous bred
Oni girl rocks
>>101675045Thanks
>>101675016>>101675031what did she mean by this
just wanted you guys to know that i have a boner
would you like help with that, anon
>>101671423>But for hobbyist use for "fun" (porn), I think imagegen is way more popular than anything people do with LLMslmao no, ERP is addictive as crack, at least the first few monthsthe reson you barely see less loras on LLMs is because the smallest LLMs are 4 times as big as the biggest image gen models