Discussion of free and open source text-to-image modelsRecently kneaded dough: >>103132365Theory of Ghosts Edition>Beginner UIMetastable: https://metastable.studioEasyDiffusion: https://easydiffusion.github.ioFooocus: https://github.com/lllyasviel/fooocus>Advanced UIForge: https://github.com/lllyasviel/stable-diffusion-webui-forgereForge: https://github.com/Panchovix/stable-diffusion-webui-reForgeAutomatic1111: https://github.com/automatic1111/stable-diffusion-webuiSwarmUI: https://github.com/mcmonkeyprojects/SwarmUIComfyUI: https://github.com/comfyanonymous/ComfyUISD.Next: https://github.com/vladmandic/automaticInvokeAI: https://github.com/invoke-ai/InvokeAI>Use a VAE if your images look washed outhttps://rentry.org/sdvae>Model Rankinghttps://imgsys.org/rankings>Models, LoRAs & traininghttps://civitai.comhttps://aitracker.arthttps://huggingface.cohttps://tensor.art/modelshttps://liblib.arthttps://github.com/Nerogar/OneTrainerhttps://github.com/kohya-ss/sd-scripts/tree/sd3https://github.com/derrian-distro/LoRA_Easy_Training_Scripts>SD3.5L/Mhttps://huggingface.co/stabilityai/stable-diffusion-3.5-largehttps://replicate.com/stability-ai/stable-diffusion-3.5-largehttps://huggingface.co/stabilityai/stable-diffusion-3.5-mediumhttps://huggingface.co/spaces/stabilityai/stable-diffusion-3.5-medium>Sanahttps://github.com/NVlabs/Sanahttps://sana-gen.mit.edu>Fluxhttps://huggingface.co/spaces/black-forest-labs/FLUX.1-schnellhttps://comfyanonymous.github.io/ComfyUI_examples/fluxDeDistilled Quants: https://huggingface.co/TheYuriLover/flux-dev-de-distill-GGUF/tree/main>Index of guides and other toolshttps://rentry.org/sdg-linkhttps://rentry.org/rentrysd>Try online without registrationtxt2img: https://www.mage.spaceimg2img: https://huggingface.co/spaces/huggingface/diffuse-the-rest>Related boards>>>/aco/sdg>>>/aco/aivg>>>/b/degen>>>/c/kdg>>>/d/ddg>>>/e/edg>>>/h/hdg>>>/tg/slop>>>/trash/sdg>>>/u/udg>>>/vt/vtai
Mochi 1 waiting room
>>103143847i'm getting bored with mochi bro, how was the cogvideo 1.5 or w/e? i'll prob give kijai a week to iron it out hopefully before i mess with it myself.
Blessed thread of frenship
>>103143967me and anons on our way to glass the planet with Nvidia HQ on it (we were intercepted by Amazon PMCs)
>debo rentry missingMhhhh
>>103143967awesome
>>103143894I didn't have a chance, i kept looking at the repo and seeing update after update trickling in and couldnt summon the effort to get it done as im kinda burned out on ideas that appeal and have been v low energy this weekend. >>103143967Nice.
>>103143978There's no better ostracism than oblivion, also it's off-topic. I'd rather anons discuss worthwhile things, instead of wasting time on schizo lore.
>103144014>nigboGo back
gentlemen and ladies show your skill at AI video generation.
>>103143994>im kinda burned out on ideassame here man. i've been hitting up chatgpt for ideas but i can't get mochi to do any good horror ones.
If you need ideas, start mixing and matching themes from whatever texts of culture you've recently encountered. Anything goes, films, videos, music lyrics. Bonus points for direct referances to your environment.
>>103144102can it do lora's? i bet it can't, i messing today with animatediff and using loras, shit works pretty decently, perhaps better than those so called animation loras.
>>103144136>>103144102like can take a latent from some source and de-noise it?
>>103144151if so then perhaps feed it with reference frames or image, controlnet an option? I've not even looked into these new video models i doubt my 12 G will work.
>>103144136>can it do lora's?i think they said they were hoping people would train loras for it but who knows if anybody will.>>103144164sadly no image2vid yet.
>>103144192>sadly no image2vid yet.not even experimental? Bummer real bummer because it would enable a lot of cool things.
>>103144202>We also anticipate that the community will fine-tune the model to suit various aesthetic preferences.nvm, that's what they said about anime, etc. who knows if that will happen though.
>>103144192>My view as i wake up in China, another drain exploding.
>>103144192>OMG111!! BREAKING TACTICAL NUKE JUST DROPPED ON UKRAINEWe're soon gonna see fake news taken to the next level aren't we, and all that slop news channels on youtube do by automated scripts. YAS, indoctrinate me harder daddy
>>103144228>We also anticipate that the community will fine-tune the model to suit various aesthetic preferences.no one has the fricken money for it yet, we don't even have a working animatediff for pony or a decent sdxl... It would take a lot of money to train such things.
>>103144281yeah and it's not even the HD model. i doubt anybody is going to waste the time and money to do it.>>103144161>>103144198>>103144222kek, i got this with a serial experiments lain prompt.
>>103144281I thought about maybe it can be cloud sourced between many many ordinary user machines but i don't even know if that would be possible due to how GPU work.
>>103144331plus its enormous amounts of data to say the least... Whos is gonna let their machines do all that work 24/7
>>103144192Without a proper toy dataset with training code with example captions they're asking a lot for people. Pixart will forever be the gold standard of how to release a community model. The included everything including the code they used to train, a dataset with captions, and even their hyperparameters.
>>103144014heckin basedI'm finally welcome here
>>103144339its the holy grail anon, it would open up creativity, imagine an AI that can read and entire book , generate the prompts for all the scenes and create a movie out of it. Or a video game. I'm sure there are a look of eyes watching and waiting, people with a lot of money to invest but they want something that actually works, they not gonna do the hard work.
>>103144331The only real solution that is feasible is a weird hybrid training where people with different datasets train synced models that are merged every X hours. But it still requires everyone to be on relatively the same hardware.
>>103144387Why the same hardware though? In any case I like pondering, there has to be a better way, video games already have realistic 3d environments.
>>103144424Because you need the weights relatively updated the same instead of some slowfag regressing the model.
>>103144424So what if we just create a game engine that can be played out by AI, something like secondlife but way better.
>>103144447Attention is luck and engineering, but mostly luck. You need to be producing and you need an idea to resonate.
>>103144440ah, really? I wondered if hardware would do things differently from other hardware when it can to training. This is why its not feasible then, hmm is it really like that?
fuk my spelling... Look, its down to us, because they not gonna hand it to us on a silver plate. So we must ponder and brainstorm and figure a way.
>>103144471It's not feasible because it's like a group project in school that requires both a brain and lots of effort, ultimately you have a bunch of people who say we should do something but don't actually care to contribute or take any initiative. You'll notice everyone talks about datasets and even though you don't need any special hardware to create a dataset you see no real effort there either.
interpolation works, we have as in in reference from one frame to another.>>103144527I will effort alright, I just need a time to get real life shit done first.
>>103144546I'll make the logo
>>103144527>I will effort alright, I just need a time to get real life shit done first.and real life shit is stressing me out which is why i am drunk... But I've lots of free time other wise, plus i'm autistic as fuck and do not give in easy. I've had lots of time to experiment with these AI models and i learn a lot about their limitations.
>>103144572and i have a if they can do it, then i can do it mentality.
>>103144589if you really wanted to do something useful and interesting make a website that lets people caption images and vote on the best captions
>>103144600Yep, I get you there, building datasets, i should start today, I wanted to make a program for aiding in that.
>>103144237sniff
>>103144615this sounds great
>>103143810>not including the preggie robo genDemoralizing
>>103144237bite
>>103144811if it's any consolation after seeing that pic i genned a few videos like this >>103144102
i give up, fucking mochi won't do a bird's eye/aerial view and gopro POV on the other side at the same time.
civitai fucking sucks
civitai fucking sucks [2]
>>103145009tell me about it.
Janny trigger happy ITT
>>103144900Blood for the blood god
>>103145051omg finally
>>103145009>>103145041it sometimes makes me want to punch babies (not really)it randomly breaks, its a laggy horrible piece of shit, it censors models that don't even need censoring...
>>103145059>censors models that don't even need censoring...There is this though, it just restored it self... https://civitaic.com/however you have to pay it seems, don't know if I trust it.
>103144161>103144198>103144222>103144237>103144248>>103144339>>103144389>>103144447>>103144510>>103144566>>103144588>>103144737>>103144830what's wrong with these posts?
>>103145059but hey, at least they "fixed" the advetisements and waste their resources on encouraging buzz begging
>>103145072jannie is gonna jannie and is triggered by gosh let me guess images of fully clothed children again? Give us a break janitor, i can make very tasteful images that have kids in them and its not against the law to do so... So can hollywood, so can every TV channel or magazine and who the fuck ever...
>>103145086as in...
>>103145072well what was it because I don't see those posts, I want to be the judge on whether they were bad or not and just take your word for it.
>>103145072Sometimes when you do something exceptionally bad you get a permaban that includes all your posts.
>ran out of memory when regular vae decoding. retrying with tiled vae decodingany way to stop this on reforge? because of it, it's taking 50 secs instead of 30 in an old a1111 version. ticked the "enable vae tiling", but it keeps happening
>>103144811thanks for the support anon, i published the lora for the atomic heart robot you can try it here https://civitai.com/models/933419/flux-atomic-heart-robot-lora-nsfw
>>103145127right, but what were they? I R curious, I don't want to get banned for the same thing, did they post tits or ass or something? i know that is not allowed here.
>not using 4chanX in Current Year
well anyway i am gonna attempt to gen an animated kot
>>103145146Usually it's related to US law, temp bans even for tits doesn't usually nuke your history. They probably posted AI CSAM somewhere.
>>103145164hmm, but what about /b/? I guess they have that same retarded shit like if its realistic its wrong but if its semi realistic its ok... Its all kind of bad really... Personally I don't care so long as its not posted here.
>>103145192Photorealistic is considered illegal by the FBI.
>>103145199Aye I know that, recently it was made illegal in the US.
>>103145199but what I meant was its all the same no matter what filter they use because ultimately one can switch to photo real model at 2 clicks of the mouse...
messy bread
>>103145221The difference is what you post. Don't post photorealistic. And don't post outside of the containment thread on b.
sloppy job
>>103145161Good luck anon, please show us when it's done.
>>103145231exactly, its hardly unlikely they could ever police AI, desu I don't know why these people think they can just post stuff like that.
>>103145242well i am trying, i will try only txt2vid here because i don't want to cheat.
All these post were deleted except the one with avatar/signature. Strange.
>>103145242to big to post here but its rubbish because i'm feeding in video frames of time lapsed storm clouds brewing but at 100% denoise just to see what it doeshttps://files.catbox.moe/cye9y4.webmnow feeding in empty latent images, it should in theory be better next time.
Very strange.
>>103145304so drunk i uploaded the wrong bloody filehttps://files.catbox.moe/7w5c9m.webmthe first one i posted is only 16 frames, nope this is no working at all...
>>103145329and i know why because i am 100% denoise i need more than 3 steps lcm sampler that is why.
>>103145337>>103145329and then i still have the controlnet depth map, lol, this is the drunkenness... I will probably wake up on the floor in the morning...
wow its really not fucking working at all, just rainbow colours, so fucking odd, do i really need to disconnect the controlnet nodes and not just set them to 0 or bypass, really odd.. Something is fuck here
hmmm cfg might be to high for lcm in this instance perhaps...
I am about to start a new project, but thought I would ask if it exists. I want a database of model + settings + lora + keywords and rated success. There is too much shit to keep track off with x/y grids. Does this exist in any form?
oh now i am getting something on cfg 2, so it was that, i'm meant to have a cat walking around a pole and i see the cat and the pole but this cat is tripping on lsd still
https://files.catbox.moe/29w7gd.webmLSD KOT!https://youtu.be/Uo-yc7JA5Ic?feature=shared&t=1139
Local Blog General
>>103145416>>103145431>>103145450>>103145475nobody is going to understand you if you post every 3 minutes with your random drunken shit.
>>103145501guess you will sit there in silence then won't you with your fucking 1girl posts ever few minutes you prick. sad little prick you are prick... fucking kys see if i care.
>>103145501you don't want to pick a fight with a druck autistic, take it from me. it might save you a lot of bother in future. because when we kick of we become extremely violent.
i am really struggling with image prompting, it is WILDLY different to LLM prompting where i can just tell it to make x component different. is there such thing as an llm frontend for any of the diffusion UIs that will take my prompt and translate it into a better diffusion prompt? its known in llm land using an llm to make your prompt better is significantly more effective than prompting yourself
>>103145605I was honestly just trying to get him to slow down so I could make sense of it. Seems he chose violence. He will be back in a few days complaining the jannies are unfair and he did not deserve to be reported.
its now looking more promising, definitely you need a total different setting from vid2vid when doing txt2viid using lcm model and weights.
>>103145072any ban for spamming gets automatic deletion of all posts. I got banned for "spamming" once because I QTd many posts at once (I was linking my favorite gens from the previous ten threads) and it deleted the mechanical keyboards general thread as well because I had posted the OP
>>103145625he is still talking you idiot it was me dumb ass. fucking end
Are there other ip adapters for sd1.5/sdxl besides the ones listed here?https://github.com/cubiq/ComfyUI_IPAdapter_plus
This is way to crazy, but i know way, its context and context overlap, so i need to up context from 8 to 32 and overlap to 24, it will lose sharpness but its the only way, then i might want to upscale or run through ksampler at very low de-noise high number of steps low cfg and rife interpolation
>>103145622it doesn't exist. It varies from model to model. Specifically if the model uses bouru tags. https://danbooru.donmai.us/wiki_pages/tag_groupsAitrepreneur had a flux one. It was kinda mid, but the structure was alright. >>103145654too many to list. Either rip them off pinned comments off of youtubers "How to controlnet video" or check huggingface.
>>103145622only with models trained on llm captions (fux for example) just ask your local llm for a pretty pretty prompt and youll be fine
>>103145685cool stuff
>>103145622You can use llm but it doesn't make the prompt "better" except in the case where you translate to chinese for a more chinese tag trained model (or vice versa from chinese to english).Llm or wildcards can however just add more tags so more is going on. That can be better.
>>103145622if the model doesn't suck ass, the creator tells you how to prompt for it. if you're using anything descended from pony, then yeah it's booru tags.
It can't do kot walking around poles anons, but it can do kots, this is too fast, its because well its set up for 30 fps video frames i can fix it, give a me few.
>>103145790>30 fps video frames i can fix it, give a me few.nah forgot it txt2vid is absolute shit, that thing does not even look like a cat. I am done, back to lurking and that faggot can stop crying, for no fucking reason at all...report me all you want dick face, i hope you fucking die.
>>103145622image prompting is more like an image search where you find an image based on a caption, tags, etc, depending on how the model was trained
>>103145815Chill mang
>>103145815you really shouldn't drink alcohol, you clearly can't handle it, probably mixing it with your meds
Hey guys, got a new laptop and would like to try local gens to more directly emulate certain artists that Bing can't do. I'm especially interested in emulating the style of old school anime like Leiji Matsumoto and Mamoru Nagano. What would be the best method to go about this? Any good guides I can use? Thanks.
>>103145918check OP
>>103145871i shouldn't but i'm ok until someone has a problem with me, isn't it interesting that one idiot can ruin it for everyone.... in a fight i would knack you seconds because your a beta ass cuck that not everyone follows your train of thought, not everyone is the same as you that likes to sit in silence autistic people use alcohol to relax and be them selves but oh no that is not allowed in this fucking shit world now you know why Trump won in a fucking land slide and we are gonna do the same all over the west because autism is sanity, its happy people not ugh you better shut up because i don't like you... you fucking miserable sad fucks, we are done with your types. No you will see the return of were pussy ass bitches get put back in their idiot boxes so the rest of humanity can actually enjoy being alive.
i will join reform party 100% we gonna crush labour you watch.
As the saying goes if you having nothing nice to say then fucking stfu
>>103145948OP has lots of great resources to be sure, I guess I'm just looking for someone's recommended front-end program to use with, for example, NoobAI-XL (NAI-XL) off civitai. Is there a UI that works best with that model? Does it really matter that much?
>>103146007that is just a question to prompt an UI war. Choose one of the beginner ones and then you can ask a question for your tastes. Some UIs favor customization and others "just work". I like recommending Foocus for first time because a bunch of stuff is hidden in the background. You will find that you outgrow it quickly, but at least you should have ideas about CFG, upscaling, loras and checkpoints.
because all my life you people bullied me... no more!
>>103146031Thank you.
>>103145918>Leiji MatsumotonoobaiXLNAIXL_epsilonPred10Version knows 1980's retro styles pretty well. 0.5 version might be better tho.I use https://github.com/Panchovix/stable-diffusion-webui-reForge and I can recommend it
>>103145654Bumping with my desire to know more. Share, anon.I know there's been an attempt to make a dedicated Autismmix finetune of it, it's on HF iirc but haven't tried it. Now obsolete due to Illustrious which doesn't work with ANY FUCKING VERSION. AAAAAA
>>103143810>>103137332Catbox for bottom-left?
wish mochi could do text like flux
>>103145969anon you're going to end up in prison because you can't handle your liquorchill the fuck out
>>103146261i'm ok you chill, i am chill. i gave you my workflow in thread already because its shared on catbox
>>103146281>i gave you my workflow in thread already because its shared on catboxbut its power is in vid2vid desu
>>103146261I do approach your call for calm, unfortunately i'm also ex military and did not know i was autistic, i have a high tendency to just rage when someone provokes me, but never violent. The police know about me, they know all to well what happen if drunk and mix with people, so these days i don't mix with people... And despite what i say i only attack when attacked first, its programmed into me.
but if you was to attack me physical in real life god help you i am a nasty bastard.
Trump voters are so fucking mentally ill.
well i am a trained fucking killer, hand to hand combat i will leave it at that.
>>103146400>Trump voters are so fucking mentally ill.i lie the uk hun, i am ex military yes, are you lonely?
this guy needs some 1girls fast
my bty when special forces just before i left so i gain a lot from that also infantry in jungle with was actually fun i enjoyed that but i was to get out because i didn't want anything more to do with what you see now... but my God i would kick you8r sorry ass to pieces if you piss me off enough but it would take a lot like harming my family etc.
On the off chance this is a well known and easily resolved bug: does anyone happen to know what "None type object is not iterrable" errors when trying to run controlnet could mean? I'm assuming I have the wrong models or something, because the preprocessor can generate a preview image just fine
>>103146281i will try meowing in the prompt see if it undestands that, these animation models are mostly trained on 1 girl dacing which is sad... It understand cut cat, but does not understand cat walkng around a pole which where we need to fill the void, why they never continued i will never know...
>>103146502it is a generic error, but I would check your resolutions match in input image, latent setting and checkpoint/controlnet recommended resolutions. Also, checkpoint matched controlnet base model.
>>103146502anon, these errors usual mean the model you are trying to use is not compactable, change them aground till they work? It is annoying but at some point your a feeding it some its not coded to do as in too large an image
>>103146502this>>103146543consider force resizing if its not an option consider learning how to impose your image on a background solid color, there are nodes for this actually i forgot there name if you need i can find them. Because they is what i do among other tricks such as background removal
>>103146502tell us what you are trying to do at least so we can help. Just the node dear and i will know, i will know for sure.
>>103146543So for instance, lets say I want the output as a generic 832x1216, I should just resize the input image to that resolution and then try it?>>103146604See that would be the smart thing to do, and I was considering it, but at this point I've been banging my head against it long enough I'm probably gonna call it and try tomorrow. [spoiler]Also what I was trying to do wasn't blue board appropriate[/spoiler] so I'll find a generic image of some guy t-posing to try and work on so I could actually post examples here I guess
>>103146502ok, i will tell you how and you must listen very carefull because i do not often do this you know because you should learn you self but i know how hard this one is..next post i will tell you but i want to know if you are still hear listening but i will tell you its not that hard.
>>103146502right what you need to first do is reliase we work in square format and the rest of the internet does not! So your image is not gonna fit inside the models we use, so you need a way to first downscale the image so that it will overlay over a solid background that is 1024 x 1024 id sdxl or pony, and 512 x 512 if SD 1.5. So how do you do that?you must use an image resize that keeps proportions, down scale that image so that it will fit onto a back screen of 1024 x 1024 for sdxl/pony or 512 x 512 sd 1.5does that make sense? i will now go digging if you can't figure out how, this does get a little trick actually.
>>103146733and le care, i am not doing it right now so you will have to say something or i will not bother...
and it is understandable he/she is a dog shit peabrain that would find even the most simplest instruction hard.
but i should continue anyway as that is the right thing to do. i'll tell this basic. Just trying to find the nodes, this is the problem with comfyui it does not take long to become cluttered. there is a node i have that takes one image as source and another as overlay image and it places the overlay image over the first, its really good. This node is really good for this use case among others such as placing a 1girl over a still background for her dance to make the animation clean... But also for this anon in attempting to import image that is not square anyway i think i said already enough smart anons will learn from that enough.
>>103146896let me expand a little, you have to remove background right but its not working? then you need to learn to remove alpha channel, invert mask, the place it over the new background, learn to do that you will be like wow, she is dancing holy shit...
most people do not get it, you look at it like its shit i'm just hinting why not get you it all? because money...
background removal is the hardest part, because not work like you think it works hint alpha channel
>>103146646>So for instance, lets say I want the output as a generic 832x1216, I should just resize the input image to that resolution and then try it?yes, at the very least it helps with the controlnet reliability. Depending on your UI you may need to tell it not to resize. I believe the default in A1111 is 512>>103146964can you go be schizo somewhere else?
nah fuck it i do not like blue balling people i will just give you my work flow for it, give me a few to sort it out.
do you think /sdg/ would be willing to trade their schizo anon
first i have to select a toktok video from youtube without stupid text infront of one girl
>>103146491Love these
how am i supposed to deal with LoRAs that don't have trigger words, do i just play with the clip strength until i get what i want?
>>103147052Jesus what a head ache for a hothead.
Alright let me start overI'm using reforged, so this is coming from the generic controlnet integrated rather then the extension you'd normally download. I'm told everything should work the same but I figured I'd mention it just to be safeI downloaded the models from https://huggingface.co/ckpt/ControlNet-v1-1/tree/main sticking to mostly the production .safetensors files. For the sake of this discussion lets say I'm going with openposeSo in the UI, I enable the controlnet dropdown, select openpose and it defaults to a preprocessor. First thing to note, the model dropdown is always empty at first, I have to refresh and then everything shows up.The preprocessor itself is working, see the example of some generic a-posing man I found and resized to 1024x1024, just as a test. If I set the model to the openposer one and try and generate I get the error.If there's something really really stupid I'm missing please let me know. Failing that I might just download A111 and try it over there to see if it's consistent
>>103147191Usually yeah
this is not 2016, this is the end of you shit forever so stfu moral faggot.
hmmm, baking dance video lets see how it comes out from depth map alone.
>>103147493yes back to 1girl dance
>>103147505yes
>>103146128Can you link that AutismMix IPAdapter finetune? I can't find it.
https://files.catbox.moe/du2cqh.png
>>103147684https://github.com/gustproof/ComfyUI_IPAdapter_plus_Style_ComponentsAlmost gave up on finding it again because I didn't bookmark or download it. Should be easier now though, do your thing Google search indexing!>Style Components is an IP-Adapter model conditioned on anime styles. The style embeddings can either be extracted from images or created manually. This repo currently only supports the SDXL model trained on AutismmixPony.>The model is trained on AutismmixPony. Functionality on other Pony derivatives is purely coincidental and not guaranteed. The adapter is not trained with Pony tags (source_anime, score_9, ...), so these tags can be omitted when prompting.Autism Autismmix Pony
Was away for a bit. 4500 gens to comb through, took me a while to look at them all. How long has the schizo guy been here? His posting style sounds familiar, I think we've had him before.
>>103147953could you tag nsfw so people don't think you are sharing a workflow?
>it's another OOM episodegod damn it, 12 GB really isn't enough, is it
>>103148311Don't worry, the 5060 Ti will have 18GB
>>103147191It depends on how the loras were trained. If it was trained with only one word as the caption, then the trigger word would always work. If it was trained with lengthy captions, then the trigger words have to be determined through guessing. Usually, it's several words among the captions. If the author doesn't mention how to activate the lora, then it's a red flag that the author has no idea what he's doing and the lora is low quality.
>>103148467Incredible as always, upscaling-chad
>>103148553thanks anonsome ComfyUI refactoring broke the tiledKsampler extension, which was an important part of an intermediate step in the workflowonly just got around to finding a replacement node that works well (tiled diffusion)
How do you judge the power level of other posters?
>>103149326by the cuteness of their 1girl. that's it.
>>103149354
>>103149406Cooked to shit. Turn your guidance down.
>>103148159Thank you.
>>103149426it's already at 1, I can't go any lower!
>>103149451Then you'd better consider using a different model.
>>103149481>FluxCan't win
Am I cooked?
>>103149489then you're probably confusing cfg and guidance.
>>103149406>>103149451>Buttchintrash
>>103149326details. I get not fixing hands. Eyes, buttons, duplicate parts/items lowers the evaluation of the power level a lot. Failing to not understand 240p vs blur/distort/artifacts drops their level to the floor. High level shit is successfully denying the AI the normal quirks like background crowds that make sense, forest paths/roads that have splits, or "rare" trees like pine. Character placement that also. Everyone walks in the road for some reason. Even a character that isn't centered is nice.
>>103148759How's that different from using a tile controlnet model without any additional nodes?
>>103149497>>103149597I give up then.
>>103149674they don't like my stuff either. I wouldn't sweat it too much.
>>103149653at extremely high resolutions you have to* process the image as broken up tiles even if you have enough vram to process the whole thing in one go, because the models don't think very well at 4k and make too many little mistakes*caveat: if you are doing a 1girl portrait or similar very simple composition, or using extremely low denoise, you may not have to
any news on this? maybe the most promising txt2img optimization in a while
>>103150081Eternal waiting room for the New Thing
>>103150081I asked about it yesterday and no-one @'ed me on this dead board
>>103150710can't wait
>>103150710damn, that's fugly
>>103149674Nah that's just fine
>>103149894Oh yeah, makes sense. Thanks, I gotta try again then. My issue was that I didn't want to set it up because as I understand you need to add a vision model to caption each tile individually. And while I do already use Florence, it's too tiny and didn't reliably work for captioning background SD hallucinations and the mood, in my experience.
>https://civitai.com/models/937345/proteussigma?modelVersionId=1049316interesting project
>>103152631What's interesting about it?
>>103152685He used a setting
>>103152718breathtaking
>>103152685eternal sdxl copium, nothing to see here
>>103152735I wish I could use a setting
We live in a setting.
>>103152718sets it for fast training, low quality output, equal gen speed. I was trying to figure out why this wasn't a article or something like that and realized that civit is the only place to post this kinda stuff if you don't want on social platforms. Kinda depressing.
I want to make high/medium resolution pixel art characters and non pixel, anime styled art for the cover/thumbnail of my game. What models should I use?
https://github.com/aigc-apps/EasyAnimate/tree/main/comfyuinew video model with native comfy support, wut
>>103152864https://civitai.com/models/478196/pixel-art-sakuemonq-artist-style-pony
>>103152880https://github.com/aigc-apps/EasyAnimateexamples look pretty shit tho
>>103152845Well it's better than nothing
>>103143810can you post some commentary on this? what is free? what does video? do any allow image upload?
>>103152930what?
>>103152930>commentaryThe OP is a long standing tradition in forum based social media sites. It does serve a purpose, but many detractors of the tradition say that people often don't read it and say that it only encourages nonsense reposes.
>>103152930>spoon feed me because I'm completely dysfunctional>your time is less valuable than mine, even though I'm clearly a low IQ, lazy moron
the chinks have invaded my robo dragussy gens, send help
mixing on new noob vpred & itercomp worked out well
How good are Apple chips and Snapdragon X for AI? Are their "neural engines" helpful or just a scam?
>>103153362Normal merge or did you have to shuffle text encoder?
>>103153401perpendicular component of vpred & itercomp = Padd_difference(vpred, P, 2.1) = ADslerp(vpred, AD, 0.5) = modelnot swapping TE, i think the extra epochs since they re-froze TE have sufficiently stabilized it
>>103153425>2.1whats that?
>>103153466the alpha of add_difference, it was added at 2.1 weight, then i slerped back with the original model at 0.5 weight. in some schizo a/b i thought the slerp made a smoother mix
>>103153482What the hell? I gotta try it, I've just been using DARE lately. Did you upload the mix?
>>103153528https://civitai.com/models/906562/chadmix-noobai-illustriousxl
>>103153582wait, I remember that dante gen, lmao, good stuff
>>103153582I'll give it a try
>>103152880I'm not enthused or vitalised by their samples.
ok anons, i have spent all afternoon trying to figure out the best img2img upscaling settings for Flux Dev, so might as well share my findings in case it helps anyone herein comfyui, after the base gen, I used a "upscale latent by" + "KSampler" + "VAE Decode" chain in order to achieve img2img upscaling. tried about every combination and here are the values you want in order to get the best end result.upscale_method : hardly matters but bislerp is the best imoscale_by : keep between 1 and 1.8, above 1.8 it's gonna start fucking up details. For a 1024x1024 base gen, 1.6 is the sweet spot (provided you have the vram).seed : unimportantcontrol_after_generate : keep on randomize or fixed if you chose 0 for seed.steps : 10 is the most efficient value, less than 10 and it fucks up, over 10 and results don't improve in any meaningful way.CFG : keep between 1.0 and 2.5, above that and it's deep fry land. I mostly use 1.5 to 2.0 because smaller details are more coherent. 1.0 runs twice as fast but some small details can fuck up sometimes. Your choice.sampler_name : deisscheduler : simple, beta is slighty worse in img2img upscaling imodenoise : keep between 0.5 and 0.8, with a sweet spot between 0.6 and 0.7. Prefer the lower end of that range for characters and images with text, prefer the higher end of that range for landscapes and anything without characters or text.thanks for coming to my reTardED talk
>>103153659agreed, they look like they have minimal movement, shitty fps, I won't even bother trying it, plus it has been made to run on low-end cards so I can imagine the quality, is just bread crumbs for vramlets
>>103153799kinda remidns me of the early days of animatediff
>>103153088yeah, I am asking to be spoon fed. Obviously, the OP is familiar with the sites. You think it's hard to write a couple sentences or adding commentary to something you are familiar with?Yeah, I am busy. Sifting through 30 fucking sites trying to figure them out for a unclear reason is something i'm probably not going to do when I pay a simple $8 a month and it just works. >online autists still not seeing the value of a simple thesis statement. I relate to my high school teacher more every fucking day because autists never seem to the see the value in providing a little context. Assumption of mind reading ability, intentional cryptic communication, and unwillingness to convey simple belief in concrete terms undermines and subverts the purpose of communication.
>>103153528it looks like someone else had a similar idea, with some extra schizo operation on the model, minus the extra traininghttps://civitai.com/models/935739/noobai-vpred-05-itercomp-fix?modelVersionId=1047459idkhttps://en.wikipedia.org/wiki/Procrustes_analysis
>>103154230>https://github.com/ljleb/sd-mechathis looks pretty sweet
>>103154317i used the comfy nodes of that libraryhttps://github.com/ljleb/comfy-mecha
>>103154335Cool. I hope there's Forge integration some day
>>103153754I'll save this.
>>103154719https://youtu.be/E8H-67ILaqc?si=_U202sHBNNGOm2QA
>>103143810does anyone how a recommendation on which tool to use for making an ongoing story? can any tool re-use images to create a different vantage point of anything?
>>103154776When it comes to keeping consistent features, your two ways to go about it would be custom trained loras and controlnet.
>>103154776https://github.com/NVlabs/consistorythere's this if it's what you're talking about but there's still no flux support yet.
>>103153754>seed : unimportantit must be different>>103154201>I relate to my high school teacher more every fucking dayYou must be 18 to post here. You could be 18 and in high school. Your post and language indicate otherwise. Go spend your parents money on online resources. >>103155082I was going to suggest controlnet reference image. This seems great. Thanks.
>>103149674do you use the special flux-specific nodes for guidance? check comfy's example workflow for flux and you'll see what I meanguidance in my experience should be set between 1.3 and 1.5 but you can get away with higher in some art styles
>>103155201are you retarded? he obviously means his teacher when he was in high school, ie, in the past. If he was referring to a teacher in the present he would say "my teacher" not "my high school teacher".
>>103154475>>103154497neat style
>>103155239>are you retarded? I assume the people don't have an attachment disorder for people from their past. Today's educational system would suggest that you would only have one type of teacher. I briefly forgot where I was. If you are so upset then you help him. I am pointing that he shouldn't be acting the way he is if he wants help. Of course that would require that "he" doesn't mean "me"
do i need to stick to these resolutions? will the image start to get messed up or is this just recommended for performance? i noticed once you to to gen really big it will start really fucking up like 2 belly buttons, 3 ass cheeks, 4 knees,
>>103155448>do i need to stick to these resolutions?for sdxl yes
>>103155448yes.backfeed a controlnet to upscaling if it is getting bad or lower your denoise and use a deterministic scheduler.
>>103155201>You must be 18 to post here.i'm detecting the projecting high schooler. I'm 36, and I relate to my high school teacher, who was a playwright, because he was seething at reading mountains of text without a fucking thesis statement. someone convince me why this thread is worth my time when I pay ideogram.ai $8 a month to make good images. I don't have a local GPU seemingly capable. I have a rtx 2060.
>>103155623NTA but you sound like a bitch, anyway a 2060 is fine for non-flux local >someone convince me why this thread is worth my timewhy?
>>103155623>someone convince me why this thread is worth my timethat sounds like a lot of work, i'll pass.
>>103155082very cool thanks.
>>103155762>>103155769I'm just asking for some coherent overview and reasoning on why all these tools are worth the time over paid services that aren't expensive. if you all want to be spergs in the club that's fine by me. I'll keep using paid ai generators that seem way better anyway. i've gone to a couple of the sites and the readme files don't even have the basic needed documentation.
Welcome to open sauce, anon
>>103155820keep using the paid ones dude since it sounds like it works for you
>>103155820i dont like paying someone to do something when i can do it myself, like having a local music collection instead of using spotify, or cooking yourself instead of ordering takeout - is one of many reasons for me
>>103155762i dont think a 2060 has nearly enough throughput to be tolerable, i think my current gen speed of 30-40s for a 1536x1536 on SDXL is too slow
>>103155885i had a 1060 when XL dropped and i suffered through it, it sucked but you CAN do it if that's all anon has right now, it's better to try and see what he can do with it rather than not
>>103155820local is uncensored, you can also train it and do all kinds of stuff with it that a paid service doesnt offer.
Need a Local Gen Manifesto in OP frfr
so how much would it be to rent a proper "local" instance remotely? the idea being to do flux-dev fp8 gens at 10s and high res SDXL gens at 5-10s
>>103155201>it must be differentit needs to be in the base sampler, in img2img it really doesn't matter unless you're doing some very veyr specific stuff
>>103155935It's already there, the very first line.
>>103155769>>103155623>>103155762>>103153088made this for you all with like 20 cents of credit and 30 seconds of my time. keep seething.
>>103156080now give him booba and vagene
>>103156080sounds like a lot of work
oooo i hope next collage is a good 'un
>>103156125
A complete guide for >>103155820is at https://www.patreon.com/SECoursesThanks for your comment!
>>103156169kek
Based Turkfag
>>103155942looks like ultra on this site is the only option that's actually an upgrade from my current machinei think ill pass for now
>>103156188can you use a shadow PC and and run shit from that? might be cheaper.
Local imggen is free only if you do not value your time
>>103156247what's that, botnet someone's gaymen rig?
>>103156267https://shadow.tech/it's a cloud pc gaming service. you get a desktop environment to install shit and they give you a top end GPU.
I'm confusing myself trying to use controlnet open poseSo I get an image that looks at least vaguely like what I want, send it to control net, run the preprocessor to get a skeleton, then edit the skeleton in the addon thing, I've done all thatWhat do I do if I just want to gen images off a skeleton control image I've already downloaded without a base image go between? I notice there's a "Upload independent control image" check in the img2img tab specifically, do I just leave the base image blank and upload my skeleton in the contronnet tab?
>>103156315Try it and get back to us, or try one of DrFukans excellent courses!
>>103156315yeah you just then select model and leave preprocessor empty
Git' yer buns over to:>>103156340>>103156340>>103156340
>>103152930There are tutorials out there, although admittingly barebones, that teach you how to do local gens. Afterwards, you need to figure it out mostly by yourself. If you have no desire to tinker, then it's not for you.