Previous /sdg/ thread : >>103464526 >Beginner UI local installEasyDiffusion: https://easydiffusion.github.ioMetastable: https://metastable.studioSwarmUI: https://github.com/mcmonkeyprojects/SwarmUI>Local installForge: https://github.com/lllyasviel/stable-diffusion-webui-forgeComfyUI: https://github.com/comfyanonymous/ComfyUISD.Next: https://github.com/vladmandic/automaticInvokeAI: https://github.com/invoke-ai/InvokeAI>Use a VAE if your images look washed outhttps://rentry.org/sdvae>SD 3.5 info & downloadhttps://rentry.org/sdg-link#sd35 https://civitai.com/models/896953/stable-diffusion-35-mediumhttps://huggingface.co/city96/stable-diffusion-3.5-medium-gguf---https://civitai.com/models/878387/stable-diffusion-35-largehttps://huggingface.co/city96/stable-diffusion-3.5-large-gguf>Try online without registration sd3.5-medium: https://replicate.com/stability-ai/stable-diffusion-3.5-mediumsd3.5-large: https://replicate.com/stability-ai/stable-diffusion-3.5-largesd3.5-turbo: https://replicate.com/stability-ai/stable-diffusion-3.5-large-turboflux-dev: https://huggingface.co/spaces/black-forest-labs/FLUX.1-devtxt2img: https://www.mage.space>Models, LoRAs & upscalinghttps://civitai.comhttps://huggingface.cohttps://aitracker.arthttps://openmodeldb.info>Index of guides and other toolshttps://rentry.org/sdg-linkhttps://rentry.org/rentrysd>View and submit GPU performance datahttps://vladmandic.github.io/sd-extension-system-info/pages/benchmark.html>Share image prompt info4chan removes prompt info from images, share them with the following guide/site...https://rentry.org/hdgcbhttps://catbox.moe>Related boards>>>/h/hdg>>>/e/edg>>>/d/ddg>>>/b/degen>>>/vt/vtai>>>/aco/sdg>>>/u/udg>>>/tg/slop>>>/trash/sdg
>mfw Resource news12/09/2024>LiFT: Leveraging Human Feedback for Text-to-Video Model Alignmenthttps://codegoat24.github.io/LiFT>IV-Mixed Sampler: Leveraging Image Diffusion Models for Enhanced Video Synthesishttps://anonymous.4open.science/w/IVmixedSampler-A646>ComfyUI Nukes & Embeddings for Text Encodershttps://github.com/zer0int/ComfyUI-Nuke-a-Text-Encoder>CLIP finetune: SAE-informed adversarial traininghttps://github.com/zer0int/CLIP-SAE-finetune>OpenAI's Sora video generation AI model arrives globally later todayhttps://www.engadget.com/ai/openais-sora-video-generation-ai-model-arrives-globally-later-today-182613208.html12/08/2024>ComyfUI HelloMemeV2https://github.com/HelloVision/ComfyUI_HelloMeme>SVDQuant adds ComfyUI supporthttps://github.com/mit-han-lab/nunchaku>Elon Musk’s X gains a new image generator, Aurorahttps://techcrunch.com/2024/12/07/elon-musks-x-gains-a-new-image-generator-aurora>Black Ops 6 Zombies Voice Actors Quit In Protest Amid Excessive AI Usagehttps://tech4gamers.com/black-ops-6-zombies-voice-actors-quit-ai-use>L4GM: Large 4D Gaussian Reconstruction Modelhttps://github.com/nv-tlabs/L4GM-official12/07/2024>Golden Noise Generator for ComfyUI (SDXL)https://github.com/DataCTE/ComfyUI_Golden-Noise>Negative Token Merging: Image-based Adversarial Feature Guidancehttps://github.com/1jsingh/negtome12/06/2024>ComfyUI: Masking and Scheduling LoRA and Model Weightshttps://blog.comfy.org/masking-and-scheduling-lora-and-model-weights>Switti: Designing Scale-Wise Transformers for Text-to-Image Synthesishttps://github.com/yandex-research/switti>SwarmUI 0.9.4 Releasehttps://github.com/mcmonkeyprojects/SwarmUI/releases/tag/0.9.4.0-Beta>VisionZip: Longer is Better but Not Necessary in Vision Language Modelshttps://github.com/dvlab-research/VisionZip>MEMO: Memory-Guided Diffusion for Expressive Talking Video Generationhttps://memoavatar.github.io/
>mfw Resource news12/10/2024>Sana-ComfyUIhttps://github.com/NVlabs/Sana/blob/main/asset/docs/ComfyUI/comfyui.md>Diverse Score Distillationhttps://billyxyb.github.io/Diverse-Score-Distillation>StableAnimator: End-to-end ID-preserving video diffusion frameworkhttps://github.com/Francis-Rings/StableAnimator>You See it, You Got it: Learning 3D Creation on Pose-Free Videos at Scalehttps://vision.baai.ac.cn/see3d>Gen-3Diffusion: Realistic Image-to-3D Generation via 2D & 3D Diffusion Synergyhttps://yuxuan-xue.com/gen-3diffusion>Normalizing Flows are Capable Generative Modelshttps://github.com/apple/ml-tarflow>No Annotations for Object Detection in Art through Stable Diffusionhttps://github.com/patrick-john-ramos/nada>iLLaVA: An Image is Worth Fewer Than 1/3 Input Tokens in Large Multimodal Models https://github.com/hulianyuyy/iLLaVA>GraPE: A Generate-Plan-Edit Framework for Compositional T2I Synthesishttps://dair-iitd.github.io/GraPE>Enhanced 3D Generation by 2D Editinghttps://jahnsonblack.github.io/GE3D>BiDM: Pushing the Limit of Quantization for Diffusion Modelshttps://github.com/Xingyu-Zheng/BiDM>Open-Source Acceleration of Stable-Diffusion.cpphttps://github.com/SealAILab/stable-diffusion-cpp>Surrey announces world's first AI model for near-instant image creation on consumer-grade hardwarehttps://www.surrey.ac.uk/news/surrey-announces-worlds-first-ai-model-near-instant-image-creation-consumer-grade-hardware>DiffusedTexture: AI-Powered Texture Generation for Blenderhttps://github.com/FrederikHasecke/diffused-texture-addon>Open Preference Dataset for T2I Generation by the HuggingFace Communityhttps://huggingface.co/blog/image-preferences>ComfyUI nodes to use MMAudiohttps://github.com/kijai/ComfyUI-MMAudio12/09/2024>LiFT: Leveraging Human Feedback for Text-to-Video Model Alignmenthttps://codegoat24.github.io/LiFT>ComfyUI Nukes & Embeddings for Text Encodershttps://github.com/zer0int/ComfyUI-Nuke-a-Text-Encoder
Thanks for baking the thread anon
>mfw Research news12/09/2024>Perturb-and-Revise: Flexible 3D Editing with Generative Trajectorieshttps://susunghong.github.io/Perturb-and-Revise>Birth and Death of a Rosehttps://chen-geng.com/rose4d>MotionFlow: Attention-Driven Motion Transfer in Video Diffusionhttps://motionflow-diffusion.github.io>Expanding Performance Boundaries of Open-Source Multimodal Models with Model, Data, and Test-Time Scalinghttps://arxiv.org/abs/2412.05271>Mind the Time: Temporally-Controlled Multi-Event Video Generationhttps://mint-video.github.io>From classical techniques to convolution-based models: Review of object detection algorithmshttps://arxiv.org/abs/2412.05252>DreamColour: Controllable Video Colour Editing without Traininghttps://arxiv.org/abs/2412.05180>LoRA.rar: Learning to Merge LoRAs via Hypernetworks for Subject-Style Conditioned Image Generationhttps://arxiv.org/abs/2412.05148>The Silent Prompt: Initial Noise as Implicit Guidance for Goal-Driven Image Generationhttps://arxiv.org/abs/2412.05101>SLayR: Scene Layout Generation with Rectified Flowhttps://arxiv.org/abs/2412.05003>UniMIC: Towards Universal Multi-modality Perceptual Image Compressionhttps://arxiv.org/abs/2412.04912>SleeperMark: Towards Robust Watermark against Fine-Tuning T2I Diffusion Modelshttps://arxiv.org/abs/2412.04852>Customized Generation Reimagined: Fidelity and Editability Harmonizedhttps://arxiv.org/abs/2412.04831>Addressing Attribute Leakages in Diffusion-based Image Editing without Traininghttps://arxiv.org/abs/2412.04715>Parametric-ControlNet: Multimodal Control in Foundation Models for Precise Engineering Design Synthesishttps://arxiv.org/abs/2412.04707>Latent Space Characterization of Autoencoder Variantshttps://arxiv.org/abs/2412.04755>ReF-LDM: Latent Diffusion Model for Reference-based Face Image Restorationhttps://chiweihsiao.github.io/refldm.github.io>Action-based image editing guided by human instructionshttps://arxiv.org/abs/2412.04558
>>103477309np
>mfw Research news12/10/2024>Efficiency Meets Fidelity: A Novel Quantization Framework for Stable Diffusionhttps://arxiv.org/abs/2412.06661>[MASK] is All You Needhttps://compvis.github.io/mask>Visual Lexicon: Rich Image Features in Language Spacehttps://arxiv.org/abs/2412.06774>Proactive Agents for Multi-Turn Text-to-Image Generation Under Uncertaintyhttps://arxiv.org/abs/2412.06771>Ranking-aware adapter for text-driven image ordering with CLIPhttps://arxiv.org/abs/2412.06760>MAVias: Mitigate any Visual Biashttps://arxiv.org/abs/2412.06632>MVReward: Better Aligning and Evaluating Multi-View Diffusion Models with Human Preferenceshttps://arxiv.org/abs/2412.06614>PrEditor3D: Fast and Precise 3D Shape Editinghttps://ziyaerkoc.com/preditor3d>MoViE: Mobile Diffusion for Video Editinghttps://arxiv.org/abs/2412.06578>UniPaint: Unified Space-time Video Inpainting via Mixture-of-Expertshttps://arxiv.org/abs/2412.06340>See Further When Clear: Curriculum Consistency Modelhttps://arxiv.org/abs/2412.06295>Rendering-Refined Stable Diffusion for Privacy Compliant Synthetic Datahttps://arxiv.org/abs/2412.06248>Towards Long Video Understanding via Fine-detailed Video Story Generationhttps://arxiv.org/abs/2412.06182>ASGDiffusion: Parallel High-Resolution Generation with Asynchronous Structure Guidancehttps://arxiv.org/abs/2412.06163>Latent-Reframe: Enabling Camera Control for Video Diffusion Model without Traininghttps://latent-reframe.github.io>FlexDiT: Dynamic Token Density Control for Diffusion Transformerhttps://arxiv.org/abs/2412.06028>Track4Gen: Teaching Video Diffusion Models to Track Points Improves Video Generationhttps://hyeonho99.github.io/Track4Gen>Post-hoc Probabilistic Vision-Language Modelshttps://aaltoml.github.io/BayesVLM>Nested Diffusion Models Using Hierarchical Latent Priorshttps://arxiv.org/abs/2412.05984
>mfw more Research news>Accelerating Video Diffusion Models via Distribution Matchinghttps://arxiv.org/abs/2412.05899>3D-Consistent Image Inpainting with Diffusion Modelshttps://arxiv.org/abs/2412.05881>MotionStone: Decoupled Motion Intensity Modulation with Diffusion Transformer for Image-to-Video Generationhttps://arxiv.org/abs/2412.05848>Self-Guidance: Boosting Flow and Diffusion Generation on Their Ownhttps://arxiv.org/abs/2412.05827>SILMM: Self-Improving Large Multimodal Models for Compositional Text-to-Image Generationhttps://silmm.github.io>Language-Guided Image Tokenization for Generationhttps://arxiv.org/abs/2412.05796>BudgetFusion: Perceptually-Guided Adaptive Diffusion Modelshttps://arxiv.org/abs/2412.05780>A Tiered GAN Approach for Monet-Style Image Generationhttps://arxiv.org/abs/2412.05724>Evaluating Hallucination in Text-to-Image Diffusion Models with Scene-Graph based Question-Answering Agenthttps://arxiv.org/abs/2412.05722>Remix-DiT: Mixing Diffusion Transformers for Multi-Expert Denoisinghttps://arxiv.org/abs/2412.05628>Do We Need to Design Specific Diffusion Models for Different Tasks? Try ONE-PIChttps://arxiv.org/abs/2412.05619>Uncovering Vision Modality Threats in Image-to-Image Taskshttps://arxiv.org/abs/2412.05538>Enhancing Sample Generation of Diffusion Models using Noise Level Correctionhttps://arxiv.org/abs/2412.05488>Multi-Armed Bandit Approach for Optimizing Training on Synthetic Datahttps://arxiv.org/abs/2412.05466>MotionShop: Zero-Shot Motion Transfer in Video Diffusion Models with Mixture of Score Guidancehttps://motionshop-diffusion.github.io>The Role of Text-to-Image Models in Advanced Style Transfer Applications: A Case Study with DALL-E 3https://arxiv.org/abs/2412.05325>Sound2Vision: Generating Diverse Visuals from Audio through Cross-Modal Latent Alignmenthttps://arxiv.org/abs/2412.06209>Generative Lines Matching Modelshttps://arxiv.org/abs/2412.06403
what's the correct version of anytest controlnet?
https://tensor.art/models/802224248199386282/Ca3-style-Illustrious-1.0can you guys generate some pictures with this?I want to see your best waifu prompts with this LORA.:)
>>103477326>>103477306>>103477319So cute!
>>103477378I'm using CN-anytest_v4-marged_pn_dim256.safetensors for Pony.
>>103477594This honestly.
>>103477604ty
>>103477681Have fun with it. It's fucking voodoo magic. By far the most powerful tool available for this sort of thing. It's the reason I'm still using Pony - if I can't use Anytest with the newer models, that's a non-starter.
>>103477765well im using anytest v1 with noob vpred and saw loads a new versions
>>103477783>well im using anytest v1 with noob vpred...that works?And I've never even seen v1 of it.https://huggingface.co/2vXpSwA7/iroiro-lora/tree/main/test_controlnet2 does not have a v1 of it.
>>103469735what is hyper/turbo? qrd?
>>103477923models that diffuse very rapidly and converge images in 1-6 steps
>>103477296Luigi Mangione Model dropped today: https://civitai.com/models/1025986/luigi-mangione-or-fluxd
>>103477993based
>>103477929where can I get one?
>>103478116https://huggingface.co/models?search=turbo
>>103478216begun, the maid wars have
>nylon jacketgreat token
>>103478161thanks, but which model did that guy use?
Is it normal for prompts to not look the same if swapping from reforge to comfyui?
>>103478831yeah. they use different encoding algorithms. they also generate seeds differently; comfy defaults to cpu while forge uses gpu
>>103478851Gotcha. I'll read some more rentry links and practice thanks.
>Flux in OP image>Stable Diffusion General?
behold, my latest masterpiece>>103479224we welcome all models here
>>103477296
>>103479508your magic android is on fire yo
Not related to Stable Diffusion but how do people usually do deepfakes nowadays?
>>103479754b/ has a celebrity thread going on right now. they can help
>>103480507Lookin good
Revisiting my oldest prompts from 2 years ago is kinda nostalgic
>>103478216nice maid
I miss schizoanon
Getting some interesting results using upscaling mid generation.
G'mornin Anons, have a great day!
>>103481750gm
>gm
>>103482375gm
>>103482407Morning!
gm
/ldg/ has lost the plot. Maybe I return to avatarland
>>103483416welcome home, avatar man
Gm anons
>>103483394gm>>103483416it was inevitable
Morning anons
>>103483416too much videos being posted?
>>103483738Gm Anon
>>103483519>>103483738>>103483768gm
>>103483738gm. I thought you were double-dipping the gms but I realize now that >>103481750 is a hampster, not a quokka
>>103483898Gm
>>103483919>quokka>hamsterwhat's next, groundhog anon?
>>103484094we have a mouse anon too
>>103484145the rodent mafia is taking over /sdg/
>>103484145He banned himself for three months instead of unfiltering you tho
>>103484174>rodent mafiaehrmaachkullya quokka is a marsupial
>>103484306>marsupial>related to kangaroosweird>In 1696, Dutch explorer Willem de Vlamingh mistook them for giant rats, and renamed the Wadjemup island 't Eylandt 't Rottenest, which means "the rat nest island" in Dutch. Vlamingh had originally described them "as a kind of rat as big as a common cat"close enough to rodent lel
>>103484328>1696 explorationprobably as close as humans will ever get to exploring alien planets. imagine rolling up to some untouched island and seeing a bunch of weird rats and weird trees and weird birds. must have been pretty cool
>>103484518that's how all tourists feel when travelling to NYC
mfw
>>103484944would love to see a bunch of eevee evolutions. I wonder if you could even get the AI to invent some new ones
>>103484929different seed, different wildcard
>>103485002i queued up 100, using an eevee and its evolutions wildcard, have to 'run an errand'
>>103485061>have to 'run an errand'ominousI hope no innocent health insurance CEOs get hurt
>>103481746you reinvented kohya deep shrink?
>>103485641sexo
>>103485665
>103485665>debo
>>103485780Fran is best girl
>>103485907correct
>>103485888checkedsaddest fucker on /g/ central
>debo going anon to compliment franpathetic
>>103486003why wouldn't I just compliment fran on main? its not like its some secret that I think he's one of our most talented posters
>>103485150https://files.catbox.moe/7iiij2.jpgit was much more boring
>>103486479>https://files.catbox.moe/7iiij2.jpghuge, both figuratively and literally
>>103486595https://files.catbox.moe/prh911.jpggrid from this morning, each grid took like an hour 47 minutes, i sometimes run them while away, lots of wildcards on the tag strengths
>>103486650pls post 88
>>103486650what a cool series. some of them came out super stylish
>>103486658https://files.catbox.moe/fejg3b.pngif you mean this one>>103486773i suppose i could pick out the better ones, and see which wildcarded strength tags were more in effect, too much effort rn
>>103486819thx, i dig the curves and shapes in two colors
a.i. toes... good reason to prompt socks
>>103486986That's radical
>>103487024>molten cheesethat's gotta burn
I should probably get something to eat
>>103487090burgers?
>>103487191I can practically hear the nom
>>103487191How many boyfriends did you and your mother share?
>julien
>trani
>>103487191I got tacos, also nice burger
ざあ—こざあ—こ
>>103487612dark migu
> >103487191Don't forget the flip
easydiffusion is literal dogshit. Anything actually good for historical battle scenes and ruined cities like Stalingrad during ww2? I need something advanced af and also please provide the optimal image settings if possible.
debo the pedo
>>103487798hangry
starting to think we'll never get a General A.I. board
>>103487949AI will continue to consume more and more of our lives and culture that it will either 1) become a necessity for there to be an AI board or 2) AI is so ubiquitous that it can't be contained in a single board
>there will either be an AI board or there won't be an AI boarddeep
>>103487949>>103487984>>103488016i just want to post 1girl
>>103488100we noticed
>>103488144( ͡° ͜ʖ ͡° )
>>103483919>a plump Syrian Hamster
debo-esque situation
>>103487363>>103487351>>103487147Best one's I've seen for a while (I waited 15 minutes to post this)
man, my own custom LORA is so great it looks 100% hand made and not like a generic AI render.
>>103488186this is amazing.
>>103488186i tried something similar a long time ago
brand new ntr versionhttps://civitai.com/models/926443?modelVersionId=1152565
>>103488321Why does it ask for two quality prompts?
>>103488431double the quality
>>103488431nice
why do all the gens itt suck?the only anon doing creative things is koff
>>103488673hello koff
>>103488692not koff im just telling the truth
>>103488703who is koff?
https://youtu.be/WhTUL3PEfis?si=8dFYV_HlTavH_ynK
>>103488635Thanks, cute catgirl
Next Thread>>103489402>>103489402>>103489402
feel her up
fill*
>>103478481>>103478057What model and / or lora is this?
>>103485780>>103485641What are you using to get those multiple views / comic format? Just tags or a specfic lora?
>>103489600ntr mix
>>103489642What tags are you using?Does that count as pixel art?
>>103489672>pixel art, pixelated, pc98, ditheringI'd add in a pixelation post-processing step too if I wanted to clean it up but I was just experimenting with the tokens to see how they'd come through
>>103489618Multiple views, panels, and speech bubbles.
well well well, what do we have here? Fran? Debo? Koffing? some guy posting eggs or whatever? kino
>>103489488needs bigger boobs. and possibly bunny ears
apropos nothing i beat my machineit's a part of me---i beat my machinei beat my machinei beat my machinei beat my machinei beat my machine
recommended programs for preparing a LoRa dataset? surely there's something to do cropping/masking/captioning quicker than doing it manually with regular photo editors and text editors