[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/g/ - Technology

Name
Options
Comment
Verification
4chan Pass users can bypass this verification. [Learn More] [Login]
File
  • Please read the Rules and FAQ before posting.
  • You may highlight syntax and preserve whitespace by using [code] tags.

08/21/20New boards added: /vrpg/, /vmg/, /vst/ and /vm/
05/04/17New trial board added: /bant/ - International/Random
10/04/16New board for 4chan Pass users: /vip/ - Very Important Posts
[Hide] [Show All]


Janitor applications are now being accepted. Click here to apply.


[Advertise on 4chan]


Discussion of Free and Open Source Text-to-Image/Video Models and UI

Prev: >>106647201

https://rentry.org/ldg-lazy-getting-started-guide

>UI
ComfyUI: https://github.com/comfyanonymous/ComfyUI
SwarmUI: https://github.com/mcmonkeyprojects/SwarmUI
re/Forge/Classic/Neo: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassicneo
SD.Next: https://github.com/vladmandic/sdnext
Wan2GP: https://github.com/deepbeepmeep/Wan2GP

>Checkpoints, LoRAs, Upscalers, & Workflows
https://civitai.com
https://civitaiarchive.com/
https://openmodeldb.info
https://openart.ai/workflows

>Tuning
https://github.com/spacepxl/demystifying-sd-finetuning
https://github.com/Nerogar/OneTrainer
https://github.com/kohya-ss/sd-scripts/tree/sd3
https://github.com/derrian-distro/LoRA_Easy_Training_Scripts
https://github.com/tdrussell/diffusion-pipe

>WanX
https://comfyanonymous.github.io/ComfyUI_examples/wan22/
https://github.com/Wan-Video

>Chroma
https://huggingface.co/lodestones/Chroma1-Base
Training: https://rentry.org/mvu52t46

>Neta Lumina
https://huggingface.co/neta-art/Neta-Lumina
https://civitai.com/models/1790792?modelVersionId=2122326
https://neta-lumina-style.tz03.xyz/

>Illustrious
1girl and Beyond: https://rentry.org/comfyui_guide_1girl
Tag Explorer: https://tagexplorer.github.io/

>Misc
Local Model Meta: https://rentry.org/localmodelsmeta
Share Metadata: https://catbox.moe | https://litterbox.catbox.moe/
GPU Benchmarks: https://chimolog.co/bto-gpu-stable-diffusion-specs/
Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-one
Txt2Img Plugin: https://github.com/Acly/krita-ai-diffusion
Archive: https://rentry.org/sdg-link
Bakery: https://rentry.org/ldgcollage

>Neighbours
>>>/aco/csdg
>>>/b/degen
>>>/b/realistic+parody
>>>/gif/vdg
>>>/d/ddg
>>>/e/edg
>>>/h/hdg
>>>/trash/slop
>>>/vt/vtai
>>>/u/udg

>Local Text
>>>/g/lmg

>Maintain Thread Quality
https://rentry.org/debo
>>
>>106650916
are the collages THIS bad on purpose?
>>
File: littlemisty33.mp4 (401 KB, 556x582)
401 KB
401 KB MP4
blessed thread of frenzone ;3
>>106650933
its not that serious mate
>>
Someone is having trouble with my lora, I never bothered to test it on anything but Q8 and now I realize it's horrendous garbage on fp8_scaled
What is causing this massive discrepancy in quality?
Same settings 2.2 lightx2v 6 steps 3/3 split
Catboxes
fp8
https://files.catbox.moe/myltvc.mp4
Q8
https://files.catbox.moe/ayo5gw.mp4
>>
>>106651004
The problem is that Chroma excels at nsfw, faces, and soul. It does not excel at being detailed or coherent. Depending on what I'm doing, I just can't use chroma. Maybe I can inpaint the faces, though. Qwen has a ton of slop issues though. I feel like I'm getting somewhere but I need to train more.
>>
>>106651055
>It does not excel at being detailed or coherent.
Have you tried doing a second pass? I've good results with ironing out the scuff with it
>>
Chroma HD is coherent if I use speed loras, but then it has those flux lines across it if you go to 1024x1024 or above. There's no winning.
>>
>>106651055
>The problem is that Chroma excels at nsfw, faces, and soul
so it's the best?
>>
>>106651075
Gen at ~1024 and then upscale ?
>>
>>106651090
It does it at 1024, that's why I said "1024 and above".
>>
>>106650960
>>106651004
nice
>>
>>106651036
Something is probably wrong in how the lora was trained with some settings that "overoptimized" to a specific version of the model it was trained on
I remember some chroma loras for example having horizontal line artifacts on specific, more complex prompts, when inferenced with even the Q8 versus bf16 version of the model

Post lora metadata
And what is high_noise_model.safetensors in the lora loaders anyway?
>>
>>106651073
How exactly does that work? Just img2imging on lowish denoise? I've tried using it as a second pass for Qwen that way and it destroyed fine details.
>>106651080
It's probably best in terms of actual final images it can produce, but there are styles and concepts it can't do. I'm on an autistic Orientalism wildcard kick and I'm doing shit that would never work in chroma. I wish I could get a little less slop and a little more soul, but Chroma just doesn't have the adherence.
>>
File: Capture.png (58 KB, 717x999)
58 KB
58 KB PNG
>>106651118
Trained on fp16
>And what is high_noise_model.safetensors in the lora loaders anyway?
Lightx2v
https://huggingface.co/lightx2v/Wan2.2-Lightning/tree/main/Wan2.2-T2V-A14B-4steps-lora-rank64-Seko-V1.1
>>
I wonder if wan animate can be used for static images to drop in and replace characters?
>>
File: file.png (662 KB, 2443x1838)
662 KB
662 KB PNG
>>106649859
you've got a low noise lora connected to the high noise model.
>>
File: AnimateDiff_00327.mp4 (990 KB, 1280x720)
990 KB
990 KB MP4
>>
File: sp.png (277 KB, 1057x996)
277 KB
277 KB PNG
>>106651140
Just don't use Chroma HD and you should preserve or enhance detail. 0.60-0.65 for no or minimal changes.
>>
>>106651204
lol is that a psych bed? I've been using Chroma HD because I thought it would handle small details better. I'm surprised to hear that isn't the case. I'll add that to my list to try. I'm attempting to start a lora training before bed.
>>
>gen with illustrious
>5-8 seconds with upscaling
>gen with chroma
>40-70 seconds
>8 minutes with upscaling
what the FUCK kind of vram do i need for this? i only have 16gbs
>>
>>106651230
>I've been using Chroma HD because I thought it would handle small details better
https://files.catbox.moe/qphnpf.jpg
A common mistake.
>>
>>106651234
Chroma is just retardedly slow
>>
>>106651234
>8 minutes with upscaling
Nigga what are you doing
>>
is it possible to use chroma adequately on very low vram (8gb)? i've got chroma-base q4_km, and using chroma-flash-lora_r32-fp32 but results are very artifacty like previous guys said, vertical stripes and all. am i missing something, or should i just give up trying
>>
File: 1731191973249308.mp4 (1.08 MB, 640x640)
1.08 MB
1.08 MB MP4
I can get her to show tits easily but can't make that light in the alley go dark
>>
>>106651249
Wow, I guess some new versions have come out? I've been out of the loop. It's hard to say much when the style for the examples isn't very detailed, but 2kdc looks pretty solid.
>>
Does Chroma have a dmd2 lora or something alike?
>>
>>106651278
I'm the guy from earlier in the thread complaining about the stripes and they actually get worse with the full version lol. You can either live with it or set up a workflow to use tiled diffusion to rescale it with 768x768 tiles, since that resolution seems to avoid the striping.
>>
>>106651310
It has a flash version which generally sucks, and there are a bunch of experimental loras for it that speed it up and work okay but add artifacts.
>>
Can we just not talk about chroma? I find it terribly boring
>>
>>106651380
learn what a filter is newnigger luddite brownzoom
>>
File: 00458-197389854.png (1.86 MB, 1248x1848)
1.86 MB
1.86 MB PNG
>>
>>106651380
yeah, everything must revolve around you, sorry
>>
Is Chroma 2K exclusively meant to be used at 2K resolution or can do it everything in-between as well?
>>
>>106651401
It's not 2k resolution but total pixel size
>>
>>106651234
chroma is 512x512 fluxslop. a complete waste of $150k
>>
File: laura.jpg (51 KB, 493x464)
51 KB
51 KB JPG
>>106651380
No. Speaking of, chroma is too much fun, especially with lighting prompts + cycling through various lighting loras.
>>
So T5XXL's max token count is 512, but if I concat two conditionings and feed it to Chroma does that give me 1024 tokens or is there still a 512 limit?
>>
>>106651556
What novels are you prompting that you'd ever run out of tokens?
>>
>>106651563
I'm not but I'm curious.
>>
>>106651556
obv it will either error out or cut it off
>>
>new lora comes out
>you gotta rerun all images of hot women x you with it
>you discover a new hot woman
>you gotta rerun all loras with her x you
At some point a new model drops and everything begins anew
>>
>>106651708
And by lora I was obviously talking about a new video gen porn lora
>>
>>106650933
Researchers are not often known for their aesthetic sensibilities
>>
File: 1730819204000017.mp4 (675 KB, 640x640)
675 KB
675 KB MP4
>>
>Chroma-2K-QC.safetensors
what the heck does QC stand for and how is this one different?
>>
>>106652155
quality control
>>
>>106652170
aight, ill give it a go then
>>
>chroma takes too long on my GPU
>leave it genning overnight
>use the images to make a lora for illustrious
>now i have chroma at home
i am literally a genius
>>
should've made a sd1.4 lora instead BAKA
>>
dont ever call me a baka again, or i'll suck your ochinchin
>>
File: ComfyUI_00908_.png (1.37 MB, 1024x1024)
1.37 MB
1.37 MB PNG
chroma radiance is pretty interesting.
>>
File: 1727605749444250.png (685 KB, 768x512)
685 KB
685 KB PNG
>>106652195
So true, 1.4 is still king
>>
File: 1751522822892401.png (1.67 MB, 1536x1152)
1.67 MB
1.67 MB PNG
>>106652240
>>
File: 1741220722407031.png (3.65 MB, 2304x1152)
3.65 MB
3.65 MB PNG
>>106652254
>>
>>106651556
The model's text conditioaln is like a cup of water. You can't put more water in a cup than its capacity. The text encoder on the model is the actual hard limit and most models don't even train that high for tokens because the cost per token is quadratic.
>>
>>106652155
>>106652170
What about DC? There's a DC there as well. I can only assume this information is on a discord somewhere
>>
>>106652216
That looks really nice. I need to give Chroma another go.
>>
File: 1754137115064086.mp4 (878 KB, 640x640)
878 KB
878 KB MP4
>>106652111
>I'll be taking that
>>
Speaking of the old days of sd1.
I remember being able to create some really nice abstract and painterly stuff by genning at very low cfg and steps, then inpaint with higher values.
It's impossible to do with newer checkpoints. Anyone know how I can recreate that workflow with the new stuff?
>>
>>106652295
detail calibrated
2k-dc or 2k-qc are merges
>>
>>106652418
where are you getting this information for the chroma variants? am i missing something obvious?
>>
>>106652448
it's been explained in these threads before
>>
File: RA_NBCM_00001.jpg (600 KB, 1872x2736)
600 KB
600 KB JPG
>>
>>106652448
Probably devs discord or twitter
>>
Does srpo work for neoforge?
>>
>spent hours trying to get those gradio text to speech/clones to install
>ask ai for help
>says i have to downgrade pyshit from 3.12 to 3.11
>try chatterbox tts for comfyui
>follow the instructions slowly and carefully
>only the voice changer kind of works while the rest of the workflow errors

I'm not downgrading or having 2 versions of pyshit (been there, done that)
>>
>>106652216
>chroma radiance
Is there a definitive model and workflow? Looks like there's a million snapshots
>>
File: 1733926622219073.png (3 MB, 3008x992)
3 MB
3 MB PNG
>>
File: 1736938560677999.jpg (1.14 MB, 1664x2496)
1.14 MB
1.14 MB JPG
>>
File: 1742081555381374.png (156 KB, 1642x605)
156 KB
156 KB PNG
Are you ready for China to save local once again in 3 days?
>>
File: 1750854089091436.gif (112 KB, 220x220)
112 KB
112 KB GIF
>>106652709
is it censored?
>>
>>106652717
it's gonna be HunyuanImage Edit and Tencent always makes the least uncensored models of them all
>>
File: 00000-2502409727.png (1.36 MB, 896x1152)
1.36 MB
1.36 MB PNG
>>106652483
>after like 2 minutes of loading and unloading the image comes out

Extremely basic prompt. It truly is over for instagram thots.
>>
>>106652709
This dude is full of shit.
>>
>>106652767
you need to lurk more
>>
>>106652736
what about onlyfans thots
>>
>do my usual upscale at 2x with 4 gens
AUGH

This is like 6 seconds on weeb checkpoints.

>>106652773
Didn't taxes eradicate them?
>>
File: IMG_4914.png (1.89 MB, 1284x2778)
1.89 MB
1.89 MB PNG
By the way you can use tellif.ai to detect if images or videos are real or AI generated or modified. It is free and pretty accurate
>>
>>106652791
>consistent pixel patterns
I guess that's because of the VAE right? try with a chroma radiance image instead
>>
File: RA_NBCM_00006.jpg (926 KB, 1872x2736)
926 KB
926 KB JPG
>>
>>106651278
It's fine. It's not a fast model anyways. My old PC has a 3070 and I let it run a q8 of Base.
>>
>>106651401
It's in megapixels. It can handle lower res just fine.
>>
Is there a way to do pic related with Forge or ComfyUI?
>>
File: 00004-2482050065 copy.jpg (1.33 MB, 3584x4608)
1.33 MB
1.33 MB JPG
Meh, not impressed with the upscaling. Just a quick double 2x.

Also, there seems to be a severe issue with eye contact to with the camera/viewer. Probably because the portrait photographs that's been used for the training are with models looking off into the distance.
>>
am i having a skill issue or do teeth always come out fucked with loras? i even added a few images of just the teeth from multiple angles and they still come out like a piece of unmolded clay.
>>
File: 00011-777536207.png (1.89 MB, 896x1152)
1.89 MB
1.89 MB PNG
I take it that chroma only works in comfyui with the torch updates? Is this update going to brick my usual comfyui workflows or is it just a separate feature?
Neoforge doesn't like it.
>>
>>106652834
remove a background? Yes, dozens.
>>
File: Wanimate_00017.mp4 (154 KB, 480x832)
154 KB
154 KB MP4
>>
>>106651055
>The problem is that Chroma excels at nsfw, faces, and soul. It does not excel at being detailed or coherent
How many times do we need to discuss this trade-off and why it exists? We've been saying the exact same shit over and over since the SD1.5 days.

Model A, trained on extremely heterogenous data:
- broadly capable
- natural-looking
- "sovl"
- fun to prompt
- if you gen 400 images from the same prompt all 400 are different
- "bad hands" "limbgore" "incoherent" etc
- cannot produce a perfect image

Model B, finetuned from A, trained on very tightly-curated and heavily tagged data:
- not able to do much outside of what it was specifically trained to do
- "good hands"
- prone to sameface
- needs to be prompted a certain way
- prone to looking plastic or otherwise uncanny
- no point in genning 400 images from the same prompt because they'll all look more or less the same
- one anon makes a post every thread to say some variation of "Model B blows Model A out of the water and proves the Model A team didn't know what the fuck they were doing", posts completely soulless plastic gens which he thinks are unimpeachably perfect state-of-the-art
>>
Why is it Chroma is the only model that people crawl out of the woodwork and write entire essays when someone mentions its shortcomings?
Is there some kind of Chroma derangement disease?
>>
>>106653036
>- one anon makes a post every thread to say some variation of "Model B blows Model A out of the water and proves the Model A team didn't know what the fuck they were doing", posts completely soulless plastic gens which he thinks are unimpeachably perfect state-of-the-art
anons confidence knows no bounds truly
>>
File: Wanimate_00018.mp4 (232 KB, 832x480)
232 KB
232 KB MP4
>>
>>106653057
lmao this is pretty good
>>
>>106653057
the merge was real...
>>
File: tellif sucks.png (101 KB, 1129x298)
101 KB
101 KB PNG
>>106652791
Nope. As always, turning your cfg a bit lower with a proper model is all it takes to completely fool these garbage "AI detection" websites. Most people in this thread can do a better job detecting AI with their eyes than any of these websites can.
>>
>>106653133
>Natural imperfections detected
it meant shadows? kek :^)
>>
>>106653139
it meant, the human
>>
>>106653036
>discuss this trade-off and why it exists
you did not explain why Model A HAS TO BE incoherent and bad at hands
>>
File: moy3_00079.webm (3.4 MB, 720x800)
3.4 MB
3.4 MB WEBM
>>
>>106653191
because it tries to show hands in the extremely broad and unlearnable variety of ways they appear. Hands are complicated 3D objects that articulate in many places and can be turned in many directions, making them, from the AI's 2-dimensional perspective, an unlearnable random mess

Any of these models can do good hands easily if you do something like "her hands are in her pockets" or "she's holding up her phone for a selfie", etc. As soon as you highly constrain it so hands can only appear a small number of ways, the problem goes away.

This is why "lying in the grass" was such a nasty prompt to test a base model with: it is far less tightly constrained than e.g. gens of people standing, photographed from the front. You could easily 'fix' this by training a model on a small handful of gens where people are always lying a particular way in the grass, but you are 'training' it to stop trying to represent the other 99% of valid possibilites. And in general this extreme narrowing is how most "Model B" type finetunes accomplish their high level of "perfection".

I have typed out posts 99% identical to this one far too many times over the past three years, and I'm getting tired of doing it. We should all know this shit by now.
>>
File: Wanimate_00019.mp4 (249 KB, 832x480)
249 KB
249 KB MP4
>>
>>106653219
sounds like a small model problem, not a general model problem
>>
>>106653235
What a fucking retarded response, why did I waste my time replying to you
>>
I don't understand what I'm doing wrong.
And how do I update neoforge? There's no instructions at all on its page.
>>
>>106653245
you'll grok it one day, buddy
>>
Wanimate points editor is the buggiest garbage piece of shit node I've ever had the displeasure of using
>>
>>106653246
You have to do a git pull on the repo.
If you don't know how, just put this into an update.bat file outside the folder, replace FOLDERNAME with your exact neoforge folder name.

@echo off

git -C "%~dp0FOLDERNAME" pull 2>NUL
if %ERRORLEVEL% == 0 goto :done

git -C "%~dp0webui" reset --hard
git -C "%~dp0webui" pull

:done
pause
>>
>>106653246
Also is that reForge you're using? Updating from that to neoforge won't work, totally different repos.
>>
>>106653305
>Wanimate points editor
I don't know why you're attributing that to Wanimate and not Kij/Meta
>>
>>106653317
I gazed into my crystal ball and he will appear again in 10 minutes asking why he's getting an error about X package not being found or a previously working package no longer working.
>>
>>106653013
who's the slut on the left
>>
File: Wanimate_00022.mp4 (250 KB, 832x480)
250 KB
250 KB MP4
>>106653367
idk, I just saw it in a thread about fat people.
>>
>>106653325
Yeah, neoforge. I did git pull and apparently I'm already at latest version.

Fuck it, I'm not into realistic stuff anyway.
>>
>>106653317
why would you automagically delete alterations made by anon?
>>
>>106653305
You don't fucking need it retard, stop using the wrapper nodes and shite workflows ffs.
>>106653158
>>106653252

>>106653329
THIS.

Its not the models fault its the absolute bloody retards that don't know what the fuck they are doing using every shit node possible for absolutely no reason at all. Just pushing shit out there and trying to ram it down other peoples throats. then people come here and shit this thread up everyday complaing about problems from what ever shitty workflow they be using.

Thank God the native nodes just bloody work and don't OOM. Embrace the spaghetti, it is the only way! Its more important to understand how everything works and not rely on others.
>>
>>106653416
I'm about to catbox the native workflow i'm using, its still a mess but it should be clear how it works.
>>
File: ComfyUI_18071.png (3.39 MB, 1280x1600)
3.39 MB
3.39 MB PNG
>>106649455
lol, the look of total disbelief on their faces at where they appeared is adorable!

>>106651234
>40-70 seconds
That's not so bad... with my "heavy" Flux wf that's about what I get per pull (73-75s) with a 4090.

>8 minutes with upscaling
How big are you making these?
>>
>>106653528
holy grain
>>
i'm still impressed by what pony can do in realism as long as you give it a Canny reference for pose and anatomy. doesn't feel so bad being a vramlet. landscapes and artistic stuff look great on chroma though from what i see here
>>
>>106653748
Chroma is pretty much trash
>>
>>106653748
i think bigasp merged with small amount of lustify is probably the best sdxl can offer in realistic porn realm
>>
>>106653779
I also do live tinkertrannying workflows with sdxl, how could you tell??
>>
>>106650954
catbox?
>>
File: wan_animate_00003.mp4 (1.82 MB, 480x832)
1.82 MB
1.82 MB MP4
>>106653305
>>106653416
>>106653489
With the audio, workflow is embedded in the video, just drag to comfyui or open it as workflow.

https://files.catbox.moe/gr9h0j.mp4

Its native comfy with out all that points editor and brapper nodes and just works.
>>
>>106654001
still has a slight issue with bad first frame between the chunks, could be shift value. hmm I will work on fixing it and improve quality now.
>>
File: image.png (252 KB, 460x460)
252 KB
252 KB PNG
*taps*
Now, you are thinking about him
>>
>>106654034
yup, george nader
might watch duplicators again
>>
File: 1167752492635496112-NEO.png (3.48 MB, 1080x1920)
3.48 MB
3.48 MB PNG
Mommy Marika
>>
AniStudio adding audio gen today anons
check his repo for the update
>>
buy an ad
>>
>>106654248
kek julien can keep lying about UI updates forever since literally nobody checks his repo or actually uses it anyway
>>
>>106650954
>>106653212
>>106654034
>>106654262
you seem to be confused or, at least, a bit lost,
friendless avatartroon general is this way >>106652275
>>
Julien really using reverse psychology moves?
>>
File: 1167752492635496138-NEO.png (3.79 MB, 1920x1080)
3.79 MB
3.79 MB PNG
>>106654148
>>
File: Preview.jpg (60 KB, 452x409)
60 KB
60 KB JPG
>>106650916

He trusted you
He believed in you
He wanted to impress /ldg/
He felt he belonged to /ldg/
With limited knowledge but unlimited heart, he built a UI with love rather than expertise
He dedicated time, energy, sleepless nights trying to give a gift to his beloved community
But that community ignored him, mocked him, cast him aside
The old saying became reality: "A child who is not embraced by his village will burn it down to feel its warmth"
Now he relentlessly haunts this community and others
Seeking recognition
Seeking love
But above all, seeking acceptance

Shame on you, /ldg/.
Your technical elitism and lack of human compassion created your own nemesis.
You have your villain now.
>>
>>106654148
>>106654304
>>106654148
>mommy
Oh, hi newfag!
Fetish posting is in this direction!
>>>/e/edg
>>>/h/hdg

Also there is an AI general in /vg/ ^^
>>
File: 1167752492635496146-NEO.jpg (570 KB, 1920x1080)
570 KB
570 KB JPG
>>106654304
eat shit nogen faggot
>>
>>106654387
Thanks for the unnecessary insult but this is unrelated to local diffusion talk or topic.
We are not your personal "art" gallery.
You have to go back, mewfag
>>
File: 1167752492635496159-NEO.jpg (588 KB, 1920x1080)
588 KB
588 KB JPG
>>106654304
>>106654397
Cry me a river
>>
>>106654387
How is this related to local diffusion? Can we discuss actual local diffusion concepts instead? Your post is just mindless coomer gens.
>>
>>106654415
Can we discuss actual local diffusion concepts instead? I don't care about your fetish
>>
>>106654418
>>106654430
shut up, bitch
>>
>>106654415
bruh we already got like 5 coomer AI boards listed right in the OP for lost newfags like you. if you're not here to actually discuss local models then stop being a tourist and gtfo
>>
File: 1167752492635496166-NEO.png (3.85 MB, 1080x1920)
3.85 MB
3.85 MB PNG
>>106654415
>>106654436
dont bother about nogens, dont feed the trolls
>>
>>106654436
My penis is larger than yours.
>>
>>106654449
no workflow, no details, nothing useful somehow you calls ME the troll
mfw when asking for actual local diffusion talks makes you a troll
this is why these threads are garbage,
>>
File: 1167752492635496157-NEO.png (3.76 MB, 1920x1080)
3.76 MB
3.76 MB PNG
>>106654449
>>
>>106654449
Are you the Sunday shizo? Get in the queue, plenty of other schizos waiting to shitpost here
>>
julien save us
>>
why does everyone ITT just go completely off topic? stop the BS >>106654468, are you here to actually discuss or just dump your mommy fetish shit on us?
>>
Hi! Has anyone tried Krita Diffusion? Is there a detailed tutorial? I saw it on YouTube, and I'm very interested, but I would prefer a written tutorial instead of watching a vid
>>
>>106654534
Ask this guy, he wants proper discussion happening here. I'm sure he's going to help you anon.
>>
>>106654468
your shit is messy and looks distinctly AI with that greasy layer on everything
>>
>>106654498
Anon, help him, please. >>106654534
>>
>>106653036
>we
>sd1.5 days
>never posts a gen
anon
>>
>>106654544
it's just a julien bootlicker here to shit up the thread out of spite
>>
Anon, >>106654468 talk to the guy. You using Krita with diffusion models for your fetish? Show him the ropes, explain how local diffusion works and how you generated your image. This is how we do local diffusion posting around here.
>>
>>106654553
oh we know.
>>
>tfw you hit a jackpot with mixing loras

EXHILIRATING
>>
>>106654544
Okay, >>106654534 sorry , I use SDXL and Forge WebUI. If you're new, I recommend Forge because it's simple and quick to use. Krita uses ComfyUI as a backend, which requires complicated node wiring setup. I can't help you with your specific issue, but you asked in the right place! ^^
>>
post 1 more mommy for good measure
>>
>>106654540
Man, it's not meant to make you angry. It's so that together with your gens, you share something and give depth to the general and elevate its quality. Nothing less, nothing more. If you want to call me a troll, that's okay and up to you. I just wanted to give a little more depth to your posting.
>>
>>106653219
I think we're closer than not to a breakthrough.
diffusion models are capable of high-level superficial copying governed by small amount of guided reasoning. they're currently less like skilled art forgers and more like counterfeiters who don't know anything about art.
but if we can teach the process of art through action in the way that we can model or simulate CAD as python in a 3D environment, recording the stroke by stroke inputs and interactions from software and combining it with the artists' discussion of what they're doing as captioning from, say, "let's draw" videos should teach it "reasoning" in the same way that that's working for what they call PhD-level reasoning in LLMs, then apply the method to a video model where the final frame is the completed work?

1) capture raw inputs from software and tools
2) record drawing area
3) sync with transcript of artist discussion
4) train on instruction-visual-reasoning
5) mode learns how to art?
>>
File: 1758453489363982.jpg (1.89 MB, 1792x2304)
1.89 MB
1.89 MB JPG
>>
>>106654611
Current AI art is just fancy copy paste. Real solution is teaching AI the artistic reasoning process itself.
>>
https://github.com/FlyMyAI/flymyai-lora-trainer

is this bullshit? because it looks and readss like vibe-coded total nonsense.
and on that matter, what are the bare min requirements to train a qwen edit lora?
>>
File: 1738791533569357.png (124 KB, 294x388)
124 KB
124 KB PNG
kek, ai really brings out the worst in you guys

>constant bickering
>shilling and spamming
>knowledge gatekeeping
>scoffing without actually doing any experimenting
>elitists and contrarians
>>
>>106654643
stick to onetrainer
>>
>>106654534
i use krita exclusively for ai. it's really good and also really straight forward.
>>
File: 00085-2635032247.png (2.83 MB, 1248x1824)
2.83 MB
2.83 MB PNG
>>106650933
80% of time the collages are absolute trash. stopped care about them and post my gens else where when it gets too spergy here. Don't take these threads serious at all or you will end up disappointed and burnt out.
>>
File: ComfyUI_00435_.png (1.59 MB, 1024x1024)
1.59 MB
1.59 MB PNG
Change my mind: Chroma1 HD Flash is the best model to use for inference of a LoRA trained on Chroma1 HD
>>
>>106654690
Yea I can tell
>>
I'm a 12GB vramlet but lora training for Chroma is supposedly possible. When I choose the training setting in OneTrainer the computer basically freezes and becoems unresponsive when loading the model. Is it supposed to be this way and it then gets better with offloading again or is something wrong here?
>>
>>106654690
the fuck is that
>>
>>106654715
>When I choose the training setting in OneTrainer
the 8GB training setting
>>
File: 1729089904279777.mp4 (91 KB, 640x346)
91 KB
91 KB MP4
tfw spent 20hrs doing ai fuckery
>>
>>106654650
i do, but for qwen edit i will prob have to use ai-toolkit.
>>
File: ComfyUI_00442_.png (1.59 MB, 1024x1024)
1.59 MB
1.59 MB PNG
>>106654697
That LoRA was trained on Chroma checkpoint 36. picrel is on HD

>>106654716
>the fuck is that
Promise not to freak out? It's a woman
>>
>>106654747


>HD
>>
>>106654747
>a woman
gross
>>
>>106654747
are you using the wrong sampler, the pic looks garbled
>>
>png/jpg
skip
>mp4/webm
watchie
>>
when will 4chan support webm, the superior format?
>>
File: 1745666949842933.webm (888 KB, 464x472)
888 KB
888 KB WEBM
>>106654804
?
i'm pretty sure 4chan supported webm before mp4
>>
>>106654811
meant webp
>>
File: ComfyUI_00444_.png (2.38 MB, 1024x1536)
2.38 MB
2.38 MB PNG
>>106654715
You're probably running out of RAM if the system locks up during model loading. It shifts blocks from system RAM to VRAM when you can't fit it all at once. Increase RAM (fast) or make virtual memory/swap at least the size of the model (slow).

>>106654787
>are you using the wrong sampler
simple/heun 8 steps 1cfg. beta gives worse likeness and cleaner output. I should test training only 1024px to see if it makes a difference
>>
>>106652111
Huh. Kind of cool.
>>
>>106654690
any non-HD + flash lora is infinitely better
>>
>>106654846
What RAM size do you need at least? I have 32GB, which I assumed should be enough, but I only did XL trainings so far.
>>
>>106654871
nta, but what scheduler/sampler/steps/cfg should i use with flash lora? been using euler beta 12 steps 1cfg, but it gens come out too ai-ish
>>
>>106654898
heun, res_2s, the 2s, 3s, 3m group. Also you can let it bake for more than 20 steps because chroma outputs varies depending on the number of steps
>>
File: 00010-2477897022.png (3.59 MB, 1536x1536)
3.59 MB
3.59 MB PNG
>>
>>106654846
(long back:2)
>>
Why haven't you, anon, switched to AniStudio?
>>
>>106655060
that's not my name
>>
>>106653036
First of all, no-gens, I never said my soulless plastic gens were perfect or even good. Those were tests of my wildcards that were undersized and use the turbo lora, and I still can't get the style right. I said they are better in certain ways that Chroma can't emulate. If you'd ever generated an image in any model you would understand that. Secondly, I'm not so certain that the soul/coherence divide is as necessary as you make it out to be. Have we ever actually got a model trained on high-resolution non-slop data at the training length and dataset size that the slop models are trained at?
>>
why would you reply to the chroma schizo
>>
>>106655261
I am duly chastised.
>>
WAN Animate would be glorious if we get local model.
>>
>>106654884
From what I've understood, you need to have more RAM than the physical size of the model.
>>
>>106655318
and how physically big is a model? like my ssd is tiny, only a couple inches long
>>
>>106655185
Shut the fuck up, no one fucking cares you mentally ill faggot. We are here to read and talk about local models and you are filling up the thread with your whinging about how anon was mean to you, just fuck off you boring brain damaged cunt.
>>
>>106655330
this, I'm just here to see if there's some news or cool video memes, the rest is noise
>>
File: 1731416934341507.png (157 KB, 1653x573)
157 KB
157 KB PNG
https://xcancel.com/LodestoneE621/status/1968976726714699812#m
that's cool, I wished he would also show the time with the whole model being on the VRAM too
>>
this thread is lacking in mikus
>>
Is there anyone ITT that isn't a schizo for this or that?
>>
>>106655324
18gb, which in 3.5" floppy disks is 12.5k
so, 43,750,000 inches long
>>
>Struggling to get a wan 2.2 anime gen to do what I want after 5-6 tries.
>Decide fuck it I'm just gonna try a regular gen with the official settings and without the light loras.
>Despite taking 4 times as long they come out better, especially less yapping (obviously).
>Notice that that annoying color shift/distortion at the start of the animation also doesn't happen so it loops better.

Bros... I think using the speedup loras is what's causing the shitty color shift.
>>
>>106655318
Yeah that's why I was wondering, there's enough free RAM to load the entire model plus text encoder and vae into RAM but my PC still shits itself when trying to load the model
>>
Maybe if you people weren't such all such insufferable cunts it would be easier to distinguish which of you were schizo thread personalities and which were regular old insufferable cunts. Jesus fucking christ, I post gens and actually talk about different models and lora training and workflows but if I respond once to the wrong person I'm "filling up the thread". Don't fill up the thread bitching about it, then.
>>
my armpits smell like heaven right now
>>
*sniff*
>>
*ah-CHOO*
>>
I can run wan animate on a rtx 6000 using the native github code and those results are great but is slow af.

Using the kj nodes and workflow the results are pretty bad. And I don't even understand if its doing the "animation" mode or the "sustitution" mode. In the github repo you have both options

Will comfy native support save us?
>>
>>106655580
>native github code
those are raw command lines without UI? how much slower?
>>
Is there any proper ui which is not written in a toy language like python?
>>
melonpuff or yesman_(cvbnyesman) for lora for chroma. I finally understand how to work this thing properly now
>>
>>106651197
I noticed that after I posted it and removed all of the loras i added, but my prompt is still being ignored 90% of the time. In the other 10% it actually works despite changing nothing, so I am pretty confused right now. Not sure what to do.
>>
>>106655580
https://github.com/kijai/ComfyUI-WanVideoWrapper/issues/1262#issuecomment-3314926799
sounds like kijai still has quite a bit of work besides even this so who knows how much longer
>>
>>106655663
How do you tag it?
>>
>>106655686
Joy caption to add both natural language and booru tags, it doesn't hurt anything and it works.
>>
all i want for wan 2.5XL-supreme-fp8-scaled is that it can do 30 second videos.

FIVE SECONDS isn't enough anymore, even if it means i have to buy a a6000pro
>>
>>106655608
swarmui or join the anistudio waiting room
>>
>>106655708
It seems to be the way to go
>>
>>106655758
It's pretty easy but sometimes it decides to add the booru tags in a new paragraph which is fucking annoying and requires review
>>
Did anyone try this? https://github.com/komikndr/raylight The guy is pretending you can now run big model in vram with 2 cards
>>
File: ComfyUI_temp_lmnzu_00001_.jpg (1.22 MB, 1728x2304)
1.22 MB
1.22 MB JPG
1girl, bad_hands
>>
>>106655378
Does this work for comfy? Has anyone here tried this yet?

>>106655418
Speed up loras for 2.2 are fucked kinda.

>>106655529
*imagine*

>>106655728

Can't see them doing that for this release, some anon before mentioned they would need to rework the entire architecture or something, fuck knows. But yes, 30 seconds would be perfect. A man can dream.
>>
>>106655930
>>106655728
Isn't it locked to 5 seconds because the dataset is made of 5 seconds clips?
>>
File: ComfyUI_temp_atppa_00001_.jpg (559 KB, 1024x1536)
559 KB
559 KB JPG
Low denoise on overbaked models always produces fun results
>>
>>106652659
really love these
>>
>>106655006
Pig trying to use a water pokémon attack?
>>
>>106655968
That almost looks cool, just need the brushstrokes and paint streaks to look much less randomized
>>
Hear me out.
Low res, low quality, but long video with context. An openpose controlnet is run at the same time to get the movement.
From there, upscale it in batches.

Wah lah.
>>
>>106656072
you really are just the worst. so glad i have you and all names filtered.
every now and then i peek at the posts and everytime i regret it.
>>
>>106655940
I believe so but best to read their official wan 2.1 paper https://arxiv.org/pdf/2503.20314 gonna give it a read my self when I get the chance
>>
>>106656099
Mad because you're a no-genner.



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.