[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/g/ - Technology

Name
Options
Comment
Verification
4chan Pass users can bypass this verification. [Learn More] [Login]
File
  • Please read the Rules and FAQ before posting.
  • You may highlight syntax and preserve whitespace by using [code] tags.

08/21/20New boards added: /vrpg/, /vmg/, /vst/ and /vm/
05/04/17New trial board added: /bant/ - International/Random
10/04/16New board for 4chan Pass users: /vip/ - Very Important Posts
[Hide] [Show All]


[Advertise on 4chan]


Discussion of Free and Open Source Text-to-Image/Video Models and UI

Prev: >>106494102

https://rentry.org/ldg-lazy-getting-started-guide

>UI
ComfyUI: https://github.com/comfyanonymous/ComfyUI
SwarmUI: https://github.com/mcmonkeyprojects/SwarmUI
re/Forge/Classic/Neo: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassicneo
SD.Next: https://github.com/vladmandic/sdnext
Wan2GP: https://github.com/deepbeepmeep/Wan2GP

>Checkpoints, LoRAs, Upscalers, & Workflows
https://civitai.com
https://civitaiarchive.com/
https://openmodeldb.info
https://openart.ai/workflows

>Tuning
https://github.com/spacepxl/demystifying-sd-finetuning
https://github.com/Nerogar/OneTrainer
https://github.com/kohya-ss/sd-scripts/tree/sd3
https://github.com/derrian-distro/LoRA_Easy_Training_Scripts
https://github.com/tdrussell/diffusion-pipe

>WanX
https://rentry.org/wan22ldgguide
https://github.com/Wan-Video
https://alidocs.dingtalk.com/i/nodes/EpGBa2Lm8aZxe5myC99MelA2WgN7R35y

>Chroma
https://huggingface.co/lodestones/Chroma1-Base
Training: https://rentry.org/mvu52t46

>Neta Lumina
https://huggingface.co/neta-art/Neta-Lumina
https://civitai.com/models/1790792?modelVersionId=2122326
https://neta-lumina-style.tz03.xyz/

>Illustrious
1girl and Beyond: https://rentry.org/comfyui_guide_1girl
Tag Explorer: https://tagexplorer.github.io/

>Misc
Local Model Meta: https://rentry.org/localmodelsmeta
Share Metadata: https://catbox.moe | https://litterbox.catbox.moe/
GPU Benchmarks: https://chimolog.co/bto-gpu-stable-diffusion-specs/
Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-one
Txt2Img Plugin: https://github.com/Acly/krita-ai-diffusion
Archive: https://rentry.org/sdg-link
Bakery: https://rentry.org/ldgcollage

>Neighbours
>>>/aco/csdg
>>>/b/degen
>>>/b/realistic+parody
>>>/gif/vdg
>>>/d/ddg
>>>/e/edg
>>>/h/hdg
>>>/trash/slop
>>>/vt/vtai
>>>/u/udg

>Local Text
>>>/g/lmg

>Maintain Thread Quality
https://rentry.org/debo
>>
All round town anons ask me if I know of the "blessed thread" and do you know what I tell them
>>
Comfy needs WebUI style workflow or node improvements for users. Swarm and SDNext aren't solutions. People want Gradio.
>>
File: Qwan_00006_.jpg (546 KB, 1984x2976)
546 KB
546 KB JPG
>>106497283
Please tell me
>>
>>106497293
people dont know what they want until you give it to them desu
>>
>>106497293
ComfyTip:
Group reusable nodes using Nest Node Builder.
Load/ungroup them when needed instead of adding individual nodes.
Save time with saved nested nodes (ControlNet, Prompt, Upscale, etc.).
Export as JSON to transfer/share.
ComfyUI is faster and more convenient than Forge once a workflow is finalized, trust me.
>>
Blessed thread of frenship
>>
>>106497316
Comfy It's trash, a big pile of ugly frustrating trash, and why the hell is it called comfy ui? it anything but comfy its awful. I'm will never using it again, fuck !
>>
File: ComfyUI_00167_.png (1.32 MB, 1280x720)
1.32 MB
1.32 MB PNG
Graphical UIs suck. Why can't we write code directly?
A basic workflow is less than 10 lines of code:
text_encoder = load_text_encoder("te.safetensor")
model = load_model("model.safetensor")
vae = load_vae("vae.safetensor")
model.load_lora("lora.safetensor")

latent = new_latent(width=1024, height=1024)
latent = latent.ksampler(steps=20, cfg=4, seed=12345)
image = vae.decode(latent)
image.save("out.png")


Ability to write functions make abstraction possible:
latent = new_latent(width=1024, height=1024)
latent = my_custom_first_pass_sample(latent, steps=20, cfg=4)
latent = my_custom_hires_fix(latent, steps=30, cfg=4, denoise=0.7)
latent = my_custom_adetailer(latent, ...)

No more spaghetti. No more janky node to workaround node UI deficiencies like reroute, switch, number to string to number, numeric operations on numbers, etc.

Not to mention with loop statements it's easy to make custom XYZ plots, which node UI just can't do.

With a real programming language, it has the power of both forge-style UI and node-based UI do while also better than both in flexibility and cleanness.

Pair it with a Jupyter notebook style UI, you can nicely iterate and inspect results anywhere in the middle of generation and nicely iterate until you get a good result.
>>
>>106497325
>Comfy It's trash
Can I push a big orange button to convert a "comfy" workflow to python or some real scripting language and hack that?
>>
>>106497316
What is killing me on my Comfy are the updates that and other shenanigans that keeps breaking constantly , like, the project is great but it's so fragile and takes a misplaced period to almost make you pc explode.
>>
>>106497334
Easy peasy, lemon squeezey ! What imports do I need?
>>
>>106497334
this is the best
>>
>>106497345
Seriously, they should have guideline like Forge and enforce them before allowing them to be integrated into WebUI. Seriously, that's what is keeping me from using it as my main stable diffusion ui.
>>
>>106497334
the only thing i want to write is a prompt
>>
File: ComfyUI_00200_.mp4 (861 KB, 640x832)
861 KB
861 KB MP4
>>
>>106497298
that gap man
THAT FUCKING GAP
it does things to me
>>
>>106497334
do it faggot
>>
After all this time trying comfy, I still absolutley hate it's fking guts. I tried, I learned, I made mistakes, I studied, I failed, I learned again. Debugging and debugging and debugging... I'm so sick of it. I hated it from my first git clone up until now, with my last right click delete of the repository. I have been using A1111, reForge, and Forge as my daily before Comfy. I tried Invoke, foocus, and SwarmUI. Comfy is at the bottom. I don't just not enjoy it, it is a huge nightmare everytime I start it. I wanted something simple, plug n play, push power button and grab a controller, type of ui. Comfy is not only 'not it' for me, it is the epitome of what I hate in life.

Why do I hate it so much? Here's some back ground if you care. When I studied to do IT 14 years ago I had a choice to choose my specialty. I had to learn everything from networking, desktop, database, server, etc... Guess which specialties I ACTIVELY avoided? Database and coding/dev. The professors would suggest once every month to do it. I refused with deep annoyance at them. I dropped out of Visual Basic class because I couldn't stand it. I purposely cut my Linux courses because I hated command line, I still do. I want things in life to be as easy and simple as possible.
>>
>>106497385
okay .. well you're kinda retarded so you're not helping the rest of us with our case
>>
>>106497385
Comfy is like browsing the internet in a browser with html format only. Imagine a wall of code, a functional wall of code. It's not really the spaghetti that bothers me, it's the jumbled bunch of blocks I am supposed to make work. The constant scrolling in and out is annoying but the breaking of comfy from all the nodes (missing nodes) was what killed it for me. Everyone has a custom workflow. I'm tired of reading dependencies over and over and over again.

I swear to Odin I tried my best. I couldn't do it. I just want to point and click and boom image. I don't care for hanyoon, huwanwei, whatever it's called. I don't care for video and all these other tools, I really don't. I just want an outstanding checkpoint and an amazing inpainter.

Am I stupid? yeah sure call me that if you want. I don't care. I open forge. I make image. I improve image. I leave. That's how involved I am in the AI space. DESU, 90% of the new things, cool things, new posts in this sub is irrelevant to me.

You can't pay me enough to use comfy. If it works for you great, more power to you and I'm glad it's working out for you. Comfy was made for people like you. GUI was made for people who couldn't be bothered with microscoptic details. I applaud you for using Comfy. It's not a bad tool, just absolutely not for people like me. It's the only and the most power ui out there. It's a shame that I couldn't vibe with it.
>>
>>106497392
I use comfy for video, forge for the images. I do a fair amount of nodes in unreal and blender at work, so it's not too scary, but I won't add another node system I need to master unless I have to!
>>
>chatgpt write schizobabble from the perspective of someone who hates comfyui
>>
>>106497385
>. I tried, I learned, I made mistakes, I studied, I failed, I learned again.
Engaging but not empowering. sigh
>>
>>106497392
I completely get where you are coming from. I have tried ComfyUI too, and while it is insanely powerful, it always feels like I am stuck fixing broken LEGO instructions made by 10 different people. Every time I load someone’s workflow, half the nodes are either deprecated, custom, or renamed, and I end up spending more time debugging or hunting missing certain type of files than generating.
>>
File: ComfyUI_00160_.mp4 (803 KB, 480x832)
803 KB
803 KB MP4
>>
>>106497428
You ruined that skindentation with that ugly face.
>>
>>106497428
dat phase shift
>>
>>106497385
That is why I stick with Stable Diffusion Forge for images. It just works. PNG Info gives me everything I need at a glance, even if the image was not made in Forge. Prompt, model, LoRA,
>>
File: Qwan_00010_.jpg (578 KB, 1984x2976)
578 KB
578 KB JPG
>>106497334
Can't you just use the tensor library in python and kinda just do that?
Implementing/porting all functionality of custom nodes might be a pain, though.
That being said, I don't mind ComfyUI. I kinda like the previews/compare nodes. If you'd want that in pure code it'll require a UI again, anyway.

>>106497375
This is a blue board, this is all friendly fun.
>>
>>106497392
ComfyUI has its fans and massive support and I respect that, but it is clearly not built for everyone. If you like clean workflows and hate excessive tinkering, Forge hits the sweet spot. So you are not alone!
>>
blue board, blue balls
>>
did I tell you before? I don't like comfyui
I know there are other webuis, but I don't care about them, I don't even actually gen, I just want to say I don't like comfyui
>>
>>106497436
that was actually on purpose and it took many attempts to pull it off
>>
File: 00038-1125287675.png (1.8 MB, 1216x832)
1.8 MB
1.8 MB PNG
>>
>>106497316
You'd be surprised at the number of nodes people include in their workflows that are completely unnecessary. I've even seen requests to install a custom node package to set an integer or just to hide 2 spaghettis.
>>
It's all well and good to want to write high-level code and click run but it does have some problems. Main one being that importing PyTorch and then loading your models takes forfuckingever so you're going to need to have some sort of server that holds onto that stuff for you and somehow knows what models to cache when, otherwise every time you fiddle with your gen you have to wait the best part of a minute just for it to get going. Maybe you could do it with some combination of Jupyter notebooks idk, but I've never used Jupyter and you'd also need to implement some way to easily embed those notebooks into your outputs and then convince everybody to adopt that standard.
>>
with all this fresh pasta i dont need the noodles XD
>>
To AntiComfySchizo:
You don't hate comfy, you hate python. And rightly so, it's a fully trash peer dependency environment.
>>
>>106497486
It's very annoying when workflow use random exotic nodes while perfectly fine versions exist in core or even well known packs.
I just convert them usually.
>>
File: 1738286329311319.jpg (1.64 MB, 2016x1152)
1.64 MB
1.64 MB JPG
>>
I HATE PYTHON
>>
>>106497451
>This is a blue board, this is all friendly fun.
it makes me horny in a friendly way
>>
I LOVE SNAKES
>>
>>106497502
FUCK OF NETA LUMINA SCAMMER
SHARE WORKFLOW OR GTFO
>>
File: 1739989832704588.jpg (646 KB, 1416x2120)
646 KB
646 KB JPG
>>
File: 00096-1738961045.png (2.48 MB, 1248x1824)
2.48 MB
2.48 MB PNG
ah, friday
>>
File: 1745784200408034.png (1.31 MB, 1104x1328)
1.31 MB
1.31 MB PNG
>>106497518
>>
give me prompt suggestions pls
>>
>>106497503
>I HATE PYTHON
I like python. I hate python package management and dependency hell.
>>
>>106497534
A cute 1girl sleeping under the sun.
>>
>>106497534
"An old man punching a horse"
>>
>>106495675
anyone? though reading past the last thread I’m gonna assume that even if it is possible, it’s not as easy as I think it is.
i’m basically modifying an image of a hand pulling something out of a box, and I want to use the reference image as the item being pulled out.
Still using an unmodified workflow from the guide here.
>>
>>106497534
1gil, looking at viewer, waving
>>
>>106497534
an indian uncle harassing a hot girl for sexi sex
>>
>>106497545
>>106497551
>>106497557
>>106497566
i could inpaint all of these into a single awesome gen but i use comfy so you know im not fucking around with that shit kek
>>
>>106497556
use vace
>>
I've started to notice some similarity on the wan 2.2 faces, is there anything I can do about it without prompting face traits or ethnicity?
>>
>>106497600
use loras, or use i2v instead
>>
File: Qwan_00011_.jpg (591 KB, 1984x2976)
591 KB
591 KB JPG
>>106497504
I'm not going to give you a friendly handshake, that's for sure.
>>
>>106497672
that's ok anon
>>
>>106497525
You asuka sucks you know?
>>
tfw ani was right
>>
>>106497525
pretty good Asuka
>>
k enough of the discord invasion
>>
File: 1755517650261282.jpg (501 KB, 1416x2120)
501 KB
501 KB JPG
>>
File: 1745460281544883.png (29 KB, 735x284)
29 KB
29 KB PNG
Am i really the only cumfartnigger who sees this in vram usage during every vram allocation until it stabilizes and works normally? Started happening in the last couple of days
>>
File: 1746400062064082.jpg (539 KB, 1416x2120)
539 KB
539 KB JPG
>>
>>106497882
Maybe comfyui should be dragged on the streets and ACKed
>>
man we really need better interpolation software
>>
>>106497892
pretty damn good
>>
>>106497901
film vfi is good enough for 16 to 32, and we just need some cracked topaz node connection in comfyui for slightly higher quality to 32 and for anything more
>>
File: ComfyUI_00203_.mp4 (775 KB, 832x480)
775 KB
775 KB MP4
>>106497502
>>
File: ComfyUI_00134_.mp4 (534 KB, 480x832)
534 KB
534 KB MP4
>>106497882
>mfw ram usage expands
>>
i want these goddamn snakes out of my goddamn computer
>>
File: ComfyUI_00018__.png (1.99 MB, 1328x1328)
1.99 MB
1.99 MB PNG
>>
File: 1746717988902374.png (3.57 MB, 1416x2120)
3.57 MB
3.57 MB PNG
>>
File: ComfyUI_00214_.mp4 (598 KB, 480x832)
598 KB
598 KB MP4
>>
>>106497325
>I'm will
you will be sorely missed, frenchy.
>>
File: ComfyUI_00215_.mp4 (576 KB, 480x832)
576 KB
576 KB MP4
>>
>>106498141
why is that man drooling like a retard
>>
File: ComfyUI_00462_.png (1.72 MB, 1328x1328)
1.72 MB
1.72 MB PNG
>>
File: ComfyUI_00216_.mp4 (711 KB, 480x832)
711 KB
711 KB MP4
>>
File: AnimateDiff_00301.mp4 (2.59 MB, 576x1024)
2.59 MB
2.59 MB MP4
>>
File: 1725303580349732.jpg (207 KB, 801x853)
207 KB
207 KB JPG
>>106498206
kek'd
>>
>>106498231
hors
>>
File: ComfyUI_00219_.mp4 (1.56 MB, 720x720)
1.56 MB
1.56 MB MP4
>>
Don't get all the shit Comfy gets, when the entire community has had to rely on Civit for far too long. We're talking actual tangible, long term damage from such a garbage platform thriving.
>>
File: ComfyUI_00224_.mp4 (918 KB, 720x720)
918 KB
918 KB MP4
>>
>>106498295
Yet people here rarely talk about real problems in need of an urgent solution.
>>
>>106497275

whoever genned this, please PLEASE catbox
>>
>>106497467
Not if you visit /adt/
I mean, it's crazy what they get away with posting there.
>>
>>106498325
PLEASE get better taste
>>
File: ComfyUI_00231_.mp4 (418 KB, 720x720)
418 KB
418 KB MP4
>>
File: AnimateDiff_00302.mp4 (2.45 MB, 720x1072)
2.45 MB
2.45 MB MP4
>>106498172
c'mere anon
>>
>>106498378
bisgustin
>>
File: 00226-2997302013.png (302 KB, 448x576)
302 KB
302 KB PNG
>>
>>106498439
Typical DiT manlet (because manlet syndrome isn't unique to kontext or QIE, it is some kind of transformer-wide intrinsic property, noticable in Dalles, too. Young meatbag artists also share it with transformers when they run out of bottom margins on their real world paper but still need to draw feet.)
>>
File: ComfyUI_00239_.mp4 (491 KB, 720x720)
491 KB
491 KB MP4
>>
>>106497428
2000 years post wall
>>
File: ComfyUI_00245_.mp4 (1.75 MB, 720x1280)
1.75 MB
1.75 MB MP4
>>
>>106497672
catbox?
>>
Can any of you anons make me a believable black-and-white photo of Sigmund Freud mixed with George Floyd? I want to print it and
>>
>>106498708
...and put it on my best friend's wall.
>>
Is GenJam never coming back?
>>
>>106498746
it will return
>>
>>106498746
Just say it with, move your lips when reading it so you can feel how good it feels saying.

"GOON JAM"
>>
File: ComfyUI_00251_.mp4 (1.75 MB, 720x1280)
1.75 MB
1.75 MB MP4
>>
wan 2.2 vace WHEN
>>
File: 00678-678471360.png (2.14 MB, 1248x1824)
2.14 MB
2.14 MB PNG
>>106498746
>didn't participate
>somehow miss it
feels weird
>>
>>106498878
I'm curious of the prompt
>>
>>106498923
A scenic view of an old bicycle in a field. The camera pushes out to reveal scraps of torn and ripped clothing strewn about and broken beer bottles on the ground. There are smears from bloody handprints on the tree. There is a puddle of of blood in the grass and on the ground near the torn clothes. It looks like a murder scene.
>>
>>106497872
Model?
>>
How much RAM you guys have? I plan to upgrade from 64 to 128GB, but not sure if it's worth the upgrade
>>
File: AD_00003.mp4 (610 KB, 480x720)
610 KB
610 KB MP4
>>106497872
kek, fucked up the heels
>>
File: 1734114167111416.png (3.7 MB, 1416x2120)
3.7 MB
3.7 MB PNG
>>106498936
noob and a lora trained on 80 imgs curated from https://x.com/schauermannx2
i think it can be much better tho
>>106498961
perspective wise i think the video makes more sense heh
>>
>>106498958
96 GB
>>
>>106498958

Downloaded RAM from 64 to 128 and it reduced my Wan2.2 gens by 80 seconds. Worth it.
>>
>>106499032
where can u download ram from???
>>
File: ComfyUI_00255_.mp4 (1.59 MB, 720x1280)
1.59 MB
1.59 MB MP4
>>
Are there wan2.2 vace or alternatives yet?
>>
File: WanVideo2_2_I2V_00331.webm (615 KB, 1248x720)
615 KB
615 KB WEBM
>>
File: ComfyUI_00263_.mp4 (2.43 MB, 720x1280)
2.43 MB
2.43 MB MP4
>>
>>106498309
Badass
>>
File: ComfyUI_00265_.mp4 (3.23 MB, 720x1280)
3.23 MB
3.23 MB MP4
>>
File: 00181-3215316192.jpg (140 KB, 832x1248)
140 KB
140 KB JPG
>>106499026
no screenshot? comeon hommie, flex that dick
>>
>>106496741
>Now I kind of want to make a Suno.ai song with this phrase as the chorus:
>>he made his own bed
>>by forcing the model to run faster starting at v30, >it all went downhill from there
got ya senpai
https://vocaroo.com/1loPzeLJD8qK
>>
File: AD_00012.mp4 (460 KB, 480x720)
460 KB
460 KB MP4
>>106498961
>>106498996
true, i think i got it right this time though
>>
File: freem.png (41 KB, 729x358)
41 KB
41 KB PNG
>>106499215
here you go
>>
>>106499218
kek
>>
>>106499237
dawg I got a 5090, thought you had that blackwell pro, tryin to jerk off over here
>>
>>106499247
no 96 gig ram not vram
>>
File: ComfyUI_00267_.mp4 (864 KB, 1280x720)
864 KB
864 KB MP4
>>
odd how onetrainer wont let you define sampler/scheduler/etc for training samples. i wonder how training with comfy compares in general
>>
>>106497264
Middle right is amazing
>>
>>106499300
Ikr, I didn't find any flaw on that one, I kinda expect Wan 3.0 to have this kind of quality consistently
>>
>>106498206
LMAOOO, this is a gem
>>
File: 1747620345646949.png (286 KB, 408x632)
286 KB
286 KB PNG
>>106499113
now I get why Microsoft wants to shut this model down kek
>>
>>106499113
https://github.com/paperwave/VibeVoice
doesn't look taken down to me?
>>
File: 1742614760393807.png (79 KB, 1859x641)
79 KB
79 KB PNG
>>106499405
the corrected 7b model is "on the way", it'll be a lobotomized version of the one we already have lol
>>
File: ComfyUI_00143_.mp4 (472 KB, 592x816)
472 KB
472 KB MP4
I just realized WAN's generation are tuned for 16 fps. Increasing fps makes the motion fast most of the time.
>>
>>106499405
because that obviously is not the original microsoft repo
>>
>>106499455
yep, I hope their next version will be at 24 fps, that's the threshold where it doesn't look chopped as fuck
>>
https://www.theverge.com/anthropic/773087/anthropic-to-pay-1-5-billion-to-authors-in-landmark-ai-settlement
Anthropic to pay $1.5 billion to authors in landmark AI settlement
holy fuck dude, this is bad, like really really bad
>>
>>106499455
>I just realized WAN's generation are tuned for 16 fps.
I have no idea where people got the idea that it isn't. It's been a really hard myth to dispell.
>>
>>106499505
>I have no idea where people got the idea that it isn't.
that's because the 5b version is actually working at 24fps, so I also thought the 14b version would be too
>>
>>106499502
That's like 80% of what I've spent gooning to opus 4.1
>>
>>106499502
>Let's fuck with the development of this groundbreaking technology because some fat bitch wants money for her chad thundercock schlock novel
Grim. I hate the antichrist.
>>
>>106499502
It's actually good in the sense that they're only paying for illegally downloading the books, not for using them in training. 1.5b is nothing to them but (like everyone's been saying) this case further widens the gap between the big guys and the little guys (which is what the authors proclaim they are against kek).
>>
>>106499455
The frame interpolators are decent for solving this issue. I use GIMM-VFI, which you can search for with comfyui custom nodes manager and install it.

I believe these are quite a bit better than the old interpolated frames you'd get with TVs. Though they're obviously not perfect, as they're only spending like 60 seconds to generate the interpolated frames for your entire video.

The way it works is your frame count gets increased (like if every 2nd frame is interpolated then it goes from 81 --> 160 or whatever), so now your video is in slow motion, but then you fix this by increasing your FPS to make it faster. And then it'll look right.
>>
File: radiance_test__00012_.png (1.39 MB, 1024x1024)
1.39 MB
1.39 MB PNG
Testing Chroma Radiance
I think the output is pretty interesting. It has sort of a weird mottling pattern that seems unique. It's probably something that should go away with further training progress, but I actually like it.
>>
>>106499502
>>106499556
The judge explicitly ruled that training on protected works is fair use which is the biggest win for AI bros
>>
https://vocaroo.com/1eCSSHLSRPJ0

You know, for the first time using vibe voice it's actually pretty good. I was expecting another so-so model but it's actually pretty good.
>>
>>106499585
yeah we won but it'll be reported as a loss.
as usual, we pretty much have to just wait about 12 fucking years until the whiners die out and the AI zoomers take over, then they'll start making a bunch of youtube videos finally correcting the record for all the misinformation being spread.
>>
File: rad_test__00023_.png (1.68 MB, 1024x1024)
1.68 MB
1.68 MB PNG
>>106499580
Poor anatomy and other Chroma problems seem about the same as the regular models
>>
>>106499604
Problem is radiance is slow. And raising the resolution increases memory requirements like crazy.
>>
>>106499613
Whoa. It's almost like the Vae exists for a reason.
>>
>havent updated in months because comfy runs fine
>start it up, suddenly it crashes when loading a specific controlnet
>switch to different controlnet and it works fine
Wha
>>
>>106499592
Yeah the 7B model is crazy good.
>>
>>106499585
>The judge explicitly ruled that training on protected works is fair use which is the biggest win for AI bros
then why do they have to pay to use it? that's the fucking problem
>>
https://voca.ro/13lAHtQGa9KR
>>
File: radiance_test__00029_.png (1.13 MB, 1024x1024)
1.13 MB
1.13 MB PNG
>>106499604
Unfortunately it looks like the biggest thing hoped to improve with Radiance isn't any better. Small, high frequency details are still get melted and deformed. Not using VAE isn't helping. I think it was already doomed when they decided to train with 512x.

>>106499613
I am overly GPU-rich so I didn't really notice, but I thought skipping the VAE could have lead to speed improvements eventually?
>>
>>106499624
>Whoa. It's almost like the Vae exists for a reason.
for edit models, vae is a disaster, you want the model to only modify certain parts of the image but with a vae you have a compression loss on all the pixels, vae-less is the way to go for edit models
>>
>>106499596
I cannot think of a bigger win for training other than forcing those included in datasets to pay the trainers. It's that big.
>yeah we won but it'll be reported as a loss.
True but in time, it won't.
>>106499641
>then why do they have to pay to use it?
Not to use it, anon. To obtain it. I agree it's still gay and retarded but until the entire copyright apparatus is taken down that's the way it'll be.
>>
>>106499665
>Not to use it, anon. To obtain it. I agree it's still gay and retarded but
I don't think you realize how fucked up this is, every uncoming company will need billions of dollars to get the data needed to train their models, it'll kill everything, only giant companies will afford to do that, the US eldorado is over
>>
File: 1749452196972887.png (2.13 MB, 1664x1248)
2.13 MB
2.13 MB PNG
>>106497892
same seed and prompt, different epochs of a new version
need to figure out why noob lineart cnet crashes comfy tho
>>
>>106499678
>I don't think you realize how fucked up this is,
See >>106499556
>this case further widens the gap between the big guys and the little guys (which is what the authors proclaim they are against kek).
It sucks but I'm a half glass full kind of person. It does prove that the artists and authors suing are either 1. hypocrites or 2. being fooled by large copyright holders but we've all suspected as such already.
>>
https://vocaroo.com/1EeqoY7Nm8wo
>>
>>106499556
>It's actually good in the sense that they're only paying for illegally downloading the books, not for using them in training.
3000 dollars for a single book? really? they're not paying to buy a book, they're paying the extra to use it for training, how is that "fair use"? the judge is fucking RETARDED
>>
File: comfyui_3432.jpg (235 KB, 905x992)
235 KB
235 KB JPG
>>
>>106499723
What I've been seeing going around is the first frame being the image and the second frame being the box display, so like the character walks up on to the desk.
>>
>>106499722
Put yourself in Anthropic's shoes. Spending 1.5b to save 183b is a steal.
>the judge is fucking RETARDED
Pretty sure that number was reached between the two parties. Actually I think it was Anthropic that came out and said "that's fine we'll pay it.
>>
>>106499740
for anthropic it's fine, but this sets a precent, now every company that wants to replicate their success know they will have to first need billions of dollars to make their first model, this will be impossible for almost everyone, the US is dead, China has won
>>
>>106499648
>VAE could have lead to speed improvements eventually
Theory is supposed to learn better so technically but in practice I doubt it. We only have one pixel model in the wild and it's really meh cause it was undercooked, which I am assuming is from the high training resource this technique requires.
>>
>>106499740
$183 billion, but they are still operating at a loss, they are still not profitable.
>>
>>106499714
It kinda starting speeding up like crazy near the end but speech was pretty natural until then. Don't know who it is though.
>>
>>106499648
>Not using VAE isn't helping.
it is >>106499655
>>
I bet if you're able to obtain your dataset via the clear web and not torrents or other illegal means it'd be fine.
>>
>>106499756
It's Dagoth Ur from morrowind. It sounds pretty much exactly like him but then again I think he's one of the easiest voices to replicate.



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.