[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip / qa] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/g/ - Technology


Thread archived.
You cannot reply anymore.


[Advertise on 4chan]


File: the longest dick general.jpg (2.69 MB, 3264x1562)
2.69 MB
2.69 MB JPG
Discussion of free and open source text-to-image models

Previous /ldg/ bred : >>102956911

Phrenologic Exploration Edition

>Beginner UI
Fooocus: https://github.com/lllyasviel/fooocus
EasyDiffusion: https://easydiffusion.github.io
Metastable: https://metastable.studio

>Advanced UI
Forge: https://github.com/lllyasviel/stable-diffusion-webui-forge
reForge: https://github.com/Panchovix/stable-diffusion-webui-reForge
Automatic1111: https://github.com/automatic1111/stable-diffusion-webui
ComfyUI: https://github.com/comfyanonymous/ComfyUI
InvokeAI: https://github.com/invoke-ai/InvokeAI
SD.Next: https://github.com/vladmandic/automatic
SwarmUI: https://github.com/mcmonkeyprojects/SwarmUI

>Use a VAE if your images look washed out
https://rentry.org/sdvae

>Model Ranking
https://imgsys.org/rankings

>Models, LoRAs & training
https://aitracker.art
https://huggingface.co
https://civitai.com
https://tensor.art/models
https://liblib.art
https://github.com/Nerogar/OneTrainer
https://github.com/kohya-ss/sd-scripts/tree/sd3

>SD3.5
https://huggingface.co/stabilityai/stable-diffusion-3.5-large
https://replicate.com/stability-ai/stable-diffusion-3.5-large

>Sana
https://github.com/NVlabs/Sana
https://sana-gen.mit.edu

>Flux
https://huggingface.co/spaces/black-forest-labs/FLUX.1-schnell
https://comfyanonymous.github.io/ComfyUI_examples/flux
DeDistilled Quants: https://huggingface.co/TheYuriLover/flux-dev-de-distill-GGUF/tree/main

>Index of guides and other tools
https://rentry.org/sdg-link
https://rentry.org/rentrysd

>Try online without registration
txt2img: https://www.mage.space
img2img: https://huggingface.co/spaces/huggingface/diffuse-the-rest

>Maintain thread quality
https://rentry.org/debo

>Related boards
>>>/aco/sdg
>>>/aco/aivg
>>>/b/degen
>>>/c/kdg
>>>/d/ddg
>>>/e/edg
>>>/h/hdg
>>>/tg/slop
>>>/trash/sdg
>>>/u/udg
>>>/vt/vtai
>>
i fucking hate flux fuck fuck fuck thread those fucking bfl snakes
>>
File: image (9).png (435 KB, 1024x1024)
435 KB
435 KB PNG
>>
File: 00007-2059676223.png (1.28 MB, 832x1216)
1.28 MB
1.28 MB PNG
>>
yep, it's sloppin' time
>>
File: image (10).png (378 KB, 1024x1024)
378 KB
378 KB PNG
>>
>>102964636
don't be that jealous anon, it's not good for your mental health
>>
File: image (5).png (407 KB, 1024x1024)
407 KB
407 KB PNG
>>
File: 1727216024.png (825 KB, 1024x1024)
825 KB
825 KB PNG
>>
>>102964712
jealous of licking bfl's dirty butthole, hoping, no, PRAYING, that one of the two fossilized shits they shat out will be usable if we clean it with enough money?
>>
File: 00037-1868180960_cleanup.png (1.21 MB, 832x1216)
1.21 MB
1.21 MB PNG
>>
>>102964744
>one of the two fossilized shits they shat out will be usable if we clean it with enough money?
isn't that what happened with SDXL though?
>>
>>102964753
yes the 2.6b model, wonder how much it will cost to unshit a 12b
>>
Why is ldg quality so low now? What changed?
>>
File: 1727215035.png (1.19 MB, 1024x1024)
1.19 MB
1.19 MB PNG
>>
0/10
>>
flux/10
>>
File: 1726986290.png (1.54 MB, 1024x1024)
1.54 MB
1.54 MB PNG
>>
>>102964760
>unshit a 12b
there's not much to "unshit" though, flux is already great as a base model, it just needs more concepts, the work is less significant than SDXL where we had to unfuck the anatomy
>>
File: image (2).png (394 KB, 1024x1024)
394 KB
394 KB PNG
>>102964786
>>
>>102964789
>flux is already great as a base model
flux dev is, problem is the flux dev licence is dogshit, any work you do on flux will be legally owned by bfl. if anyone's going to pour money on flux it will be schnell
>>
File: 1729643888_0001.png (1.16 MB, 1024x1024)
1.16 MB
1.16 MB PNG
>>
>>102964789
Any new concept requires thousands of steps and you have to do it as part of a full train with hundreds of thousands of images, it's not cheap or easy
>>
>>102964807
>flux dev is, problem is the flux dev licence is dogshit,
I don't want to sound like a doomer, but we'll never get a great base model with the goated Apache 2.0 licence, even if it's the case for Mochi, it's only the 480p version, the HD version will probably have a shit ass licence too :(
>>
File: 1729309735.png (1.63 MB, 1024x1024)
1.63 MB
1.63 MB PNG
>>102964804
>>
>>102964818
my point is that it's more trivial to add Gwar Gura onto a model as a concept rather than trying to improve a fucked up anatomy, yeah you need money but it's not like everyone that train model do it for the money, some people just want to improve a model and share their result with everyone
>>
>>102964820
yeah.. but the sd 3.5 license is pretty decent, research use and non commercial is free and you only need to pay for commercial if you make more than $1 million in revenue annually, so fingers crossed medium turns out to be good
>>
>>102964849
Loras only get you so far
>>
>>102964856
>yeah.. but the sd 3.5 license is pretty decent
not decent enough for the pony fag, if he turned down SD3.5 it means that this fucker made more than a million out of ponyXL, which is absolutely insane
>>
>>102964862
I never said anything about a lora, could be a finetune that simply adds a shit ton of characters from danbooru or some shit, you know it's gonna work fine on Flux because adding characters is easy enough if you have enough pictures of them, fixing the anatomy on the other hand is another story, and thankfully we don't have to deal with the hard concepts on flux, flux is good on the hard concepts, but not so good at trivial ones
>>
>>102964868
ponyxl was a shitly trained model anways, i'd rather someone like the illustrious try a new finetune instead. and that's why lower param count models are better, because it makes finetuning more accessible to smaller groups and you don't have to rely on retard grifters like astralite for finetunes are much. big parameter counts aren't everything
>>
>>102964868
>if he turned down SD3.5 it means that this fucker made more than a million out of ponyXL
Or he's just happy continuing with whatever he does
>>
File: file.png (271 KB, 1633x1555)
271 KB
271 KB PNG
>>102964897
I get that, I wished we would get Flux dev quality on a 0.1b model, but that's just utopia at this point, the scaling law is a thing and a small model's limit will always be lower than a bigger one
https://dynomight.net/scaling/
>>
>>102963916
This will be great for merges. Seems way better than the previous version. I wonder if it could be used as base for lora training
>>
File: 02163.jpg (1.73 MB, 1792x2304)
1.73 MB
1.73 MB JPG
>>
>>102964917
you need to remember that image gen is nowhere near the limits of the transformer architecture like llms are, the text gen space is filled with mega corporations pouring billions into squeezing out every possible point out of whatever benchmarks they are testing on, openai, anthropic, meta, qwen, whatever the fuck. image gen meanwhile is a much smaller market, filled with much smaller companies fiddling around with much smaller models and comparatively tiny compute. i don't see why we couldn't get flux level quality in a way smaller parameter count. maybe not 0.1b obviously but 5b? sure.
>>
>>102964980
Isn't DALLE something like 4B?
If so, knowing how good it can look, pretty sure we didn't see nothing yet.
>>
>>102964980
>i don't see why we couldn't get flux level quality in a way smaller parameter count. maybe not 0.1b obviously but 5b? sure.
I agree with that number, a perfectly trained 5b model could match flux dev, and that's why I believe Sana is gonna miss the mark, SD3M too, we need something medium, or else we have little shits or else we got this giant Flux model, there's still improvements to be made, and I won't forgive SAI for giving up on SD3-4b
>>
File: image (3).png (436 KB, 1024x1024)
436 KB
436 KB PNG
>>
>>102964988
>Isn't DALLE something like 4B?
no clue, but openai barely cares about dalle because their main product is gpt and image gen is just too much legal trouble for little potential profit, and it still mogs everything else out there. so i think there's way more potential optimizations and research to be made for image gen, flux is just the tip of the iceberg.
>>
>>102965024
>openai barely cares about dalle
that's funny you say that because yesterday they improved the turbo distillation
https://xcancel.com/OpenAI/status/1849139783362347293
>>
>>102965024
>it still mogs everything else out there
it's probably because not only they didn't curate nsfw from their dataset, their synthetic captions are probably good because they used a good vlm
>>
>>102965041
obviously they are going to make optimizations to make running dalle cheaper, but it's not their main focus is what i meant.
>>
>>102965055
>curate nsfw from their dataset
still pissed off that it seems all current release start with a "safety" paragraph just meaning "ok so we got rid of all the useful anatomy data that is porn, because who knows people could generate that"
>>
File: 02166.jpg (2.1 MB, 1792x2304)
2.1 MB
2.1 MB JPG
>>
>>102965088
what if you went outside and saw this thing up in the sky trying to inhale the earth
>>
File: image (16).png (438 KB, 1024x1024)
438 KB
438 KB PNG
>>
File: 00040-1254627563.jpg (438 KB, 896x1152)
438 KB
438 KB JPG
>>
>>102965078
>still pissed off that it seems all current release start with a "safety" paragraph just meaning "ok so we got rid of all the useful anatomy data that is porn, because who knows people could generate that"
I was born too soon desu, in 20 years it'll be the norm to release uncucked models because there's alaways someone with balls who will dare to do it first and show to the masses that it's not the end of the world
>>
>>102965127
It's mostly a matter of compute.
And time is with us on that.
>>
The number of steps definitely have an impact
>bf16 64 steps
https://files.catbox.moe/z5r6yu.webm
>bf16 200 steps
https://files.catbox.moe/qvy7jx.webm
>>
>>102965160
https://github.com/genmoai/models/issues/2#issuecomment-2434139372
>While in preview, Mochi 1 only supports text-to-video. As a quick hack, you could describe the image extensively as the T5 XXL encoder supports long prompts, but that is a suboptimal solution since it'll lose most of the detail. We know I2V is important for the community so stay tuned.
nice, a model that can't do image2video is shit desu
>>
>>102965160
https://www.reddit.com/link/1gbg4ot/video/dm9ktw05cswd1/player
30 minutes
https://old.reddit.com/r/StableDiffusion/comments/1gbg4ot/mochi_1animation_24gb_vram_fp8/
I was planning suring today to test Q8 but my day has started early :(
>>
File: file.png (43 KB, 810x573)
43 KB
43 KB PNG
ok so I think if we want local Mochi to be as close as possible as the demo, you should go for 163 frames, and 30 fps, looks like this is the settings the model has been the most trained on, I'll provide a video once it's been done
>>
>>102965286
If you can do multigpu, maybe it's possible with 2x3090.
>>
>>102965195
good that they work on that, it's so much easier so start with an image you already genned or you just want to animate
>>
File: file.png (88 KB, 1489x540)
88 KB
88 KB PNG
>>102965299
it's already possible if you go for Q8, it asks for 19gb of vram
>>
File: 1729835388.png (565 KB, 522x672)
565 KB
565 KB PNG
>>
File: file.png (42 KB, 741x521)
42 KB
42 KB PNG
>>102965286
just as a comparison, Minimax videos are running on 25 fps + 141 frames
>>
>>102965286
>Mochi
>163 frames, and 30 fps
5.43 sec
>>102965337
>Minimax
>25 fps + 141 frames
5.64 sec

really small difference kek
https://www.youtube.com/watch?v=ekkSkdWt0mQ
>>
>>102965286
30fps seem to be what people render at, wether is the temporal speed of the model or not we don't actually know.
>>
>>102965376
>wether is the temporal speed of the model or not we don't actually know.
it is the temporal speed of the model, because the script tested a video straight from the API demo of Mochi
>>
>>102965383
Thanks. I didn't know that.
https://files.catbox.moe/5uu88o.mp4
>>
File: 00098-1254627564.jpg (330 KB, 1152x1536)
330 KB
330 KB JPG
asp2 wrangling
>>
File: 00170-1254627563.jpg (504 KB, 1728x1344)
504 KB
504 KB JPG
>>
I've been generating a lot of images and I'm pretty sure messing with the Flux sigmas just gets you less intelligent results (more mistakes).

I don't think I'll be touching that stuff anymore.
>>
The most annoying thing about genmo is you can absolutely see the potential but it's just not good enough
>>
>>
This one is almost good

>Under the ultraviolet pink and blue glow of holographic advertisements, two young Russian teenage girls in sleek, angular white outfits stand and pose on a high-rise rooftop. The sprawling cityscape sprawls beneath them, filled with pulsing neon lights. The camera captures their movements and beautiful face in slow motion as rain begins to fall, creating a shimmering effect on the rooftop

Maybe "teenage" is what their captioner used
>>
>>102966406
perhaps smaller resolution gives better results?
>>
Is it correct to say that RTX 3090 (24GB) and RTX 4090 (24GB) cannot run the original 24GB flux1-dev.safetensors at full speed?

Because I don't get full speed on my 3090, I have to use the 12GB quant. Is this just a problem with my setup?
>>
>>102966437
>perhaps smaller resolution gives better results?
I wouldn't know, I'm using the website so I can't customize anything, but there's no point in generating anything smaller than 720p for my usecase anyways
>>
>>102966578
have you tried this yet? requires pretty beefy gpu tho
>https://github.com/kijai/ComfyUI-CogVideoXWrapper
>>
I replaced "angular white outfits" with "blue leotards" and I got belly. Looks like ages are still inconsistent though


>>102966598
CogVideoX doesn't gave high enough framerate unfortunately
>>
>>102965708
I don't know it works pretty good for 1seagull gens
>>
File: 2024-10-25_00003_.png (1.58 MB, 1024x1024)
1.58 MB
1.58 MB PNG
>>
>>102964980
>image gen meanwhile is a much smaller market, filled with much smaller companies fiddling around with much smaller models
Because size does not matter, Dalle 3 outclasses Flux at 1/3 of the size.
>>
File: 1725174399046033.png (812 KB, 883x911)
812 KB
812 KB PNG
sd 3.5 likeness seems very bad compared to flux
>>
File: 2024-10-25_00004_.png (1.75 MB, 1024x1024)
1.75 MB
1.75 MB PNG
>>102966720
>>
>>102966821
could be user error
>>
>>102966856
Also, Emma has changed a ton. She's unrecognizable now.
>>
LOAD ALL THE LORAS
>>
File: 2024-10-25_00005_.png (2.13 MB, 1024x1024)
2.13 MB
2.13 MB PNG
>>102966845
Dishonesty at -.05 obliterates the image. Actually, it looks like a Canon 1DX at expanded iso.
>>
File: 00350-1254627563.jpg (489 KB, 1120x1440)
489 KB
489 KB JPG
>>
>>102966659
why is it so BLURRY
>>
31 minutes on 4060ti 16gb GGUF_Q8_0 mochi 67 frames 64 steps.

https://files.catbox.moe/0fkt54.mp4

Going to use this prompt and see how it comes out >>102966406
>>
File: 2024-10-25_00008_.png (1.14 MB, 720x1280)
1.14 MB
1.14 MB PNG
>>
File: 00147-1620015057.jpg (1.16 MB, 1408x2064)
1.16 MB
1.16 MB JPG
Mixuco!
>>
File: 2024-10-25_00010_.png (1.71 MB, 720x1280)
1.71 MB
1.71 MB PNG
>>102967195
>>
>>102967116
>why is it so BLURRY
Because videos are blurry
Maybe it's overtrained on bokeh like base flux is
>>102967122
>Going to use this prompt and see how it comes out
Since it's been more than 31 minutes now I'm assuming your result wasn't good enough to post (or too young)
>>
I've seen some pretty good results coming out of minimax/hailuo. Is there a way to run it locally in a similar fashion to comfyui?
>>
>>102967505
To get something as good as minimax/hailuo, you'd need 100GB+ vram, but also the model.
There is nothing as good locally, nor the ability to run it, and so far I've not been impressed by the blurry mess posted here for Mochi 1.
>>
File: ComfyUI_02636_.png (1.25 MB, 1024x1024)
1.25 MB
1.25 MB PNG
>>
File: ComfyUI_02598_.png (1.56 MB, 1024x1024)
1.56 MB
1.56 MB PNG
>>
>>102967486
nah i had a shower and made homemade pizza!
Turned out reasonably ok, for what it is, Q8 and all.

here you go:
https://files.catbox.moe/0et1cj.mp4

Prompt executed in 1811.06 seconds
>>
File: ComfyUI_02651_.png (1.55 MB, 1024x1024)
1.55 MB
1.55 MB PNG
>>
File: 1418041423282.jpg (20 KB, 306x306)
20 KB
20 KB JPG
>thought I'd have more control over facial features to be be able to better create individual looking characters in Flux
>nah
>>
I want my specific fetish to be in the next big AI model.

I have 1000 pics manually tagged dataset.
Which website do I upload it to make sure it will be added in the next databas
e scrap?
danbooru.donmai.us ?
>>
File: 3024981597.png (1.36 MB, 1344x768)
1.36 MB
1.36 MB PNG
>>
File: ComfyUI_01596_.png (1.27 MB, 1024x1024)
1.27 MB
1.27 MB PNG
>>
File: 2024-10-25_00013_.png (1.4 MB, 720x1280)
1.4 MB
1.4 MB PNG
>>
i demand a chibi migu in a suit
now
>>
File: ComfyUI_01600_.png (1.82 MB, 1024x1024)
1.82 MB
1.82 MB PNG
>>
File: 1876002013.png (1.35 MB, 1536x640)
1.35 MB
1.35 MB PNG
>>
File: 1712539816.png (1.07 MB, 1536x640)
1.07 MB
1.07 MB PNG
>>
File: ComfyUI_01607_.png (1.02 MB, 1024x1024)
1.02 MB
1.02 MB PNG
>>
File: file.png (491 KB, 448x544)
491 KB
491 KB PNG
>>
cozy thread
>>
>>102964600
>Model Ranking
>https://imgsys.org/rankings
wow SD3.5 not even on the list, it must be really shit compared to flux then
>>
File: 306218340094574596.webm (899 KB, 1248x720)
899 KB
899 KB WEBM
>>
>>102968542
it's okay to be gay
>>
>>102968530
cozy thread to be mass reported
>>
File: file.png (307 KB, 448x544)
307 KB
307 KB PNG
>>
>>102968633
clowns r funny :)
>>
>>102968577
total janny death and so on and so forth
>>
File: what_is_it.png (54 KB, 759x421)
54 KB
54 KB PNG
I am going to see if I can re-inject some images back into mochi. It says models on one side and samples on the other. It doesn't appear to be latents (as I have thrown VAE decode node on it and failed horribly). Has anyone tried this and got anywhere?
>>
File: Mochi_00002.webm (192 KB, 856x480)
192 KB
192 KB WEBM
>>
File: monroe.jpg (42 KB, 400x600)
42 KB
42 KB JPG
is there a word for this? trying to hold dress down while wind is blowing it up
>>
>>102968799
You're gonna need a lora for that I think
>>
>>102968799
Billowing up/upwards
>>
File: 00531-1615470503.jpg (1.09 MB, 1260x1620)
1.09 MB
1.09 MB JPG
>>102968694
It's so tiresome

>>102968799
Drop that image to interrogator, should be doable with just prompt
>>
File: 1708550557580330.png (467 KB, 399x399)
467 KB
467 KB PNG
back from the three day ban after posting a "body horror" image (pic unrelated)
what did i miss?
>>
>>102968903
Sana and SD 3.5M waiting room.
>>
>>102968903
>467 KB PNG
>back from the three day ban after posting a "body horror" image
What the fuck?
>>
can you catbox the body horror i wanna see it
>>
File: Mochi_preview_00002.webm (726 KB, 856x480)
726 KB
726 KB WEBM
>>102968708
I guess I will jump into the code. I think just a few keyframes will clean this up.
>>
File: 00549-2887602201.jpg (402 KB, 1344x1728)
402 KB
402 KB JPG
>>
>>102967728
>homemade pizza!
gonna need a finetune for that I'm afraid
>here you go
Thanks for sharing anon they look stylish. Good to see that maybe genmo just needs good prompts + luck because the quality didn't degrade very much at all going from 16 to 8 bit and from the websites model to the 480p one

>>102968542
Did you change the prompt for ages at all for this gen? This is close enough to the age I'm trying to get so maybe you cracked the code or maybe just got lucky
>>
File: 00175-1733960784.png (1.04 MB, 896x1152)
1.04 MB
1.04 MB PNG
>>
>>102968937
so nothing new?
>>102968944
>>102926927
>>
>>102969048
didn't change it
>>
File: 2x_upscale_video_00003.webm (1.75 MB, 1712x960)
1.75 MB
1.75 MB WEBM
Prompt executed in 1286.83 seconds.
Q8, 65 steps, 73 frames
Ran upscaling afterwards
>>
File: 3723796886.png (1.01 MB, 1344x768)
1.01 MB
1.01 MB PNG
>>
File: 00574-2887602198.jpg (349 KB, 1344x1728)
349 KB
349 KB JPG
>>102969193
Looks very nice after rife interpolation
>>
File: 00579-2887602198.jpg (430 KB, 1344x1728)
430 KB
430 KB JPG
>>
File: 00163-3004118374.jpg (850 KB, 1280x1720)
850 KB
850 KB JPG
>>
File: lmao.png (862 KB, 1049x778)
862 KB
862 KB PNG
I wonder how often that happens on civitai
>>
>>102969370
SHUT IT DOWN NOW
>>
>>102969370
BRO
>>
>>102969370
kek, close civitai
>>
>>102969370
HOLY SHIT AN AI GENERATED NAKED CHILD JUST FLEW OVER MY HOUSE
>>
Is it possible to use a diffusers model with ComfyUI? I trained a LoRA with SimpleTuner and it only works with the diffusers library...
>>
I've been genning some cute libtarts
>>
File: ComfyUI_Flux_14574.jpg (226 KB, 832x1216)
226 KB
226 KB JPG
>>
File: ComfyUI_00850_.png (1.62 MB, 1280x1024)
1.62 MB
1.62 MB PNG
>>
>>102969664
Very cool. I like the painted metal texture. From the neck down, at least. The head... not so much.
>>
Best prompt?
>>
>>102969140
>didn't change it
Understandable

>>102969193
Yeah this prompt is the best one so far. gonna work on variations of it when I have more time next week
>>
>>102969370
lmfao it's also really low quality and slimy. I remember whenever we had some pedo spamming in here they'd always be the worst gens you've ever seen even from a technical point of view. Either it was a case of debilitating mental handicap leading to both behaviors or it was bad actors who don't give a shit about AI image gen and just copied their settings from a civitai workflow.
>>
>>102969664
predicting with 99% confidence this one will be prominent in the collage
>>
>>
>>102969193
Try and run frame interpolation on it too with RIFE. The frames are all fake generated anyways so what is a few more?
https://github.com/hzwer/Practical-RIFE
>>
>>102968876
lol, good gen. chainmail didn't quite work out though.

The inability of these models to do keyboards and chainmail and things of that nature seems like the next big limitation I'd be trying to solve if I were a researcher

>>102968526
>448x544
based, looks good
>>
>>102969994
Chainmail fail is user error. Testing merges with Asp2.
>>
>>102970103
ok true it usually looks better than that. Nonetheless regular repeating linked geometry like that isn't the strong suit of these models
>>
>240p upscaled to 480
Prompt executed in 261.31 seconds, 5x faster than 480p

You can kind of get something at low resolutions. 240p is probably too low though.
>>
>>102969994
Yeah just playing with Pixart
>>
>>102969994
Chainmail and similar things are logic problems, same reason why rooms are nonsensical. The AI needs reasoning to understand why things are the way they are, why light switches are placed where they are, for example. It's a complex problem.
>>
Starting to hit my monthly limit on some accounts
>>
File: ComfyUI_00856_.png (1.36 MB, 1280x1024)
1.36 MB
1.36 MB PNG
>>
>>102970279
>some accounts
Hackermens!
>>
>>102970290
I see a futuristic plasma rifle in it
>>
>>102968708
Can Mochi be used as an image generator?
>>
File: Mochi_00006.webm (597 KB, 376x688)
597 KB
597 KB WEBM
Accidentally flipped the width and height, let it finish anyway, surprised it kind of worked.
>>
The sd3.5 branch of kohya appears to work for lora training. I mean, it's running, and the loss looks decent. Anybody know if you're supposed to set the weighting_scheme to something other than the default? There's a bunch of different options that affect both the timestep sampling and the loss weighting.
>>
>>102970444
flipper like that dolphin
>>
now that google colab is not collaborating anymore, is there any way to use free sd2.5? (creating 100000 fake accounts on their site kinda sucks)
>>
>>102970321
>Hackermens!
I was able to make more accounts thanks to my holiday SIM card
The retarded part is that it says 30/30 monthly gens when there's literally only 17 gens including moderated ones visible in my account
>>
File: Mochi_00009.webm (394 KB, 640x376)
394 KB
394 KB WEBM
Last one for now
>Prompt executed in 583.43 seconds
I think this resolution strikes a nice balance, between speed and quality for my 16gb card. Will test more later to see if I just got a lucky seed
>>
>>102969965
Link the custom node, but it only has support for up to 4.10
https://github.com/Fannovel16/ComfyUI-Frame-Interpolation
>>
>>102970629
Yeah I tried it, it's good for CogVideoX which is 8fps not much need for it with Mochi at 24fps
>>
>>102970477
>>102970477
>Anybody know if you're supposed to set the weighting_scheme to something other than the default?
Welcome to the bleeding edge, where you already know more about this than everyone else in the world.
And when you're here, you see what you get and experiment by changing the defaults and tell other people about it, and they will know, because you told them.
>>
>>102970648
24 FPS is enough to get stable output for 60 FPS video if that is what one wants. Not sure if the glitches would be amplified or not doing that.
>>
>>102970538
>is there any way to use free sd2.5?
If you mean sd3.5 have at it:
https://huggingface.co/spaces/Nymbo/Stable-Diffusion-3.5-Large-Serverless
>>
File: tmp24sm323i.png (1.16 MB, 1232x920)
1.16 MB
1.16 MB PNG
>>
so is Sd 3.5 a noticeable improvement to XL apart from the ability to do text?
>>
>>102970849
T5 is much, much better for prompting.
>>
>>102970648
Pretty sure mochi is 30fps?

>>102970675
I don't see a reason why interpolation would amplify glitches, it just tries to figure out an intermediate frame. Maybe if you're doing abstract stuff or explosions/particle effects
I have no idea what I'm talking about though
>>
>>102970405
it only has 4 nodes. I am looking into to cracking one open to get to the image gen piece.

short answer, no, it can not
>>
I have all the nodes. Assume the position
>>
Any video model can be used as an image generator, you just extract whatever frame you want from the MP4 with ffmpeg and save it
>>
File: download.png (116 KB, 750x690)
116 KB
116 KB PNG
>>102970849
Is there a way to have a "general coloration", like this image is blue. I tried "color fog", but it doesn't work, with midjourney it were easy and worked quite nicely
>>
File: tmp6mqcli57.png (1.28 MB, 1152x896)
1.28 MB
1.28 MB PNG
>>
It seems sd3 does not recognize anymore "by author"... So sd1 is still king after all
>>
>>102971023
feed it a colored latent, lower your denoise slightly and maybe increase steps
>>
I'm getting extremely confused with training. I've trained around 8 character loras and only 3 of then are generating decent results. I think it might be an issue with my captioning. How do you guys caption datasets of people? I've read mixed results and tried a few but the end result is always subpar...
>>
>>102970988
Yeah, it does. I think it's because the workflow default goes to 24 FPS which is why it is confusing but it is quite clear from the blog post announcement.
https://www.genmo.ai/blog
>Motion Quality: Mochi 1 generates smooth videos at 30 frames per second
>>
>sd3.5_large.safetensors
>sd3.5_large-Q[458]_[01].gguf
>sd3.5_large_fp8_scaled.safetensors
>above ones with "_turbo"
which one should I pick?
>>
File: ComfyUI_Flux_14590.jpg (213 KB, 832x1216)
213 KB
213 KB JPG
>>
nice
>>
very opposite opinion
>>
File: image (18).png (435 KB, 1024x1024)
435 KB
435 KB PNG
>>
"model" seems to help keep them older than pubescent but my sample size is only 2
>young Russian teenage model girl

>>102971230
>Yeah, it does. I think it's because the workflow default goes to 24 FPS which is why it is confusing
Shitty defaults and their consequences have been a disaster for the generative AI community
>>
>>102971677
what model anon? I like the water effect
>>
>>102969889
Probably to ensure that it was unmissably AI.
>>
>>102971068
can't people just make loras for it?
>>
File: 00357-2335055315.png (963 KB, 896x1152)
963 KB
963 KB PNG
I've been trying flux models for a while but I still don't get the difference between distilled CFG and CFG, what is each one for
>>
>>102971873
the other does the other thing
>>
>>102971068
They had pictures on their dataset with "by author".
They threw their tags to the trash.
They used a Vision Language Model to create new tags, and it doesn't know any authors.
I claim they did this because they are retarded and don't know their value could have been outdo midjourney, which is still king at this.
Clit Eastwood was preferable.
>>
File: 00358-3991456649.png (1.06 MB, 896x1152)
1.06 MB
1.06 MB PNG
>>102971891
>>
>>102970849
The ability to tell it where you want the things on the picture instead of relying on noise looking for what you asked for.
>>
>>102971917
CFG does not exist in Dev because they distilled it and made it 1.0 the default and unchangeable.
Dedistill brings CFG back at the cost of everything else worth using Flux for.
>>
press button to continue
>>
File: 00360-2759304127.png (1.02 MB, 896x1152)
1.02 MB
1.02 MB PNG
I'm trying out Verus Vision and holy fuck it's slow
>>
File: 306243625737261058.webm (1.35 MB, 1264x720)
1.35 MB
1.35 MB WEBM
wtf, I didn't type any prompt and it turned the fighter jet into a paper plane
>>
File: 00365-2397504837.png (991 KB, 1152x896)
991 KB
991 KB PNG
the results are good but damn is it slow
>>
>>102968031
Maybe wikimedia commons? Upload them as public domain images?
>>
File: ships_1.webm (2.13 MB, 856x480)
2.13 MB
2.13 MB WEBM
>>102966229
the movement is so fucking good, and in my 100% expert anon opinion the problem's with the VAE which is the one that turns the latents into the image making the finer details and implementing them coherently.
I can also see the potential, in this gen you can see the cup of coffee, the coffee stains in the borders of the mug, the coffee moves realistically but then the ships, the part that moves and is detailed looks like shit, that's why I think it's a problem with the temporal part of the VAE. Im disgusted at how close to impressive these look. So close yet so far, how sad.
>>
>>102972222
It looks like they are predicted based on motion vectors kind of like DLSS, probably not the VAE
>>
File: 00361-2486161067.jpg (451 KB, 1280x1720)
451 KB
451 KB JPG
>>
Man, not being able to start with an already existing image in mochi 1 makes awful results.
And worst of all, you can only see that after 30-40mn of inference.
>>
File: 00367-3805991773.png (886 KB, 1152x896)
886 KB
886 KB PNG
>>
File: 00062-547987379.png (1.38 MB, 832x1216)
1.38 MB
1.38 MB PNG
>>
>>102972254
thats a real drwawing
>>
File: 00372-2471372132.png (880 KB, 896x1152)
880 KB
880 KB PNG
>>
>>102972262
>30-40mn
lucky, I wait 1.5h to get 121 frames at 80 steps
>>
>Even Verus Vision is like 35gb
Please tell me there's a way to run this shit on CPU. My poor 3060 can't take this...
>>
>>102972573
I wish it was worth it
>>
>>102972603
8 or 12GB?
>>
>>102972626
12gb.
>>
>>102972654
I'm also running Verus on 12GB, but with a 4070 tho. At least memorywise it's doable
>>
>>102972677
Oh, shit, really? Is it FP16 or FP8? How slow?
>>
>>102972573
Using a 3090@300W, I get 5s at 24fps after 40mn.
>>
>>102972716
FP8, genning time is a bit more than a minute at 1152x896
>>
File: 2639166284.png (982 KB, 832x1216)
982 KB
982 KB PNG
>>
>>102972759
Mmm, I see. I hear the details suffer pretty noticeable changes going from FP16 -> FP8, is that true here, too? I've seen some comparisons with other models, it does seem like a bit of a drop.
>>
>>102972817
I haven't tried out FP16 Verus and don't have that much experience with Flux yet so I can't tell really. But Verus seems to have fixed the plastic skin issue that plagues all other Flux models I've tried so far, but it's also super fucking slow compared to the ones I've tried before.
>>
>>102972262
>Man, not being able to start with an already existing image in mochi 1 makes awful results.
Is it difficult to "patch" in image2video? I know that image2image from a text2image model is relatively simple since instead of just a random latent image full of noise you just start the gen with the image itself as the latent space. Why couldn't you just set the latent space of the first frame of the video model to an image in the same way?
>>
>>102972988
>Why couldn't you just set the latent space of the first frame of the video model to an image in the same way?
Because nobody has produced code for this, of course it's doable, we need a coding genius to figure it out.
And this proves ChatGPT is a meme or you could ask it.
>>
>>102972603
What? Flux on CPU? Willing to wait 2 hours for a 512x512 image?
>>
>>102972039
Imagine it in reverse.
>>
>>102972025
Only 3 images and I already hate the guts of that cat.
>>
>>102971917
Cute!
>>
seems like 100 steps is a solution for a lot of mochi quality ....issues

https://github.com/kijai/ComfyUI-MochiWrapper/issues/21
>>
>>102973372
oh, I'll try it asap
>>
>>102973372
>200 steps makes it even better
there is no cope settings with this one huh?
>>
>>102973372
>>102973438
It's over...
>>
is it possible to get any dynamic pose with flux? i wanna do fight poses with it but i always get static standing poses.
>>
>>102973449
What? No, this is great, you can get the quality of the unreleased HD already at 480p, you just need to up the steps.
Imagine if you could get Flux quality on SD3.5 in 100 steps, people would abandon Flux in a heartbeat, but it reaches its limits in 40, maybe low steps video generation is just like Schnell, so it needs 100 steps to shine.
>>
>>102973495
Um, not really, try the Kolors model for that, add the text in postproduction.
>>
>>102973518
>45 minutes to generate a 4 second video on a 3090
>this is great
>>
>>102973518
I think we all fell into the trap that 50 steps was the optimal point and what the model was good at, well, at least i did.
We've been spoiled with SD having small parameter ranges, it seems as you suggest, video models may require more iterations per frame.
>>
>>102973495
try repeating yourself with different words and using full sentences.
>>
>>102973550
It is what it is, I just batch them before going to sleep and that's it.
Until someone inevitably figures a way to accelerate that.
>>
>>102973495
The model was fed with shit captions around movement/pose etc, so it's mostly chance based.
>>
>>102973588
Text2video is just too random for it to be worth it, for me at least. If it was img2video that would be another story
>>
>>102973610
Oh I agree, I think the mochi devs said something about img2video, if they release an updated code with it, it'll be so much simpler to prompt it.
>>
>>102973550
>oh god, I wish this technology never existed and that there was no way for me to do this because now I have to wait 45 minutes.
>>
>>102973495
controlnet
>>
>>102973628
>if they release an updated code with it, it'll be so much simpler to prompt it
Neither img2img or ControlNet weren't added by the people that released Stable Diffusion, they were added by the community, if instead of waiting for mochi devs to implement it someone did it we'd already have it.
It's rocket science but we have rocket scientists.
>>
File: 1722630127970299.png (101 KB, 2015x675)
101 KB
101 KB PNG
>>102973667
see picrel
>>
Is there a browser for the rereleased LAION dataset yet?
https://laion.ai/blog/relaion-5b/
>>
File: Mochi_00004.webm (1.15 MB, 864x488)
1.15 MB
1.15 MB WEBM
>Prompt executed in 1444.05 seconds
100 steps. Only 60 frames tho
16gb VRAM
>>
>>102973702
Interested in Mochi stills generator. anon said he's looking into how to do it.
>>
Man, seeing more big youtubers make anti-ai videos is so weird when AI capability is constantly improving. It's like two parallel worlds.

Godot reddit community just banned AI too.
>>
>>102973806
If reddit is against it, it's probably the truth.
>>
File: ComfyUI_02642_.png (1.31 MB, 1024x1024)
1.31 MB
1.31 MB PNG
>>102973806
why are so many people against AI and getting so Butthurt over it?
did people also chimp out like that when Photoshop became a thing?
shouldnt people be happy that more tools exist now that can help one make great content?
>>
>>102973847
>why are so many people against AI and getting so Butthurt over it?
They're uninformed mostly.
>>
>>102973847
>why are so many people against AI
It's scary.
>>
>>102973847
>why are so many people against AI and getting so Butthurt over it?
Every other video is about moral panics around ai, no surprise the result is this + extreme way over the top hype that sold something that doesn't exist, aka perfect flawless ai.
And fucking openai that cannot stfu about selling their stuff like it's nuclear weaponry to increase the hype and get more investors.

>did people also chimp out like that when Photoshop became a thing?
Yes but it was way more subdued since social media weren't amplifying the thing left and right.

>shouldnt people be happy that more tools exist now that can help one make great content?
Obviously, but just look at how misinformed the average normie is around this.
What a normie sees every day are videos on how ai is evil, and another gazillion shitty ai made youtube videos/shorts with synthetic voices there to cash in easy money.
>>
File: ComfyUI_temp_vctya_00004_.jpg (851 KB, 2496x1924)
851 KB
851 KB JPG
im finally satisfied with the state of a gigachad lora, just training the base illustrious version (i gen on noob 0.5 and trained first for it) and ill have a full post in an hour or so
>>
>>102973930
i was really quite pleased with how it stabilized dynamic poses, which made me put in extra effort to deliver a clean 1.0 version
>>
>>102973847
Yes people really were butthurt when people started using tablets to digitally paint. They used the exact same "you're not a real artist" argument, said to professional artists that were switching from easel to iPad.
>>
>>102973914
>What a normie sees every day are videos on how ai is evil, and another gazillion shitty ai made youtube videos/shorts with synthetic voices there to cash in easy money.
it makes perfect sense to see people holding the simultaneous idea that ai is super dangerous to everything while at the same so useless no one will use it lol
>>
>>102973955
>Everything that was made before I was a teen is normal and part of life, everything invented after is dangerous and should be heavily controlled or banned.
Classic.
>>
>>102973978
I just think it's hilarious that these people write about ebil AI on the internet, the single most disruptive thing to ever exist that has cost millions of people their jobs.
>>
>>102973955
That's different since it's direct input vs directing. You aren't the artist, the AI is, you are merely trying to guide it to get the output. Using AI to generate images is more like you comissioning an artist, only that the artist is an AI in that case.
>>
smells like 2023 in here
>>
>>102974004
that was my fart
>>
>>102973995
Distinction without a difference, same arguments every time, I don't care after that. The fact is digital painting made art more accessible which means less commissions and less art supplies being bought thus costing people jobs. It's just a removal of a huge barrier to entry because traditional media costs a shit ton of money thus if any retard can buy a $200 tablet and paint it threatens the old guard. The end.
>>
>>102974030
There are literally professional artists working right now that have never used traditional media. Let that sink in. Just as there will be AI artists that will work professionally that have never picked up a digital pen. Don't seethe because people have removed the need to use a crude tool to draw a picture.
>>
I literally am an artist tho
>>
>>102974030
Current AI generators don't allow the translation of ideas and the level of fine tuning that a skilled artist has by simply drawing by hand.As long as AI does not allow the same amount of fine control it will not a replacement of an actual hands on creative process. Once we have art generators that allow this level of fine tuning in real time it will be superior/equal to hand drawn art, but the current descriptive approach will not be able to do that.
>>
>>102974083
If I generate a 1000 images refining a prompt until I get something I like, I'm an artist because I ultimately made the final picture. Feel free to seethe all day about it.
>>
>>102974022
open a window the smell is getting worse
>>
>>102974097
I do both genning and drawing and they are different things, and genning does not allow the amount of creative expression that drawing by hand does. It's its own thing, but as of now it's closer to photography than drawing and painting, as in taking a rough concept and the nrearranging it unti lyou get something you like.
>>
>>102974122
As someone with an art degree and have done multiple paintings, I completely disagree. And as the AI gets more complex in the ability to take direction and refine a picture through words or annotations, even that point won't be valid. But it doesn't matter, we've already established that randomness is art. If you throw paint randomly until you have a final composition that you like, that is art. No different than hitting the generate button.
>>
File: u4874.png (154 KB, 860x997)
154 KB
154 KB PNG
>>102974154
>As someone with an art degree and have done multiple paintings,
>>
>>102974097
>I downloaded a book I'm an author
>>
Trying that 100 steps for a video idea, my gpu is a space heater constantly spewing out 300W, thankfully it's becoming cold so it's kind of useful.
>>
is there any recommended lora for flux?
>>
did someone piss of /ic/ or something
>>
>>102974185
>I sprayed diarrhea from my asshole on a canvas. I'm an artist.
>>
>>102974199
I had marginal success on 1 test, it's better but my scene was inhently blurry due to the prompt, it's late here and i've gone stupid.
I will have a crack at it tomorrow if someone hasn't coded img2vid by then lol
>>
File: neverever.png (14 KB, 457x121)
14 KB
14 KB PNG
>>102973702
Let me show you picrel from here:
https://huggingface.co/playgroundai/playground-v2.5-1024px-aesthetic
That was promised in February 16th, and we are still waiting.
Don't trust anyone when they talk about their plans about the future, eat your chocolate now because the scientist that promised you two pieces may never come back.
People should assume they'll never provide image-to-video for mochi and do it themselves, or we could be asking where is it 8 months from now.
>>
>>102974241
>someone hasn't coded img2vid by then lol
sometime autism is surprising so who knows
>>
>>102974206
I strongly recommend an artstyle lora (just download a bunch and pick your favorite) because Flux sucks at style. Also loras decrease generation speed quite a lot so you might want to instead merge it or use someone's checkpoint.

The Turbo Alpha lora is good for speeding up your gens while you refine your prompt.
>>
>>102973847
>did people also chimp out like that when Photoshop became a thing?
There's the concept of hard work, that anybody deserves what they have if they worked hard for it, learning how to use Photoshop requires skill so it's respected, someone outdoing a skilled photoshopper by clicking a button will not.
The same happened when photography was invented, because people assumed making a great photo didn't require hard work.
>>
>>102974266
thanks anon
>>
>>102973761
local video models should be focused in img2video and deliver quality in smaller resolutions to save computing power, that is the only way to compete against cloud services


>>102973847
AI feels bigger than photoshop, the photographic camera, any new technology so far
>>
>>102974244
sure but I still hope they deliver lol
either way I'll be happy
>>
>>102974122
>genning does not allow the amount of creative expression that drawing by hand does
It does, you're just not using the tools right.
>I can't figure how to express my ideas creatively with AI generation, so nobody else can, so nobody does
>>
>>102974266
>The Turbo Alpha lora is good for speeding up your gens while you refine your prompt.
I should've added - it's good for speed but it does reduce quality a little.
So it's for writing your prompt, but not for finalization.
>>
>>102974257
If it's not there tomorrow morning EU hours i'm going to frustrate myself using a new install and env and gpt4o and try to lolzcode it. I've not coded anything since 68000 assembly. so like, don't depend on me.
>>
>>102974308
How much faster is it anon? Compared to just go with fewer steps and a smaller resolution?
>>
>>102974288
>AI feels bigger than photoshop, the photographic camera, any new technology so far
Only for people that spent decades of their life mastering their craft only to see other people outdoing them with some text and a click.
This isn't really bigger than mass production of chairs, that you spent so much time learning hoe to make the most comfortable one and now nobody will buy it because the fabric made 100 ones at 1/10 of the price and people don't know the difference because they haven't sat in yours, and even then, they'd rather save the money to get something else, so you wasted your life.
It's just that most people didn't live in the industrial mass production revolution, so this feels larger than life.
>>
>>102974221
True anon if you put it this way then yes you're also an artist
>>
>>102974313
well good luck, someone having fun with this will hopefully figure it out, the model without img2video is kind of sad after all
>>
>>102974221
Using ai is like being a manager. It's not easy, but it's not really art. Only metaphorically.
>>
>>102974397
It can be art, but you're closer a movie director than a graphical artist. I don't get why some people are so adamant about pitting it against handdrawn art, despite some outputs looking similar they're entitely different processes
>>
goddamn you are fucking retarded anon
>>
>>102974475
gonna cry?
>>
is brap posting still a thing?
>>
>>102974495
Brapposting is eternal
>>
>>102974397
You can use AI to enhance what you created manually or as an assistant.
>>
>>102974505
promptlet
>>
File: NotArt.png (1.92 MB, 1360x768)
1.92 MB
1.92 MB PNG
>>102974457
I don't get why people care so much about being considered artists, hampart is a kind of art done exclusively for profits, the whole NFT Monkeys fiasco was the peak of it, so not all art is good and not all artists are good so it's not necessarily a good term to be associated with.
I spent hours working on picrel, it's not art? Good! Who cares about that.
>>
>>102974530
i've always seen it just as gacha rolling for cool looking pictures.
>>
It doesn't matter what someone else thinks. All that matters is how you feel on the inside. How do you feel on the inside, anon?
>>
>>102974515
???
>>
>>102974550
>felt recoil
>recoilless
>soulless space shooter
>>
File: 2024-10-25_00021_.png (982 KB, 720x1280)
982 KB
982 KB PNG
>>
>>102973942
Not bad
>>
>>102974530
As said it can be art, but it's not hand painted art just as it is not a photography orca theater play
>>
>>102974530
I mean, I could doodle, paint, produce music, etc. before AI so when I tell someone that (in real life) they usually stop saying "you're not an artist" since I mainly use genai now.
>>
>>102974606
until the computer can make pretty pretty pictures without needing a human to prompt it the human is the artist simple as
>>
>>102974606
art really requires an artist, and ai art is artistless. therefore ai art is artless. They are images. This distinction is important to understand, but while philosophically accurate, what I said won't shape the vernacular, people will say ai artist and ai art, though neither exist.
>>
>>102974642
I used an LLM to generate this prompt:
>>102974576

>... a still life featuring a wicker basket overflowing with a variety of juicy, ripe apples, set against a warm, golden background that evokes a sense of autumnal coziness. The apples should be the main focus of the image, with some of them spilling out of the basket and onto the surrounding surface. Incorporate some subtle, natural textures and shading to give the image a realistic, tactile feel. The overall mood of the image should be inviting and appetizing, making the viewer want to reach out and grab an apple.
>>
>>102974642
When you comission someone to draw a picture for you are then you the artist?
>>
>>102974656
you still had to give the llm instructions, no matter how vague
the image didn't arise out of the ether
>>
File: 2024-10-25_00023_.png (1.01 MB, 720x1280)
1.01 MB
1.01 MB PNG
>>102974576
>>
>>102974664
it can't be like commissioning a human if the computer is not sentient. again, we had this discourse over a year ago once AI started to actually look like what it was trained on.

if you say "prompting is just commissioning" they you must also say "the computer is just a human"
>>
>>102974697
>muh sentience
meaningless word. you can absolutelly make prompting an ai be like commissioning a human, without making the ai sentient.
>>
>>102974697
The computer isn't a human of course, but it is the agent that creates the image. It fulfills the same role as a human artist here, it takes your instructions to turn them into an image. You are not involved in the creatipn of the image besonders giving instructions. So, if you aren't the artist when comissioning a painting why would you suddenly be the artist when comissioning it from an AI?
>>
>>102974643
Nothing you said is accurate. The prompter is the artist.
>>
>>102974753
>you are not involved in the creation of the image besides being involved in the creation of the image
Your concession has been accepted
>>
>>102974737
>>102974753
i dont want to accuse you of not interacting with AI at the same level as myself and others ITT but i am struggling to come up with another explanation for this phenomenon
if you are under the impression that the only way's to manipulate latent space are through words then you have much to catch up on
in the end, >>102974550 is correct and none of this matters. if one person says they're an artist when they use it and other says they themselves are not - neither can exude that onto the other.
just gen, anon
>>
>>102974762
So someone who comissions a painter to draw a painting for them is the artist who painted the image?
>>
New

>>102974813
>>102974813
>>102974813
>>
>>102974796
Did you accidentally include my post in there? You did not bolster your use of the concept of sentience.
>>
>>102974803
A traditional comissioner? No
A comissioner who is sitting right there the entire time directing the painter? Yes, how is this even a question.

Are directors not artists?
>>
>>102974844
>you can absolutelly make prompting an ai be like commissioning a human, without making the ai sentient.
okay, so do it tell me how
>>
>>102974869
>Technology will never be like that because I can't imagine it
And other things people have said throughout history, only to be proven wrong every time.
>>
>>102974666
No, you don't. It can generate text based on the random seed.
>>
>>102974886
And monkeys can write shakespeare. What the fuck is your point?

You do realize that the technology is irrelevant to the fact that it does in fact translate human intent into something else?
>>
>>102974876
never said that it wont be in the future. i said it's clear that it's not right now. so answer my question
>>102974886
fair, however i'd argue that your choice of LLM, image model, etc still means you are the ultimate decider. i'll only agree with you once we have literal robots who are functioning members of society
>>
>>102974918
>so answer my question
Ask one
>durr how do am thing work
is not a question its a piss poor rhetorical
>>
>>102974935
>Ask one
>>102974869
>>you can absolutelly make prompting an ai be like commissioning a human, without making the ai sentient.
>okay, so do it tell me how

im beginning to think you're being disingenuous, anon
>>
>>102974943
retard lmao
>>
>>102974918
And what if humans cease to be functioning members of robot society?
>>
>>102975103
come back when you learn how to do more than just prompt
>>102975115
good question
>>
Can Flux do actual recognizable people yet? I tried like two weeks ago and it couldn't even get a celebrity's hair color right. Sure as fuck knows what Geralt looks like though.

I'm currently convinced they nuked all popular celebrities from their training sets to dodge controversy.
>>
>>102975136
reading comprehension
>>
File: image (1).png (518 KB, 1024x1024)
518 KB
518 KB PNG



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.