Everyone Is Giving You Wrong Answers EditionDiscussion of Free and Open Source Diffusion ModelsPrev: >>107784474https://rentry.org/ldg-lazy-getting-started-guide>UIComfyUI: https://github.com/comfyanonymous/ComfyUISwarmUI: https://github.com/mcmonkeyprojects/SwarmUIre/Forge/Classic/Neo: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassicneoSD.Next: https://github.com/vladmandic/sdnextWan2GP: https://github.com/deepbeepmeep/Wan2GP>Checkpoints, LoRAs, Upscalers, & Workflowshttps://civitai.comhttps://civitaiarchive.com/https://openmodeldb.infohttps://openart.ai/workflows>Tuninghttps://github.com/spacepxl/demystifying-sd-finetuninghttps://github.com/ostris/ai-toolkithttps://github.com/Nerogar/OneTrainerhttps://github.com/kohya-ss/musubi-tunerhttps://github.com/kohya-ss/sd-scriptshttps://github.com/tdrussell/diffusion-pipe>Z Image Turbohttps://huggingface.co/Tongyi-MAI/Z-Image-Turbo>WanXhttps://github.com/Wan-Video/Wan2.2https://comfyanonymous.github.io/ComfyUI_examples/wan22/>NetaYumehttps://civitai.com/models/1790792?modelVersionId=2485296https://nieta-art.feishu.cn/wiki/RY3GwpT59icIQlkWXEfcCqIMnQd>Chromahttps://huggingface.co/lodestones/Chroma1-Basehttps://rentry.org/mvu52t46>Illustrioushttps://rentry.org/comfyui_guide_1girlhttps://tagexplorer.github.io/>MiscLocal Model Meta: https://rentry.org/localmodelsmetaShare Metadata: https://catbox.moe|https://litterbox.catbox.moe/GPU Benchmarks: https://chimolog.co/bto-gpu-stable-diffusion-specs/Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-oneTxt2Img Plugin: https://github.com/Acly/krita-ai-diffusionArchive: https://rentry.org/sdg-linkBakery: https://rentry.org/ldgcollage>Neighbors>>>/aco/csdg>>>/b/degen>>>/r/realistic+parody>>>/gif/vdg>>>/d/ddg>>>/e/edg>>>/h/hdg>>>/trash/slop>>>/vt/vtai>>>/u/udg>Local Text>>>/g/lmg>Maintain Thread Qualityhttps://rentry.org/debohttps://rentry.org/animanon
>>107787932thanks for the bread anon
>>107787928What the hell, are you running the full fp16?
>hates debo>is just a worse version of him
>>107787967no, ltx 2 at fp8 + the text encoder at fp16, why the fuck is comfyui asking for so much memory just to load them
>>107787932>TFT unironically using pony v7 in 2026How the mighty hath fallen
>>107787984135GB for FP8?? wtf
Let's pretend LTX2 is awesome so wan niggers open source 2.5
>>107787932>a new video model with sound just came out>no video collagesltx 2 is that bad? lool
>>107787932>>Maintain Thread Quality>https://rentry.org/debo>https://rentry.org/animanonhey is there a reason these are in the OP? it's just off topic schizobabble
i did not think this was gonna turn out at all
>>107788014make the same without tattoos
I waited for LTX2. I wanted something to be better than wan 2.2 so badly, and what we got was something censored at its core and even more resource hungry than wan already is. My smile and optimism: gone.
>>107788014what? you thought z-image couldn't do 3girls, sitting?
>>107788013the schizo ban evading baker will have a meltdown if we try to get rid of them. it's a reminder that /ldg/ is monitored by someone we originally left /sdg/ for but couldn't handle being an undesirable
>>107788014now make them look regular and not like turbo whores
>>107788032it's too big and the censorship part ia making me feel I'm using an API model, I don't want to feel like a cuck on local, fuck this modelhttps://files.catbox.moe/mgkbzy.mp4
>>107788032wait until training is possible, we'll see if censorship can't be beaten at least partially
>>107788048I like this style.
>>107787897Yes the finetune was the issue. lexivisionII is better.
>>107788051why can't he just take the hint already and fuck off?
ltx2 kissing audio quality?
>>107788090he insists on posting hit zit dogshit 2girls that look like a shitty photoshop layer filter
okay thats enough samefagging for now anonie
>>107788070>lexivisionIIwhat's that?
>>107788013He is a literal rdrama fag. He used to link to rdrama threads way back when there was just one /sdg/ thread, thinking anyone actually gave a literal fuck about that juvenile bullshit.
>>107788103is the donkey gonna be okay? please gen a follow up where the donkey is okay
>>107788102at least it isn't uggo gilfs he normally makes
>>107788108https://civitai.com/models/1607200/lexivision-ii-and-lexivision-z
>>107788029>>107788034>>107788053i happen to like turbo whores
>>107788115what a disgusting cretin. nice image btw
>>107788127>an experiment with generation of synthetic data using SDXL LexiVision AIO VNSFW >This is another merge of two private models that are being used to generate synthetic training data for other models.jeet levels off the charts in orbit even
who are you guys talking about? I'm new here
>>107788069James Jean. Picked it up decently, but could be better.
Has anyone tried the new version of Zeta-chroma?https://huggingface.co/lodestones/Zeta-Chroma/blob/main/zeta-chroma-x0-proto.safetensors
>>107788162Purchase an advertisement
>>107788145the baker that the mods refuse to do anything about
>>107787984mine loads with about 10gb free, I think it just highballs it and asks for 90% of your ram regardless of what it actually needs
>>107788162lodestone talked some techno babble i didnt understand other than it's in "pre training". i think this is from the latest version
>>107788140idc what works works
>tfw 32 stars and failed
>>107788184I don't have 150 gb of ram though, only 64, it's using the pagefile, comfyui's memory's management is so fucking ass when it comes to ram
>>107788192>idc what works worksbased philosophy
>>107788133much better
every time i want to post new loras civit is down, this is horseshit
>>107788195what failed?
>>107788133>turbo whoresoh that's why it's called Z-image turbo!
>>107788210tell me about it
>>107788226it's not down
>>107788162So is there a workflow for this or can we use the regular chroma workflows? And no, not joining or ever using discord
>>107788190desu I think we have more chance to undistill turbo and make it good than finetuning the base and make it as good as turbo, they've hidden the RLHF process details on their paper and that's for a reason, it's an important secret sauce
>>107788269>finetuning the basewhat base?
>>107788116he's just taking a nap
>>107788269>>107788285people are still hoping for base to be released in the year of our lord 2026
>>107788190>random 'ecker cameo
does anyone know how to re-enable automatic vhs preview? after the update, the vhs preview no longer appears automatically. fucking comfyui team
>>107788054https://files.catbox.moe/wdipti.mp4
>>107788236hell yeah brother
>>107788241models can't be rated/releasedhttps://civitai.com/changelog?id=129
>>107788338i'm sorry for lying then :( get well soon
5070ti with 64gb of ram here. default settings comfy workflow ltx2 took 225 seconds from button click to "It's raining it's raining", subsequent gens take 104 seconds (the text encoder offloading is painful)>>107788154>James Jeanbased. now that you mention it i recognize it. i almost bought a bottle of johnny walker blue label just because it had a dragon drawn by him on it
Does anyone have a workflow to generate a character in different outfits and poses?
>>107788305and they removed the search bar on the settings so you can't even write "preview" to find what you want quickly, those Ui jeets are so incompetent I swear to fucking god
>>107788305>vhs previewi dont know what that is but you activate preview in the settings now
Blessed thread of frenship
>>107788393nigger
>>107788398kek
>>107788348ty fren
>>107788013those are our lolcows / schizos and there is a rich history as to why those rentries are needed newfren
>>107788359I still see it
>>107788133>>107788323https://files.catbox.moe/ufzzwb.mp4those powerpoint (((glitches))) are so annoying, it removes the fun of that model
https://www.reddit.com/r/StableDiffusion/comments/1q5jgnl/ltx2_runs_on_a_16gb_gpu/it is indeed possible on 16gb
>>107788420i'm gonna wait at least a week before i bother with that model
>>107788435>powerpoints on 16 gblet's goo!!https://files.catbox.moe/bf7l3w.mp4
and there you have ithttps://files.catbox.moe/yuswxf.mp4
>>107788409nice try lolcow
>>107788477>in the soonlmaooo
>>107788477>in the sunare you sure this is a good idea Kim Jong Un? :d
>>107788477>re-erased soon
>>107788495that I can believe
>>107788477his mouth is blurrier than his eyescan you feed this thing encoded audio?
>>107788477https://github.com/huggingface/transformers/pull/43100you have no idea how much I want GLM image to be actually good so that we can move on from (((Alibaba))) once and for all
>>107788434very wet
>>107788477>even gets the engrish rightgood model.
>>107788477In the end, he can't even pretend not to find it funny lool.
>>107788477that's why I lurk this place, it knows the news before anyone else and there's some hidden gems here and there
Hey guys, I'm looking into some Wan2.2 video gens. I'm looking at the list of diffusion models and there's like 30 different ones. Is there any info anywhere regarding what the differences are?Like: high noise vs low noise? bf16 or fp8?Please and thank you.
>>107788560>Is there any info anywhere regarding what the differences are?>bf16 or fp8
>>107788477lolcatbox?
>>107788560there used to be a wan 2.2 rentry on the OP to help newfags out, don't know why it got removed though
>>107788585it already is a catbox anon
>>107788560If you're using ComfyUI then start with the example workflow and adapt it to your needs once you get it working. Wan 2.2 consists of a high noise model followed by a low noise model so you need both. Just download what the example workflow tells you to download.Hope you have at least 16 GB VRAM.
>>107788574Okay how do I go about recreating the image for ZIT? Just do it in a straightforward way and that'll be fine?
Is Nikolas' relative here? I have a question.
I am new to ComfUI and installed everything from the wan22ldgguide and it works fine. I used Wan2GP before that and there was a continue video option. I have no clue how to do that with ComfyUI. Is there workflow for that? Can someone please point me in the right direction? Any help would be appreciated.
Why is kij saying his 24gb 64gb ram setup is not eating pagefile when it is for basically everyone else?Like I don't care. I can just make a fuck huge pagefile on my nvme but why?
could someone share a working ltx 2 folder for comfyUI portable?
>>107788657linux vs windows memory mgmt maybe?
>>10778859732GB, I'm looking at this page, https://comfyanonymous.github.io/ComfyUI_examples/wan22/ from the OP, and it quickly puts me in a page with a lot of different wan2.2 models. I see, high noise to give it some broad strokes then low noise to refine the high noise output, in a nutshell?Reading between the lines, use fp16 if I can.What are the inpaint, camera, and controls models for?wan2.2_ti2v_5B_fp16.safetensorslooks like a non-specialised but easy to begin model?
https://huggingface.co/lightx2v/Qwen-Image-Edit-2511-Lightning/tree/mainslop vs slopitty slop
>>107788477https://files.catbox.moe/eukq6x.mp4Come on dude, they even powerpoint the jeets
>>107788699Don't use 5B, it'll output crap. Especially since you have a 32GB card.Open ComfyUI, click Templates, select Wan 2.2 14B Image to Video. Download what it tells you to download.Don't do anything else until you get that working.
>>107788587it was dated and poorly written
https://files.catbox.moe/ill7c8.mp4a checkpoint in $current_year that can't do text really feels like a regressionI mean the video is no wan2.2 either, but at least it's fun
>>107788734I answered my own questions after looking at the templates list (i must have an old comfy its only got 2.1)inpaint - fill in the gap from a start and finish image,control net, use an existing clip as a guide for the prompt,camera? don't know
>>107788760https://files.catbox.moe/ovc59x.mp4yeah the text is absolute dogshit, at least the audio is consistent
another Q, if i bash a folder with the comfy github will it erase missing files from the source (delete all my models)?
>>107788734Thank you by the way, you've actually been a great help!
>>107788552close but no cigar
>>107788809btw, who does the image input has to be jpeg compressed for it to work? that's dumb
>hear about ltx2>open comfy>check if there's a ltx2 template already>oh cool there is>click>pic relatedWhere do I find an actual workflow?
>>107788886update, it was released today
>>107788886what? I'm using the template and it's not using api nodes
>>107788760Are you prompting anything related to style or era? Zoomer here, it looks like it could've come out of the 2000's or something.
How do i get free ComfyCredits (tm)?
>>107788880kek
LTX2 is the only model I can think of that actually cannot generate a miku t2v.
>>107788898nope, pretty barebones prompt you can check yourselfI get the feeling that a lot of the training data was sitcoms, it kinda pulls it in that direction
>>107788906kandinsky can't either
>>107788032They released the full model didnt they? You can probably train it in just need a horny chink or saudi oil baron lol
>>107788906the size, the powerpoint (((feature))) and no Migu is gonna kill the model, I feel I'm doing an humiliation ritual, having to spare more than 100gb of pagefile just to run this garbage
https://files.catbox.moe/unkrlq.mp4
>>107788950Lol
Zit is actually inferior to Qwen/Flux 2 if you're trying to do anything multi-subject. It's retarded like SDXL
>>107788967when you go for NAG + boomer prompt it works fine
>>107788950migu :(
sometimes you need to dust off chroma and get some 1girl asian footpics to remind you what it's all about
>>107788936That's a lot to criticize the model for, but the powerpoint thing is a skill issue.
https://files.catbox.moe/o5et5j.mp4
>>107788988
>>107788999https://files.catbox.moe/93vp5x.mp4went from 0.5 megapixels (49 sec) to 0.92 megapixels (1.23 mn) and it's definitely better, the biggest strength of that model is the speed
>>107788973>NAGI need the node link, cant find it
>>107789048https://github.com/scottmudge/ComfyUI-NAGgo for those settings for Z-image turbo, it's really important
Is the MultiGPU anon a schizo trying to rugpull with a crypto miner? Comfy already has off-loading to RAM.
>>107789076>Comfy already has off-loading to RAM.and it's shit, and when an automatic feature is shit, you go for the manual one
I have no fucking idea why my RAM and pagefile is doing when running LTX 2. Sometimes it's comfortably around 40-60gb and other times it's maxxed out and raping my nvme drive and there is no pattern to it.
>>107789031we need an AI board more than ever, it's gonna be annoying to always have to upload videos through a catbox, c'mon 4chan, get your shit together
https://files.catbox.moe/9mgyue.mp4
Anon talk about nsfw T2V but don't use Kandinsky 5 pro? X)Kijai made a cumfart versionNo LoRA, 24fps, 100% nsfw.
it's crazy how useless this thread is when you need help...
>>10778918230% of the time, it is because you are bad at asking for help and no-one is bothered to figure out what your problem even is
just woke up, is it still not possible to run this ltx2 shit on a 4090?
>>107789182Most of the questions asked here could be answered by an LLM
>>107789165>100% nsfw.I don't believe you, show a video then
>>107789161>>107788760literally no different than "gibs buzz pls". it's tired.
>>107789196this was generated on my 4090https://litter.catbox.moe/vehmzaqnfbs4xlxj.mp4
>>107789196you can run on a 24gb vram card, but you have to do some manual offload to make it work, like--reserve-vram 4
https://files.catbox.moe/kpo5em.mp4Having to upload this shit to catbox is annoying.
>>107789208How much time?
>>107789241NTA but shit is ludicrously fast. Under 2 minutes on a 3090 at 720p and 200+ frames.
>>107789182let me guess you were the anon who didnt know how to pip install things
pip install
>>107789212wish i knew about this earlier. i have it at 1.5 but i can surf 4chan, watch porn and youtube videos while running heavy models all at the same time
>>107788886>see where ltx/light tricks is based>see piclol, lmao
>>107789241nta but on my 3090 with the distilled version I have this time>0.92 megapixels>125 frames>8/8 [01:22<00:00, 10.33s/it]it's really fast, the slow part is loading/unloading the models unfortunately :(
>>107789216Nice.
>>107789249you also have multiGPU if you want to some manual offloading https://github.com/pollockjj/ComfyUI-MultiGPU
https://files.catbox.moe/q44ttz.mp4
>>107789252>>107789245Pretty good!
>>107789280my ears...
>>107789280kekd
>>107788936>no migui2v miku in then
It's cool that with edit model, you can make your character do an "A-pose", edit to get the other sides, and then generate a 3D model. I wish it came with textures though...
https://files.catbox.moe/ub810c.mp4
>>107789303yeah but even on i2v I had my fun by making miku appear on the screen, can't do that with that model, sad
audio + image to video with ltx2 is fucking insane btw. And these take less than 2 minutes to gen. Could be ever better with more time im sure.https://files.catbox.moe/eea5wn.mp4https://files.catbox.moe/wunip1.mp4https://files.catbox.moe/m3tt74.mp4https://files.catbox.moe/k29y60.mp4
>>107789307kek
>>107789308loras work, there is nothing stopping a lora from doing off screen mikus once they are made
>>107789271I thought about running multiple GPUs like with an RTX 4000 pro but I would need a new PSU
>>107789309>https://files.catbox.moe/wunip1.mp4ok that's pretty cool
>>107789309so that's a yes to >>107788510 ?
>>107789309I had thought sora 2 would be some massive 1T model or something, but ltxv2 has made me think it could be closer to like 50-100B
>>107789305Which image edit model and which model for 3d?
>>107789325WF: https://files.catbox.moe/f9fvjr.json
>>107789208do I need to modify the template workflow a lot to make it work?>>107789212>--reserve-vram 4how much vram would that reserve?4gb or 400mb?
>>107789326sora 2 is much better though, and with way more pop culture knowledge, and what LLMs made me learn is that you need for the model to be big to remember all those concepts
>>107788133based turbowhoremaxxercatbox/prompt?
lol, comfy fixed the previewshe disabled them
Is there a general on this site for video generation? I remember a few threads on /gif/. Or is this the only thread?
>>107789329I use my finetune to get a specific type of non-flat color 3d model (but its Qwen Edit so 30~GB of GPU RAM). Alternatively use Grok Edit, that works too and you get like 1000 images per hour?
>>107788585based 4chanXtard
>>107789326No it could still be an autoregressive LLM model of like 1T, but probably MoE so it doesn't cost so much it's impossible to run.
>>107789341He thinks on a level beyond the average programmer. He finds solutions for problems others deem unsolvable.
>>107789332>how much vram would that reserve?4gb, you can take my workflow if you want, I removed a lot of useless bullshit from the official template (fuck upscalers)https://files.catbox.moe/lclc9t.json
>>107789344I meant which model that generates 3d models from images. All the ones I've seen before weren't very good and I don't think I've seen any that take 4 images as input.
We are living in the future.https://files.catbox.moe/t3emm7.mp4
>>107789309>And these take less than 2 minutes to gen. Could be ever better with more time im sure.Since the current workflow can't use sageattention somehow>Error running sage attention: Input tensors must be in dtype of torch.float16 or torch.bfloat16, using pytorch attention instead.yeah it can be even better
>>107789362spell it phonetically, smoke-you-lease
>>107789355Hunyuan3DMV. It generates voxels then smooths them out, so you might holes. I then go sculpt mode in blender to fix it up.
>>107789363I meant the quality could be better with way more steps but that too. Still a lot of speed that could be had as well. LTXV2 is black magic in that regard
>>107789371>LTXV2 is black magic in that regardthey tried to get speed by compressive the vae latents on some shit on ltvx1 but it didn't work out, glad to see they didn't give up on that idea, it looks better than Wan 2.2 while being much faster, that's what I love to see, actual architecture improvements, and not just "stack moar layers bro"
https://www.reddit.com/r/StableDiffusion/comments/1q5k6al/fix_to_make_ltxv2_work_with_24gb_or_less_of_vram/there is a fp8 gemma encoder too it seems, 12gb
Still remains to see how it trains but for some reason people last night were freaking out because it kept generating indians.
https://files.catbox.moe/qoooe4.mp4
>>107789387fp8 gemma encoder only works on the LTXV2 nodes WF btw, naitive comfy does not support it yet, it just loads it as fp16 anyways
>>107789396>"death to all kikes">uses a model made in Jerusalemkek
>>107789309desu if that model can be pruned to 14b and have its censorship layers removed it can definitely be a Wan 2.2's replacement
>>107789182Yeah, its pretty terrible. Its usually filled with combative elitist contrarians who'll tell you how something they never built actually works. The only helpful advice from ldg was a suggestion to use chroma last year. Mind you, its the same place who will throw an isle 7 tantrum when their favorite variation of a model doesn't get released on time. I come here for the keks now.>>107789192>>107789201kek
>>107789387>>107789400fp8 on text encoders have never been a good idea, I'm waiting for its Q8 gguf
>>107787989>falling for obvious shitposting
>>107789406https://files.catbox.moe/41k7n2.mp4
https://files.catbox.moe/s6k434.mp4it can do trump's voice really well lol
>>107789458lmao, true true
>>107789458I assume everyone who's ever asked for a nsfw TTS is either retarded or pretending to be retarded
https://files.catbox.moe/jmus4n.mp4it's so biased towards indian shit :(
>>107789483you can prompt the accent, dialect, even the toneit's actually an impressive tts on its own. In my limited testing it knows angry, sarcastic and worried
>>107789352thanks bro, I'll try it out.
The LTXV WFs are WAY better btw https://github.com/Lightricks/ComfyUI-LTXVideo/tree/master/example_workflowsYou get much better results, something in comfy is brokenhttps://files.catbox.moe/mpd5u9.mp4https://files.catbox.moe/lvnhqk.mp4https://files.catbox.moe/htkh8y.mp4https://files.catbox.moe/55tof3.mp4https://files.catbox.moe/jgj4md.mp4
how much better would the tech have been if everything wasn't python?
>>107789540I dunno, save an extra litre of water per thousand gallonsit adds up
>>107789518the ltx workflows all crap out at the text encoder stage for me.
>TXVGemmaCLIPModelLoaderNo files matching pattern 'tokenizer.model' found under E:\ComfyUI_windows_portable_nvidia-latest_20\ComfyUI_windows_portable\ComfyUI\modelsyou piece of shit, I have gemma_3_12B_it.safetensors
ltx gens regularly have extra limbs, this was never an issue with wan 2.2
>>107789540pyTorch is C++
>>107789547use the fp8 encoder, and switch out the vae decode, their WF is made for 5090 but can work on 4090
>>107789518:(
workflow says to use:text_encoders/ comfy_gemma_3_12B_it.safetensorsand the workflow link doesnt have it:https://huggingface.co/google/gemma-3-12b-pt/tree/main
>>107789518>something in comfy is brokenpottery
>>107789548you have to download the entire gemma 3 folder and put that inside of it.https://huggingface.co/google/gemma-3-12b-it/tree/mainyou dont need the checkpoints from this, just all the smaller files
>>107789555all niggardry that surrounds it is python. 3/4 of pytorch is python abstractions. it's so bad comfy made a separate module just for skipping over python abstractions
>>107789540not better, everything actually important is not in python
>>107789568That's just the API. The heavy work is done in C++ and Cuda.
>>107789560>>107789567the ltxv2 WF's need the full gemma folder and the checkpoint put inside of it
>>107789547I had to start comfy with "-reserve-vram 4" flag
>>107789567>inside of it.inside of what? I make a new folder on the text_encoder folder?
>>107789562How do I do that on windows? It only says git
>>107789578then it's not pytorch anymore, it's libtorch
>>107789584no, put the comfy_gemma_3_12B_it.safetensorsinside of the https://huggingface.co/google/gemma-3-12b-it/tree/main folder
>>107789562I use this: https://huggingface.co/GitMylo/LTX-2-comfy_gemma_fp8_e4m3fn/blob/main/gemma_3_12B_it_fp8_e4m3fn.safetensorsyou can also use https://huggingface.co/Comfy-Org/ltx-2/blob/main/split_files/text_encoders/gemma_3_12B_it.safetensors
>>107789592no
>>107789572>>107789578why is the API 3x more code than the actual kernals?
>>107789601>https://huggingface.co/google/gemma-3-12b-it/tree/mainis this a fucking joke?
>>107789567thanks, was worried I had to download 30gb of shit again
>>107789583That works fine for the cumfart workflows. Not so much for the ltx ones.
https://www.reddit.com/r/StableDiffusion/comments/1q5r23b/comment/ny3fedo/?utm_source=share&utm_medium=web3x&utm_name=web3xcss&utm_term=1&utm_content=share_button>Create a subfolder inside the text_encoder folder, for example "gemma3fp8", move the gemma12b fp8 file inside the newly created subfolder, and download all the json files from this this link (google's huggingface page) inside the newly created folder where the fp8 file is.
>>107789621try with the fp8 text encoder?>>107789603
>>107789613yes, now choose your preferred pronoun and safe word anon
>>107789609Doesn't matter.
>>107789382>it looks better than Wan 2.2 while being much faster, that's what I love to see, actual architecture improvements
I've never seen an application that makes people jump through so many hoops like this before. it's truly aweful
>>107789613get the files here I guesshttps://huggingface.co/unsloth/gemma-3-12b-it/tree/main
>>107789629You could find any number of high motion wan frames to clown on the model too.
>>107789615and now:ValueError: Cannot use chat template functions because tokenizer.chat_template is not set and no template argument was passed! For information about writing templates and setting the tokenizer.chat_template attribute, please see the documentation at https://huggingface.co/docs/transformers/main/en/chat_templatingwtf
>>107789628you don't find it off the python wrapper API is more bloated than the actual inference kernals? I thought python was supposed to be convenient
only anons who used early LLMs ooba would understand actual annoyanceinstalling ltx2 is nothing
>>107789629I already answered to you about that >>107789031
>>107789609jeets like this >>107789628
>>107789641ah, disable the prompt enhancement node, you gota do a whole load of stuff to get that working
>>107789641nm it's the stupid enhancer node.
>>107789643It doesn't matter the heavy lifting is in the kernals. Everything else is just API. There's nothing that can't be provided by Python fast enough.
>>107788435>16gbit workshttps://files.catbox.moe/mqkjkn.mp4
>>107789659>>107789661it just works for me with 24GB vram
>>107789636>https://huggingface.co/unsloth/gemma-3-12b-it/tree/mainit should be working now if you do this shit, btw the "model.safetensors" is this filehttps://huggingface.co/Comfy-Org/ltx-2/blob/main/split_files/text_encoders/gemma_3_12B_it.safetensors
Oh. AND USE Res_2s NOT ULER!!!
Do I just have some problems with my workflow's prompt adherence or is LTX just not trained on choking? Or gut punches?
can I get a basic workflow without all this enhancer garbage by these retards who think a BASIC WORKFLOW should add all this crap, it wont work with it bypassed.
>>107789678bro why is it so slow to load the encoder, it's taking ages to fill up my ram
>>106977329is it possible you could reupload this? sorry for replying to something ancient, i've been looking for something that works to do this
https://files.catbox.moe/sipc7m.mp4
>>107789702>>107789352
Is there a point in using gemma3 12b abliterated?https://huggingface.co/mlabonne/gemma-3-12b-it-abliterated
>>107789722if we're lucky, using abliterated would remove the censorship and the powerpoint shit lool
>>107789702>LTXVGemmaEnhancePromptCannot use chat template functions because tokenizer.chat_template is not set and no template argument was passed! For information about writing templates and setting the tokenizer.chat_template attribute, please see the documentation at https://huggingface.co/docs/transformers/main/en/chat_templatingand if I bypass the enhancer:CLIPTextEncodemat1 and mat2 shapes cannot be multiplied (1024x62208 and 188160x3840)I wonder how retarded normies figure out comfy
>>107789712kek
>>107789736>I wonder how retarded normies figure out comfyThey don't, they use grok.This stuff is too technical and cutting edge for most normies.
>>107789664you keep saying this but it's literally a pile of shit all over and on top of it. just make bindings because we don't need any of this fucking python shit.
>>107789748@grok is this true?
>>107789664>It doesn't matter the heavy lifting is in the kernalsif it doesn't matter why force python in order to use it? why make the utilities for processes in python? you say it doesn't matter but that would mean using python at all is pointless
>>107789352this is loading/working normally unlike the shit template workflow in comfy, waiting to gen but it's working, along with --reserve-vram 4 in launch optionsupdate it comfy anon
>>107789768it's for researchers first
>>107789768What a pointless discussion.Because researchers in ML use python since forever, it's the de facto language used, and no amount of bitching from a random anon will change that.If you are trying to argue that without python the generation would be faster, that's just false.
late to the party, can I run ltx 2 with 16gb vram and 32gb ram?
>>107789518>RuntimeError: Expected all tensors to be on the same device, but got index is on cpu, different from other tensors on cuda:0 (when checking argument in method wrapper_CUDA__index_select)sigh... I guess it won't work if you use --reserve_vram right?
>This model will only work on a 5090 at the bare minimum>Actually err, a 24gb gpu will work. >Well maybe 16gb>okay 8gb is fine toohow does this happen every time?
>>107789795>>107789798if researchers put out a model worth a damn they wouldn't need to waste time learning python. that's how you know a code assist model is garbage, it's training and inference is written in python
fresh when ready >>107789820>>107789820>>107789820>>107789820
>>107789798you are clearly are a junior intern or something
>>107789822why do you bake early? what the fuck is wrong with you?
>>107789823I'm sure I am anon, secretly we all use cobol.
>>107789847obsession over his male crush
>>107789854kek
>>107789854cuda uses fortran not cobol>>107789861samefag. it really isn't that funny when he had the low hanging fruit
>>107789794the anime girl runs towards the camera and says "miku miku miku" in a cute anime style voice.sticky anon's workflow cause the comfy template one sucks dick and this works, also this model is pretty fast, 16GB 4080 64GB ram and it works just fine.proof it works: https://files.catbox.moe/v1qdpo.mp4
>>107789823>>107789865mr catjak doesn't have a junior position, he can't get a job
>instant crying
Qwen Edit 2511"Remove the jacket. Keep everything else the same"It doesn't remove the jacket but it zooms out and put the girl at half scale in the middle. What causes this?
>>107789880hope Mr catjak feels better soon
>>107789918if you read the last 15 minutes of posts it appears that someone is crying because of "Mr catjak"
>>107789912
i remember hearing news over the past few days about a z-image lora fix inside of either z-image or ComfyUI. not the lora that fixes other loras, but something inherent that was resolved in either the model or the GUI. i was working when i read about it. anyone know what it was, or am i hallucinating?