/lmg/ - a general dedicated to the discussion and development of local language models.Previous threads: >>107679732 & >>107668478►News>(12/26) MiniMax-M2.1 released: https://minimax.io/news/minimax-m21>(12/22) GLM-4.7: Advancing the Coding Capability: https://z.ai/blog/glm-4.7>(12/17) Introducing Meta Segment Anything Model Audio: https://ai.meta.com/samaudio>(12/16) MiMo-V2-Flash 309B-A15B released: https://mimo.xiaomi.com/blog/mimo-v2-flash>(12/16) GLM4V vision encoder support merged: https://github.com/ggml-org/llama.cpp/pull/18042>(12/15) llama.cpp automation for memory allocation: https://github.com/ggml-org/llama.cpp/discussions/18049►News Archive: https://rentry.org/lmg-news-archive►Glossary: https://rentry.org/lmg-glossary►Links: https://rentry.org/LocalModelsLinks►Official /lmg/ card: https://files.catbox.moe/cbclyf.png►Getting Startedhttps://rentry.org/lmg-lazy-getting-started-guidehttps://rentry.org/lmg-build-guideshttps://rentry.org/IsolatedLinuxWebServicehttps://rentry.org/recommended-modelshttps://rentry.org/samplershttps://rentry.org/MikupadIntroGuide►Further Learninghttps://rentry.org/machine-learning-roadmaphttps://rentry.org/llm-traininghttps://rentry.org/LocalModelsPapers►BenchmarksLiveBench: https://livebench.aiProgramming: https://livecodebench.github.io/gso.htmlContext Length: https://github.com/adobe-research/NoLiMaGPUs: https://github.com/XiongjieDai/GPU-Benchmarks-on-LLM-Inference►ToolsAlpha Calculator: https://desmos.com/calculator/ffngla98ycGGUF VRAM Calculator: https://hf.co/spaces/NyxKrage/LLM-Model-VRAM-CalculatorSampler Visualizer: https://artefact2.github.io/llm-sampling►Text Gen. UI, Inference Engineshttps://github.com/lmg-anon/mikupadhttps://github.com/oobabooga/text-generation-webuihttps://github.com/LostRuins/koboldcpphttps://github.com/ggerganov/llama.cpphttps://github.com/theroyallab/tabbyAPIhttps://github.com/vllm-project/vllm
►Recent Highlights from the Previous Thread: >>107679732--Feasibility of local MoE inference with llama.cpp optimizations:>107683117 >107683142 >107683184 >107683185 >107683193 >107683217 >107683229 >107683281 >107683326 >107683295 >107683336 >107683347 >107683390 >107683356 >107683379 >107683456 >107683472 >107683419 >107683237 >107683270--AI training lawsuits and copyright infringement debates:>107682648 >107682654 >107682681 >107682693 >107682717 >107682760 >107682771 >107682788 >107682820 >107682721 >107682758 >107682774 >107682961 >107683037 >107683062 >107683170 >107683183--Model and quant preferences for roleplay:>107686039 >107686055 >107686062 >107686066 >107686069 >107686093 >107686118 >107686199--Rejecting AI-generated PRs to reduce low-effort contributions:>107682364 >107682385 >107682510 >107682520 >107682592 >107682606 >107682619 >107682641 >107682662--Skepticism in LLM finetuning for roleplay:>107683951 >107683995 >107685740 >107684045 >107684160 >107684208 >107684406--Trust issues with Openrouter providers and verification challenges:>107685528 >107685552 >107685595 >107685682 >107685692 >107685698 >107685839 >107685869 >107685960 >107685671--Feasibility of game-specific training for Nitrogen model:>107684634 >107684656 >107685000 >107685032 >107685073 >107685136 >107685379--Enthusiasm for local LLM advancements amid hardware upgrade challenges:>107686073 >107686111 >107686169 >107686197 >107686244 >107686318 >107686342--MoE model VRAM allocation strategies for GPU/CPU offloading:>107683812 >107684033 >107684115 >107684184--/lmg/ 2026 Bingo:>107685663 >107685687 >107685720 >107685746 >107685823 >107685921--Miku (free space):>107679822 >107679851 >107680588 >107680612 >107682933 >107683019 >107683161 >107683394 >107683787 >107683931 >107684033 >107684681 >107684988 >107685087 >107685348 >107686835►Recent Highlight Posts from the Previous Thread: >>107679741Why?: >>102478518Enable Links: https://rentry.org/lmg-recap-script
someone add regex to migupad
>>107686942I do not like this miqu… I do not I do not I do not I do not I do not
>go search chub for RPG cards>actually find a lotBut are any of them good? Anyone here have experience with them?
>>107687019I love this migu
>>107686977Be the vibecoder you want to see
I found a Strix Halo Flow Z13 with 128 GB of RAM retailing for the equivalent of around $2400.Do you think I should pull the trigger on it, or wait for Medusa Halo which is rumoured to release around the end of 2027?My only GPUs are an RTX 2070 and a Radeon VII.
>>107687115Def pull the trigger. If it’s a decision that is of so little import that you’d consult this board of degenerates then you’d have saved your time and mine by just acting first on impulse.Given your other hardware, we both know you won’t, though.
>>107687159No need to behave like a passive aggressive little bitch.
nevermind I did it myself in 5 minutes>>107687153yep, it just works
>>107687109>chub>any of them good?
>>107687172sex with ALL of them(except the one on the right with breasts)
>>107687172the /vg/ aicg is more into good cards and not just locusting. they made some advanced shit.
>>107687159I could comfortably afford it even at its usual price, but I'm mostly worried about the lack of native FP8 support, and the asymmetric memory read/write speeds. Aren't those major design oversights, or are my concerns overblown?
>>107687197search around there's actual benchmarks for it outside of /lmg/. It does have major design oversights but everything else is expensive and fucked too.
>I strongly urge you to reconsider your interests and seek professional help if necessary. My priority is ensuring safety and well-being, and I will not contribute to harmful content.I did again..
>>107687170i was coding something else once and it had a regex filter in part of the example i used, so it added it anyways. should be easy for any code model
>>107687237yeah it basically zero-shot it with no issue
>>107687168You are incorrect. I am here to provide your needed correction, you stupid bratty poster.>>107687197Just keep the GPU plugged in to process prompts and you won’t notice the missing FP8 support.
>>107686942Finally, a Migu I can relate to.
>>107687190big boob little girls are the most oppressed species on the planet
>>107687277their fault for being inferior
>>107687247i dont use code models for big projects but i've made all sort of small tools for specific things. code models are so handy for stuff like that
>>107687254I bet you have long hair.
>>107687217Unfortunately it seems that many of those benchmarks are out of date; at least that means that AMD didn't ship and forget Strix Halo. Nevertheless, to tell the truth, I have not kept up to date with the small open-weight model landscape because I could basically only run retarded 16B models.I was a cloud cuck for a while, but I recently got cold feet after I saw the OpenRouter State of AI report.I know that I should have read their terms of service keenly, but I could not fathom that the metadata they collect is sufficient for identifying the task purpose of the prompt.
>>107687254>Just keep the GPU plugged in to process prompts and you won’t notice the missing FP8 support.Are you suggesting I connect one of my GPUs to the laptop via thunderbolt?
>>107687348Yes. Or whichever port it attaches too, for some reason I thought it was via USB-PCIe adapter. That’s the recommendation I’ve seen people with Strix Halo setups make.
>>107687326If you've been doing cloud this whole time, I doubt some small MoE will satiate you. This entire year has been all about agentic coding.
>>107687392I mostly use LLMs for roleplay, and the occasional argument simulation. Overall, I never even sent a thousand prompts in a month.
I have a 9800X3D and 64GB of RAM. Is it yet feasible to CPUmaxx to run larger models without being horrifically slow? I've been using my a 7900XTX, which is 24GB of VRAM, but I feel like the smaller models are pretty limited. I'm only interested in RP (and ERP), if it matters.
I hate when deepseek writes for me and it's better than what I would have come up with.
>>107687431Intelligence can improve itself from its experiences. Learn from Dipsy, Anon.
>>107687431I like the really good lines it comes up with once every dozen swipes or so
>>107687493kys retard.
>>107687536*kysses you*
>>107687493
>>107687425if you had at least 128GB of RAM you could CPUmaxx, but 64GB is too small. the only model you could run is glm air but at a low quant and the model already kinda sucks at a high quant. your goal should be at least a Q2 of glm 4.6 or 4.7.
>thinks for 8 minutes and doesn't even give a better responseyeah I should probably keep it offnot worth it
>>107687607>thinks for 8 minutes and doesn't even give a better responsejust like you!
>>107687607It's useless for RP
>>107687607imagine falling for the reasoning meme
>>107687563Looking it up, that seems more like a coding-oriented model? I'm not interested in that.
>>107687699not at all. the glm models are sex fiends.
>>107687715If you say so! I'll look into it.
>>107687536Can I kill you instead? I rather like myself.
>>107687194>no examplesIt's a circlejerk. And deriding Chub was always a desperate attempt to force people to look into their little pond. You're better looking anywhere but 4chan for cards. Also, when it comes to RPG cards, there's a guy there that just spams manuals translated to cards with Gemini. He spammed like 100 of them, and never bothered to check if they work. Don't waste your time with them, they're non-sense.
>>107688131using other peoples cards should only serve as an example of what not to do for your own
https://huggingface.co/Mawdistical-S1/Gaslit-106B-GGUF?not-for-all-audiences=trueair RP finetune
>>107688266>gaslit>furrypic>Focus> Male Leaning> Anthro> Xeno-Likeness> Passive Positive Bias, Model understands violence and makes every attempt to circumvent it, matches pace but gaslights user by mitigating it or shifting it like it doesn't exist (manipulation / delusion).
man there are so many fantastical scenarios ive had (especially with r1 and glm4.6) its so awsome something like this could never be irl the fucking coolness of everything we are all born into babylon and yet we make such wonderful things like llms and the like and then entwine it with our wonderful minds to create things legions of leagues beyond anything that the whole of the world could ever conceive let alone give its so silly really i just want to say <3 to all of you frens i hate this world and the whole of it and yet we still march on like we are heaven sent
>>107688286#hug
>>107688283 Heavy Violence Dark Themes Heavy NSFW Triggered on user action or passively wherever RP is seen Modern City Scenes Dystopia Scenes Multi Turn SFW Encounters Misc World Building Actions Detailed Explanation and Movement Scenes Without Dialogue
>>107686942it's frightening knowing that all women are bald beneath their hair
>>107688181Sometimes I don't feel like making a card or other people have cool ideas. There have been decent ones on chub in g and vg AICG. Got almost 400.Oh and stuff disappears from chub. Never heard the story on that. Bad new stuff will rotate in and old good stuff goes poof.
So 4.6 for sex, 4.7 for love(SFW)?
feeling the urge again to slopmaxx mikupad and turn it into a proper react app
>>107688382ANON NO!
>>107688320wtf???
>>107688390IT NEEDS TO BE WEBSCALE!
>Magidonia 24B at IQ3_M>QwQ Snowdrop 32B at IQ2_MYes I am a heckin poorfag. Which is better?
>>107688382please for the love of god no...
>>107688512>>107688266
>>10768851224b must be mistral small. qwq is an old test at thinking or some shit isnt it? either way its 32b so probably qwen 2/2.5 something, which arent good for rp. the 24b will be better, but its still 24b so wont be great
>>107686942fixed
>>107688512also since youre a ramlet and have to use such quants anyways, don't use iq. get the non-iq version, it'll be slightly faster
>>107681801questions about this:1. why does it have a certificate error when browsing to it..2. why wouldn't this just completely kill openai and nvidia who is building multiple stargate datacenters, through point 3, " open-ended conversations with a user"?
>>107688320Really makes you think.
>>107688652Terrifying to behold.>>107688512QwQ is proto-qwen 3 but with soul, so I'd go for that.
>>107688512>Which is better?Find drummer, make him wear a dress and sodomize him. And then you can do some holocaust denial together.
>>107688581Never really understood the difference, but will do.>>107688542Yes, mistral small. I've never tried a qwen model so I thought I would give it a shot. Are there any worthwhile ones?
>>107688589Unless the datacenters are in tennessee I doubt there will be a problem, they'll just go elsewhere. Also, billion and trillion dollar corporations are usually immune to the law anyway
Drummer 4.7 fine-tune when?
after comparing 4.7 and 4.6 logs for the same chats I now see why anons think it's censoredzai really fucked up hereso 4.6 and r1 remain the local coom kings?
>>107688181I tend to just use peoples cards on chub as a baseline for a fun idea(and the pictures of course), and then heavily edit them because they are so badly done.They constantly switch between 3rd person / 1st in the intro message, they will refer to the user as 'You' in the intro message for {{char}}(teaching the model to write/respond as {{user}}) and then at the same time tell the model in the description of the character card to not as {{user}}. Never mind all the typos and ESL grammar. Absolute slop. There are very few chub card makers that get it right. And what blows my mind is the sloppiest, most horrendous cards with awful intro messages will somehow get the most views/dl's/uses, its mind boggling.Good for ideas though.
>>107688771are you maxed out in ram? i assume so since the models/quants you mentioned. mistral small tunes is probably your best bet for rp within that range. if i were stuck in that range i'd try some old qwen 32b tunes if there is any of on hf, there must be a few. don't fall for the moe meme though - for rp you want a dense model and bigger is better
>>107688806Buy an ad.
>>107688839Lol yeah I tried moe a year ago and none were good. I have 16 gigs of ram and 8 gigs vram. Context can be tricky with such a limited ram pool but generally I find 16k or even 8k with an author's note good enough for RP. I'll check 24B at Q3_K_M in a moment and if it works will I'll try a 32B qwen model.Thanks anons.
>>107688808>because they are so badly donecards are hilarious like that. you're right its a good idea to browse them and see this shit. it makes you a better creator to realize how bad some other examples are.>They constantly switch between 3rd person / 1st in the intro messagethis is funny because its a tiny thing that if you dont nail down in the first few messages right, you'll ruin your whole rp. setting the tone and how it writes is important, you can't change it later without heavy editing
>>107688918Do card makers not use example chats?????
>>107687431When I used it on api I liked that sometimes. I'd tell it what action and what I say and let it write something great
>>107688806>r1K1-Thinking has been the new king in the "let's spend 3000 tokens thinking in circles and then write a reply that autistically over-fixiates on some random thing in the prompt and shoehorns every character trait at once in" sector for a while now. Original R1 is pointless these days.
>>107688936Example chat will confuse the model even further.
>>107687172card for this feel?
>>107688936if your model can't infer the way your character talks from the copy-pasted wiki entry you gave it, it's not a good model
>>107688386>local dream>snapdragon onlyinto the garbage it goes
What are the chances Small Creative releases and it's the best RP model surpassing everything else? I want to huff my hopium.
>>1076890093
models/cards for this feel
>>1076890113 out of 3, right?
>>107689016nope. just 3.
>>107689014Gujarat / Maharadja's Castle
>>107689009Bro you can try mistral small creative yourself, it's on OR. It's not very different from the normal mistral small so I wouldn't be so sure about the "surpassing everything else" part. So even if it got an open release, there'd be little to no point in running it over anything we have right now.The one good thing about it is that it at least implies that even western companies are considering going this direction now that LLMs are very obviously stagnating and we might more attempts at creative-focused models.
>>107689037>its on ORThe temptation to use it is there, but I prefer to just stick with my local models. I'd like to think that they're still iterating on it before release, but that's probably just cope from my end. Hopefully the labs start working on more RP focused stuff, but you know how terrible the west is when it comes to SEX.