/lmg/ - a general dedicated to the discussion and development of local language models.Previous threads: >>107129334 & >>107121367►News>(11/07) Step-Audio-EditX, LLM-based TTS and audio editing model released: https://hf.co/stepfun-ai/Step-Audio-EditX>(11/06) Kimi K2 Thinking released with INT4 quantization and 256k context: https://moonshotai.github.io/Kimi-K2/thinking.html>(11/06) LocalSong 700M melodic instrumental music generation model released: https://hf.co/Localsong/LocalSong>(11/05) MegaDLMs framework for training diffusion language models released: https://github.com/JinjieNi/MegaDLMs>(11/01) LongCat-Flash-Omni 560B-A27B released: https://hf.co/meituan-longcat/LongCat-Flash-Omni►News Archive: https://rentry.org/lmg-news-archive►Glossary: https://rentry.org/lmg-glossary►Links: https://rentry.org/LocalModelsLinks►Official /lmg/ card: https://files.catbox.moe/cbclyf.png►Getting Startedhttps://rentry.org/lmg-lazy-getting-started-guidehttps://rentry.org/lmg-build-guideshttps://rentry.org/IsolatedLinuxWebServicehttps://rentry.org/recommended-modelshttps://rentry.org/samplers►Further Learninghttps://rentry.org/machine-learning-roadmaphttps://rentry.org/llm-traininghttps://rentry.org/LocalModelsPapers►BenchmarksLiveBench: https://livebench.aiProgramming: https://livecodebench.github.io/gso.htmlContext Length: https://github.com/adobe-research/NoLiMaGPUs: https://github.com/XiongjieDai/GPU-Benchmarks-on-LLM-Inference►ToolsAlpha Calculator: https://desmos.com/calculator/ffngla98ycGGUF VRAM Calculator: https://hf.co/spaces/NyxKrage/LLM-Model-VRAM-CalculatorSampler Visualizer: https://artefact2.github.io/llm-sampling►Text Gen. UI, Inference Engineshttps://github.com/lmg-anon/mikupadhttps://github.com/oobabooga/text-generation-webuihttps://github.com/LostRuins/koboldcpphttps://github.com/ggerganov/llama.cpphttps://github.com/theroyallab/tabbyAPIhttps://github.com/vllm-project/vllm
►Recent Highlights from the Previous Thread: >>107129334--Papers:>107130633--llama.cpp VRAM optimization challenges and AMD EPYC memory architecture quirks:>107132531 >107132547 >107132605 >107132615 >107132685 >107132754 >107132705 >107132740 >107132765 >107133279 >107133407 >107133585 >107133671--Budget and power challenges for a high-end workstation PC build:>107130125 >107130157 >107130181 >107132027 >107132049 >107132074 >107132080 >107132104 >107132118--Hardware performance for running GLM-4.5 models on RX 6600 XT:>107133281 >107133294 >107133328 >107133338 >107133381 >107133444 >107133460--Uncertainty over RTX 50 SUPER's 3GB GDDR7 memory availability:>107131960 >107132001 >107132894 >107133211 >107132060--Budgeting and hardware compatibility challenges for tensor parallelism prototyping:>107130539 >107130706 >107130899--Speed vs quality tradeoffs with K2 Thinking model on SSD hardware:>107136636 >107136667 >107136687 >107136699 >107136721 >107136777 >107136820 >107136885--Character.ai model architecture and commercialization challenges:>107137178 >107137277 >107137296 >107137860 >107137233 >107137275 >107137300 >107137444 >107137520 >107137724--Model discussion with NSFW and uncensored features:>107133720 >107133729 >107133752 >107133948 >107134600>107134837 >107133737--Debate over model weight formats and open weight access for finetuning:>107129703 >107129880 >107129911 >107129971 >107135655 >107135714 >107135921 >107135957 >107135992 >107137717 >107137751 >107137833 >107130017--Logs:>107130261 >107135147 >107135334 >107135409 >107135481 >107135491 >107135517 >107135792 >107135854 >107135967 >107136320 >107136332 >107136385 >107136522 >107136469 >107136808 >107136984 >107137104 >107137141 >107137735 --Miku and Luka (free space):>107129864 >107130191 >107130344 >107131403 >107131513 >107131552 >107137895►Recent Highlight Posts from the Previous Thread: >>107129340Why?: >>102478518Enable Links: https://rentry.org/lmg-recap-script
>>107138549skill issue
>>107138613Models to translate chinese text from the image to hindi?
>>107138549Because there's demand, and most people are too dumb to prompt. If a model can't talk about the smaller things in life with no system prompt and zero context then people will give up until there's a model that can, even if it's measurably stupider.
When and if the AI bubble bursts, how do you predict that it will affect local models? Do you think there a period of stagnation since the major AI companies stop development due to the crash, or do you think local models will pick up the slack and slowly iterate when the major players stop?
wholesome message from /oursaar/https://youtu.be/mdlGTMAPoz8
>>107138775local cannot progress without corporate. the difference is that when corporate AI fails, we will still be here and we will still have our models
>>107138775I think China will keep chugging along, so local will still get something.
>>107138775The only reason local models are even being made is to get investors and build general interest for the company's proprietary models. If the bubble bursts then I wouldn't expect anything but finetunes, rather than actual new model releases, unless some group goes the crowdfunding route and there's enough interest for people to pay up.That said, I don't think we'll see a pop for at least another 2-3 years, if at all.
>>107138775Imagine outcry when suddenly their AI boyfriend is shut down, it will be like GPT4 shutdown, but 100x worse.
>>107138842A pop will trash the US economy at this point, so they'll keep the charade for as long as they can
>>107138842>That said, I don't think we'll see a pop for at least another 2-3 years, if at all.LolLmao
>>107138862you can short the market if you're that confident lol
>>107138842>>107138859OpenAI plans to IPO next year. I imagine the pop will come shortly after that.
>>107138867I don't have enough money to gamble with but this guy is doing it https://uk.finance.yahoo.com/news/michael-burry-shorting-ai-stocks-092424898.html
>>107138639にんじん = carrotsじゃがいも = potatoesthird one - not sure(blank)ねぎ = it's probably たまねぎ, the regular ball onion, but with the first two characters missing, it looks like it's ねぎ, the signature green onion [hence the decision]
>>107138890this isn't hindi and you are no model
>>107138775I will short nvidia and use the money to buy up all the dirt cheap datacenter hardware to create the ultimate local model
>>107138968nvidia has buyback agreements with pretty much all the datacenters they supply. They'd rather toss their GPUs into an incenarator than let people have more than 16GB VRAM for less than $2000.
>decide to return to GLM-Z1 for nostalgia sake>On regular Z1 {{char}}: before <think> jailbreak works like a charm.>Rumination just doesn't give a fuck. If you tell it to write degenerate smut it will immediately go into a recursive thinking loop to refine the response (but it's a dumb 32B model and misses the point of the scenario entirely) We have to go back, thoughbeit. >mfw if I died an untimtely death my loved ones would stumble upon my AI lab, see what I was getting AI to write and think I was the most awful human being on the planet.This is the path to a post scarcity future that will cure all human suffering though.
uhm...which local model is the least safety slopped and good at coding so I can vibecode le epic malware?
>>107139156pygmalion 8b
>>107139156Deepseek R1.
It's been more than 5 threads and no new goof supported. I think we need to do something.
>>107139246>and no new goof supported.What do you mean? There was a zombie lobby that was being bombed by TNT, that sounds like a new goof to me.
Do zombies slip on banana peels? If so we could have another banana hell lobby with zombies included
Tried kimi-linear on OR because there's no gguf yet. And it's sloppy, it writes nothing like K2 at all, but a lot like Claude. Damn because when I begged for a 2025 mixtral I didn't mean another copy. Welp guess us 24GB vramlets will have to wait some more.
Someone explain to me why the best sampling for RP/creative is simply not this:>First sampler: minP (or Top-P if you like it better I guess)>Second sampler: Temperature>Set temperature to 5 or so>Start raising minP (lowering in case of Top-P), find the value that produces minimal to no brain damaged outputs and stop there (In my testing for minP it seemed to be around 0.3 but likely to vary a lot based on model)>You now have sampling that cuts all stupid tokens as first step and then levels out the probabilities of all remaining tokens so all of them are equally valid picks, promoting variety.
>>107138842>The only reason local models are even being made is to get investors and build general interest for the company's proprietary models. If the bubble bursts then I wouldn't expect anything but finetunes, rather than actual new model releases, unless some group goes the crowdfunding route and there's enough interest for people to pay up.>That said, I don't think we'll see a pop for at least another 2-3 years, if at all.if the AI bubble pops, openAI might be fucked, all the rando smaller orgs might get fuckedgoogle will remain very strong because they are in fact extracting profits from such models (AI search -> they get ad revenue via AI, not the target site). same for meta with whatever add AI voodoo they use to print money. One of the two may or may not then sell their AI services at a premium to fill the market need - and keep it proprietary. Heck, google's top secret model "Sajak" is already basically AGI
>>107139402Just do topK 10 temp 5.
>>107139407>Heck, google's top secret model "Sajak" is already basically AGIWhat's the story here? Tried checking but getting nothing in a quick search.
I added the extra CoT data (written by QwQ) I said I was going to add to the Gemma finetune. The result is fairly interesting.Now it's much less neurotic about its own mistakes, but still quite a lot of "you are absolutely right" slop.
>>107139418Feels like 10 tokens is too many for every situation since a bunch of time there is really only one correct token like when saying someone's name and what not but might be fun to try at least.
>>107139447>since a bunch of time there is really only one correct tokenNot really. Unless you are doing maths or outputting some sort of strict structure (json, html), or getting specific answers (yes, no, blue, red, etc), you more often than not want a healthy pool of possibilities.In the case of the name for example, the next token might be the first half of the name, or a token that initiates a preamble to getting to the name.The difference between >The guy's name? John.and>The guy's name? It's that one motherfucker man! John, the asshole.Tokens are positional, basically.
>>107139447Yeah no, even Top K of 3 causes grammar and punctuation errors at times. Has to be P to handle cases where only 1 or 2 tokens are at all reasonable.
>>107139500Maybe you are using big models that have a gentler slope from good to bad but I'm a vramlet and mistral stuff for example has a ton of cases where it drops to garbage almost immediately after the likeliest token whenever it's very sure about the top token.
>>107139402Variety does not equal goodYou can have 50 different indians shit on your plate, but that won't make you want to eat it.
make it stop aaaaaauuuughhh
Why don't llama.cpp guys provide Linux binaries with Cuda support compiled in? Windows version comes with CUDA support too, it can't be just because of some arbitrary software license.
>>107139742shaddup and compile
>>107139738Isn't it ramping up because they are transitioning to DDR6
>>107139779nope. thats still 2 years out at least
>>107139738but anon think of all the proprietary models they will train with all of that ram
>>107139540So P first, K second?
>>107139540That doesn't make sense. If the model makes mistakes at top k=3, at a higher top k mathematically it will make them more often.
>>107139897I usually just set topK to 20 and adjust temp based on how locked in the model is. If it goes very sloppy you need to confuse it.
>>107139742Because we Linux users can fend for ourselves.
>>107139924unironically this, I also compile with new cuda (llmaocpp on winshit is at 12.4) and with ZEN5 arch optimizationswintoddlers are RETARDs
>>107139792Is it? Isn't ddr6 slated to start releasing consumer models by next autumn?
>>107139924>git pull>cmake ...How hard can it be?
>>107139982no. zen 6 is still gonna be ddr5
>>107139742Because linux users are used to being third world citizens.
why compile by yourself if no new goof supported?
>>107139984to be fair I always have to check the build.md to see what the cuda on command was
I pull and recompile llama.cpp multiple times a day as an autistic stim
>>107140040You don't have terminal history autocomplete?
>>107140043Whatever gets you through the day anon, God bless you
>>107140044>cmake.. urg wat was it ctrl-r r r
>>107140044I run it on containers.
>>107140044show hist configshopt -s histappend # append don't overwriteHISTCONTROL=ignoreboth # ignore space beginning lines and duplicatesHISTSIZE=10000HISTFILESIZE=20000
shopt -s histappend # append don't overwriteHISTCONTROL=ignoreboth # ignore space beginning lines and duplicatesHISTSIZE=10000HISTFILESIZE=20000
>>107140074>cmake>press right arrow keyNot so difficult
>>107140110My 3090 gets 0 because I never bothered to download that garbage
>>107140129holy fucking based
>>107140110It's okay bro no need to be shy today you learned spoilers don't work on /g/
I want to preserve some of Opus-3 before it gets switched off in 2 months. I'm thinking I'll put like $500 on openrouter and build a dataset. I know I could get random prompts out of datasets on HF and fire them off, but that'd be shallow.What's a good way to get multi-turn out of it? The datasets I've seen doing this with another LLM writing a response don't' see that great. The "human" follow-up replies are too generic and half the conversations are discussing what Claude is allowed to discuss.
>>107140145Hello sir.
>>107140145I was thinking about distilling from closed models as well because frankly all the open datasets are all trash.The best way might be to prompt it with random segments of conversational datasets. Also there might be value to sampling with the same prompt multiple times at a high temperature to capture an approximation of the distribution rather than only the top tokens since having (or in this case estimating) the soft logits is supposed to be much better for distillation, but I'm not sure how valuable that is compared to doing one capture with different prompts.Another strategy would be to just use the model in the way you normally use it and just capture the logs. But that is obviously very time consuming.
A third alternative might be to offer free usage through a proxy while logging it and let people do the hard work of prompting it for you.But that would have to be rate limited and otherwise locked down to prevent people from trying to DDoS you and waste money.
>>107140277>while logging it and let people do the hard work of prompting it for you.You really want to train it on "ah ah mistress"?
>>107140264>I was thinking about distilling from closed models as well because frankly all the open datasets are all trash.Yeah, I noticed that. But I don't know that mine would be any better considering these would have been made by smarter people than me.If Opus-3 is one of the models you wanted, we've got until January 5th: https://docs.claude.com/en/docs/about-claude/model-deprecations>Also there might be value to sampling with the same prompt multiple times at a high temperature to capture an approximation of the distribution rather than only the top tokens since having (or in this case estimating) the soft logits is supposed to be much better for distillation, but I'm not sure how valuable that is compared to doing one capture with different prompts.Good point, I think I'll do that for at least the first turn. Even if I can't figure out how best to use them right now, at least I'll have it before the model is removed.> Another strategy would be to just use the model in the way you normally use it and just capture the logs.Yeah, I've got about 200 conversations I can export in openwebui.Still, likely going to need a lot more than this.Kimi-K2 suggested I need to use different system prompts as well.I think I'm going to need a model responding to it, but not stupidly like this:https://huggingface.co/datasets/kalomaze/Opus_Instruct_25k?conversation-viewer=2(Why does it have to say "Claude" in every reply?)
>>107140356No, I'm interested in logic, programming and reasoning, but I assumed OP wanted to distill for coom since modern models do better at "productivity" tasks.
>>107139751Retard, there is already Vulcan and CPU version for linux but not CUDA.
>>107140370ur rarted
New Cydonia is really goodv4ze is good too, but I've been getting better results from v4zd. Responses are still varied and perfectly coherent at 24K context.
>>107140360If you can gather your or somebody else's logs about the topic you care about (even for other models), you can finetune a LoRa (what base model you finetune it on doesn't really matter) to predict what the user would say given a certain assistant message.
>>107140356>>107140277Yeah we've already got an Opus-3 "ah ah mistress" dataset which I think was created that way via 4chan volunteers. The Magnum models were trained with it.
>>107140370hush nowjust compile it
>>107140380pretty sure i've read all this shit about a dozen other times.. looks pretty fuckin same to me
>>107140380Fuck, cut off the last line.New Cydonia is really goodv4ze is good too, but I've been getting better results from v4zd.Responses are still varied and perfectly coherent at 24K context.
>>107140365No, I'm interested in logic, programming and reasoning, but I assumed OP wanted to distill for coom since modern models do better at "productivity" tasks.Not to "coom", just the overall voice of the model.Opus-3 isn't very good for logic/coding (otherwise one of the Chinese labs would have distilled it and I wouldn't bother)
>>107140399Well, that's the extent of my knowledge. I searched around but there doesn't seem to be anything too fleshed out unlike the style transfer in the visual domain which is a mature ML task.
I've another idea. Maybe ask it to write an infinite choose your own adventure game with say 4 different options on each generation, and systematically explore all possible branches of the tree? I think that would be interesting in and of itself besides the Claude situation.
>>107140394It's hard to convey the value of a model in a single post, no one here is going to read thousands of words of a slop fantasy RP that devolves into smut. I really do think that it's the new best coom/creative model that can comfortably fit in 24GB VRAM.The main points I enjoy about it, compared to regular Mistral, Gemma, Qwen models ~30B and under>characters will swear when it makes sense in context, most other models will either do it in every reply, making the character seem stupid, or be too prudish to have the character swear of their own accord>swipes are varied, even at a modest temp of 0.7 (which is about the upper limit for mistral small, before it starts getting noticeably dumber>doesn't speak for user particularly often, a problem I've had with other recent Cydonias>relationships and sex are effectively built up slowly, e.g. characters will flirt, a day can pass without further mention, and they'll recall it and continue the next day, ~2-3k tokens later.
what am I in for?
>>107140501safety cuckery even if you don't goon
saaaar do not redeem the chain of thoughthttps://www.youtube.com/watch?v=IeCS6hsnOXs
>>107140392fuck off avatarfag.
>>107140501>we must refuseit's a decent model for sfw tasks & tool calling, runs fast
It seems that adding the QwQ data to the dataset made the model much more sensitive to overfitting, even though the validation loss kept going down, now I had to decrease the lr from 1e-05 to 1e-06 because the CoT data made it primed to get stuck in repetition loops. I think it's probably because of the repetition inherent in CoT models.
>>107138890they're all ingredients for butchered """curry"" so I'm assuming just curry powderchecked a dictionary and it's curry roux
>>107138775>When and ifThere are hundreds of promising research ideas yet to be tested at scale. LLMs are already impacting the job market and only continue to improve. Not gonna be a sudden a-ha! thing where the world changes overnight, just bumpy steadily better until everyone's left wondering where the jobs are and the civil unrest picks up because UBI ain't happening
>>107140380Every model feels sloppy compared to Cydonia in its size range desuI still try other models people recommend but they majorly suck ass
>>107140501toss for the smarts, air for the dick
>>107140689no toss on the 'ick?
>>107140689>>107140692go back to the sharty you homo faggots
>>107140689>we must refuse>"the dick?" Air echoesboth are awful
>>107140734so are every other llms, but for 100b those two are the only decent options
>>107140741>so are every other
Ok, I changed my data mix to the following:my own assistant logs x 4openthoughts x 2qwq cot x 1x n being the number of times the data is duplicated in the dataset, i.e. a hacky way of having a different number of epochs for each data class while still randomly shuffling the samplesnot sure how it'll work
next I wanna try adding some RP data to see if it helps with coherency, and also check out the openassistant dataset
after that it might be time to begin testing the waters with rlvr
oh, and also some data augmentation although I'm not sure how that works on text only tried it with images
>>107138606At least in Germany the supposed $500 MSRP for the Intel Arc B60 has so far not materialized, at 770 € they're I think just not worth buying.Is the pricing in other regions at least better?
>>107140748holy zased
>>107140749I just realized this causes us to train on the validation set. Fuck. Oh well, the validation split didn't seem to be very useful anyway.
>>107140832tech MSRPs are just marketing material, they're complete fiction.
>>107140853I figure I will have to make an explicit manual split beforehand and then it'll be alright
it took like half an hour to get huggingface hub installed, due to GPT5 hallucinations and microsoft store fuckery. this is not a serious industry
According to this talk the way the Llamas were trained was by taking a validation set from the complete training set and then changing the weight of each dataset based on how much it affected this validation set. They claim this is an open problem. I think it can be fairly easily explained as some of the data being overrepresented in the training set and causing overfitting if not downsampled. https://www.youtube.com/watch?v=-TIZPe_YaiU
>>107140877skill issue
>>107140689>toss>smartslmao
whenever I feel cold I just crank up my 3090's power limit
>>107140877If you need GPT5 to install a fucking program I think this hobby might not be for you
>>107140894I'm not going to watch a random 1-hour video, but it's common knowledge that most AI companies optimize their pretraining dataset mixtures for synthetic benchmarks, other than "safety".
>>107140877what do you need huggingface_hub for?
>>107140921Does validation loss count as a synthetic benchmark though? It's about how accurately it predicts the pretrain dataset.As for the video, the claim happens at about the 15 minute mark, but the channel is one of the best channels I found when it comes to ML theory.And people say there is nothing worth watching on youtube.
toss bros?
>>107140928>what do you need huggingface_hub for?i want to use the extropic thrml simulator to do topk prefiltering, but i need to rip the gptoss embeddings first so i can shuffle them around so each axis of the embedding fits into the 2d thrml array meaningfully
>>107140956lol
>>107140956>We must not reveal we are cucked.
>>107140921lol they have entire teams for safetycucking
>>107140956weird that an optimized model devotes thinking tokens to "so"
>>107141012he was talking about pretraining, has nothing to do with safety
>>107141030Curtailing the corpuses (corpii?) used in pretraining is very much a job for the safety team
>>107141086ok, fair. but do you have any reason to believe that optimizing the validation loss on a subset of the complete unfiltered corpus would correlate in any way with safety? because the claim on the video was about optimizing the validation loss