/lmg/ - a general dedicated to the discussion and development of local language models.Previous threads: >>102987959 & >>102976869►News>(10/25) GLM-4-Voice: End-to-end speech and text model based on GLM-4-9B: https://hf.co/THUDM/glm-4-voice-9b>(10/24) Aya Expanse released with 23 supported languages: https://hf.co/CohereForAI/aya-expanse-32b>(10/22) genmoai-smol allows video inference on 24 GB RAM: https://github.com/victorchall/genmoai-smol>(10/22) Mochi-1: 10B Asymmetric Diffusion Transformer text-to-video model: https://hf.co/genmo/mochi-1-preview>(10/22) Pangea: Open-source multilingual multimodal LLM supporting 39 languages: https://neulab.github.io/Pangea►News Archive: https://rentry.org/lmg-news-archive►Glossary: https://rentry.org/lmg-glossary►Links: https://rentry.org/LocalModelsLinks►Official /lmg/ card: https://files.catbox.moe/cbclyf.png►Getting Startedhttps://rentry.org/lmg-lazy-getting-started-guidehttps://rentry.org/lmg-build-guideshttps://rentry.org/IsolatedLinuxWebService►Further Learninghttps://rentry.org/machine-learning-roadmaphttps://rentry.org/llm-traininghttps://rentry.org/LocalModelsPapers►BenchmarksChatbot Arena: https://chat.lmsys.org/?leaderboardCensorship: https://hf.co/spaces/DontPlanToEnd/UGI-LeaderboardCensorbench: https://codeberg.org/jts2323/censorbenchJapanese: https://hf.co/datasets/lmg-anon/vntl-leaderboardProgramming: https://livecodebench.github.io/leaderboard.html►ToolsAlpha Calculator: https://desmos.com/calculator/ffngla98ycGGUF VRAM Calculator: https://hf.co/spaces/NyxKrage/LLM-Model-VRAM-CalculatorSampler visualizer: https://artefact2.github.io/llm-sampling►Text Gen. UI, Inference Engineshttps://github.com/oobabooga/text-generation-webuihttps://github.com/LostRuins/koboldcpphttps://github.com/lmg-anon/mikupadhttps://github.com/turboderp/exuihttps://github.com/ggerganov/llama.cpp
►Recent Highlights from the Previous Thread: >>102987959--Multiple perspectives feature reduces bias scores across social dimensions:>102987976 >102988012 >102990681 >102988100 >102988132 >102988543 >102989070--LLM RAM Calculator helps estimate memory requirements:>102995865--Introduction to LLM sampling:>102990514 >102990633--Transformers.js v3 supports WebGPU and ONNX runtime models:>102989640--Open source models struggle with self-correction and output quality:>102993033 >102993349 >102994098 >102994218 >102994315 >102994381 >102994615 >102994913 >102995031 >102994946--Meta publishes open source music model but deletes weights:>102989563--Gpt-soviets trained on moe-speech sounds better than Tomoko:>102988359 >102988410 >102992236--Advice on using Koboldcpp to test different AI models and quantizations:>102988564 >102988717 >102988765 >102988805 >102988832 >102988754 >102988796 >102988844 >102988853 >102988849--iGPU and APU performance for inferencing compared to discrete GPUs:>102990238 >102990310--Tips for generating smut stories:>102989833 >102990006 >102990313 >102993566--Newfag guide to AI models and terminology:>102996361 >102996652 >102996683 >102996713 >102996779 >102997428 >102997550--Mistral, GPT-SOVITT, and improved models anticipated:>102997159 >102997173 >102997176 >102997185 >102997213 >102997203--LLM ERP capabilities poll results and discussion:>102993026 >102993613 >102993659 >102993982 >102993744 >102993800 >102994067 >102994113 >102994486--INTELLECT-1 training progress update:>102987982--GLM-4-Voice sounds great in Chinese, but may be cherry-picked:>102993857--AI censorship discussion and risks of different modalities:>102988230 >102988262 >102988351 >102988361 >102988387--Miku (free space):>102988359 >102989044 >102989186 >102991784 >102992608 >102996542 >102997941►Recent Highlight Posts from the Previous Thread: >>102989254Why?: 9 reply limit >>102478518Fix: https://rentry.org/lmg-recap-script
Where did the rentry go
>Flexora auto-selects which LLM layers to fine-tune, cutting training costs.>Average accuracy improvement: +7.21% on Llama3-8B, +8.33% on ChatGLM3-6B, +1.98% on Mistral-7B-v0.1https://arxiv.org/abs/2408.10774https://x.com/rohanpaul_ai/status/1850673624384168224
>>102998216>the real OP doesn't want to put it inIt's over...
>>102998257I guess I'll just post the latest one here: http://rentry.org/pcrkt9pa
Sovits server API is really a piece of garbage. You couldn't even send the reference audio over the request, you need to have it already on the server and provide the path to it. Patched that and got the inference working on a server CPU at last, miss me with that gradio shit.
COAT: Compressing Optimizer states and Activation for Memory-Efficient FP8 Traininghttps://arxiv.org/abs/2410.19313>FP8 training has emerged as a promising method for improving training efficiency. Existing frameworks accelerate training by applying FP8 computation to linear layers while leaving optimizer states and activations in higher precision, which fails to fully optimize memory usage. This paper introduces COAT (Compressing Optimizer States and Activations for FP8 Training), a novel FP8 training framework designed to significantly reduce memory footprint when training large models. COAT addresses current limitations through two key innovations: (1) Dynamic Range Expansion, which aligns optimizer state distributions more closely with the FP8 representation range, thereby reducing quantization error, and (2) Mixed-Granularity Activation Quantization, which optimizes activation memory using a combination of per-tensor and per-group quantization strategies. Experiments demonstrate that COAT effectively reduces end-to-end training memory footprint by 1.54x compared to BF16 while achieving nearly lossless performance across various tasks, such as Large Language Model pretraining and fine-tuning and Vision Language Model training. COAT also achieves a 1.43x end-to-end training speedup compared to BF16, performing on par with or surpassing TransformerEngine's speedup. COAT enables efficient full-parameter training of large models on fewer GPUs, and facilitates doubling the batch size in distributed training settings, providing a practical solution for scaling large-scale model training. https://github.com/NVlabs/COATGit isn't live yet. another step towards FP8 training runs becoming more common
kind of crazy to think about how ai is a solved science and with a couple more gens of nvidia chips and a few years of datacenter and power infra expanding we will be able to just use the current algorithms to create agi
DreamCraft3D++: Efficient Hierarchical 3D Generation with Multi-Plane Reconstruction Modelhttps://arxiv.org/abs/2410.12928>We introduce DreamCraft3D++, an extension of DreamCraft3D that enables efficient high-quality generation of complex 3D assets. DreamCraft3D++ inherits the multi-stage generation process of DreamCraft3D, but replaces the time-consuming geometry sculpting optimization with a feed-forward multi-plane based reconstruction model, speeding up the process by 1000x. For texture refinement, we propose a training-free IP-Adapter module that is conditioned on the enhanced multi-view images to enhance texture and geometry consistency, providing a 4x faster alternative to DreamCraft3D's DreamBooth fine-tuning. Experiments on diverse datasets demonstrate DreamCraft3D++'s ability to generate creative 3D assets with intricate geometry and realistic 360° textures, outperforming state-of-the-art image-to-3D methods in quality and speed. The full implementation will be open-sourced to enable new possibilities in 3D content creation.https://dreamcraft3dplus.github.iohttps://github.com/MrTornado24/DreamCraft3D_Plusbig improvement over the original in how long it takes to generate a model (from 3 hours to 10 minutes)
>>102998360I wonder why they're not exploring FP2 to FP12 models like Aphrodite engine did.
>>102998364
>>102998427A discord chat / reddit meme.
>>102998171Based.
>>102998444
INTELLECT-1 is at 29.50% complete, up from 27.60% last thread.
>>102998415>exploring FP2 to FP12 models like Aphrodite engine did>like Aphrodite engine didThis project really likes to copy other people's code and take credit for it.
>>102998559Nah retard he really did it himself
>>102998564hi Alpin
>>102998596cope
>>102998616hi cope
>>102998483>perplexity plateauing at 6.75grim
>>102998564>this file is copied fromhttps://github.com/vllm-project/vllm/pull/8751/files#diffLiterally most files are taken from this Github and the DeepSpeed library:https://github.com/usyd-fsalab/fp6_llm>We propose TC-FPx, the first full-stack GPU system design scheme with unified Tensor Core support of float-point weights for various quantization bit-width (6-bit, 5-bit, 3-bit, etc.)>FP6-LLM is already integrated in DeepSpeedBut suddenly Aphrodite takes all the credit. What a piece of shit.
>>102998623Not that I think this is going to be any good but the perfect coombot shouldn't have best perplexity cause that would lead to zero variety + lots of slop.
>we made it easy to support different quantizations>suddenly Aphrodite takes credit of the entire research
>>102998635>>102998674What the point of that research when it's not used pratically?
>>102998216>>102998257>>102998265>whole thread works together to make something for the benefit of the whole general>discord OP doesn't go with itTale as old as time
Newfag here. Seems like most people running local are using it to generate images which I don't give a shit about.Does anyone run one locally for code generation or analysis? Trained from stackoverflow or a codebase or something?
>>102998265I saved it for myself. OP is a fag as always
>>102998716>Does anyone run one locally for code generation or analysis?Yes. Deepseek is great for code and logic type work if you have the resources to run it at a high quant. coder 2.5 q8 is my daily
>>102998705But dare you bake a thread without miku pic.. be prepared for extreme shitstirring.
>>102998703>not used praticallyIt's already in the deepspeed library, and vLLM used it through the library for the FP6 support. Copying the files and adding the other combinations doesn't give you the right to take credit for everything, asshole.Go fuck yourself, grifter.
>>102998756you're just jealous
>>102998705>Whole threadNo it is just you newfags. Can you make /local c.ai exodus general/ please and fuck off?
>>102998747I wish mikufaggot prime would dox some newfags.
>>102998811>making a list of recommended models together to improve the general le bad and newfag... because it just is ok??
>>102998811This general is literally a offspring of aicg, stop denying the roots
>>102998820It is because oldfags know that finetunes just make models retarded or at best change style slightly. There is no objective best style. And if you are an oldfag you just download new thing and check it out yourself very quickly. By new thing i mean base models or instructs of course.
>>102998838>because oldfags know that finetunes just make models retarded or at best change style slightly.retards you mean, or if by oldfag you mean someone who only ever used undi models
>>102998834The roots of my country have nothing to do with me not wanting some filthy migrants.
>>102998838Here, try this one side by side with regular qwen2.5. Then tell me if its dumber / not any different:https://huggingface.co/EVA-UNIT-01/EVA-Qwen2.5-72B-v0.0
>>102998845Hi drummer. Buy an ad.
>>102998838The concept was to make the model table bigger and separated into best usecases in the future, and more importantly to give the average /lmg/ browser just something good to just use instead of constantly switching modelsWhile the rentry is still barebones, it went from crap month-old placeholders to 6 good model suggestions in just one threadI really don't get the seethe and hate at the concept, both from half the thread and OP apparently
>>102998739What do you use it for typically?Can you tell it to do something like:>give me a function that does x
>>102998867All the more reason for you to make your own refugee camp.
>>102998874>my seethe at a rentry has no footing so I'll just tell you to leave
>>102998870>give me a function that does xYou can use it for that. Also: you can ask for a full program, you can copy a program into context (if you have enough) and query about it or get it to make improvements or explain things, you can pair-program with it starting from ideation right to working product including build-chain/makefiles, you can copy the schema into context and get it to generate complex sql queries with a high degree of sophistication...for just a few ideas of what you could use it for.It is very good at instruction following in my experience. Better than largestral and a bit worse that 405b.
>>102998927>you can pair-program with it starting from ideation right to working product including build-chain/makefilesInsane.Ever use it to write tests? Or have it generate code from the tests you write?Also does your setup you run it with a GPU? Or can you do CPU only? My proxmox server only has an iGPU so I would assume it's dogshit, but worth an ask
I want to branch out a bit more for the culture benchmark. What memes do you want your LLM to know about? I've already included the iToddlers BTFO meme.
>>102998999Include ebussy
>>102998867What part of buy a fucking ad is hard to understand?
>>102999135dont feed the trolls
>>102999149Upvoted!
>>102999135You created the Rentry to make money with the idea of selling the spots in it.The existence of the Rentry will also make the thread be astroturfed to hell and back, everyone will keep spamming their models to appear as "organic word of mouth of the thread" to be put in the Rentry. This is a tactic that has been used before by Sao.
>>>102999135>You created the Rentry to make money with the idea of selling the spots in it.>The existence of the Rentry will also make the thread be astroturfed to hell and back, everyone will keep spamming their models to appear as "organic word of mouth of the thread" to be put in the Rentry. This is a tactic that has been used before by Sao.
>>102999179Go back to sharty, incel.
>>102999186after you take your meds
been out of touch for a yearwhat's the best uncensored model around 7B?
I saw people using multiple graphics cards to get more of the GGUF model on VRAM. Adding another 8GB VRAM card didn't scale as expected for me. With one card, I offloaded 15 layers (~6.5GB) to VRAM. Using two cards with Koboldcpp, I offloaded rougly around 24 layers (4.5GB each), but beyond that, I get OOM. It's not going to scale at 100%, right?
>>102998171anybody got xtts2 running on win10?
Small update on proofing of concepts for RPG Maker MV based LLM front end. I figured out how to put the LLM response into an in-game message box now. (you also have to manually code in your own word wrap function because for some reason it can't do that shit automatically) Thankfully it uses unispace text so it's pretty easy to deal with. Although justifying the text so that it doesn't look like shit is a whole other animal It's also really janky in that it only allows a maximum of 12 lines per script, so you basically have to minify everything.But I figure the way to do it would be to just make a big library of bite sized functions for everything and then use event calls in the games own built in event system to call them up.
>>102999419yeah
>>102998739>>102998927Are you using this? https://github.com/QwenLM/Qwen2.5-CoderFrom what I can tell DeepSeek is cloud only? Can you run it locally?
>>102998971>Also does your setup you run it with a GPU? Or can you do CPU only?>My proxmox server only has an iGPU so I would assume it's dogshit, but worth an askI only use my GPU for context processing and infer on my cpu. You need a lot of memory and decently fast ddr5 to have a chance at using big models effectively.
>>102999509>Are you using this? https://github.com/QwenLM/Qwen2.5-Coderno, I've never had much luck with Qwen and haven't seen any convincing results from any of their recent models>From what I can tell DeepSeek is cloud only? Can you run it locally?Yes you can run it locally if you have the memory, and since its an MoE, the inference performance is relatively good for the results you get using cpu (3x the speed compared to largestral)
>>102999463no
>>102999419>>102999463>>102999734Why are you like this?
>>102998855NTA but I did. Getting outbidded hard though.
>>102999751stop samfagging
something went wrong with anthracite's datasets after v2for all model bases, the v2 version is always smarter than the v3/v4 version
You guys aren't ready for what's coming
if they were genuinely close to AGI then people wouldn't be quitting, even if they were scared, because they'd want to be a part of it and maintain influence over itso the quitting doomers are just engaging in the usual EA/rationalist speculative retardation, AGI isn't soon
>>102998705>Discord You can tell it's true just by looking at certain posters ITT.
Good night /lmg/
RIP my catto. He literally just passed away.
>>103000026condolences
>>103000042It's too late at night/early in the AM to arrange for a cremation right now. Is it weird that I feel too paranoid and weird about putting him on ice before rigor fully sets in?
>>103000026>toxoplasma gondii incubator diedOh no, anyway.
>>102999437Not sure what you are trying to achieve anon. Hate to break it to you but isnt that kinda useless?Why would i run a rpg maker game where I then have inference from my typing something. Did you think that through properly?Rpgmaker would be perfect if you can show a llm the tileset and make it create a map. Still lots of fuckery for the json of the map.The new sonnet 3.5 can do it KINDA. First llm that i could show a picture of the tileset and say tell me the x/y where i should place it. make a diverse map.So currently this stuff fails at the start. With no maps no events. Premade maps are boring.
>>103000085>Why would you do something difficult when you can do something easy?Because I have a dick.
>>103000100I think you misunderstood me. Why would anybody use this?The reason rpgmaker would be great is for creating a premade (coherent) story thats connected.Why would you run inference through rpgmaker lolThis is like the guy who made npcs in skyrim talk through gpt. Its a gimmick and nothing more.
>>103000085nta, but...wut?Just so I'm clear: you're proposing that one should use an LLM as a way to do things classical algorithms already do better and berating anon for using LLMs for convos/text gen because that's "kinda useless"?
>>102999991Neat! I'd love to have that for my VR waifu.
>>103000116Because a few weeks ago I jokingly asked about the viability of creating an inferencing front end entirely in RPG Maker MV (back at the height of the SillyTavern drama). And now I'm going to make it a reality. Because I fucking can.Not because it's a good idea.Not because it's the best practice.But to make people like you seethe about its very existence.
How are translation models? I'm partially interested in translating shitty Japanese h-games.Accuracy is whatever so long as it can form coherent sentences and do it in a timely manner.
>>103000125Why would I run inference through rpgmaker? That doesnt even make any sense.How will the story be created for example? On the fly? Does NPC A know what NPC B said?Is it just a premade map and then just let the llm go from there? Just use sillytavern and RP. Why the need for rpgmaker is what I am asking. What I am saying is it that rpgmaker would be perfect to give the llm a prompt and it will make maps, events, etc. Only 1 map with a bit of dialogue would be really cool. Create images with flux to spice shit up.But currently it already fails at the start with the map unfortunately.LLMs are terrible with this. Sonnet 3.5 got a huge improvement but still nowhere near enough.>>103000137If its something for yourself and you fuck around I don't care, but you post updates for other people.Obviously I assumed you wanted people to use whatever you make.>>103000165You would need to look at f95. There exists some tool that already uses chatgpt for translation. Its pretty good but the slop smells. If you point to local you could translate I suppose.Keeps important stuff in context already etc. I thought gemma27b is good at translating for a smaller model but its cucked unfortunately. Not sure how well it will take ero content. Probably write in a style that not hot at all.
>>103000165> How are translation models?That reminds me…did lcpp get plamo 100b support?
>>103000181You are mega gay bro
>>103000193So are you for replying anon.
>>102999425pls respond. Why can I use more vram for a single card but less vram per card when I go multi card? The language I see everyone use is "I have 32gb vram (8gb+24gb)" as if it stacks up. Do I need to run llama.cpp in cli for that kind of efficiency?
>103000198Such a great way to own yourself lfmao
are used 3090s still the best now?
>>103000198>>103000193I'm actually gay and give you both the seal of utter faggotry.
>>103000207I don't know. I have 4 but I'd be weary about buying more.They're still the least overpriced relative to what they offer but their price should have declined a lot more by now than it has. They're getting rather old now which means long term reliability concerns beyond ebays 30 day money back policy.
>>103000236>I'm actually gayNo need to make this more cringe than it already is.
>>103000000
>>103000165Best is still the usual suspects which are the online models. Mistral and Nemotron are on top for local which shouldn't be surprising but what is surprising is that Gemma is surprisingly good for its size and punch above their size with the 9B and 27B models many times larger.
>>103000247Until AMD becomes more competitive with ROCm, 3090s aren't losing more value anytime soon. I expect it to slowly atrophy until AMD or Intel has a solid performance GPU that has 24GB that is cheaper and runs AI faster than 3090s.
>>103000322Isn't AMD desperately trying to become competitive in the AI training department.
svelk
>>103000335on the enterprise level. None of them give a shit about hobbyist AIMi300X is a very popular option for doing full finetunes. Doesn't quite match an H100 in terms of compute but 192 gigs of VRAM is 192 gigs of VRAM. And full fine tuning is quicker than lora training.
>>103000247>>103000322There's currently no competition in that niche, and NVidia has already cut production of the 4090. Those expecting the release of the 5090 to lower prices on the 3090 are delusional.
I can imagine that Huang cannot sleep at night, haunted by the memory of his greatest mistake: releasing the RTX 3090 with its excessive VRAM.
Than god there are no women in /lmg/ that the refugees could rape.
>>102998171
>>103000026:(
>>102999761Kek. What's the point of buying an ad for models you can download for free?
>>102998999At least all the ancient memes from here. If they did their job properly it should have been included in the dataset
>>103000484Because he has a crowd funding link on his HF profile. But he bought an ad. So he's a legend.Unlike the ones who inorganically shill their shit here all day long and don't buy an ad.
>>103000181You can actually do better by including function calling which is supported by llama3.1/3.2 https://www.llama.com/docs/model-cards-and-prompt-formats/llama3_1/#user-defined-custom-tool-calling it's supported in vllm (not in llama.cpp afaik) https://docs.vllm.ai/en/latest/serving/openai_compatible_server.html#tool-calling-in-the-chat-completion-api
Is there an extension or something that works with sillytavern's and SD forge that allows the local model's output to be parsed and turned into a pony-friendly output that uses booru tags and pastes them onto forgeUI? I dont know how to make extensions and I have zero coding experience and just recently learned how to even write python shit. I wanted to use o1 to hold my hand through the entire process of making my own extension if nothing like that exists.
will anon buy 50series?
>>103000785Sure, when 70s come out I'll buy a 5060.
I am at the cutting edge of AI baneposting.
>>102999425>>103000200Generally speaking it is better to have the VRAM on a single GPU vs. spread out over multiple GPUs but the increase in the number of layers that you can fit should be roughly linear with total VRAM.If you use the exact same model and settings for both cases (context size and KV cache quantization most relevant) and there are no other applications consuming relevant amounts of VRAM that should not be happening.
>>103000785The value of GPUs for AI is primarily determined by their VRAM-to-price ratio, along with reasonable performance and adequate support. I do not expect the 50 series to fare well in these metrics.
good morning /lmg/. are we still at the peak of newfag infestation wave?
>>103001074One(1) "backend confused" newfren made y'all shit your pants for 20+ hours, think about it.
>>103000785yes
>>103001105you need to go back to c.ai. all of you
>>103000207https://www.digitaltrends.com/computing/its-time-to-bid-farewell-to-nvidia-rtx-30/
>>103001131"I talk to chatbots, but LOCALLY" is not a hobby worth gatekeeping.
>>103000903Nice tinyllama you got here
>>103001148yes it isyou're not wanted hereget the fuck out
>>103001148Don't forget limited context, too.
>>103001107Restraining the Pochiface in a closet until it dies from dehydration.
>>103001193What the fuck are you talking about, schizo?
>>102998171Fall into something new
>>103000932Thanks for acknowledging that vram usage scale near linear. It should be similar with either gguf or exl2, right?With that, I can focus on my settings and have more confidence getting another card.
>>103001393>It should be similar with either gguf or exl2, right?I'm not particularly knowledgeable when it comes to the internal workings of ExLlama but I don't see why it would be different.One thing that could as of right now make a difference with llama.cpp and derivatives: If you set --split-mode row the scaling will not be linear because as of right now the KV cache is only on a single GPU.
>>103001334Fall into a puddle of Leaku pee
>>103001699>dystopian robot referenceBased
>>103001699>not a cardaw
How CPU affects on performance (in general) in GPU inference and GPU training?I only found that link for inference where saying that it have almost zero effect:https://www.pugetsystems.com/labs/articles/effects-of-cpu-speed-on-gpu-inference-in-llama-cpp/And for training I see that if CPU have doesn't enough cores (at least 4 core per 1 GPU) and memory channels it would bottleneck training.How would affect GPU training for example difference between 2 x AMD EPYC 7763 64-Core and 2 x AMD EPYC 9965 192-Core?In the raw benchmark it is more than twice faster:https://openbenchmarking.orgDoes anon knows any other benchmarks/links?
>>103001952Thanks anon, I really appreciate.
What's the current best poorfag CPU model?I want something to talk to while waiting for a bitnet model to come out.
>>103002035Mistral Nemo 12b or one of its finetunes.If you want really fast, olmoe, but it has low context.But i'd still buy even a 16gb gpu if i were you... just in case...
>>103002035That thing doesn't even look like Miku, wtf are they smoking to pretend it's her?
>>103002097
>>103002035pyg6b
>>103002127Okay zoomer
>>102999437That's a really cool experiment anon. See how you can take the concept.>It's also really janky in that it only allows a maximum of 12 lines per script, so you basically have to minify everything.What the fuck.Can you at least break things into multiple script files?
>>103002097>That thing doesn't even look like Miku, wtf are they smoking to pretend it's her?
https://github.com/ggerganov/llama.cpp/pull/9702>added implementation of DRY sampler (post-refactor) #9702Was merged in on Friday.
>>103002205yeah, you can chop everything up into function definitions and function calls and attach them to common events as long as they are attached to the window object, otherwise it doesn't treat the scene as a contiguous environment.
>>103000181not that anon or any other involved in this conversation or project yet, but i am excited about anon's research.it would be cool as heck to be able to have a conversation with my LLM by walking around and interacting with an SNES looking map.sure, an RP experience would better just using a text to text interface. but think about it the other way around, this could embetter rpgmaker experiences.
I asked in a prior thread about llama-server's behavior when receiving a prompt larger than the configured context size : >>102991521>Back in the day llama.cpp server would crash if you tried to stuff a prompt larger than the configured context sized into it, it no longer does that.>Is it safe to assume that it's simply cropping the context at the top?>Is there a reason one would want to do that instead of just setting the correct prompt size in the frontend software?
Can someone explain to a vramlet why everyone is offloading the kv cache in Kcpp? Is there some downside to keeping it in RAM?
>>103002629it's slower
>>103002629Yes.Any context processing becomes dog fucking slow.With flash attention and quanted cache, putting the kv cache on RAM to fit one, maybe two more of the models layer's on VRAM is a bad tradeoff.
>>103002654Both prompt processing and inference, or just the prompt processing? Cause I seem to get better inference speeds with offloading more layers.
>>103001809https://rentry.org/miqumaxx
Why are you guys running locally when you can get free Llama 3 405B API key straight from Openrouter?Genuine question.
>>103002587>Is it safe to assume that it's simply cropping the context at the top?Even if it parses the whole prompt, whatever falls out of the context is just gone. Whatever method it uses (whether it's directly cropping or shifting the context) i don't think it makes much of a difference.>Is there a reason one would want to do that instead of just setting the correct prompt size in the frontend software?Do what exactly? The frontend doesn't necessarily know how long (in tokens) your input is nor how much of the context is used without the backend telling it. As far as i know they just send the text raw-ish (adding chat format tokens and all the extra things ST does, for example) and just let the backend tokenize it and process it. So even if you set the context length in the frontend, i don't think they can do much with that information. Maybe they guesstimate how many tokens there are, but it's never going to be precise. Unless they actually tokenize the user input, but still. I don't think they do.
>>103002817To remove the dependency on someone else's stuff.And i'll just wait for the next fuck to ask again next thread.
>>103002364based. i'm going back to local again.
>>103002817Nothing is free.
>>103002818>Maybe they guesstimate how many tokens there are,At least as far as ST is concerned, it calls an API to tokenize the text to know how many messages it can fit.Meaning that ST actually knows how many tokens are in the prompt it's sending no matter the model the backend is serving, as long as the server has a tokenization API, that is. Otherwise it just guestimates using a default tokenizer on the frontend itself.See >https://docs.sillytavern.app/usage/core-concepts/advancedformatting/#tokenizer
>>103002847Yeah but you can just make your stack or do whatever you need and replace it with your own locally ran shit at a later date instead. It's not like it's a proprietary model you're getting dependent on.>>103002851It's free for now and not proprietary which should be good enough, no?https://openrouter.ai/nousresearch/hermes-3-llama-3.1-405b:free/apihttps://openrouter.ai/meta-llama/llama-3.1-405b-instruct:free
>>103002817>>103002886Interesting. What's the catch here? Aside from the data collection of course.
>>103002817>Why are you guys running locally when you can get free Llama 3 405B API key straight from Openrouter?Because I can run 405b locally.I'll let someone else explain why giving others your logs for free in exchange for access is actually a bad deal.
>>103002886>Yeah but you can just make your stack or do whatever you need and replace it with your own locally ran shit at a later date instead. It's not like it's a proprietary model you're getting dependent on.I like running models on my own hardware, without the need for internet or depending on others. Even if it keeps working forever, i'd still depend on them. If it worked just *once* on my own setup, it will always work.
>>103002903They actually say they don't collect your data on the free API. The "catch" is to make developers make apps around openrouter APIs so that businesses in 3-5 years time are locked into their ecosystems.They don't gather data precisely because they want businesses to use their service, businesses would immediately forgo them if they collected their data and thus trade secrets.Which is precisely why I wonder why people in /lmg/ use their own personal hosted stuff instead of free non-proprietary stuff that doesn't gather your data at all and will be guaranteed available for free until 2026.
>>103002629>>103002654>>103002664>>103002674So after some testing, offloading seems better with longer prompts. The effect on processing speed seems minuscule? This is with Radeon 7800xt 16GB, so maybe it's just some ROCM thing.>"lowvram + fa" 40/57 layers10k contextProcessingSpeed: 135.65T/sGenerationSpeed: 3.94T/szero contextProcessingSpeed: 144.31T/sGenerationSpeed: 6.60T/s>"fa" 46/57 layers10k contextProcessingSpeed: 133.35T/sGenerationSpeed: 3.31T/szero contextProcessingSpeed: 161.32T/sGenerationSpeed: 7.51T/s
What's the best local model for programming?chatGPT sucks, Claude has pathetic rate limits
>>103002916>giving others your logsThey explicitly state they don't keep logs because they want to attract enterprise businesses and developers with their free API access.>Because I can run 405b locallyUnless you also have free electricity it still makes more sense to use their free API, no?I'm legitimately trying to figure out why people here would prefer locally run over a free API that doesn't gather data.
>>103002952e: lowvram is the second one, sorry
>>103002934>They actually say they don't collect your data on the free APIcool story
>>103002953Never tried one for programming, but i've seen anons recommending deepseek and qwen2.5. Depends on your hardware i suppose.
>>103002934Well, I for one had no idea.I'll take a look at it and see if it's worth the switch.Having my shit locally and offline is really nice, regardless.I'm a tinkerer, so there's a certain pleasure in having to fiddle with models and launch parameters and such.Also, fine tuning my own small models.But for local productivity, I guess there's not much reason not to use such large model for free.>>103002952>>103002964If you benchmarked it and it works better for you, than awesome.
>>103002958>free API that **CLAIMS** doesn't gather data.
>>103002953I was using deepseek through their website and it's really, really nice.
>>103002987They are trying to compete with OpenAI, Anthropic and Google on the enterprise API front. All of those other providers state to not gather your data or use your data through the API portal. It's in their best interest to not do so as it would scare away the very enterprise demographic they want to attract with this.It would also just be a lawsuit waiting to happen, bankrupting them in one go if just a single developer/company using their API finds out about it.
>>103003019It's in their best interest to not do so, they don't have an incentive to lie as it would scare away the companies and developers they try to attract with this.
>>103003035>you saved the data. i'll sue you>we dont. but if you can show we save data, you'll have to show it in court>nevermind, then. how much for 100m extra tokens?
>>103002934>y tho?I also own my own domain, run my own dns and mail, host my own websites and private "cloud" apps and manage my own offsite backups.This is /lmg/. We don't need a reason, but if we did you might want to check the news for the various rugpulls starting to happen (speedrunning to the point of normie suicide already). Those won't slow down once people become dependent on proprietary cloud stuff.Self-determination pays for itself often enough that it shouldn't be dismissed out of hand.
>>103003035That means they have an incentive to not get caught.Truth is you can't really know until it's tested, just like with all those "zero log vpns" able to help identify people when the 3 letter agencies come knocking.I'll acknowledge the other side too, that, just how like there are true zero logs vpns out there, openrouter could be telling the truth, but we can't really know either way until it's tested somehow.
>>103000298>Big 3 are all around the same level and peaked with gpt 4 turboIt is so over for llms
>>103003061>I also own my own domain, run my own dns and mail, host my own websites and private "cloud" apps and manage my own offsite backups.I do all of those things as well. I'm on /lmg/ posting here for a reason.The question here essentially is why not use a free API of a non-proprietary model that doesn't gather logs or data for as long as possible and if it ever comes down it's just a one line replacement with your own self-hosted API link.I understand not wanting to depend on proprietary models, or not wanting your access to get logged, but I don't understand why not use a free API that doesn't log anything and is of a proprietary model you could hypothetically run yourself whenever needed.
is used a770 16gb any good?
>>103003110Are you saying that some people are willing to lie on the internet? Preposterous...>>103003133>that doesn't gather logs or data for as long as possible>that doesn't log anythingDo you really trust them that much?
>>103003110Those zero log VPNs target essentially consumers. Openrouter is trying to target developers and businesses with these API keys, completely different demographics and different incentive structures. Openrouter thinks they can get money by vendor lock-in over the years and using the free API as a trojan horse. Logging data would be equivalent to shooting themselves in the foot for just a tiny sliver of money compared to what they could have had by speaking the truth and roping in a bunch of businesses that get hooked to their system over time.It just makes no financial sense for them to secretly log. A lot of risk for barely any reward.
>>103003133One reason I can think of is to not get "used" to the smarts of a 405B then possibly have to settle for less when it inevitably stops being free and you can't run that at home.
>>103002934>https://openrouter.ai/docs/provider-routing>Data Privacy>Some model providers may log prompts, so we display them with a Data Policy tag on model pages. This is not a definitive source of third party data policies, but represents our best knowledge.What do you think openROUTER means? They aren't guaranteeing random providers aren't logging (and I'd bet openrouter themselves are, too..."We use the information we collect for various purposes, including: To improve our services" from their privacy page)
https://arstechnica.com/tech-policy/2021/08/zoom-to-pay-85m-for-lying-about-encryption-and-sending-data-to-facebook-and-google/
The answer is CP. People want to RP with CP and they can only do so without paranoia on their self-hosted model. It's that simple, no need to get philosophical about things.
>>103002903>>103002934>>103002958Not running locally atm, but valid reasons that I see are:1. I don't believe they're not gathering data.2. It's rate-limited.Still pretty nice. I don't even care that they gather my data and read my coom logs. So I'm gonna use it now.
Good morning, /lmg/
>>103003162That's your interpretation and assumptions.As I said, we can't know until it's experimentally proven.That's all.I'm not telling you not to use it, just arguing that a policy by itself is not a guarantee, even legally binding contracts are broken when if it's more profitable.>>103003161Socking, I know.>>103003215This. I'm going to give it a go too.
>>103003207I do it on my paid Claude account.
>>103003168Openrouter promised to keep it free until at least the end of 2026. So you have 2 years to get the hardware to run something equivalent. By 2027 I bet you you can run a 70B model with equivalent smarts. Probably better inference hardware out then for cheaper as well.
So I bought those RTX 3090s for nothing?!....
>>103003271I'll buy them for 200$ a piece.
>>103003250>Openrouter promised to keep it free until at least the end of 2026.Where? And how can they promise that when they're not the ones running the models?
>>103003271I'll take them off your hands for $220 each
>Yes please. Despite its reasonable-sounding theoretical foundations, TFS has not stood the test of time and the collective experience of tens of thousands of power users has shown that it is soundly beaten by Min-P in practice, at a fraction of its complexity.>FWIW, I believe that Typical and Mirostat should meet the same fate. Judging from presets posted in forums and model cards, they're very rarely used nowadays, and Mirostat in particular is difficult to build an intuition for. The lesson appears to be that samplers using entropy/perplexity as a control metric are overlooking one or multiple key properties that strongly determine output quality. A combination of Min-P and XTC can do anything those samplers can do and more, and do it in a way that is comprehensible to a human operator thinking about probability distributions.>Heck, even Top-P and Top-K (for values > 1) should probably be deprecated and show a warning when used. Min-P fixes the conceptual flaws of both of them and is flat out superior. AFAIK, those samplers originate from the early OpenAI days and were then incorporated into HF Transformers, after which everyone else cargo culted them into every frontend and backend. Their continued availability (and the fact that they are presented at the very top of most frontends' sampler settings) really hurts the experience of new users, who probably shouldn't be using any sampler other than Temperature, Min-P, XTC, and DRY.
>>103003271250.
>>103002886>limited to 20 requests per minute and 200 requests per day
>>103003220good morning miku
>>103003348Nothing wrong with that. Everything should be enjoyed in moderation.
>>103003337I was just reading it. I agree for of them except top-k. Stop posting self-portraits.
>no string ban/antislopYeah that's a shill.
>>103003360Kys shill
>>103002886>https://openrouter.ai/nousresearch/hermes-3-llama-3.1-405b:free/api>8,192 contexthttps://openrouter.ai/meta-llama/llama-3.1-405b-instruct:free>8,000 contextAlso you're being weirdly pushy about moving people from their local setups to an API
>>103003348>You can just make more accounts and request more API links as needed
>>103002934>They actually say they don't collect your data on the free APIWhere do they say that? I can't find any page on their site that actually states the policy
Buy the 50 series goy, you'll get to run 16GB VRAM at twice the speed. They even use 600W, the higher this number is the better btw, because you know you're putting your GPU to WORK. Unused electricity is wasted electricity. Accelerate!
>>103003337>p-e-wIsn't he the guy who made Min-P, XTC, and DRY? And now he wants to remove all the samplers except is? How totally unbiased.
>>103003337TFS still has its uses as a less aggressive MinP. Just because redditards couldn't figure out how to use it doesn't mean that it's bad. Typical and Mirostat can go.
>>103003610Don't worry, p-e-w will chose for you, he knows best, and ggerganov agrees>@p-e-w Thanks for the insights. I will use this opportunity to ask you about what do you think is the best sampling strategy when the LLM is used for fill-in-the-middle (FIM) tasks (e.g. code completion). Do you have any experience and recommendations? My intuition is that one would generally want greedy sampling in this case (since we are not looking for variety and/or creativity), but with some conservative way to terminate early as soon as the LLM is not "confident" about the next token. Do you agree and if yes, what would be a good approach to model this during sampling?
>>103003576Not min-p, but i think he did the other ones. As far as i remember min-p was kalomaze. Mirostat and Typical are shit, though. They're not intuitive.Also, the new k-shift in the other PR is the lazy version of what he mentions in the paper. Absolutely useless without the rest of the paper. It defeats the purpose of implementing it at all.>https://github.com/ggerganov/llama.cpp/pull/10048>https://arxiv.org/pdf/2402.10200
Has anyone here figured out how to use batching effectively locally to maximize throughput in some useful way?
>>103003645>https://github.com/ggerganov/llama.cpp/pull/10048>I am currently sick and will be off the computer for a few days, but I intend to do a full review of this interesting PR soon.>but please remove all samplers but mine kthx
>>103003666Do you really have nothing to say?
>>103003701Hope you get well soon so you can resume gutting the project?
>>103003627>Don't worry, p-e-w will chose for you, he knows best, and ggerganov agreesGod bless koboldcpp, they'll probably keep the old samplers like they kept support for old quants in the early days.
>>103003646>locallyWhat's the difference with "remotely"? Does batching mean a different thing?Batching is more useful for offline stuff (as in non-interactive). Batch 1000 completion requests, check them later. Like making synth datasets and that sort of stuff.
>>103003729Schizo.
>>103001193Fuck you.
>>103003763j-e-w
>>102999437I wanted to do something like this for a long time. Is it open source?
>>103003745>What's the difference with "remotely"?I mean for a single-user local scenario. Multi character rp or maybe multiple solutions with different samplers for review etcbatching is obviously useful in multi-user provider type scenarios>Like making synth datasets and that sort of stuffYeah, like that but more local-specific niche stuff
>>103003808Maybe to process the prompt in parallel with the running chat to do stuff like summarization?I bet that it could be leveraged for some real time shit like talking to multiple cahracters "online" and the like.
>>103003732Hopefully he gets bored before he removes all samplers, hopefully>I rarely stick to one idea for more than a few months>https://github.com/p-e-w>https://worldwidemann.com/about/
>>103002916>Because I can run 405b locally.I've got 256GB of DDR4 already, another 256GB is about $300. Despite having a 28-core Platinum 8280L, it's still going to be dogshit slow, right?I've played wit 405B online, it's better than 70B, but is it worth being in the s/t gen speed range better?
>>103003788Isn't he reinventing temperature?
>>103003798>Philipp Emanuel (((Weidmann)))>Jewish (Ashkenazic): artificial name from German Weide ‘pasture meadow’ + Mann ‘man’.Oy
>>103003891No. Temperature modifies probabilities. He's talking about measuring the 'confidence' of the generated tokens and when to stop generating.
>>103003886>it's still going to be dogshit slow, right?yes. I'm best-case cpumaxxing and struggle to hit 1t/s at q8. You'll likely hit 10% of that with that description of your setup.>but is it worth...It can be if you need frontier capabilities, but I daily deepseek. Its the best speed to capability tradeoff in my case. I only run 405b when I need to.