[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/g/ - Technology

Name
Options
Comment
Verification
4chan Pass users can bypass this verification. [Learn More] [Login]
File
  • Please read the Rules and FAQ before posting.
  • You may highlight syntax and preserve whitespace by using [code] tags.

08/21/20New boards added: /vrpg/, /vmg/, /vst/ and /vm/
05/04/17New trial board added: /bant/ - International/Random
10/04/16New board for 4chan Pass users: /vip/ - Very Important Posts
[Hide] [Show All]


[Advertise on 4chan]


File: 1764936420039709.jpg (496 KB, 896x1200)
496 KB
496 KB JPG
/lmg/ - a general dedicated to the discussion and development of local language models.

Previous threads: >>107686942 & >>107679732

►News
>(12/26) MiniMax-M2.1 released: https://minimax.io/news/minimax-m21
>(12/22) GLM-4.7: Advancing the Coding Capability: https://z.ai/blog/glm-4.7
>(12/17) Introducing Meta Segment Anything Model Audio: https://ai.meta.com/samaudio
>(12/16) MiMo-V2-Flash 309B-A15B released: https://mimo.xiaomi.com/blog/mimo-v2-flash
>(12/16) GLM4V vision encoder support merged: https://github.com/ggml-org/llama.cpp/pull/18042
>(12/15) llama.cpp automation for memory allocation: https://github.com/ggml-org/llama.cpp/discussions/18049

►News Archive: https://rentry.org/lmg-news-archive
►Glossary: https://rentry.org/lmg-glossary
►Links: https://rentry.org/LocalModelsLinks
►Official /lmg/ card: https://files.catbox.moe/cbclyf.png

►Getting Started
https://rentry.org/lmg-lazy-getting-started-guide
https://rentry.org/lmg-build-guides
https://rentry.org/IsolatedLinuxWebService
https://rentry.org/recommended-models
https://rentry.org/samplers
https://rentry.org/MikupadIntroGuide

►Further Learning
https://rentry.org/machine-learning-roadmap
https://rentry.org/llm-training
https://rentry.org/LocalModelsPapers

►Benchmarks
LiveBench: https://livebench.ai
Programming: https://livecodebench.github.io/gso.html
Context Length: https://github.com/adobe-research/NoLiMa
GPUs: https://github.com/XiongjieDai/GPU-Benchmarks-on-LLM-Inference

►Tools
Alpha Calculator: https://desmos.com/calculator/ffngla98yc
GGUF VRAM Calculator: https://hf.co/spaces/NyxKrage/LLM-Model-VRAM-Calculator
Sampler Visualizer: https://artefact2.github.io/llm-sampling

►Text Gen. UI, Inference Engines
https://github.com/lmg-anon/mikupad
https://github.com/oobabooga/text-generation-webui
https://github.com/LostRuins/koboldcpp
https://github.com/ggerganov/llama.cpp
https://github.com/theroyallab/tabbyAPI
https://github.com/vllm-project/vllm
>>
File: threadrecap.png (1.48 MB, 1536x1536)
1.48 MB
1.48 MB PNG
►Recent Highlights from the Previous Thread: >>107686942

--Implementing character roleplay with system prompts in Open-WebUI to constrain LLM responses:
>107697050 >107697202 >107697800
--Quantization quality thresholds in Llama.cpp for large language models:
>107694813 >107695987 >107696086
--Quantized model optimization under RAM/VRAM constraints:
>107688512 >107688542 >107688581 >107688771 >107688839 >107688911 >107689227 >107689299
--Feasibility of running 4.7 model with 128GB RAM and 32GB VRAM at 3T/s speed:
>107694348 >107694375 >107694574 >107694605 >107694687
--RWKV.cpp as Microsoft's on-device AI implementation:
>107697911 >107698596 >107698971
--Open-source model GLM-4.7 achieves top ranking on benchmark index:
>107689325 >107689538 >107689545
--M2.1 model performance and roleplay evaluation:
>107698092 >107698194 >107698171 >107698182 >107698198
--Hardware selection dilemmas for local LLM enthusiasts:
>107687115 >107687159 >107687197 >107687217 >107687326 >107687392 >107687421 >107687348 >107687388
--Quantizing Llama model with bf16 tensors:
>107696219
--Tennessee AI training restrictions on emotional relationships:
>107698160 >107698180
--Gaslighting language models to bypass censorship:
>107692222 >107692310 >107692314 >107692485 >107693252 >107695957 >107696118 >107696260 >107696518
--Browser-specific performance differences in ComfyUI workflows:
>107695920
--Anticipation and skepticism around Small Creative:
>107689009 >107689037 107689080
--AI tech for authoritarian parenting in China:
>107698384 >107698600 >107698985
--AI as interactive fiction game director:
>107690487 >107690540 >107690553
--Proposing Cockbench update with chat templates for training insights:
>107698263
--EGPU scalability for local 3T parameter models:
>107692736 >107692864 >107692886
--Miku (free space):
>107688568 >107690307 >107694744 >107688652

►Recent Highlight Posts from the Previous Thread: >>107686945

Why?: >>102478518
Enable Links: https://rentry.org/lmg-recap-script
>>
Gemma
>>
>>107700893
Why was this hit?
>>
>>107700924
Canceled
>>
File: 1757922766538175.jpg (115 KB, 1142x1142)
115 KB
115 KB JPG
>>107700977
Previous thread links were wrong. Unacceptable.
>>
File: 1755632570054463.png (26 KB, 224x224)
26 KB
26 KB PNG
is local still light years behind in TTS voice cloning?
>>
File: file.png (54 KB, 677x314)
54 KB
54 KB PNG
the ai is being very kind and understanding. im inspired to embrace my inner snowflake
>>
>>107700909
Bald was better
>>
>>107701174
this is illegal in TN
>>
>>107701174
There is no wrong way to fantasize
>>
>>107701174
Yes, even the very AI the ones who make the models try to suffocate agrees with you and not them.
>>
File: 1739906856872592.gif (248 KB, 498x362)
248 KB
248 KB GIF
>>107701253
There is, it's literally part of "wrongthink".
>>
>>107701268
i want to migu migu
>>
>>107701088
>is local still light years behind in TTS voice cloning?
local TTS is light years behind

local voice cloning was perfected pretty early on with RVC models in my opinion. If you can get a TTS from some other source, turning that voice into someone else's is trivial locally at that point
>>
>>107701174
What model?
>>
>>107701310
>huggingface.co/google/switch-c-2048
>>
>>107701314
tank
>>
>>107701280
migu is not for migu
>>
>>107701310
there aren't a lot of models that allows you to fuck the still warm corpse of a dead eight year old little girl
it's nemo 12b or specifically rocinante v1.1
>>
>>107701332
Kek, nemo is pretty horny to begin with
>>
Is Gemma 3 the best general purpose model around 30B parameters, or should I use something else?
>>
>>107701332
>corpse
How does it work? She literally can't react
>>
>>107701394
Roleplay, it can still describe you what's happening.
>>
>>107701386
Gemma3 feels so cucked to me
I prefer qwen3-vl 30b
>>
File: 1761412669869213.jpg (213 KB, 1658x1482)
213 KB
213 KB JPG
>>107700909
>>
>>107701394
i like to have another character watching me do it too, it's even better if it's a woman describing her horror over what im doing, along with the dirty details
like when the dead little girl's bladder releases
im such a sick fucker...
>>
>>107701450
Maybe Tennessee is right after all...
>>
>>107701332
>rocinante
placebo
>nemo
only reason it was good is complete lack of censorship. makes me wonder how much better glmchan would be if she had zero censorship.
>>
>>107701534
Sucks that Chinese models can never have zero censorship since they just distill it from western models. Hopefully one day they can move past that need.
>>
>>107701332
Are you using straight nemo or some kind of finetune of it? It always seemed to lose coherency for me, maybe I need to give it another shot.
>>
File: 1744111940731012.png (797 KB, 1658x1482)
797 KB
797 KB PNG
>>107701433
Fixed your glazed garbage ^.^
>>
>>107701332
>it's nemo 12b or specifically rocinante v1.1
How does it compare to normal nemo? Like what benefits?
>>
Is Medusa Halo going to save local models?
>>
File: 1745936031854034.png (737 KB, 1062x1776)
737 KB
737 KB PNG
>>107701631
UOH?!?!?
HOW DID YOU REMOVE THE POISON?!?!
>>
>>107701715
>OH NO MY ART
>I better make it look like shit, that will show the AI people!!!
>>
>>107701696
>Like what benefits?
It got astroturfed when nemo was new. People downloaded it instead or regular instruct and thought it is the magical finetune and not just the instruct model being good.
>>
>>107701696
You can use ChatML with it if you're really anal about using the format. That's about it.
>>
I actually went back to 4.6 for sex.



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.