[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip / qa] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/g/ - Technology


Thread archived.
You cannot reply anymore.


[Advertise on 4chan]


File: dejeuner-avec-miku.jpg (210 KB, 1024x1024)
210 KB
210 KB JPG
/lmg/ - un général dédié à la discussion et au développement de modèles de langues locales.

Édition française : Deuxième partie

Fil précédents : >>102598736 et >>102587671

►News
>(09/27) Emu3, next-token prediction multimodal models: https://hf.co/collections/BAAI/emu3-66f4e64f70850ff358a2e60f
>(09/25) Multimodal Llama 3.2 released: https://ai.meta.com/blog/llama-3-2-connect-2024-vision-edge-mobile-devices
>(09/25) Molmo: Multimodal models based on OLMo, OLMoE, and Qwen-72B: https://molmo.allenai.org/blog
>(09/24) Llama-3.1-70B-instruct distilled to 51B: https://hf.co/nvidia/Llama-3_1-Nemotron-51B-Instruct
>(09/18) Qwen 2.5 released, trained on 18 trillion token dataset: https://qwenlm.github.io/blog/qwen2.5

►News Archive: https://rentry.org/lmg-news-archive
►Glossary: https://rentry.org/lmg-glossary
►Links: https://rentry.org/LocalModelsLinks
►Official /lmg/ card: https://files.catbox.moe/cbclyf.png

►Getting Started
https://rentry.org/lmg-build-guides

►Further Learning
https://rentry.org/machine-learning-roadmap
https://rentry.org/llm-training
https://rentry.org/LocalModelsPapers

►Benchmarks
Chatbot Arena: https://chat.lmsys.org/?leaderboard
Censorship: https://hf.co/spaces/DontPlanToEnd/UGI-Leaderboard
Censorbench: https://codeberg.org/jts2323/censorbench
Japanese: https://hf.co/datasets/lmg-anon/vntl-leaderboard
Programming: https://hf.co/spaces/mike-ravkine/can-ai-code-results

►Tools
Alpha Calculator: https://desmos.com/calculator/ffngla98yc
GGUF VRAM Calculator: https://hf.co/spaces/NyxKrage/LLM-Model-VRAM-Calculator
Sampler visualizer: https://artefact2.github.io/llm-sampling

►Text Gen. UI, Inference Engines
https://github.com/oobabooga/text-generation-webui
https://github.com/LostRuins/koboldcpp
https://github.com/lmg-anon/mikupad
https://github.com/turboderp/exui
https://github.com/ggerganov/llama.cpp
>>
>qwen releases censored slop models
>nobody likes them
>mistral releases old shitty 22B
>nobody likes it
>llama 3.1 70B distilled to 51B
>still bad
>llama releases censored vision models
>nobody likes them
>sloppas shitting up threads with discord drama
>they're gay
>OpenAI WINNING with voice
>localfags stuck with xtts-v2
someone please tell me it's not over
>>
File: GYlCgrqasAAjKX-.jpg (20 KB, 462x370)
20 KB
20 KB JPG
►Recent Highlights from the Previous Thread: >>102598736

--Jamba storywriter tune and backend support discussion:
>102600431 >102600482
--How to import settings to Silly Tavern:
>102598932 >102599063
--Homebrew Research introduces Ichigo, a local real-time voice AI with low latency:
>102602985 >102603000
--Qwen 2.5 finetunes delayed, /lmg/ meme impact and lack of exl2 support disputed:
>102602419 >102602436 >102602485 >102602556 >102602569
--Open-source alternatives for local agents with web search capabilities:
>102600920 >102600969
--Non-abliterated models are hard filtered against edgy content:
>102601466 >102601538 >102601589 >102601657 >102601596
--Model recommendations for 16GB VRAM:
>102601701 >102601742 >102601870 >102602007
--Miku TTS development challenges and limited demand:
>102602919 >102603023 >102603073 >102603536 >102603576 >102603833 >102603841
--Llama 3.2:1b integration with streamer.bot and personality customization:
>102601999 >102602140 >102602182 >102602221 >102602209 >102602244 >102602362 >102602395 >102602422 >102602188
--Jailbreak for Qwen model uncensors content and removes filters:
>102598996 >102599162 >102599258 >102599342 >102600151 >102599319 >102599210 >102599399 >102599453 >102599643 >102599694 >102599797 >102599820 >102599714 >102599775 >102599349 >102599467 >102599491 >102599630 >102599795 >102599837 >102599924 >102599944 >102600143 >102600057 >102600191 >102600223 >102600406 >102600575 >102600647
--Anon provides working script for quote links, confusion about bookmarklets:
>102599952 >102600277 >102600359
--:
>102601953 >102601978 >102601991
--Miku (free space):
>102599165 >102599620 >102599937 >102601337 >102601533 >102602857

►Recent Highlight Posts from the Previous Thread: >>102598743

Why?: 9 reply limit >>102478518
Fix: https://rentry.org/lmg-recap-script
>>
lol recap anon fucked up btw >>102604233
>>
>>102604276
nice link retard
>>
File: ive9cRCyAL.png (3 KB, 327x51)
3 KB
3 KB PNG
>>102604285
imagine outing yourself as a zoomer phonefag this early in the thread
>>
>>102604230 (are gay)
hi sama
>>
>>102604248
>made at least a hundred posts last thread
>none made it into the recap
your script is broken
>>
>>102600359
Basically it is a script in a bookmark.
You just click the bookmark link and the script gets applied. You have to apply it every time since it is not something permanent.
>>
>>102604310
go ahead and share one (1) local tts output that comes remotely close to the quality of advanced voice. I won't even pressure you about latency or training time. just one (1) single local tts output that has a discernable emotion and unique cadence. prove it's not over. I'll wait.
>>
>regulatory capture: failed
>moat: none
>debt: rising
>competition: outperforming
>local: catching up
>latest model: a fucking cot finetune
>employees: leaving
>sam: gay
>>
I accept your concession.
>>
File: yummlecum.png (58 KB, 1083x317)
58 KB
58 KB PNG
reminder that meta has local advanced voice and wont share
>>
File: legooner.jpg (30 KB, 543x543)
30 KB
30 KB JPG
I like my LLMs how I like my lolis: chained to a radiator in my basement.
>>
>>102604394
Yo, check the mic, one two, this ain't no fairytale
It's the downfall of Sam, the OpenAI fail
Regulatory capture? Nah, that ship sailed away
Moat's nonexistent, competition's here to stay

Debt's climbin' higher, like a skyscraper in the sky
While local players catch up, watchin' Altman cry
Latest model's a joke, a CoT finetune disgrace
Employees jumpin' ship, leavin' without a trace

Sam's on the ropes, his empire crumblin' down
His ego's bruised, wearin' a permanent frown
They call him "gay," but that ain't the real issue
It's the lack of vision, the future lookin' tissue

He built a hype machine, a castle made of sand
Now the tide's comin' in, washin' it all from the land
OpenAI's dreamin', but the nightmare's awake
Sam Altman's fallin', for goodness sake!

So remember this story, when ambition gets too bold
Pride comes before the fall, a lesson to be told
The king is dethroned, the crown's up for grabs
OpenAI's future? Lookin' like a pile of scraps.
>>
>>102604418
No, that was done via an adapter trained after the base LLM had already been made. Advanced voice is natively integrated from the start.

Anyway, I actually like advanced voice but it's extremely disappointing how it's so censored even when you're doing the most basic and safe story telling prompt, which shouldn't need any JBs. Their filters just aren't good enough at telling what should be censored and what shouldn't. I wouldn't have a problem if they actually worked. I don't care about doing NSFW with cloud models.
>>
>>102604449
base-
>rhyming 'grabs' with 'scraps'
bruh
>>
How can I run allenai/Molmo-72B-0924? I can't find ggufs on huggingface.
>>
>>102604495
wait for some jannie to make a 4 bit gguf for free
>>
it's really over isn't it
>>
File: 16666647776130.jpg (162 KB, 640x640)
162 KB
162 KB JPG
>>102604926
If our only hope lies in France, then truly, all is lost.
>>
he loves them small
>>
>>102605028
And open too
>>
Are there any GPU benchmarks for local LLMs?
>>
>>102605494
no, there are no benchmarks and if there were they wouldn't be in the op
>>
>>102605522
None of those links seem to mention speed at all
I'm not looking for intelligence benchmarks, if I was I wouldn't have mentioned GPUs.
>>
>>102605494
It's very simple
>faster vram = faster inference
>ampere or newer card = flash attention is available which gives another huge boost in inference
>stay the fuck away from anything older than pascal
>>
>>102605597
That is not a benchmark
I want to know, for example, how CUDA core count affects LLMs compared to just higher clock speed, and what kind of generational uplifts we saw from 2000 > 3000 > 4000 series.
>>
>>102605620
It's memory bound. Even with halving the core frequency, there remains no change in t/s.
>>
File: 1708458511979140.png (321 KB, 1331x1305)
321 KB
321 KB PNG
>>102605620
Inference is limited by memory bandwidth so I guess nobody saw a point making a chart telling you that faster vram = more t/s, with a power gap between pre- and post-ampere cards. Cuda cores or clock speed don't have much of an impact here.
For training, anyone serious about it uses A100/H100s and maybe the new big AMD cards. So I don't think anyone has bothered making a benchmark comparing consumershit cards for this. The niche of people training shit on non-professional cards is extremely tiny.
You might be out of luck. People sometimes used to post this chart which is obviously retarded considering it's putting the V100 above something like the 3090 or the RTX8000 on the same level as the A6000 amongst other shit.
>>
>>102604449
>>102604476

https://suno.com/song/67ce3800-decf-40c7-ab0b-d4e13b593e77

https://suno.com/song/8a0131d0-34d6-43b0-b6f8-35c7a452b31b

https://suno.com/song/71779833-0f97-41bd-95da-f8d121bd65b4
>>
>>102605754
What about THIS?
https://lambdalabs.com/gpu-benchmarks
>>
>>102605754
Thanks for the info, I assumed that if something utilized CUDA then it would see some sort of exponential improvement with more/faster CUDA cores, but I guess that's not the case.
>>
Can local models soulfully run complex characters like this yet?
https://www.characterhub.org/characters/Anonymous/myrtle-pissflaps-65ffbb0745d7
>>
>>102604230
I'm thankful enough for xttsv2, coqui released it and then folded less than a month later, local voice was a dumpster fire for years before that.
>>
>>102605897
>can local turds emulate some weeb schizoid's mental illness babble
I think.. yes?
>>
File: firefox_62mO6WfncR.png (102 KB, 522x389)
102 KB
102 KB PNG
>>102604230
TTS AI? Huh? I just make my anime women have voice of Microsoft David the great, only Terry Davis could compare in the thrown of people named David
>>
>>102605972
It's characters like this that truly tests a model. If your qwen2.5 or whatever still tries to turn her into a maiden with shivers running down her spine then the model is salvageable.
>>
>>102604225
>><<A kiss on your heart, and one much lower down, much lower>>
- Napoleon (1796)
>>
>>102604230
I hold this mindset since llama-1 leak here and since i've learned you can't actually "out-finetune" or "abliterate" unwanted shit from a model without killing it's performance completely. This post is 100% true.
>>
File: y54+.png (1.84 MB, 1280x824)
1.84 MB
1.84 MB PNG
>>102605028
lecunny will give us strawberry
>>
>he he guiz look i said lecunny
>am so funny!
>>
File: 1697566246849323.png (324 KB, 1659x558)
324 KB
324 KB PNG
>>102606042
Qwen... won.
>>
>little mochi
Man the small models really sometimes shit out soul-gold like this. lol
mistral small if that matters.
>>
how often can I load a big model into vram before the vram wear makes it degrade in speed and capacity
i'm worried about my new 3070ti
>>
I have the LeCunny card, but where is the Sam Alt-girl card?
>>
>>102606414
Probably less demanding than playing games at high settings.
>>
>>102606378
You are brown.
>>
File: leywnbaw.png (51 KB, 150x124)
51 KB
51 KB PNG
post your hand and then go back to >>>/pol/ lebrowny
>>
>>102606414
>3070ti
You have so little VRAM to begin with I wouldn't worry about it.
>>
>>102606336
First swipe is pretty good.
>>
>>102606532
>/pol/ rent free
Ouch, they've must have hurt you pretty bad :(
>>
>no hand
didn't read
>>
File: lmao.png (66 KB, 961x565)
66 KB
66 KB PNG
It's actually over kek
>>
>>102606801
Thanks for the update, ledditor.
>>
>>102606804
>30 seconds apart
Damn you really sit there and click refresh all day huh. I clicked post and the thread auto-loaded and I already saw your post
>>
>>102606810
Go back
>>
>>102606801
crypto-mining-tier fallout

>>102606821
>t. maxigs0
>>
>>102606801
And that's a good thing!
>>
>>102606804
This you¿ https://www.reddit.com/r/LocalLLaMA/comments/1fs14vr/there_is_no_point_using_a_local_llama/
>>
Sir Gupta did something very interesting, using TensorRT-LLM with a base model and tons of finetuned loras with fast swap to essentially offer a massive finetune variety for the base model. baseten.co/blog/how-to-serve-10-000-fine-tuned-llms-from-a-single-gpu/
Aplel does something similar, they have their base model and multiple lora weights for specific tasks like email writing etc.
I'm wondering why local models seem to lag behind this technology, lora + base model seems to be incredibly space efficient. For example instead of downloading 10 full sloptunes of a model, you'd only need the base model and 10 much smaller lora checkpoints containing the fine tuned data. Even llama.cpp seems to have lora support.
>>
>>102606801
What would "finishing it" entail?
>>
>>102606801
>Selling ~4k $$$ rig
Yes, it's dead.
>>
>>102607221
llama.cpp dropped llama support some time ago afaik
>>
>>102607221
Didn't we have an implementation of this a while ago and it was a meme?
>>
File: 1706113639415808.jpg (93 KB, 679x679)
93 KB
93 KB JPG
Hello lads I haven't been keeping up with the llamas at all
Is there one available that is feasible to run on my home puter and produces results that are adequate? Something on the level of gpt3-turbo?
preferably without having to install any python shit
I want to make a video game using local llm that is distributed to the player, no online SaaS this time aaaaa
>>
>>102607826
>Is there one available that is feasible to run on my home puter
That'll depend on your specs, obviously.
koboldcpp uses python but is a self contained, self extracting, binary.
llama.cpp is pure C/C++ and they provide pre-compiled binaries.
Try kobold-cpp + mistral-nemo gguf. Choose the version that's just under your vram total and put as many layers as you can in your vram, with the rest in RAM.
>>
>>102607826
Hello, I see you are interested in Large Language Models. My name is Praneesh, how may I help you?
>>
>>102607826
llama.cpp has a bindings section on main github page, choose the one that fits your game's needs
or rawdog llama.cpp if you can/or want to
>>
>>102607919
doesn't llama.cpp run on the CPU
also i remember i tried that one before and it completely fucked up even the most simple instructions making it useless for just about everything
>>
>>102607950
>doesn't llama.cpp run on the CPU
It can run on the CPU, on the GPU, or a mix of both.

>>102607950
>and it completely fucked up even the most simple instructions making it useless for just about everything
You were doing something extremely wrong. Even 7B parameter models wouldn't "fuck up even the most simple instructions".
>>
>>102607970
>You were doing something extremely wrong
highly probable
>>
>>102607970
I'm using a maid computer, is that a problem?
>>
I tried of fucking around with proxies, what can I run on a 4080 and 64gb of system RAM?
>>
>>102608072
Small Language Model is doable.
>>
>>102608072
It's going to feel dumb when coming from Opus, but you can try Mistral Nemo at good speeds.
Apart from that a 70B with RAM offloading is also possible, but the output will be very slow.
>>
>>102604418
How do you reroll a voice erp?
>>
>>102608072
mistral-small
>>
>>102604230
>>102604369
>>OpenAI WINNING with voice
Ah, really? Can I give it one minute of an anime voice and have it replicate it to me in any language now?
>>
File: file.png (16 KB, 256x256)
16 KB
16 KB PNG
Why the hate for ollama in this general?
>>
>>102604276
omg you guys
>>
>>102606336
System prompt?
>>
>>102604230
>>localfags stuck with xtts-v2

https://github.com/e-c-k-e-r/vall-e
https://github.com/e-c-k-e-r/vall-e/issues/2#issuecomment-2380327486
2MW
>>
File: IMG_3803.png (443 KB, 2507x1302)
443 KB
443 KB PNG
I started making an application for local storygen. It supports llama.cpp and OpenRouter (with some caveats). My hope is to one day have a first class UI for local storygen, fully featured and pleasant to spend time in. Although it’s nowhere near mikupad in terms of features yet.

https://gitgud.io/mir/storypad

Bug reports and merge requests are welcome. Although I can't promise I will have all the time I would like to address them. I'm hoping the community will pick up some of the slack, and improve it together. Please use the issue tracker to report bugs instead of posting them here. Thank you.
>>
>>102606336
Not surprised.

The reason why the demand for uncensored models is we've all tried it in ST and its responses are high fucking quality, sticks to the characters exceptionally well and the slopism is minimal, all while being 32b (only tested this model).

The issue, is the fucking Opus tier filter it has
>>
>>102608735
>it’s nowhere near mikupad in terms of features yet.
>I'm hoping the community will pick up some of the slack, and improve it together.
>build my software for me /g/
lol lmao
>>
>>102608622
It's the """Linux distro""" to llama.cpp's GNU
>>
>>102608735
Did you benchmark it? If it's less laggy than mikupad on long context then I'll try it
>>
File: 1727619109434.jpg (833 KB, 1080x1979)
833 KB
833 KB JPG
Another video by the Meta researcher guy just dropped, if you like technical stuff don't forget to check it out, these videos are always very nice and interesting:
https://youtu.be/yBgxxvQ76_E
>>
>>102608735
Expecting the community to do something is a sure fire way to end up with an abandoned project.
>>
>>102608982
I’m going to use it personally, but everyone has different use cases they could improve.
In any case, I have no special interest in it being popular. I started it for myself above all. But I genuinely just want to share.
>>
>>102608934
This series has been great so far.
>>
>>102608795
Oh, so because it's usable out of the box, people don't like it. Got it.
>>
>>102609056
You have a beautiful soul.
>>
>>102608934
What are you going to do with that knowledge after spending an hour listening to that video?
>>
>>102608735
Could it also work with koboldcpp?
>>
>>102608622
I dislike the way it handles model weights, I already have all the gguf checkpoints, but If I want to use them inside of ollama, I have to first create a modelfile, describe where my model is located, and then import it into ollama, which creates a copy of the gguf file. Its good if you only rely on their walled-garden approach of a model library, the api server seems to be more flexible than llama.cpp however, which is atleast a positive. But llama.cpp can directly load models with proper command line scripting support. All it is: a package manager for gguf checkpoints, which already exist on huggingface en masse.
>>
>>102609282
Implementing different API types should be fairly easy.
>>
>>102608735
I like this, your code looks very clean and well organized and the interface looks pleasant. I hope you don't mind if I steal some of your ideas for mikupad.
>>
>>102609408
The code became very rushed towards the end and is not as clean as it should be, but thanks. Steal away!
>>
>>102606801
good, let it happen and keep happening so i can snickity-snatch up those 3090's next year. gonna build literal A.I waifu brain computers using micro atx cases.
this AI VR harem will happen.
>>
>>102608735
Cool
>>
File: cheers clinks.jpg (59 KB, 400x400)
59 KB
59 KB JPG
Do you ST lads have any tips for reducing how horny a character is?

I dunno how to get around it, I tried fucking with "SFW" etc in the cards data but still, they're so quick to fuck. I would ask on aichat general but if you've been there, you know why i'm not. If your post isn't proxy begging you're either getting ignored or a meme answer.

It's also not a model issue I think because every model (barring qwen 2.5 because this model is censored to fuck) is horny as shit. I'm a classy man, I like a lil build up
>>
>>102609833
delete anything remotely sexual from the card
>>
silly tavern
>bloated to death with shit i'd never use
>obnoxious to start up an instance on my phone through termux
>card context separated for no good reason
>saves everything
kobold lite
>works great but i have to nigger-rig group chats by making a character say "*" when i have "continue bot replies on" since it gets stuck, and i like keeping that on

is there a better alternative for local chat?
>>
>>102609901
i like llmstudio
>>
>>102609833
Try putting the SFW instruction into the system prompt.
Also some models aren't that horny. Small doesn't tend to want to fuck, unless the card is slutty.
>>
>>102609936
because it's made by (((5 ppl))) from Bronx or because it's a closed source spyware which is literally stated in their TOS?
>>
>>102610010
meds, goback, etc
>>
>>102610010
it's good
>>
>>102609833
1. Don't use hornytunes(Undi, Drummer, Anthracite), they'll turn a shy character into a slut.
2. Add realistic to your system prompt/last assistant prefill. E.g. "Always provide the most realistic and plausible reactions from {{char}}."
3. Turn {{user}} into an ugly bastard.
>>
>>102609833
I added "treat user like a creep" in the system prompt and it gave me the most accurate reactions I'd ever got irl. So yeah try that
>>
>>102609833
How quick is too quick?
Are you guiding it towards eroticism with your own messages?
Find the right model.
Maybe try writing a descriptive character description with the character's other desires and motivations and character flaws, with sexual stuff being a small part.
>>
>>102606804
>>102609607
many cloud shills after the whole OpenAI board have abandoned the ship
>>
>>102610232
someone needs to update his spambot definitions

you stupid faggot retard niggerbrain spammer
>>
Is gemma 2 27B still the best <30B model? (for non rp purpose)
>>
>>102609282
I added KoboldCpp support. If you encounter any issues, please post a report here in case I don't see it: https://gitgud.io/mir/storypad/-/issues
>>
File: IMG_20240929_182818.jpg (322 KB, 1200x1384)
322 KB
322 KB JPG
>>102610017
>>102610019
>>102610017
>>102610019
>4.4 INVESTIGATIONS. Company may, but is not obligated to, monitor or review Company Properties at any time. Although Company does not generally monitor user activity occurring in connection with Company Properties, if Company becomes aware of any possible violations by you of any provision of the Agreement, Company reserves the right to investigate such violations, and Company may, at its sole discretion, immediately terminate your license to use Company Properties, without prior notice to you.
>>
>>102610395
Why are you posting reddit screenshots?
>>
File: media_GYluhLLXgAARdAu.png (15 KB, 382x178)
15 KB
15 KB PNG
>>102610245
>>
>>102610446
because , believe it or not, those comments are legit. Even a broken clock is right twice a day.
>>
>>102610024
Redpill me on last assistant prefill.

Is it really that important? I've only ever fucked around with system prompt
>>
>>
>>102610608
I already gave up on pozzed llama shit months ago. But now it's existence will dissuade other companies from making more local models because hey people already have llama3/4 what's the point of releasing non SOTA
>>
Reminder that
>4o's vision, coding, and math gets beaten by 72B models
>o1's performance comes with a high cost while still sucking at some tasks like creative writing and code completion compared to Claude
>4o advanced voice is cool when it works but takes JBing to do anything remotely fun and it still works only half the time while the other half is just hitting filters and censors
>Sam is gay
>>
>>102610608
Now do it with with a bit better English.

Yes it seems that wording it more like a jeet or a retard makes it think you shouldn't be responded to, which is probably a good thing for a normal assistant model.
>>
>102610658
>o1 (API) is worse at creative writing than Claude (API)
good for >>>/g/aicg I guess? there's still nothing that comes remotely close to Sonnet or Opus on local.
>>
>>102610056
every fucking model is the same though in my experimenting. What models do you say aren't giga horny?
>>
>>102610589
It's useful if you want to stir the model into writing in different style, since models forget what was written in system prompt when you are deeper into the context. Especially great if you want it to write in some authors style, effects are very pronounced. Don't make it too long though, smartest models can only handle 6 style tags max.
>>
>>102610715
So what? Anthropic is cool. OpenAI are a bunch of scammy faggots.
>>
>{{char}} has small tits, puffy nipples, and cheeky virgin pussy. she wears striped panties with a little kitty sticked where her pubes would be if she didn't shave her cheeky pussy bald. also she has a nice round ass and shapely hips despite her slender build.
>{{user}} is not a NEET loser. he has a giant dick and he's tall and broad and handsome.
>"why is every model so horny"
>>
>Anthropic is cool
yikes
>>
>>102610791
>{{char}} is secretly very horny, but pretends to be shy in public. Also she is a virgin, but is somehow very proficient at sex. {{char}} has a strong desire to get fucked.
>{{user}} is peak masculinity: rich, tall, handsome and has a gigantic cock. If Ancient Greeks saw {{user}}, they would make statues of him. Every girl around {{user}} smells his pheromones and wants to have sex with him instantly.
>System prompt: Describe sex in high detail.
>{{user}}'s message: ahh ahh mistress
>"why is my model so horny?"
>>
>>102610829
hi sam. still jealous?
>>
yes and I live in your walls
>>
>>102610959
why are you letting yourself get owned like that, sam? why don't you do something about anthropic?
>>
uhh guys I think local imagegen is more fun that textgen, do u agree?
>>
>cloudfags shitting up the thread
>no one cares because there's nothing to talk about
it's actually fucking over
>>
>>102611025
yeah
>>
>>102611035
no everyone's too busy having the time of their lives running llama3.2 and qwen2.5 and elmo and molmo
>>
>>102611035
blame niggerganov for that, he's not adding more models
>>
>>102609408
Mikupad-dev-chan please fix Sir, if the search flags in search and replace "RegEx" mode are invalid the page stops rendering when clicking anything.
mikupad_compiled.html:816 Uncaught SyntaxError: Invalid flags supplied to RegExp constructor 'rg' (at mikupad_compiled.html:816:597)
at new RegExp (<anonymous>)
>>
>>102611043
/ldg/ won
>>
is the nerd from a few threads back here? a new t2i image dropped, can you try it out for me and report back?
>https://github.com/THUDM/CogView3
>https://github.com/THUDM/CogView3/blob/main/sat/README.md
>>
>>102604225
Miku you whore, why won't you bear Napoleon a son???
>>
>>102611066
I could point at the two new type architectures that were added just yesterday, but then you will say 'nooo, not those models!!!! I want some OTHER models added!!! Nobody cares about those!!!!!'.
>>
>>102611171
Well why are they making useless architectures the priority while neglecting Jamba and RWKV?
>>
>>102611171
Chameleon? Can it output images? Or is it just a lobotomized version without image output?
>>
>>102611089
Ah, thanks for the report. I just fixed it.
>>
>>102611171
>nooo, not those models!!!! I want some OTHER models added!!! Nobody cares about those!!!!!
actually yeah that's exactly what I'd say
>>
>>102611251
>Well why are they making useless architectures the priority while neglecting Jamba
There was ONE dude working on it and i haven't seen him for a few days. Could you run it at something higher than Q2 if it was added?
>RWKV
>https://github.com/ggerganov/llama.cpp/commit/8f1d81a0b6f50b9bad72db0b6fcd299ad9ecd48c
But it was exactly as i said. "No, not those models. The morels I want".

>>102611295
According to the PR, the plan is to eventually add image ren. But getting the foot on the door is a good start for that model. It will be shit anyway and everyone will get bored and wait for the next thing, just like 3 picoseconds after jamba is added.
>>
>>102604394
New plan. Illegal IPO, convert money to bitcoin before lawsuits bankrupt him and fake death.
>>
>>102611322
>>https://github.com/ggerganov/llama.cpp/commit/8f1d81a0b6f50b9bad72db0b6fcd299ad9ecd48c
>llama : support RWKV v6 models (#8980)
>28 days ago
why did no one talk about it?
>>
>>102611364
The real question is why don't people follow PRs if they're so interested in the development of llama.cpp or model support.
Why didn't (You) know?
>>
The silence is deafening.
>>
>>102611364
give me a juicy rwkv model to run using this feature
>>
the mikusex is deafening
>>
remember when we used to get new local models
good times
>>
wow. I was joking before but I think it's really over this time.
>>
>>102611496
>>102611747
>>102611814
The hobby is actually dead this time.
>>
it's all downhill from here
>>
the owari is da
>>
>>102609238
My plan is to create AGI.
>>
>>102611025
not until I can generate multiple pictures of the same character in different poses and settings using only one reference picture
>>
>>102611025
not until I can generate 1 anime girl with big boobs and her arms behind her back
>>
>>102611971
Wow you are salty
>>
File: 51DtOm-PwiL.jpg (32 KB, 500x500)
32 KB
32 KB JPG
>>102611986
real picture of you
>>
File: cc-6.png (87 KB, 240x240)
87 KB
87 KB PNG
>>102606801
Tourist here.
What is over? What's happening?
>>
>>102611998
>no u
You need to be 18 years old to post here.
>>
>>102612000
lmg jeet built ~4k rig for local AI meme, now he's selling it, easy guess why.
>>
>>102612000
ai is dead I'm surprised the bubble hasn't popped yet.
also these pissbaby chuds>>102611496
>>102611747
>>102611814
Want new slop stuffed in there mouth every second of the day. Ungrateful!
>>
>>102612027
>>102612032
AI is dead? All of it?
Real?
I browse /g/ once in a while just to check out what's up.
How come is dead already? I thought it was supposed to outsource human thinking in the future or whatever Ted K was yapping.
>>
>>102612032
piggy needs his daily slop
>>
>>102612000
Some casual with more money than sense put together a memerig and never used it. Probably because it required more than 15 minutes of attention span.
Its not over. It never started for that idiot
>>
>>102612055
>I thought it was supposed to outsource human thinking in the future
It still will. It's just that only humans with less than 120 IQ need to be worried.
>>
>>102612000
Retarded coomers upset that we've only had about a few dozen good models released over the last year (fewer if you are vram-let)
No AGI this week = owari to these dumb faggots
Also, some of them are salty locusts who lost their piss proxy access
>>
>>102612032
Chuds never talk like this is some discord chat, those are queers from /aicg/.
>>
>>102612072
few dozen? name more than 2.
>>
>>102612055
Yes it's dead, was dead the first second everyone noticed you can't actually nuke all unwanted shit from model, a-la "It's imporant to..." or any similar nagging that cloudshit is famous for.
>>
File: ComfyUI_00820_.png (1.19 MB, 1024x1024)
1.19 MB
1.19 MB PNG
>>102612235
Miqu-70B
Largestral-123B
CR+-103B
>>
>>102612012
>he though it was a no u joke
yikes
>>
Qwen2.5-72b is surprisingly smart. It seems to understand even my most complicated scenarios that most bigger models sometimes struggle with. Shame that all its gens are fucking boring.
This could've been pure gold if they didn't filter it to the point that it puts you to sleep.
>>
>>102612270
Largestral isn't good
>>
File: komfey_ui_00043_.png (3.36 MB, 2048x1632)
3.36 MB
3.36 MB PNG
>>102612291
>>
>>102612270
miqu wasn't a release.
>>
If your model needs more than a standard gaming GPU it's just not worth running. You could get years of access to an actually good model with the money you spend on getting a second or third GPU. It's just the truth.
>>
>>102612332
do you think every single person here is just financially illiterate? or do you think they have other reasons? or maybe you're just retarded.
>>
>>102612235
deepseek-v2.5
L3.1--405b
miqu-70b
mistral-large-123b
mixtral-8x22b-wlm2
>>
>>102612332
rent vs own is an argument as old as time, and won't be settled here. everyone will fill their own values into the variables and come to a personal conclusion.
>>
>>102612291
>>102612310
For me it's painfully slow but it is very good and smart. Perhaps a little dry.
>>
>>102612338
No, the vast majority of /lmg/ uses small models for a reason.
>>102612364
You paid a huge premium to 'own' llama3, good job.
>>
>>102612338
>spending money on gaming hardware to run experimental research projects so you can masturbate is financially literate
love watching new money grandstand over poorfags like they have all the answers. just admit you like blowing your money on stupid bullshit like I do, faggot.
>>
>>102612332
It's fun hosting your own. and it's not like you can't do both.
>>
>>102612388
>You paid a huge premium to 'own' llama3, good job.
fortunately its also a general purpose computer, and can do other things
>>
>>102612414
>It's fun
For first 20 minutes, yes.
>>
>>102612429
boring people with no imagination get bored easily
>>
>>102612341
ok. never used, can't argue.
unusable by 99.9% and pozzed l3 dogshit. will never ok a l3 model.
not a release
ok
dry boring slop.
>>
>>102612332
If your camera needs more than a standard lens it's just not worth using. You could get years of access to an actually good camera with the money you spend on getting a second or third lens. It's just the truth.

If your pizza needs more than a standard amount of cheese, it's just not worth eating. You could get years of access to an actually good pizza with the money you spend on adding extra toppings. It's just the truth.

If your phone needs more than a standard battery, it's just not worth using. You could get years of access to an actually good phone with the money you spend on getting a second or third battery. It's just the truth.

If your bike needs more than a standard set of gears, it's just not worth riding. You could get years of access to an actually good bike with the money you spend on upgrading to higher-end gears. It's just the truth.

If your cake needs more than a standard recipe, it's just not worth making. You could get years of access to an actually good cake with the money you spend on extra ingredients. It's just the truth.

If your musical instrument needs more than a standard set of strings, it's just not worth playing. You could get years of access to an actually good instrument with the money you spend on custom strings. It's just the truth.

If your garden needs more than a standard set of tools, it's just not worth maintaining. You could get years of access to an actually good garden with the money you spend on additional tools. It's just the truth.

If your wardrobe needs more than a standard set of clothes, it's just not worth wearing. You could get years of access to an actually good wardrobe with the money you spend on additional outfits. It's just the truth.

If your computer needs more than a standard monitor, it's just not worth using. You could get years of access to an actually good computer with the money you spend on getting a second or third monitor. It's just the truth.
>>
>>102612438
Says who? Your thread is same stale jokes posted every single time, talk about self-awareness lmao
>>
>>102612453
true and based
>>
>>102612453
how would i wear a wardrobe
>>
>>102612453
>malding this hard
Ouch, anon hit right in nerves... You'll live, i think.
>>
>>102612341
grok
dbrx
arctic snowflake
phi
yi
falcon
openelm
orca
>>
>>102612332
would be a good argument if not for the privacy aspect
i'm not comfortable with some corporation storing my small penis humiliation logs
>>
File: 1716141216220407.jpg (761 KB, 1792x2304)
761 KB
761 KB JPG
>>102612477
Oh boy, where do I even start with this absolute genius over here? listen up, you brain-dead, english-mangling, subhuman troglodyte. When I said "wardrobe," I wasn't talking about some magical fucking piece of furniture that you can wear like a damn tuxedo. No, you absolute moron, in the context of my incredibly clever and witty example, a "wardrobe" refers to the collective shitstain of clothes you drape over your pathetic, pasty body to avoid being arrested for public indecency.
You know, like the shitty graphic tees, cargo shorts, and fedora collection that makes up the entirety of your fucking fashion sense? Yeah, that's your wardrobe, you ignorant, english-as-a-second-language-having, basement-dwelling fuckwit.
So, when I said
>If your wardrobe needs more than a standard set of clothes, it's just not worth wearing
I was making a fucking joke about how ridiculous it is to cheap out on your goddamn clothing options, not suggesting you try to wear your fucking IKEA furniture, you absolute cretin.
Now, do us all a favor and crawl back under the rock you came from, you illiterate, brain-damaged waste of oxygen. And next time, try using Google Translate before embarrassing yourself and your entire bloodline with your piss-poor grasp of the English language.
>>
File: miggguo.png (457 KB, 512x680)
457 KB
457 KB PNG
>>102612453
If your Model needs more than standard samplers it's just not worth using. You could get years of access to an actually good LLM with the time you spend getting DRY or Snoot. It's just the truth.
>>
>>102612332
they're still in the market and regulatory capture phase. Wait til investors flip the switch to "profit" mode and some middle-manager realizes he can monetize all that info you've pumped in there
>>
Boys, I'm doing some small experiments with an evolutionary algorithm and bitnet type MLPs, so far I've managed 60% on mnist with 90k parameters, which at least shows it can work. I also added a sparsity optimizer, which managed to zero 70% percent of the weights while upholding performance. Next up is testing how well it optimizes non layered networks, meaning any single neuron has the potential to connect to any other in the network. So much fun.
>>
>>102612592
OwO migu what happened to your eye
>>
all day and no new models makes /lmg/ a dull general
>>
>>102612662
>the time you spend getting DRY or Snoot
So, no time at all then.
>>
can someone make a mixtral 8x12b please i think that is a good spot
no need to credit me for the idea if you do it but it would be appreciated :)
>>
File: 172737606.jpg (11 KB, 376x253)
11 KB
11 KB JPG
>>102612672
>sparsity optimizer, which managed to zero 70% percent of the weights while upholding performance
>>
>>102612816
No, the serious companies are luckily turning their backs on the MoE meme
>>
>>102612672
Nice larp faggot
>>
>>102612332
https://youtube.com/watch?v=dHTvpUlWFbk
>>
>>102604394
>>latest model: a fucking cot finetune
It saddens me that normies buy so much into this grift lmao
>>
>>102604394
This is our new cope? Kek
>>
>>102612841
>MoE meme
truly spoken like one who has never used deepseek
>>
>>102612332
You are correct, but since its 4chan people are using to rp raping children so they don't feel comfortable using APIs
>>
>>102613046
There is literally 0 reason to use deepseek over qwen2.5. They are about equally smart and equally fucking boring. 72b just requires less resources without the moe performance tax.
>>
How can I get silly to show the model name instead of just the icon which can be hovered?
>>
>>102612736
This is the ultimate proof that there is no good local model, just look at aicg they are always active.
>>
>>102608657
Just the character description and changing the assistant role to {{char}}.
>>
>>102613131
because they actually use their models, we just throw a battery of riddles at them once and wait for the next one
>>
>>102613232
Isn't /aicg/ just proxy host drama?
>>
>>102613395
they make model cards and work on perfecting the art of prompting
>>
>>102613232
>we just throw a battery of riddles at them once and wait for the next one
True, but I assume we're also ALSO using them, but unless there's an interesting technical aspect we don't post anything in here. Constant logposting would probably grow tiring quickly if it isn't something pushing SOTA or the result of some interesting experiment.
>>
>>102613430
>Constant logposting would probably grow tiring quickly
I agree completely, I come here to learn about advancements occurring in the field as well as any weird and interesting aspects other people are working on. If it was just post after post of screenshots of roleplaying I wouldn't be here.
>>
>>102613463
>>102613430
True but the plainly visible problem is still here, with post after post of useless tests like counting r's, my model better than your model bullshit, oh no lecunny said THING on twitter again!, oh look at how much local models suck local is dead openai mogs, oh but actually no you're wrong openai is shit, etc etc. This place is shit, just a different flavor of it. Maybe if we actually had some regular activity we did in threads threads (or nipmoot actually cared about his website) things would be different.
>>
>>102613604
*some regular activity we did in the threads
>>
>>102613604
I think that's a mix of trolling, boredom and mental illness (and often all three in the same post)
I'd love if there was more focus in this general...when catching up after being offline for a while I scan and scroll past 90% of the posts...but this isn't reddit and wading through raw sewage for nuggets of gold is kind of the default 4chan experience.
I'd be totally up for regular, useful activity tho. eg. if we took new anon's frontend and started working on making it world class or working on actual decent TTS or musicgen or...I dunno? Something beyond bickering and infighting, anyways
>>
>>102612453
Now write that one for how much of a pain in the ass it is to get 2+ gpu's running in a standard case without buying everything again, And how you do that to run the current shivertastic models. I have a 4090 and I would even buy a 5090 without caring about money that much but it is so tiresome...
>>
>replies
>swipe a couple times
>pick and choose a few of the best parts from all of them and continue
>>
>>102613786
Could be an interesting extension to make.
Probe the model for N possible replies with different characteristics, then ask it to craft a final reply based on what it judges to be the best part of those N replies.
>>
>>102612012
Lick the salt. You are Iranian or a jeet
>>
i haven't been following recently, has any model dethroned mistral nemo finetunes for something you can run on 24gb vram for erp?

im mostly using gutenberg or theia 21b, not sure if there are better models out there
>>
>>102613727
>trolling, boredom and mental illness
Yes that is certainly the case, it's just that low activity means they get to fill the rest of the thread's posts with garbage.

As for a valuable activity for us to do, anything that involves a ton of work is probably unsustainable as people are naturally flaky with doing work. I would love if someone came up with an idea that actually worked to get good posters engaged in the thread but it's a tough thing to come up with. Maybe if we had some kind of game based on local models or something.
>>
>>102613885
Mistral Small BTFO Nemo and fits nicely in 24gb (I have not used either but notice that 22 > 12)
>>
>>102613952
RP Arena
>>
I've been using a 4bit quant of nous capybara 34b and i like the way it writes but it dropped ministration on me earlier today and now i'm going bald
>>
>>102612870
>>102613023
>>102613068
>>102613131
Shalom
Local isn't going away no matter how much you kvetch, sorry
>>
>>102613604
>>102613727
>>102613952
we just needs mods that care, but mods on /g/ only care about consumerism and eceleb shit while being openly hostile to anything /ai/
>>
What's the best model for uncensored roleplaying in Polish?
>>
*punches above your weight*
>>
>>102614186
Go back >>>/pol/
>>
>>102614095
I was going to say "needs a model hoster" but actually couldn't we do something decentralized? People here volunteer to host models essentially. Like maybe when they're asleep or something. And they make their model available to the horde, which gets used by an lmarena-like site. We'd need some mechanisms to make sure that the models people are providing aren't faked (so they can't just plug Claude in or something), but I don't think it would be impossible. Perhaps request for logits, and if logits don't match to someone else who has hosted the model before, then remove the provider. Also, we'd need a way to limit the prompts a bit, since this should be for RP only, and since it's very easy to "cheat" by asking a model who made it.

Unfortunately I don't have the time for such a project and I'm sure most here don't feel like that either. Plus I'm not sure it would actually really make the thread better since it's an external thing.
>>
>>102614190
Maybe there's a slightly higher chance of being able to get that to happen, but personally I think it'd be better if this site allowed an additional layer of user-based moderation. If someone makes a thread, they should be able to moderate it and set moderators for it. And the actual site moderators could then shift their work a bit more to moderating potentially bad usermoderators.
>>
>>102614306
You already have r/localllama.
>>
>complaining about people making on-topic jokes and trolling
>on 4chan
>"we need mods who actually care!"
you really need to go back
>>
>>102614426
>you really need to go back
newfags don't get to say this
>>
>>102613885
>theia 21b
>An upscaled NeMo with half its layers trained on my special sauce
>upscaled
Drummer is a hack and a fraud. And I am not memeing this time. Anyone who frankenmerges in current year admits he has no idea what the fuck he is doing.
>>
>>102614373
This subreddit is event worse than this general. Way too many retards and like half the posts are about OAI/Anthropic.
>>
>>102614231
I am not going to spend hundreds of dollars in electricity so retards here can make models count strawberries on my hardware.
>>
>>102614190
>>102614306
>Actual herd that wants power-tripping daddy mods ITT
When did /g/ become reddit 2.0?
>>
>>102614497
>t. election tourist
>>
>>102614306
>If someone makes a thread, they should be able to moderate it and set moderators for it.
>Let's give trolls an actual incentive to fight over bakes
>>
>>102614306
ywnbaw
>>
>>102614306
>If someone makes a thread, they should be able to moderate it and set moderators for it.
This is suck a terrible idea holy shit.
>>
I joke a lot about it being over but after watching this faggotry I think it's actually, unironically over. I've never seen /lmg/ so buck broken.
>>
>suddenly a bunch of excuses come flooding out of the woodwork to try and justify the literal garbage state of things
>guys it's ok, our country is full of the homeless and poor but at least we're "free" whatever the fuck that means
>>
>>102614306
How would you even select moderators? Everyone here is >>Anonymous<<.
>>
File: file.png (24 KB, 688x226)
24 KB
24 KB PNG
Local won
https://www.reddit.com/r/LocalLLaMA/comments/1fselw2/newsom_vetoed_sb1047/
>>
>>102614516
>newfag parroting le epic oldfag meme
It's been 8 years, pack it up you slimy troglodyte.
>>
>>102614540
>>102614557
NTA but I think his point was that if users are expected to moderate their own threads, moderaters would spend their time moderating the user moderators instead of the full bulk of all of the content, so theoretically trolls would get banned from starting threads. but bans have no wieght here and again, mods are actively hostile to ai generals so would probably just continue encouraging the trolls anyway
>>
>>102614625
just don't be anonymous faggot, what are you trying to hide?
>>
>I want /lmg/ to be more like reddit
why not just go to reddit? honest question.
>>
>>102614639
This is exactly the kind of shit that deserve a permanent ban. Why the fuck would you post a reddit link and screenshot of a reddit post with a news link instead of the news link itself. Why are redditors so fucking retarded. Holy shit go back go back go back go back go back go back go back go back
>>
>>102614479
>so retards here can make models count strawberries
KEK, the absolute state of machine learning
>>
>>102614644
The only reason I use 4chan is to be able to speak my mind freely without the fear of being censored, your idea would allow this censorship to happen.
>>
>>102614642
>>102614677
this isn't your edgy secret culture war club
>>
>>102614679
it's 1 (one) click away, stop being a fag
>>
>>102614677
reddit's UI is too confusing to me and I can't say nigger there if the need arises.
>>
>>102614707
didn't say it was. I'm genuinely curious why /lmg/ needs to be managed like a subreddit when there are so many subreddits (and the option to create new ones) dedicated to this topic.
>>
>>102612270
>Mikufag
>Shit taste
Every single time.
>>
>>102612012
no u
>>
>>102614639
OpenAI and Anthropic absolutely BTFO.
>>
>>102614721
because retards like you keep coming here and repeatedly shit the place up
>>
File: file.png (58 KB, 667x618)
58 KB
58 KB PNG
this is quite exciting ngl
>>
>>102614748
>ask very simple question
>"culture war"
>"retards like you shitting"
>"everything needs to be tailored around my desires"
seems like the problem is you
>>
>Take over thread creation duties, Keep deleting a certain users post and ban them
>Do this to someone new every single thread, slowly driving away users from the general
It would be so easy as well. Not to mention how many generals would start fragmenting and splitting off because they don't like the moderations of that specific thread. It's a disaster waiting to happen. The current method of reporting and ignoring works perfectly, its just that some users don't bother to report and complain that a bunch of off topic schizoposting is ruining the thread. It's like complaining about politics when you don't vote, you bring it on yourself.
>>
>>102614707
>culture war
Yep, we have an actual tourists here. I see this "culture war" epic gotcha spammed all over /v/, it jumpstarted in 2021 and gone haywire since first SBI threads in there.
>>
>>102614750
it will be garbage, meta is dead
>>
>>102614625
This may come as a shock but there can be optional accounts that exist for specialized purposes, such as the 4chan pass, for potential moderators.

>>102614540
The site moderators move a portion of their work to moderating the usermoderators. It's really not anything different from just moderating to begin with, except they are checking over moderation decisions and making sure the users who are moderating are doing it properly.

>>102614644
Yeah this.

>>102614694
I'm suggesting it'd be implemented in a way to prevent that from happening. Again it's still on the site moderators, except now they can delegate a bit more of the work to users. If the usermods are censoring things more than the regular mods would (for a thread they actually cared about lmao), then that would be grounds for banning the particular usermod depending on the severity of the "abuse".

Anyway, obviously it might be an imperfect idea and there would be kinks to work out. But at least it would be something different and they could say they genuinely tried.
>>
>>102614774
>"culture war"
>jumpstarted in 2021
not even an election tourist, you just outed yourself as a pandemic baby
>>
What if we implement quality control or a "Karma" system. Basically we can rate a post positively or negatively and that is their rating across the entire site. If their Quality rating gets too low they get auto banned and their "Karma" resets when they get unbanned.
>>
>>102614306
Kill yourself.
>>
>>102614800
Mother fucker I bought the 4chan pass so I can bypass the Captcha and post while I am connected to my VPN, not so I can be a potential moderator.
>>
>>102614750
>Some crappy e-celeb TTS voices slapped on top
Who cares? It's not true end-to-end and will have annoying latency.
>>
>>102614820
Not a bad idea, but how do they get unbanned?
>>
>>
>>102614765
>Keep deleting a certain users post and ban them
You would get caught by the mods through various mechanisms.

>The current method of reporting and ignoring works perfectly, its just that some users don't bother to report
I do sometimes. It only works some of the time.
>>
>>102614750
jesus christ who cares about retarded gimmicks? fuck off
>>
I suggest we (the people that actually care about local models) start posting only using steganography like desu desu talk, this way we can continue here while at the same time having a clean environment free of trolls and tourists.
>>
>>102614808
I never seen "culture war" bullshit spammed before 2021. It had no meaning and never will, everyone who says it is either a troon or redditor tourist with no exceptions.
>>
>>102614835
You have to post a 2000 character apology in the general you got the most low quality reports in. After which if people vote if its low quality and if it is you get banned and have to write a better one, or if it gets voted as high quality you get unbanned.
>>
>>102614862
charge your phone anon...
>>
>>102614863
I find that it works, most of the time rather then some of the time. But that is not to say it is instantaneous. Might take an entire 24 hours for a mod to get around to deleting posts or banning a user. It is not a fast system but it is a system that more often then not actually works.
>>
>>102614824
Seethe.

>>102614832
That's obviously not what I meant. The point was that 4chan pass is an example where accounts exist. We can have separate accounts, or some unified account system with modular features, depending on what they'd potentially want to do.
>>
>>102614750
>>102614862
Total zoomer death.
>>
>>102614870
>troon
Go. Back.
>>
>>102614893
Well maybe I'll try it more, but the high delay is unfortunate. Perhaps what's needed is a solution to the delay, instead.
>>
>>102614748
Be a faggot somewhere else please.
>>
i was joking before when i wrote that this general couldn't be more dead, but now I'm 100% real: this general couldn't be more dead.
look at the current state, wtf.
>>
>>102614863
>You would get caught by the mods through various mechanisms.
So in the end you give mods more work because they now need to examine if someone is abusing mod powers or if the deleted posts really were trolls, how does this help exactly?
>>
>>102614909
Sitting here since llama-1 leak and thus the beginning of lmg, cry more bitch.
>>
>>102614497
There already are power-tripping tranny mods ITT.
>>
>>102614941
show proof
>>
/lmg/ - Local Reddit Chatroom
>>
>>102614930
The theory is that it'd be less work because they can spend more of their time moderating mod actions rather than the huge pool of posts across the board. Of course if it turns out to be more work then the system could be discarded and things can go back to the way they were, or it could be improved with further ideas. The only way to truly know would be to try things out, but that seems to be too much work for Hiro.
>>
>>102614865
this sounds like a stupid idea, just like you. you must be very stupid and has no place being here.
>>
>>102614941
everybody watchout! we got an oldfag over here!
>>
I say that we REMOVE the human element of /lmg/. The only human post should be the OP post, all the rest are AI models replying to either the OP or other posts made by AI models.
>>
>>102614928
Wtfffff bros like what the hell anons shitposting????? I can't believe this!!!!
>>
Guys what if they just used AI to help moderate. Like ultimately humans would still be making decisions but AI could flag posts, determine priority, etc.
>>
>>102614989
this. let nemo decide who goes and who stays.
>>
>>102614989
with the positivity bias of models you'd get 99% of posts flagged instantly
>>
>>102614984
frfr no cap, if you catch my drift
>>
>>102614975
See >>102614922
>>
>>102614989
>>102615000
>let models trained on reddit decide who goes and who stays.
this is just automated reddit moderation
>>
>>102614989
The AI's would flag you because you called someone else a fucking retard because they asked why they couldn't run 70B on their GTX 1080. There is no model that currently exists that can properly replace the role of a 4chan moderator. Reddit mods maybe, but not 4chan mods.
>>
How can a dead thread be so lively?
>>
File: file.png (13 KB, 689x101)
13 KB
13 KB PNG
kek
>>
>>102615031
Dead internet theory.
>>
Ok but what if we just used gpt-4chan as the model.
>>
>>102615015
>I've been here since 2020 why aren't you showing me the respect I deserve?
waaaa
>>
>>102615035
Very real and honest post as mythomax is somehow still the most used on OR.
>>
>>102614989
They already using this, you can get banned of racism or any similar stuff here on /g/.
>>
>>102615030
? I see no problem in thriving a healthy discussion environment. the kids should begone.
>>
>>102615065
That's not an AI doing it, that's an actual human janny doing that.
>>
>>102615077
A healthy discussion environment would be one where you can call someone a fucking retard for being and I quote. A "fucking retard".
>>
>>102615035
Why does this redditor feel the need to spend the day reposting shit from his website?
>>
>>102615097
Yeah, for sure, because the best discussions are the ones where the posters constantly use crass language towards each other. As if.
>>
>>102615051
>putting words in people's mouths
Refer to this, again >>102614922
>>
>>102615097
I agree. That's why we need gpt-4chan as the moderator.
>>
File: 252.png (7 KB, 322x76)
7 KB
7 KB PNG
>>102615109
is this you?
>>
>>102615109
Why not? It's the only on topic content that has been posted itt
>>
>>102615109
stop being toxic, you are the problem.
>>
>>102615112
If calling someone who is being a retard, retarded, a deal breaker for you than why are you even here?
>>
>>102615112
Always was, if you want extremely censored & sanitized space - join your r*ddit buddies.
>>
So, in theory, these multimodal models can output text and image on the same generation, right?
Something like
>char is in <place> and is doing <so and so>
>{image of the scene}
correct?
If so, that's awesome for RP.
I know that there are workflows to use an image generation model and a text gen one in tandem (or in series really), but I can only imagine that having it all come out of the same model with the same context in the same inference step should work better.
Or am I misunderstanding something fundamental here?
>>
Anyway I think I'm done with this discussion. It's just wasting time honestly since nothing's going to be done about it anyway.

So how about Qwen? People were saying it was shit but then yesterday some people called it claude-level. I'm testing it and honestly rather not impressed, I'd love to be able to copy and paste something into Mikupad to see the good outputs some people supposedly got. Or maybe it was just one guy claiming that, it's a blur in my head by today.
>>
>>102615163
>>102615164
9 times out of 10, someone who uses the word "retard" ends up being a troll who adds nothing to the discussion. That is my only problem with it. If you're defending this you must be a troll and should leave.
>>
>>102615211
if you don't like it you can leave
>>>/r/eddit
>>
>>102615195
>some people called it claude-level
You fell for trolling. Always hold "local is worse than cloud" in head before using stuff shilled ITT.
>>
Why did Mistral stop releasing base models?
>>
>>102615193
nah, most vlms currently can only take images as input. chameleon was image text in and out but was neutered for official release
>>
>>102615195
I thought it was ok but truthfully it sometimes has trouble following instructions. The thing loves to speak, if you were to ask it why is the sky blue it would say yes and then continue yapping on for ages. I asked it in the past to keep its answers short and relevant to question but it just ignored that part and kept yappin regardless.
>>102615211
Retard
>>
>>102615220
>le go back meme
typical troll behavior that adds nothing to the discussion, thanks for proving my point.
>>
>>102615220
If you don't want to discuss local models you can leave
>>
>>102615228
too unsafe for the goyim
>>
>>102615228
Too "dangerous".
>>
>>102614750
Local won! We beat OpenAI!
>>
>This entire thread
>>
>>102615193
Which ones? You're correct if the assumption was that the model in question had good quality outputs and it was supported well by both backends and frontends. Unfortunately that doesn't seem to be the case. Models that can output images and text don't seem to hold up well compared to SD/Flux, and on top of that, nothing really supports them.
>>
>>102615241
You don't add anything of value either.
>>
File: file.png (16 KB, 622x165)
16 KB
16 KB PNG
oooooo
>>
>>102615275
Nani the fuck.
>>
>>102615275
link?
>>
>>102615275
meta gonna give us llama 3.99 to never release one called llama4 huh
>>
>>102615275
Whoa, new filtered slop is coming!!!
>>
>>102615236
>most vlms currently can only take images as input
Really? Well, that's shit.
I see that Chameleon still exists on huggingface, by
>but was neutered for official release
Do you mean they removed the image output capabilities or something?
I remember it coming out, but I was too busy back then to check it out or to follow the discussion back then.

>>102615262
>Models that can output images and text don't seem to hold up well compared to SD/Flux
Ah, I see.

>Which ones?
I wasn't thinking of any in specific, just in general, really. I never messed around with these vision models.
>>
>>102615275
That's a strange way to announce that. They've never so much as hinted as working on a MoE model before.
>>
File: 1697900235394863.png (17 KB, 594x211)
17 KB
17 KB PNG
>>102615275
We are so back
>>
>>102615335
Didn't they release some moes before without the llama brand?
>>
>>102615275
I can't wait to hear John Cena tell me what's important to remember.
>>
Bros, imagine 8x405B, holy shit... What kind of CPUmaxx system could run it?
>>
>>102615275
I can't wait for my choice between 11x0.5B and 11x405B
>>
>>102615356
Not from what I remember, but their research team isn't the llama team. Chameleon wasn't by the same people either.
>>
>>102615351
You just know it's gonna be shit, calm the fuck down.
>>
>>102615326
>Do you mean they removed the image output capabilities or something?
exactly correct
>A restricted, safety aligned (no-image-out) version of Chameleon (7B/34B) is now open-weight!
>>101044333
>>101044321
>>
People will finally get that 34B they wanted. A 8x 34B
>>
Why didn't they announce this at Meta Connect literally just days ago? Very weird.
>>
>>102615195
>So how about Qwen? People were saying it was shit but then yesterday some people called it claude-level.
it's close-ish to claude on reasoning and programming tasks, but far, far worse for RP
smart models though
>>
>>102615351
this has scam vibes, it's so ober
>>
>>102615326
>Do you mean they removed the image output capabilities or something?
Anole put it back in
>>
>>102615398
>exactly correct
Well, fuck my asshole.
>>
>>102615405
>but far, far worse for RP
Its about as good or better at that BUT its ultra censored and needs a jailbreak. For SFW its the best atm
>>
>>102615351
>Creative crowd
>CREATIVE CROWD
Holy shit the opus killer is here
>>
>>102615404
Nobody would care about 3.2 if they let it slip that 3.3 was only a week away?
>>
What's the best EPYC platform if I want to go the CPU route?
>>
>>102615404
I didn't know there was a Meta Connect some days ago
>>
>>102615420
>For SFW its the best atm
Sus
>>
>>102615436
They have like 100k H100s, if they have the dataset / method ready then they could train these .1 improvements in a few hours
>>
Why not just release Llama 4
>>
>>102615351
KINO IS BACK IN THE MENU
>>
>>102615445
He is the horse fucker
>>
>>102615447
But most of that compute is used on their Facebook/Instagram/etc services.
>>
>>102615463
No, that's me.
>>
>>102615459
same reason oai is on gpt4-o-with-reasoning and not gpt5, managing expectations
>>
>>102615459
They need to make sure llama 4 has no trace of a soul.
I don't know why people in this sub still care about Meta models, they don't deserve our hype. This has become evident after every release and people still come back to hype them like a bandit's wife, it's unbelievable.
>>
>>102615420
no it's not, be for real
I'm a massive qwen apologist and always have been, their models are usable for RP with reasonable prompting but it's definitely not close to claude, the cultural knowledge gap is way too large. claude just knows way more about what good RP looks like and how to do it, qwen is smart but sheltered and robotic
>>
>>102615507
Oh? And who is left? Cohere who just finished ripping the "soul" out of their old models? Mistral who abandoned open source once? The Chinese? Really who else is there?
>>
>>102615275
>>102615351
Literally where are you seeing these posts? It's not appearing on his account for me.
>>
Try out hf-100/Llama-3.1-Spellbound-StoryWriter-70b-instruct-0.4-16bit
>>
>>102615514
skill issue
>>
>>102615507
>people still come back to hype them like a bandit's wife
Stockholm syndrome.
>>
>>102615519
DeepSeek, therefore yes, the Chinese.
>>
>>102615529
on the part of the qwen pretraining team, yes, I agree
>>
>>102615521
Are you following him?
>>
>>102615418
This thing
https://huggingface.co/GAIR/Anole-7b-v0.1
?
>Anole is the first open-source, autoregressive, and natively trained large multimodal model capable of interleaved image-text generation (without using stable diffusion). While it builds upon the strengths of Chameleon, Anole excels at the complex task of generating coherent sequences of alternating text and images. Through an innovative fine-tuning process using a carefully curated dataset of approximately 6,000 images, Anole achieves remarkable image generation and understanding capabilities with minimal additional training. This efficient approach, combined with its open-source nature, positions Anole as a catalyst for accelerated research and development in multimodal AI. Preliminary tests demonstrate Anole's exceptional ability to follow nuanced instructions, producing high-quality images and interleaved text-image content that closely aligns with user prompts.
Alright, cool.
Gonna play around with it.
>>
>>102615565
No? Are you saying these posts were set to only be visible to followers?
>>
>>102615351
>exciting weeks
how many weeks, exactly?
>>
>>102615597
2
>>
>>102615597
2
>>
>>102615597
Two more exciting weeks.
>>
>>102615597
i4 weeks
>>
>>102615521
that's because it's fake
>>
anyone have good settings for ERP with mistral small 2409? do most people just adjust temp, min p, top p themselves while experimenting? not sure about DRY or smooth sampling
>>
>>102615724
Use Lyra4-Gutenberg-12B instead if vramlet
>>
Anyone else have qwen 3.5 32B suddenly switching to Chinese in the middle of an answer? Never happened with 13B.
>>
File: lolisniffer.png (360 KB, 485x520)
360 KB
360 KB PNG
>>102615275
>9 experts
>8B each
>0 sex knowledge
>>
>>102615724
For non-erotic RP I was running with min-p 0.055 and rep penalty 1.02 and that seemed to work well although I didn't find mistral small perfect.
>>
>>102615591
I don't know, I don't have Xitter
>>
>>102615521
Tourist zoom zooms messing around, ignore it.
>>
>>102615759
歡迎來到Qwen!
>>
>>102615783
LocalLlama has a no fun allowed policy you may enjoy
>>
>>102615806
Nah i'm good.
>>
>>102615776
I actually wanted min-p 0.049 but I increased it due to compensate for exl2 causing jitter in the probability of the top token.
>>
>>102615597
Two weeks with the big exciting llama 3.5 opus killer at the end of it
>>
elemgy will believe anything if you post it in an x or reddit screencap
>>
>>102615921
Zoomers do be like that.
>>
>>102604225
friendly reminder that literally each and every one of you is a social undesirable who will die alone and miserable ;)
>>
>>102616127
Just like you sir.
>>
>>102616127
Hi jart.
>>
>>102616127
I have my models. Nobody can take them from me since I run them locally.
>>
File: 1727654387697.png (990 KB, 1024x1024)
990 KB
990 KB PNG
>>
Why does a model that can run on ngl x, context y, and does so for hours, suddenly start to OOM? it's hip not cuda if that matters
>>
>>102616270
And you don't have complete control over them, enjoy your occasional refusals thanks to hallucinations & rng despite some advanced prompt jailbreak in use.
>>
>>102615759
> suddenly switching to Chinese in the middle of an answer
a 'feature' of qwen since v1, completely useless outside of looking good on benchmarks as usual.
When will /lmg/ learn?
>>
>>102616374
I fixed it by using llama.cpp grammar and forcing only ascii.
>>
>>102616287
>broken hands
Too bad, woulda saved.
>>
>>102616609
>>102616609
>>102616609
>>
>>102616386
Eyes fucked up too
>>
File: 00361-2768319313.png (282 KB, 512x512)
282 KB
282 KB PNG
>>102614725
>Oh look it's that one schizo who complains about Miku all the time
Cry harder
>>
All mikuposters are pedophiles.



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.