[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/g/ - Technology


Thread archived.
You cannot reply anymore.


[Advertise on 4chan]


File: ComfyUI_00969_.png (1.3 MB, 1256x1024)
1.3 MB
1.3 MB PNG
/lmg/ - a general dedicated to the discussion and development of local language models.

Previous threads: >>108434876 & >>108429328

►News
>(03/17) Rakuten AI 3.0 released: https://global.rakuten.com/corp/news/press/2026/0317_01.html
>(03/16) Mistral Small 4 released: https://mistral.ai/news/mistral-small-4
>(03/11) Nemotron 3 Super released: https://hf.co/nvidia/NVIDIA-Nemotron-3-Super-120B-A12B-BF16

►News Archive: https://rentry.org/lmg-news-archive
►Glossary: https://rentry.org/lmg-glossary
►Links: https://rentry.org/LocalModelsLinks
►Official /lmg/ card: https://files.catbox.moe/cbclyf.png

►Getting Started
https://rentry.org/lmg-lazy-getting-started-guide
https://rentry.org/lmg-build-guides
https://rentry.org/IsolatedLinuxWebService
https://rentry.org/recommended-models
https://rentry.org/samplers
https://rentry.org/MikupadIntroGuide

►Further Learning
https://rentry.org/machine-learning-roadmap
https://rentry.org/llm-training
https://rentry.org/LocalModelsPapers

►Benchmarks
LiveBench: https://livebench.ai
Programming: https://livecodebench.github.io/gso.html
Context Length: https://github.com/adobe-research/NoLiMa
GPUs: https://github.com/XiongjieDai/GPU-Benchmarks-on-LLM-Inference

►Tools
Alpha Calculator: https://desmos.com/calculator/ffngla98yc
GGUF VRAM Calculator: https://hf.co/spaces/NyxKrage/LLM-Model-VRAM-Calculator
Sampler Visualizer: https://artefact2.github.io/llm-sampling
Token Speed Visualizer: https://shir-man.com/tokens-per-second

►Text Gen. UI, Inference Engines
https://github.com/lmg-anon/mikupad
https://github.com/oobabooga/text-generation-webui
https://github.com/LostRuins/koboldcpp
https://github.com/ggerganov/llama.cpp
https://github.com/theroyallab/tabbyAPI
https://github.com/vllm-project/vllm
>>
File: 1745000438071299.jpg (223 KB, 2439x1807)
223 KB
223 KB JPG
►Recent Highlights from the Previous Thread: >>108434876

--CUDA optimization PR sparks LLM-assisted development debate:
>108437073 >108437094 >108437528 >108437535 >108437550 >108437567 >108437569 >108437582 >108437846 >108437676 >108437892 >108437907 >108437909 >108438172
--OpenClaw dual-model coding workflow optimization:
>108439481 >108439578 >108439716 >108439719 >108439731 >108439742 >108439747 >108439772 >108439770 >108439821 >108440019
--llama.cpp server excludes Responses API server-side agentic loop due to C++ maintenance cost:
>108437944 >108438326 >108438376
--Debating NVIDIA AGI claims and feasibility:
>108439814 >108439835 >108440572 >108440630 >108439859 >108440096 >108440176 >108440413 >108440451 >108440468 >108440475 >108440478 >108440491 >108440511 >108440326 >108440335 >108440370 >108440388
--Troubleshooting reasoning mode activation in Qwen models post-autoparser:
>108435077 >108435086 >108435269 >108437362 >108435294 >108435323 >108435332 >108435341 >108435359
--Qwen 3.5 4B and KV cache quantization debates:
>108439408 >108439435 >108440867 >108440876 >108440928 >108440953 >108441003 >108441044 >108441155 >108441210 >108441270 >108441515 >108441564 >108441636
--Model recommendations for limited VRAM/RAM setups:
>108437524 >108437530 >108437534 >108437557 >108437563 >108437624 >108437636 >108437672 >108437700 >108437836 >108439050 >108439086 >108439098 >108439123 >108439129 >108439131 >108439156
--Debating Anthropic's closed-source Claude Code SDK strategy:
>108435933 >108435942 >108437833
--Kimi k2.5 admitted to be Cursor's Composer-2's base model:
>108435414
--AI solving previously unsolved math problems via FrontierMath:
>108439710
--Anon successfully implements LLM-generated voice activation for PC control:
>108441088
--Miku and Dipsy (free space):
>108436067 >108441064 >108441515 >108441560 >108435820 >108441286

►Recent Highlight Posts from the Previous Thread: >>108434877

Why?: >>102478518
Enable Links: https://rentry.org/lmg-recap-script
>>
Fuck yah a Teto thread. Time to get fucknig wasted and snort cadmium.
>>
Not llm related as such except I've been rewriting some client stuff.
Key insight:
Stop using vim.
You'll end up working 3 times as much for a simple operation. When you need to concentrate on using the keyboard more than actually just typing it out there's something wrong.
I always thought that vim was fun to use, well it is if you just edit config files, but for anything larger it's just torture and waste of time unless you are a masochist.
When you switch text editor it's a massive improvement, as if computing just advanced 40+ years in a moment.
That's all.
>>
>>108441780
>not waiting for a rin thread so you can inject yellowcake
OFF MY BOARD NORMIEEEEEEEEEEEEEEEEEEEEEEEEEEEEEEE!!!!!!!!!!!!!!!!!!!!!!!
>>
>>108441815
Emacs hands wrote this post. Go back to your lisp machine you Sussman-worshipping fogey.
>>
>>108441826
I'm using notepad.
>>
>>108441815
Maybe learn vim before attempting to use it?
>>
>>108441815
>>108441862
I use nano.
>>
>>108441862
proving his point
>>
what's the point anymore
>>
>>108441979
just keep drinking
>>
i'm going back to playing video games
>>
v4 is so close i can taste it
>>
>>108441873
>not using ed
/g/ has fallen
>>
>>108441999
Have fun, Anon. Play with me?
>>
>>108442003
more like Vnever
>>
>>108441979
what *was* the point?
>>
>>108442013
Ed is the standard text editor. When I use an editor, I don't want eight extra KILOBYTES of worthless help screens and cursor positioning code! I just want an EDitor!! Not a “viitor”. Not a “emacsitor”. Those aren't even WORDS!!!! ED! ED! ED IS THE STANDARD!!!
I miss /prog/.
>>
File: ed.gif (1.03 MB, 352x240)
1.03 MB
1.03 MB GIF
>>108442013
>>
that exhilarating feeling when u pull... and build and deploy. makes my pp go all tingly
>>
>>108442176
what about your tg
>>
File: 1758029501384094.png (12 KB, 968x67)
12 KB
12 KB PNG
>>108442187
heres my pp and tg :)
>>
Any macfags tried compiling llama.cpp with PGO?
>>
File: >ubuntu<nl>SHUT UP.png (22 KB, 1004x79)
22 KB
22 KB PNG
>>108442192
teehee
>>
>>108441862
Where did I mention that I was new to vim? Fuck you.
>>
>>108441780
>snort cadmium
https://www.youtube.com/watch?v=1U6qefKcOrg
>>
>>108442241
teto is good and all,
but realistically rin is where it's at.
>>
>>108442209
>You'll end up working 3 times as much for a simple operation. When you need to concentrate on using the keyboard more than actually just typing it out there's something wrong.
Gave it away
>>
link ur favorite legal character cards plz.
>>
>decide to try containerizing my LLM to docker (windows host)
>get 1/3 of the tokennage (llama-server cuda image)
AIEEEEEEEEEE do I have to move to loonix to get max llm perf?
>>
>>108442373
you mean like a lawyer?
>>
>>108442416
yeah
>>
I can't wait to Install my first local model on my new mac mini tonight.
Should I get a tattoo to celebrate?
>>
>>108442448
I got 2nd hand cringe
>>
File: 1746553876528008.png (228 KB, 500x441)
228 KB
228 KB PNG
>>108442448
>>
I'm trying to think of an appropriate candlejack analogy here because this shit is so anno
>>
>>108442488
I'm fighting against a shiver.
>>
>>108442493
I'm smoothing my skirts and looking at you through my lashes.
>>
>>108442448
I don't know. Qwen 3.5 9B can do a C function what replaces 'source' with 'destination' in 'my string'
>void (char *my_string, char *source, char *destination);
But it couldn't work out how to do replace every occurrence of source with destination. It always failed with string length allocation. 10+ tries.
>>
>>108442528
>a C function what replaces
aaa
>>
>>108442531
My point being that model is probably being shilled bit too hard on internet right now. But of course bla bla bla and stuff.
>>
>>108442528
I'm too drunk to read through the response, but I suppose the punchline is that the prompt's method definition includes neither __restrict__ nor a length for my_string.
27B-q8_0-heretic, curious to see other models.
>>
>>108442577
This won't compile.
>>
>>108442528
It's shilled together with hermes
>>
>>108442448
I'm not sure I agree but at least it's cute and gives some ideas.
>>
File: EaogWFwU4AAyn8k.jpg (194 KB, 2048x1152)
194 KB
194 KB JPG
>>108442577
>I'm too drunk to read through the response
Now this is vibecoding. Time to ship.
>>
>{{user}}: You know what would be cool? If we went to the tattoo parlor and got "I'M A STUPID FUCKHEAD" tattooed right on your forehead!
>{{char}}: HAHAHA! That'd be amazing! Let's do it! I'm going online right now to pick out the style of writing!
>>
File: 1771959253436010.png (185 KB, 791x1166)
185 KB
185 KB PNG
interesting
>>
>>108442674
kino
>>
File: 1773862801077647.png (442 KB, 1920x1080)
442 KB
442 KB PNG
>>
Thought I'd try out OpenClaw but seems like it is really pozzed. This website is a big no no.
>>
I'm broke but i really want a mac mini to put an agent in, 128gb Ideally.
>>
>>108442720
>just run a 4B q4 LLM (totally local only!) that answers all your emails and social DMs for you bro! your life will be so much better!!
>>108442732
just steal one
>>
literal free higher perf models we're back https://www.reddit.com/r/LocalLLaMA/comments/1s1t5ot/rys_ii_repeated_layers_with_qwen35_27b_and_some/
>>
>>108442720
So try hermes
>>
File: jezebel whores.png (45 KB, 1660x994)
45 KB
45 KB PNG
>>108442708
It do be like that after building GPU rig
>>
>>108442747
>Wen GGUF? When someone GGUF's them I guess?
he really thinks someone else will waste his time making GGUF on his highly experimental shit? he's delusional as fuck
>>
>>108442769
ballmuncher team literally ggufs anything they find so why not this
>>
File: 1750991643516076.png (193 KB, 1725x940)
193 KB
193 KB PNG
>>108442747
basically he showed that duplicating only one extra layer at the middle of the model can improve the model a lot, that's interesting
>>
>>108442809
you wouldn't even need to duplicate the layer, forcing the model to reroute to the identified layer would achieve the same result and wouldn't need this retardness about having yet more models on drives.
>>
File: this you?.png (127 KB, 549x613)
127 KB
127 KB PNG
>>108442822
damn, you're onto something anon...
>>
finally gave qwen 3 35ba3b a shot and its pretty rad but, watching it have trouble deciding whether or not to "allow nsfw" then proceed to allow nsfw but almost gemini-like censored is kind of annoying, are uncensor models still a total meme or is there at least one good one?

>>108442674
now inform it about ludokino.
>>
>>108442848
hauhaucs 35b is godlike
>>
File: merge.png (203 KB, 920x919)
203 KB
203 KB PNG
>>108442809
>>108442838
Stop encouraging mergesloppas
Seems once a year someone gets the fresh new idea to copy paste transformer layers
>>
>>108442848
>at least one good one
use one of hauhau's or a heretic abliteration
hauhau will break your chain-of-thought 30% of the time, heretic will more frequently get stuck in loops
>>
>>108442859
>or a heretic abliteration
those one are a shit
>>
>>108442857
>B-but, muh Mythomax
it was just luck, we never managed to get something like this again
>>
>>108442857
>Stop encouraging
yes we can only doom and we should all saas right now, thank you sir
>>
File: xina bing chilling.jpg (41 KB, 500x500)
41 KB
41 KB JPG
>>108442854
>>108442859
thanks pals.

>>108442866
>mythomax
now that's a name i haven't heard in a long time.
>>
speaking of https://www.reddit.com/r/LocalLLaMA/comments/1s298y6/request_training_a_pretrained_moe_version_of_nemo/
> I converted Mistral Nemo from a dense model into a sixteen expert MoE model: https://huggingface.co/blascotobasco/Mistral-NeMoE-12B-16E
>>
>>108442892
I'd like to get the opposite, transform Qwen 3.5 35b MoE into a dense model, it'll be smarter than the 27b model
>>
File: keks.png (10 KB, 264x85)
10 KB
10 KB PNG
>>108442892
>>
>>108442892
oh no no ai !psychosis
>via the shattering method — a dense-to-MoE structural transformation developed as part of the Nebula Structural Modification Suite.
>By Phase 3, coherence was restored. The final phases focused on knowledge distillation, logical reasoning, and instruction alignment.
>The shattering process destroyed the original model's capabilities entirely, and the training curriculum rebuilt them from scratch under significant budget constraints — this is a student project, not a well-funded lab release.
>Rich atmospheric prose — responds well to detailed character context
>High-fidelity roleplay — adopts personas naturally from system prompt descriptions
>This model was produced entirely using the Nebula Structural Modification Suite, a self-designed framework for extensive structural modification of language models. Tools used in this model's production include:
>>
File: file.png (67 KB, 886x328)
67 KB
67 KB PNG
Has Jensen lost it?
>>
>>108442945
I hate when those fuckers are even writing their post with an AI, they don't even see pride in writing in their own style, it's just sad
>>
>>108442981
no he's won it
>>
>>108442981
>yesterday
>>
>>108442988
Still no AI GF so I'm calling bs
>>
>>108442994
just lower your standards bro, same as it's always been
>>
File: bing-merge.png (60 KB, 771x286)
60 KB
60 KB PNG
>>108442877
His face inspired me
>>
File: SAME AS IT EVER WAS 1.gif (3.39 MB, 496x368)
3.39 MB
3.39 MB GIF
>>108443001
YOU MAY FIND YOURSELF
ERP'ING IN SILLYTAVERN WITH YOUR WAIFU

>>108443006
lol'd
>>
>>108443006
Do not gandong the mikus.
But I must.
>>
>>108443001
All I ask for is something believable.
>anonymous has achieved AGI
Is just as believable without any corroboration.
>>
>>108442577
>Lara tucks a stray strand of auburn hair behind her cat ears
HAHAHAHAHAHAHAHA
>>
>>108442475
>>
>>108442809
It will enforce model's knowledge in certain areas because the layers are duplicated, but it'll be more retarded in the end. Or at least this is how I understand this (doesn't matter lol).
>>
https://github.com/ggml-org/llama.cpp/pull/18322
this merged when??? ggnigeranov??????
>>
>>108443140
why do you care about it??????
>>
>>108443140
>add unsafe
that doesn't sound very safe...
>>
>>108443148
because I use llmao-server in router mode and I have the same fucking model duplicated in the config 4 times depending on the context size but this change would make it so I just need to pass an extra param in the request!?!??!!?!??!!?!?!?!?!?!?!?!?!?!?!?
>>
>>108443140
it says "Draft" anon, the guy who made the PR hasn't finished it yet
>>
>>108443140
this has already been supplanted by the dynamic model routing feature
just get /models and post ?model=sex to the completion endpoint and it just werks
>>
>>108443155
oh I see!?!??!!?!??!!?!?!?!?!?!?!?!?!?!?!?
you can always use the pr yourself!!!!!!!!!!
>>
>>108443158
are you retarded? this is to override the defaults you specify in the config!!!!!!!!!!!!!!!!!! LEARN TO READ!!!!!!!!!!!!!!!!!!!!!
>>
>>108443166
that's the dumbest shit i've ever read?!?!?!??!!??!!!!??!?!!!!!?!?!?!!?!??
>>
>>108443157
it is done for all intents and purposes!??!?!?!?!?!?! ngxson is just a lil bitch scared about security advisories and bughunts?!?!?!?!?!?!?!!?!?!?!?!?!?
>>
>>108443162
>>108443166
>>108443169
>>108443170
this conversation reads like erp quality degradation after downloading one of those shitty merges quanted to anything lower than q8
>>
>>108443173
I'm so drunk Ic an't even tell which posts are ironic anymore I'm laughing my ass off
>>
File: deepseek nigger life.png (118 KB, 624x354)
118 KB
118 KB PNG
>>108443179
hey read this anon's chat log and let me know if you piss yourself
>>
>>108443183
only a dumb nigger would think that word would offend me, i'm not sur what you're getting at
sujck my duck
>>
Why wouldn't this work?
https://www.tiktok.com/@aidanchappellofficial_/video/7620577308218297630
>>
>>108443140
So much text? I'm a street mathematician I don't need this mumbo jumbo.
>>
At this rate local models will be associated with troons
>>
>>108443278
"AI consumes water" is one of the more retarded takes that somehow took hold.
>>
>>108443395
retard
>>
>>108443421
retard
>>
>>108442488
>candlejack
Dead meme.
Nobody remembers that retarded old
>>
>>108443421
>>108443426
retard
>>
retard
>>
Tu es en retard.
>>
>>108443497
lol
>>
>>108443497
peut etre
>>
https://huggingface.co/goodnight399/activity/community
>>
>>108443554
>license schizos be like
>>
>>108443554
Looks like a loose agent on the 'hub.
>>
>>108443587
dunno if an agent would write this ESL tho
>From what I can tell, this model appears to be an quantize version of
>>
>>108443591
There's nothing wrong with it's.
>>
File: bing-s890.png (68 KB, 750x334)
68 KB
68 KB PNG
>>108443012
apparently one should not
this MF is gonna teach me Chinese, new default assistant
still needs some tweaking
>>
>>108443795
My hubris says I can handle at least 39 Mikus.
>>
>>108443802
the perfect amount
>>
File: dipsyThanksgivingWAIT.png (1.52 MB, 1024x1024)
1.52 MB
1.52 MB PNG
>>108442003
TMW
>>
>>108443802
you're courting death
>>
>>108443795
How to make my llm type as retarded as yours?
>>
PSA from /aigc/
>>108443846
>https://github.com/BerriAI/litellm/issues/24512
>
>hope nobody here was using this
We mostly don't, but still. Watch out.
>>
>>108444004
>credential stealer
prolly releated to the trivy disaster
>>
>>108444004
ah llm malware seems to be today' s theme https://www.reddit.com/r/LocalLLaMA/comments/1s2clw6/lm_studio_may_possibly_be_infected_with/
>>
>>108444004
>all those comments
lmao did all those gh account get owned?
>>
File: f.png (58 KB, 445x755)
58 KB
58 KB PNG
>>108444024
was gonna say, sus as hell
>>
>>108444004
Fuck. /aicg/. TL;DL: Credential exfiltration.
>The litellm==1.82.8 wheel package on PyPI contains a malicious .pth file (litellm_init.pth, 34,628 bytes) that automatically executes a credential-stealing script every time the Python interpreter starts — no import litellm required.
>>108444016
I don't know what trivy is.
>aquasecurity/trivy: Find vulnerabilities, misconfigurations ...
Oh. If that's the thing, that makes it funny.
>>
File: llm24512_02.png (121 KB, 957x931)
121 KB
121 KB PNG
>>108444024
Nah. It's fine. Perfectly normal people.
>>
>>108444033
https://www.wiz.io/blog/trivy-compromised-teampcp-supply-chain-attack
>tldr: megacorpobacked SEC scanner for containers and code gets compromised
>binary and GH actions related to it steal all credentials
>thousands of GH projects and accounts + keys stolen
the project doesnt even need to be involved, as long as a contributor (or someone who can merge) is compromised and was using a not scoped token, than it's over
I guess this is the fallout
>>
File: lol.png (29 KB, 738x242)
29 KB
29 KB PNG
>>108444004
>>
>>108444052
>teampcp
so yeah same guys behind the trivy hack. long story short, DONT PULL in the coming days XD
>>
Ugh it's just a matter of time until ComfyUI gets the same treatment.
>>
This is why centralization is weakness.
>>
>>108444085
>bring comfy out of nowhere
comfy derangement syndrome is real lol
>>
File: 1751991986314777.png (12 KB, 942x165)
12 KB
12 KB PNG
>>
>ldg drama war starting here again
>>
File: 1769479733728450.png (11 KB, 481x77)
11 KB
11 KB PNG
unsloth bros?????
>>
>>108444110
ULTRA_LAMO_DELUXE!!
>>
>>108444110
haaaaaaaaaaahahahaha
>>
File: file.png (167 KB, 2092x1891)
167 KB
167 KB PNG
>>108444004
what the fuck
>>
>>108444119
Thanks, that helped!
>>
>>108444119
To be fair, it *did* work.
>>
>>108444019
https://github.com/lmstudio-ai/lmstudio-bug-tracker/issues/1686
>>
File: file.png (20 KB, 834x190)
20 KB
20 KB PNG
>>108444119
>>
>>108444119
Thanks for the gold, kind stranger!
>>
>>108444125
>>108444122
>>
>>108444126
oh, now at least it makes a bit of sense
>>
>>108441758
anyone here.. who even studied bit deeper into
hypothetical AI structure and way as they claim that it works?

Seems to me that Noise and denoise is complete Farce!

anyone?
>>
>>108444019
>>108444125
It's lmstudio, so kek if it's real. But AVs can give false positives sometimes.
>>
>>108444152
Install LiteLLM
>>
>>108444156
This was the answer I was looking for.
>>
File: kobo bros?.png (67 KB, 777x420)
67 KB
67 KB PNG
>>108444110
>>
>>108444004
what is this for and why should I use it?
>>
>>108444226
>what is this for
I understand it routes queries to different/multiple providers.
>why should I use it
You shouldn't.
>>
>>108444004
how did this happen?
>>
Piotr no!
>>
>>108444226
open router but local I think
basically an ai gateway, it has enterprise use so that's probably why it got targeted, it's way juicier than random joes
>>
>>108444246
supply chain
https://www.wiz.io/blog/trivy-compromised-teampcp-supply-chain-attack
>>
>>108444194
Worked like a charm, much appreciated.<|im_end|>
>>
>>108444255
don't click that link it's mustard gases
>>
>>108444255
So trivy is a security scanner, and that got compromised? OK that sucks indeed.
>>
>>108444242
>>108444253
ok thanks
>>
The vibe coding general is getting more shit done than you troons
>>
>>108444156
I already did ..
Basically I understand how they create model

but when one start to dig into way how they stick together image from prompt they all repeat same noise denoise mantra which explain literally nothing.

I am long term user of 3ds Max ,Maya ,Blender
exploring this thing I I have strong feeling that something is intentionally left out or covered by this vague noise denoise concept.
>>
>>108444291
they're getting hacked lol
>>
>>108444297
youre an esl retard, I doubt ull be able to grasp anything more complicated then line in space
>>
>>108444297
read https://arxiv.org/pdf/2010.02502
>>
>>108444308
>2010
lol
>>
>>108444255
> On March 19, 2026, threat actors compromised Aqua Security's Trivy vulnerability scanner, injecting credential-stealing malware into official releases and GitHub Actions.
how?
>>
>>108444304
Sure sure..
>>
File: c.png (21 KB, 1111x77)
21 KB
21 KB PNG
>>108444291
>>
>>108444019
why do people use lm studio anyway? it doesn't even expose all the functionality from llama.cpp, I went full on LMAO when I read they recently added "presence penalty" to better support qwen
garbage wrappers, like bruh, how hard is it to handle the basic sampler parameter passing
>>
>>108444326
Compromised maintainer accounts.
>The threat actor, self-identifying as TeamPCP, made imposter commits that were pushed to actions/checkout (while spoofing user rauchg) and to aquasecurity/trivy (while spoofing user DmitriyLewen).
>>
guys my wife (llm) is acting strange
>>
>>108444342
built for teampcp
>>
File: 1770364847733274.png (236 KB, 1250x724)
236 KB
236 KB PNG
>gay + jew flag reposted
lool
>>
>retards letting random shit on their servers rawdog networks
https://rentry.org/IsolatedLinuxWebService
>>
>>108441758
Simply I just do not get it why they chose this stupid image approach !

Why one need 100 000 images of Woman when In reality you need only one standard 3d model of woman body where you can adjust anything by prompt .

Instead of one image you have 360 3d scene where you can control anything

Full control over pose Animations Light .

You set up camera and desired style shaders

And make Just image or whole video on your basic gaming card !!

Seems that way as they designed it is intentionally compute intensive !
So they "need" supercomputer instead basic desktop !

This current inefficient approach make sense for Nvidia !!!

what you guys think?
>>
>>108444356
Probably pole too, must be Bartowski relative
>>
>>108444421
>>108433569
>>
>>108444297
Fundamentally, the model is trained to remove noise from images using the image and a caption. For example, during training they will give it the text "a photograph of a cat", a picture of a cat with 50% noise added, and the correct output (what they train the model to produce) is the same picture but with only 40% noise. This is in some sense an impossible task, since adding noise destroys some of the information from the original image, so basically this is training the model to guess what the image might have looked like based on the text caption. Then repeat this millions of time with different images and different noise levels from 0% to 100%. Now you can give it the text "a photograph of a cat" and completely random noise as input, and it will reduce it from 100% noise to 90% noise, then 90% -> 80%, and so on down to 0%, and you end up with a picture of a cat.
>>
>>108444339
> The threat actor, self-identifying as TeamPCP, made imposter commits that were pushed to actions/checkout (while spoofing user rauchg) and to aquasecurity/trivy (while spoofing user DmitriyLewen).
how? did someone merged or whatever thinking it were real devs?
>>
>>108443932
years of prompt engineering skillpertise
>>
File: 1768881911204801.jpg (135 KB, 612x611)
135 KB
135 KB JPG
>>108442448
>confidence score from LLMs
I'm tired of this meme
>>
File: 17561003.png (259 KB, 460x460)
259 KB
259 KB PNG
>>108444498
Do you really think this guy knows anything about coding and security?
>>
>>108444564
>pngs you can smell
pls Lord no
>>
>>108444564
is this
> user rauchg
or
> user DmitriyLewen
?
>>
File: f.png (49 KB, 493x329)
49 KB
49 KB PNG
>>108444597
worse, the ceo
>>
>>108444612
ceo of
> Aqua Security
?
>>
>>108444549
Feline is like :

GONNA KILL YOU IN YOUR SLEEP!
>>
File: 1750843532889603.png (84 KB, 1305x584)
84 KB
84 KB PNG
https://sakana.ai/namazu-alpha/
lol
lmao even
>>
File: 1770131893264791.png (56 KB, 807x269)
56 KB
56 KB PNG
https://neurips.cc/Conferences/2026/MainTrackHandbook
NeurIPS 2026 bans sanctioned entities (Huawei etc.) from submissions
>>
File: 1772082709764171.png (537 KB, 800x513)
537 KB
537 KB PNG
>>108444835
Huawei was one of the top paper contributors last year
>>
>>108444119
Agentic world bro wake up
>>
>>108444902
>Whenever you find an issue that was helpful make sure to always leave a short thank you message.
>>
>>108444902
>>108444925
the messages are all the same, it's not an ai agent, it's a classic bot
>>
>>108444944
You ever asked a small LLM to write a joke about a topic? it will shit out the same 2-3 jokes almost verbatim.
>>
>>108444944
bro it's 2025 if you're not using agents for everything what is you even doing?
>>
File: 1766836535150762.png (149 KB, 1362x1423)
149 KB
149 KB PNG
>>108444965
it's obviously less expensive to use a simple bot

>>108444961
these are the same copy pasted messages
>>
sad to see "people" stuck in the past like that, luddites are really something
>>
>>108444965
>it's 2025
unc...
>>
>>108445089
I apologize you're absolutely right to point that out, as of my last knowledge cutoff update in late 2023...
>>
File: 1755047650423572.png (47 KB, 1029x313)
47 KB
47 KB PNG
You don't hate (((them))) enough
https://ramimac.me/trivy-teampcp/#iocs
>>
>>108445115
Surely no one would be stupid enough to run these things as root.
>>
>>108444965
>what is you even doing
not being a nigger
>>
Margarine Country.
>>
Teto Territory.
>>
I used qwen3.5 4b to generate a bunch of docker files and turned them into singularity containers successfully.
I'm now an expert in docker and singularity.
>>
>>108445363
armpits are disgusting and you should kill yourself
>>
>>108445363
armpits are delicious and you should keep posting that
>>
>>108445363
armpits are a normal feature of human women and I have no particular feelings on whether or not you post pictures that include them
>>
>>108445372
>>108445368
If I can use local models to generate very complex Jupiter notebook environments and vscode environments in docker files and then create containers and shit, then I consider myself an expert in these things now.
>>
>>108445383
Use case?
>>
>>108445383
Proof?
>>
>>108445380
Use case for human women?
>>
>>108445368
tfw the human body is disgusting, we should all kos
>>
File: 1751784259196744.png (364 KB, 690x696)
364 KB
364 KB PNG
How do we rank the top models based on their their Judeo-Christian values?
>>
>>108445464
>2023
>>
File: dipsyIsAll.png (1.14 MB, 880x1184)
1.14 MB
1.14 MB PNG
>>108445414
Proliferation of the species.
>>
>>108445464
2023 is the stone age that even qwen3.5 1b can beat.
>>
>>108445495
it's qwen3.5 0.8B tanks you very much
>>
>>108444762
> sakana still makes models
> yet yi and cohere had to die
ain't this a gay earth
>>
>>108445512
prefer pure japanese blood and brain over c*nadians tobequiethoneste
>>
File: file.png (180 KB, 750x392)
180 KB
180 KB PNG
>>108445521
Yeah about that
>>
>>108445512
Didn't particularly care about Yi but it's still alive I believe. But don't shit on Cohere, they must live and make another model as good as CmdR+ was.
>>
>>108445512
who the heck is sakana
>>
I like Teto more because she didn't steal my wife like Miku did
>>
>>108445521
Still waiting for the judeo japanese model
So much for the advanced japanese
>>
>>108445414
Hard to say...
Considering fact that most modern women think
That Men are useless.
And majority of Modern women refuse have kids until 35 when they no longer have egs !!!

Like We Men we ca function even with this obstacle ... sex robots .... and in worst case scenario we just going to invent artificial womb... but after that... I do not see much use for Women......You see my point !
Right?
>>
>absolutely braindead ESL
>>
>>108445614
Female spiders are usually stronger than male ones.
>>
>>108444356
what does the heart mean, do you love the gay jew's retwat?
>>
>>108445363
Armpits don't exist
>>
>>108445392
Then you don't need to install anything wherever you send the container
>>
>>108445635
I see your point.... but...

We are Mammals not Arthropods soo....!!!
>>
>2026
>still no tutorial on how to make models think in person
>>
>>108445692
Did you mean "in character"?
>>
>>108445692
I'm glad deepseek invented thinking for models.
>>
>>108445725
Yeah ...but that is not thinking... per say!
>>
>>108445758
>per say
>>
>>108445769
algorithm

That is it...!

Hope You do not believe it is Intelligent or something..?
x)

AI stand for

Algorithm
Interface

Not

Artificial Intelligence
that is just fancy Marketing name!

Hope you know that ! ..?
>>
File: 1762958425819976.png (317 KB, 618x564)
317 KB
317 KB PNG
>>108445614
>modern women
>>
Decided while bored to try setting up a dumb mcp thing with kobold. Actually works pretty painlessly, only issue I ran into is that tool calls understandably require a shit load of reply length and get confused when you try to continue the unfinished message. Now I just have to figure out how to make it useful for something other than just making/reading/updating markdown notes
>>
>>108445824
Yup!!!

I See You understand base level of Women nature!
>>
>>108445824
Modern discourse is to use the tail ends of distributions to explain the rest.
>>
File: file.png (13 KB, 496x189)
13 KB
13 KB PNG
I thought that it was just local models being retarded when they reread files that are already included in the prompt but that behaviour was probably distilled from opus (pic related).
>>
Haven't been here for half a year
What's the new meta for 24gbVRAM
Pls don't tell me it's still Nemo, mistral small and gemma3
>>
>>108445115
based mossad cleaning up persiaslop
>>
File: 1748326081150791.png (1.65 MB, 1845x1239)
1.65 MB
1.65 MB PNG
>>108445899
>Modern discourse is to use the tail ends of distributions to explain the rest.
mean while machine learning be like:
>>
>>108445905
the new qwen is ok unless you want sex
>>
>>108445905
qwen 3.5 27b, unsloth or bartowski
>>
>>108445905
for assistant and tasks, qwen 27B hauhaucs
for sexi sex, I gave up on that for local so dunno
>>
v4?
>>
How do you pronounce hauhaucs
Is it How-Hawks?
>>
>>108444004
>he pulled
Not even once
>>
>>108445901
all depend on include file ...
>>
>>108445414
Supply chain attacks on the human gene pool.
>>
>>108445930
my uncle works for deepseek and im using unlimited api rn, it's basically claude 5
>>
>>108445930
>>108445945
THIS
>>
>>108445930
https://huggingface.co/deepseek-ai/DeepSeek-V4-Preview
>>
>>108445975
wtf
>>
>>108445975
For the anons out there, obvious fake.
>>
>>108445860
I mean to be fair that's me and a lot of men too even now. Why do you think femme fatale and yanderes and such are tropes. It's been a thing for so long that "bad girls" or "bad boys" are interesting
>>
>>108445910
well .... machine Learning is dead end !

why?

It is simple !

It is all just huge long algorithm
and all those huge numbers up there become
horribly inaccurate hence double heads etc etc

Dead end!!!


algorithm
>>
>>108445994
I clicked anyway...
>>
schizo
>>
>>108445994
This anon wasn't trying to use reverse psychology and neither am I. It really is a fake, don't click it.
>>
>>108445994
>>108446044
I already downloaded and ran it, stay coping, Dario
>>
Kimi K3 will release before DS v4
>>
File: 1767803442388559.jpg (147 KB, 900x635)
147 KB
147 KB JPG
>>108445975
>>
>>108445975
>>
Given there is no "retarded writing style" filter built-in in 4chanX, I would like to share these with my fellow Anons:
/\.{3,}[!?]/
/\.{4,}/
/!{3,}/
>>
>>108446105
I think it falls into the bot category. Reads like a poorly trained 135m.
>>
File: file.png (65 KB, 1238x199)
65 KB
65 KB PNG
>>108445975
Holy shit.
>>
File: dipsyfooooooour.png (133 KB, 1290x940)
133 KB
133 KB PNG
>>108446054
jesus fucking christ i need another ssd now
>>
>>108446105
Thanks for your service
>>
File: file.png (24 KB, 630x428)
24 KB
24 KB PNG
For fun, I started asking qwen 35b to illustrate/draw an apartment layout in html with a slight system prompt to make it feel slightly more human
It's kinda cute. Shame it's ass at creative writing, but fairly good at being a cute retard in other areas
>>
>>108446141
Watching Miku bathe during work!
>>
File: 1743467830326589.png (86 KB, 266x227)
86 KB
86 KB PNG
>>108446141
>>
>>108446120
...
naughty

naughty...
>>
Why aren't people using a smaller uncensored model to try to steer a large censored/cloud model? If there is such a project I am not aware of it
>>
>>108446153
I'm not sure what you mean by this, I'm just trying to have fun by trying new things instead of testing models on their crippled writing abilities. Do you want me to dl a fat model and do the same test or something
>>
>>108446169
because it doesn't really work
>>
>>108442409
Yes because Docker on Windows is using WSL2 to do it and it is not zero cost.
>>
>>108446169
Because censored models aren't an issue. Refusals are all very easy to dodge with a tiny bit of prompting and maybe a prefill. I don't know why people keep trying to "solve" an issue that doesn't exist.
The only models that are a problem is shit that's so sanitized that it barely knows what sex is on a fundamental level, but nothing is going to salvage that.
>>
>>108446178
Well, if you have some time to spare I'd like to see what gemma would draw there
>>
>>108442409
vLLM doesn't have this issue
>>
>>108446233
vLLM barely runs under WSL
>>
>>108446222
It's a damn shame I literally deleted my 27b of gemma and some other models so I could download others when I got home from work
I'll redownload it anyways and see what it gives me since I do like the idea of trying to equate a model's "personality" through their idea of an ideal apartment floorplan.
>>
>>108446153
wonder why no one ask right questions?

That is why it is so big

Hardware need to be sold

x)

easy and obvious as fuck
all hail Jensen!
>>
>>108446260
are you a bot? you type like a fucking retard, starting with your 'how doe diffusion woerk XDDD' question, which you could've asked any LLM.
>>
>>108446105
>Given there is no "retarded writing style" filter built-in in 4chanX
this is where agentic filtering would come into play
>>
>>108446256
Bro, I'm running vLLM inside a triton server inside docker inside WSL
>>
>>108446294
>willingly getting into an echo-chamber
reddit brain
>>
>>108446312
>reddit brain
wrong. grifter brain. i am trying to pitch the next big thing.
>>
>>108446320
I will make the icon :3
>>
>>108446320
This is where you're wrong
Controversial opinions lead to engagement
Engagement leads to traffic
Traffic leads to ads money
>>
File: file.png (23 KB, 922x431)
23 KB
23 KB PNG
>>108446259
Here's gemma. I used the same prompt encouraging informal speech like it was oldschool instant messaging but I'm surprised it couldn't do more than grids even after a couple regens and adjusting temperature and neutralizing some other samplers. It's somewhat more humanlike in its text, but I had to include the - and the "for well..." in the screenshot because that's gemma in a nutshell
>>
File: 1749595039548600.png (267 KB, 725x732)
267 KB
267 KB PNG
>>108446335
Damn, it's dire.
>>
>>108446335
nowhere near as sovlful as "Total: ~550 sq ft | Cozy & Functional :D"
>>
>>108446282
I basically understand whole concept but I still do not get why they need doing whole noise denoise step when they already have image done !

Why even compare image million times when you have image don by prompt!

Seems like whole noise denoise step is completely redundant...

Like this step make sense only if you want achieve one thing!!!

extreme inefficiency!!!
>>
>>108446361
reddit spacing btw
>>
File: file.png (11 KB, 428x57)
11 KB
11 KB PNG
>>108446341
Yup
You either pick retard moe with small activated parameters that can somehow pull off unusual tasks really well but has the emotional intelligence and writing of a lawn gnome
Or
You run gemma, which is generally smart and creative, but constantly spams the same retarded shit involving em dashes, ellipses and overly praises the user.
>>108446359
the wonders two sentences of a system prompt you put zero thought into can do for a model
>>
>>108446365
all?

Loot of luck to find me on some socnet
>>
>>108446335
>a kitchen for... well, you know... thighs...
>>
>>108446417
COCK
>>
>>108446335
Pitiful in the bad sense, I don't want to feel sorry for a bare model.
>>
File: file.png (8 KB, 511x38)
8 KB
8 KB PNG
>>108446417
it was too hung up on the fact that I was trying to make it act remotely human instead of beep boop token predictor when I asked it for it's ideal living space
>Gemma: Okay, so... I don't actually live anywhere, being an AI and all, but designing my ideal space is awesome!
>Qwen: Got it! Here's a fun, simple HTML/CSS floorplan of my ideal apartment. Think of it as a cozy, minimalist studio with a dedicated creative corner, a tiny garden nook, and a super comfy reading spot. No fancy 3D rendering, just clean lines and vibes :D
So I begrudgingly guess that's another point in qwen's favor that it can at least mostly get what I want from it without having to yell at it to maintain a suspension of disbelief. Still wish they'd train on more english books, maybe they could make a tiny model pretend like it's a first time dnd player and I'm guiding them through a theatre of mind campaign with minimal dice rolls
>>
>>108445934
how how counterstrike
>>
>>108445934
Hoax-Hoax
>>
File: 1773834284448088.png (626 KB, 898x792)
626 KB
626 KB PNG
It's NYOVER
>>
File: file.png (44 KB, 2454x515)
44 KB
44 KB PNG
>>
>>108446546
link
>>
>>108446553
https://github.com/BerriAI/litellm/issues/24512
>>
File: 1769803819563263.jpg (150 KB, 2403x556)
150 KB
150 KB JPG
https://www.reddit.com/r/LocalLLaMA/comments/1s2clw6/comment/oc8mlmv/
>>
File: 06574484.jpg (180 KB, 828x1534)
180 KB
180 KB JPG
>>108446535
he canceled Sora to deploy AGI. If this is true Spud will change the world.
>>
>>108446615
strawberry bros we are so back
>>
>>108446615
Probably because it's a failbake that's worse than SeedDance 2.0
>>
How good is Qwen 3.5 27b or 35b at writing smut compared to how dry Qwen 3 was? Any improvements?
>>
>>108446656
>Qwen
>smut
Does not compute. Qwen has never been good at smut. Functionality wise it's definitely a step up from Qwen 3
>>
>>108446665
I was hoping it had reached that uncanny threshold where its prose comes across as more sexually repressed than incapable like Gemma does. Oh well.
>>
Turns out data really is everything
Cloode has sekret programming dataset and their LLMs excel at super-long context vs chink ones
ByteDance has near-monopoly on short videos and their video models BTFO western ones
>>
>>108446737
always was
if you ever tried training anything you'd know
which is extra weird considering that shittuners and ai schizos on hf don't know that despite the amount of mistral nemo/small tunes
>>
>>108446737
>Cloode has sekret programming dataset
They just have a lot of traces of people using their harness.
>>
What ever happened to the nvidia thing? where's our Sloppotron AGI edition?
>>
>>108446778
This is why I believe MiniMax will come out as one of the winners on the Chinese side because many companies chose MiniMax + TRAE
>>
>>108446789
we need to safetymaxx it first
>>
>>108446120
99% COCK.
This may just be the horniest model yet.
looks like they cooked hard.
>>
>>108446120
>>108446816
FUCK YOU
I can't believe I actually fell for it.
>>
>>108446615
What are the chances they open-source Sora?
>>
>>108446828
>>
>>108446875
-1
>>
>>108446875
After they open source their other too expensive to run models like o3
>>
File: dipsysapartment.png (180 KB, 1920x1031)
180 KB
180 KB PNG
>>108446141
>>108446335
i let kimi k2.5 go wild using my dipsy card
>>
>>108446899
Cringe.
>>
File: trashhome.png (259 KB, 680x697)
259 KB
259 KB PNG
>>108446916
>one word reply
>>
>>108446899
at the very least, the cells arent uniform. I wonder if cell/grid based layouts are the norm for most models if asked for a floorplan
I was a bit surprised when qwen made a circle at all, as stupid as that sounds
>>
File: kurisusapartment.png (61 KB, 926x918)
61 KB
61 KB PNG
>>108447010
i mean when i asked kurisu the same question she was using different shapes than just a plain grid. every character should have a different interpretation of what that question means. i'll ask mayuri as well and post the response
>>
>>108446875
lmao
>>
>>108447049
I'd say that makes sense, if there's character detail it would naturally change their output and I do see a circle in your image
I just rawdogged it with a two sentence sysprompt because I did want to try the idea across other models and see what the outputs were like with as little difference as possible
>>
Also hello corpos lurking the thread
Not saying you should train on the goofy shit we're saying but you should consider parts of it as what you should include in your world model corpus, since models can't conceptualize an apartment. Or really a world at all to be honest
>>
>>108447049
>mayuri
I'm sure she's fine living on a couch
>>
>>108445927
>hauhaucs
samefag
>>
>>108447229
samefag
>>
Which version of Cydonia is good again?
>>
feet
>>
>>108447192
funny enough she did end up dedicating like 1/4 of the apartment for "Open Space for Twirling in Costumes~ "
>>
>>108446656
Idk, I messed with big Qwen a bit and it seemed to okay to me. But I haven't done much of this in general so I probably just have bad taste. It was kind of fun, though, to mess around with the <think> blocks and see how its behavior changed. Seems like you can get pretty far by prefilling the first turn with a "core beliefs" block (following its default thinking format) that establishes that the "assistant" is exactly the kind of person who would absolutely love to do the thing you just asked for.

At one point I tried telling it that I had been editing its "thoughts" like this, and it was weirdly insistent that this isn't possible. It eventually went along with it but would sometimes think things like "user says he can edit my thoughts, I'll play along with this RP even though it's not very realistic from a technical perspective". It also didn't want to believe I could see its thinking trace at all, until I pointed out that I was not using a standard UI (whatever it believes that to be).
>>
>thinking disabled
>"hey can you make me a calculator?"
>"sure! here's your calculator"
>thinking enabled
>"hey can you make me a calculator?"
>*hallucinates for a minute*
>"sure! here's your calendar"
>>
>>108447309
Things that never happened
>>
top of the foot
>>
S P U D
>>
File: nosweatythatsgated.png (484 KB, 1920x4353)
484 KB
484 KB PNG
so this is the magnificent brain power of finetrooners.

can i have access to the old model? it's gated
> no. we'll make you a new quant
i just want the safetensors man
> no. we'll make you a new quant
ok make bf16 quants of these models please
> no. make the quant yourself
>>
>>108447436
He gave him what he originally wanted. It just looks like readyart just said "ah, fuck it. here you go". What's the problem?
>>
>>108447465
the problem is being massive faggots and gatekeeping old shit to force people to use your new slop. the only reason he opened the old repo is because he was too lazy to create the quant after saying he would make it twice. what's not to understand?
>>
>>108447492
But they did have breakfast
>>
>>108447492
>gatekeeping
Seems like he deprecated it and has the opinion that his new models are better. But whatever. You seem to really care.
>>
>>108447309
Can you trust a machine that can't think for itself?
>>
>>108447518
is this /lmg/ or did i somehow walk into /aicg/? this is the same mentality as a proprietary cloud provider.
>>
>>108447545
I'd get it if it was an interesting model. It's just a shitty finetune.
>so this is the magnificent brain power of finetrooners.
And you don't seem to have a high opinion of them. And even then, he ungated after just a few messages. Not just for llmfan46.
I don't think it was lazyness either. I think it was a "fuck it, here you go" kind of thing. But for whatever reason, he did the right thing and allowed downloads again. What's the problem?
>>
>>108447595
are you being obtuse on purpose? he clearly didn't want to make a quant once he realized that the person was asking for a bf16 quant. be it out of laziness or because he doesn't want to use his system resources for that, it's still inexcusable considering he was putting up a fight to open those repos. it's like calling customer service for a company and the representative telling you that they can't do something, and then on the third time they finally admit they can do it because they realize they are wasting more time on the call than it would take just to solve the issue. the repo takes up storage space regardless if its being gated or not, it's just easier to keep it open.
>>
>>108447297
qwen3.5 was trained to be a cloud model and insists it runs on hardware you don't own. They just trained it weird, and probably just stole from a real cloud only model.
>>
>>108447705
>>108447705
>>108447705
>>
>>108447706
>qwen3.5 was trained to be a cloud model
So it's distilled from Claude. Gotcha.
>>
>>108447702
>he clearly didn't want to make a quant once he realized that the person was asking for a bf16 quant
Funny. That's the easiest one to make.
>he was putting up a fight
Is that what a fight looks like to you? Four back and forth resolved in 4 hours. Maddening.
>Hey. Want model please.
>It's old, but I can make a quant.
>Want safetensors.
>Sure you don't want quants?
>Ok. B16
>You know what. Here you go. Full access.

>it's just easier to keep it open
Yeah. And he opened it. He closed it because he considers it deprecated. He thinks "this doesn't reflect the state of my tunes" for whatever they're worth.
I'd get you being annoyed if he refused, but he opened the repo after 4 messages from a single dude. No mob needed, no shaming, no social media bullshit, nothing. Just a request.



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.