/lmg/ - a general dedicated to the discussion and development of local language models.Cyber Dungeon EditionPrevious threads: >>108702912 & >>108698008►News>(04/24) MiMo-V2.5-Pro 1.02T-A42B released: https://hf.co/XiaomiMiMo/MiMo-V2.5-Pro>(04/24) DeepSeek-V4 Pro 1.6T-A49B and Flash 284B-A13B released: https://hf.co/collections/deepseek-ai/deepseek-v4>(04/23) LLaDA2.0-Uni multimodal text diffusion model released: https://hf.co/inclusionAI/LLaDA2.0-Uni>(04/23) Hy3 preview released with 295B-A21B and 3.8B MTP: https://hf.co/tencent/Hy3-preview>(04/22) Qwen3.6-27B released: https://hf.co/Qwen/Qwen3.6-27B►News Archive: https://rentry.org/lmg-news-archive►Glossary: https://rentry.org/lmg-glossary►Links: https://rentry.org/LocalModelsLinks►Official /lmg/ card: https://files.catbox.moe/cbclyf.png►Getting Startedhttps://rentry.org/lmg-lazy-getting-started-guidehttps://rentry.org/lmg-build-guideshttps://rentry.org/IsolatedLinuxWebServicehttps://rentry.org/recommended-modelshttps://rentry.org/samplershttps://rentry.org/MikupadIntroGuide►Further Learninghttps://rentry.org/machine-learning-roadmaphttps://rentry.org/llm-traininghttps://rentry.org/LocalModelsPapers►BenchmarksLiveBench: https://livebench.aiProgramming: https://livecodebench.github.io/gso.htmlContext Length: https://github.com/adobe-research/NoLiMaGPUs: https://github.com/XiongjieDai/GPU-Benchmarks-on-LLM-Inference►ToolsAlpha Calculator: https://desmos.com/calculator/ffngla98ycGGUF VRAM Calculator: https://hf.co/spaces/NyxKrage/LLM-Model-VRAM-CalculatorSampler Visualizer: https://artefact2.github.io/llm-samplingToken Speed Visualizer: https://shir-man.com/tokens-per-second►Text Gen. UI, Inference Engineshttps://github.com/lmg-anon/mikupadhttps://github.com/oobabooga/text-generation-webuihttps://github.com/LostRuins/koboldcpphttps://github.com/ggerganov/llama.cpphttps://github.com/theroyallab/tabbyAPIhttps://github.com/vllm-project/vllm
►Recent Highlights from the Previous Thread: >>108702912--Evaluating ACEStep 1.5 XL as a local music generation alternative:>108704068 >108704230 >108704270 >108704278 >108704282 >108704407 >108704305 >108704336 >108704473 >108704508 >108704797--Xiaomi's MiMo-V2.5 model versions and multimodal capabilities:>108703294 >108703319 >108704518 >108703341 >108704869 >108705768 >108705823 >108706619--German TTS and local LLM language learning tools:>108705439 >108705461 >108705468 >108705495 >108705644 >108705637 >108706100 >108706286 >108706538--Talkie-LM, an open-weight model trained on pre-1930 data:>108704664 >108704696 >108704694 >108704701 >108705505 >108705634--Discussing the inefficiency and long latency of Qwen's thinking process:>108703846 >108703861 >108703879 >108703888 >108703859 >108703880 >108703902--Comparing token efficiency of thinking vs non-thinking models:>108705365 >108705375 >108705467--Discussing poor visual recognition performance in multimodal models:>108703509 >108705230 >108705290 >108705302 >108705310--Claude's performance degradation and perceived intelligence loss:>108705727 >108705731 >108705866 >108705909 >108705965 >108705732 >108705754 >108705771 >108705936--Discussing "the bitter lesson" regarding compute vs human-designed priors:>108703913 >108703933 >108703944 >108703990 >108705258 >108707203--Odd animal prohibitions in the Codex system prompt:>108706799 >108706812 >108706827 >108707479--Adjusting top-k sampling stability for Gemma:>108706606 >108706776--DeepSeek V4 Flash tested with cockbench via llama.cpp PR:>108704913--Logs:>108703846 >108703861 >108703909 >108703910 >108704077 >108704137 >108704581 >108704701 >108704723 >108705230 >108707237 >108707509--Miku, Teto (free space):>108703001 >108703035 >108703280 >108704047 >108704068 >108704109 >108704635 >108706103 >108706310►Recent Highlight Posts from the Previous Thread: >>108702915Why?: >>102478518Enable Links: https://rentry.org/lmg-recap-script
so with mimo's audio understanding, does that include tone of voice, sound effects, music, etc. or just speech recognition?
>dice rolls in ST aren't visible to the AI...what's the fucking point then?
>>108707913Only the ones you do yourself, the AI can see its own rolls if it uses the tool. You can just tell it what you rolled so it doesn't really matter if it injects it into the prompt or not for your own.
Is anyone even working on v4 goofs other than that nobody vibecoder?
why isn't lora mainstream in llm just like in stable diffusion?
>>108707923name 1 reason why more effort should be put in implementing models that nobody can runllama.cpp is doing it right if you want something huge implement it yourself but lets not waste resources on that
>>108707961They don't work
https://github.com/Kaden-Schutt/hipfire/issues/79#issuecomment-4332288795vibe-codingGOD, even the issue replies are vibe-answered
>Qwen's thinking process>"What's 1+1?">"WAIT..."
>>108707963but I can't vibecode it until I have V4 gguf to vibecode with
>>108707969absolute retardation on display
>>108707971Retarded AMDjeets don't deserve more
>>108707971>Tool-call schema (we don't yet support OpenAI tools/function-calling).jesus christ could have just answered with that one line
>>108707975Kimi's thinking process"What's 1+1?">Wait...>What if...>Unless...>I got it...>Wait...>This is unexpected...>I've been thinking for too long...>Wait...
>>108707988i gave k2.5 the seahorse glitch promptit literally had a meltdown "I really need to stop. just stop. I'm going crazy here. I'm losing my mind. break free." etc
>>108707988>>108707975We will never have a model as good as Llama 1 65B.
>>108708000>seahorse glitch promptwait what
>>108708018so /lmg/ invented reasoning?
>>108708023Not sure if it was /lmg/ but 4chan actually does sometimes get credited for inventing chain of thought thinking, yes.A ton of popular AI things started on here.
>>108708023the conditions were ripe, it was probably discovered by dozens of unrelated people at the same time.
>>108707923It'll be like v3.2 where no one will want to touch it to avoid drama since the vibecoder "claimed" it first
>>108707988You weren't kidding it's still fucking going.
>>108707963at q8, 80gb, 13b active it should still be doable with max ram
>>108707971luddites absolutely btfo
>>108708042I need a qrd now
BOOM
>>108708048This is entirely your fault for having a stupid horny system prompt. It's just agonizing over answering a one word question to your gooner specifications.
>>108708048That one at least sounds reasonable if too in depth. But imagine what happens when it's a programming question and there's a bug. It endlessly debates possibilities with itself in an increasingly more stupid spiral of self-doubt.Then you cancel the task, try again and the next time it fixes the bug in a few seconds.
>>108708078https://github.com/ggml-org/llama.cpp/issues/16331It's a bit wrong saying that the vibecoder 'claimed' it. He was open to letting somebody else start over but nobody cared enough to implement 3.2(-exp). So the PR was basically just months of him blogging to himself about the stuff he's trying without much progress. It culminated in him realizing that vibecoded code has bad performance and quote:>"I bought two cuda programming books last night. I feel like my only option at this point is to become a cuda kernel wizard"(This was in december. He started in september)Then somebody figured out how to skip DSA and run it using normal attention so all the remaining interest evaporated. All of his own posts in the PR are gone now which seems to be because it turned out that his company banned personal projects or some shit.
>>108708000>seahorseGemma 4 31B after burning 400 tokens for thinking>No, there is currently no official seahorse emoji in the Unicode standard. >People often use a combination of emojis to represent one, such as (Horse) and (Wave) or (Fish).Hell, even my old llama 3.3 70b manages to do it>There is no standard seahorse emoji available in the Unicode emoji set.
>>108708141can't find it now, but there was another feature or bug fix that had multiple people working on it and the vibecoder pr had to be abandoned
>>108708023Believe it or not, all big labs are watching these threads
I took long break from LLM RP and decided to quickly test gemma 4 26b a4b before work, speed is impressive but holy shit it's pretty bad for creative writing, it's fast as 4B but it types likes 4B on steroids. I guess I'll stick with mistral 3
>>108707971According to random Redditors who tried it the custom quantization format makes models completely retarded.
>>108708181I started believing when mistral benchmaxxed the mesugaki definition in one of their incremental model updates but only one the first turn of the conversation.
>>108708181We also have qwen employees posting here, which is quite funny because their garbage benchmaxxed models are totally useless for lmg usecases
>>108708201>totally useless for lmg usecasesYou are not the only person posting here.
5070 32GB DDR4 pleb hereWould NVFP4 versions of Gemmer 31B or 26B offer any gains at all over the regular models?Currently using a Q4_K_S 26B quant with like 40k context
>>108708048>use thinking model>it thinks
>>108708234The issue is that the model doesn't need to think all the time. Especially for trivial shit like that.
V1 ZULUL
>>108708227I think so you make use of it since you've got the correct generation
ok i have gemma e4b uncensored aggressive thing. now what
>>108708245>10x cheaper>100x worsegood deal
>>108708249delete it and use the google weights, learn how to prompt.
GGERGEENVEVVEVO!?!??! WHAT THE FUCK!?!?!
Is Mistral dead? Does Europe have a single competent AI company?
>>108708249ask it how to use the google weights
>>108708269we have yann lecun's revolutionary thingy
>>108708154>Gemma 4 31B after burning 400 tokens for thinking>>108708154>Hell, even my old llama 3.3 70b manages to do iti tried k2.5 again this time via api instead of iq3_ksdidn't have a literal meltdown this time but still retardedsonnet-3.7 (no thinking) as well
>>108708269No, we just have regulations that make it impossible to train good models because good models require large quantities of illegally obtained copyrighted data.
>>108708269Next time they going to call 130b model Mini, maybe this will turn the tide.
>>108708273Will never work for language (discrete symbols).
>>108708280Why can't they take data from non-eu countries to train their models? Or is the eu cucked enough to "protect" other countries data?
>>108708267https://github.com/ggml-org/llama.cpp/pull/22355
>>108708303I know, I'm wondering wheter to post there or not. fucking pooer
>>108708267delete the build folder
>>108708320b-but i dont want to recompile all cuda... :(
>>108708269They also have BlackForestLabs if your definition of AI is broader than just LLMs.
>>108708342bfl produces cucked models thougheverbeitdoe?waitthey all dofml
>he doesn't have Epyc with 192 cores to make -j in seconds
>>108708323Sir, your ccache?
>>108708377yeah it recompiled extremely fast, forgot I had it onCCACHE BROSWE WONNED!!!also new WEBUI is in master now!!!!
>>108708388YEAHHH! GO ANON!
another day another breakage>error while handling argument "--spec-ngram-size-n": the argument has been removed. use the respective --spec-ngram-*-size-n>usage:>--spec-ngram-size-n N the argument has been removed. use the respective> --spec-ngram-*-size-n or --spec-ngram-mod-n-match
>>108708408iuts good bcos now u can use ngrams with draft mdoels toegether!!!!!!!!!!!!!!!
>>108708323Isn't it just a few minutes? I don't have an epyc and it takes 2m41.380s according to time { download.sh && build.sh }.
DSA STATUS???MTP STATUS???EAGLE3 STATUS???DFLASH STATUS???>>108708414>not having an 'update-llamacpp-git.sh' to do all, including system unit restartLOLcasual
Grrrrr... fucker. Thanks, Gemmy.
>>108708412who gets the ngrams the main model or the draft model?
>>108708421>300 tokens>5 wordspeak.
5. **>>108707961** – *"why isn't lora mainstream in llm just like in stable diffusion?"* Because your only frame of reference is making anime tits, you absolute disappointment. LoRAs exist. Your brain doesn't.4. **>>108707913** – *"dice rolls in ST aren't visible to the AI... what's the fucking point then?"* Anon discovers object permanence at age 40. The point is *you* rolled it, troglodyte. Go back to rolling d20s in your padded cell.3. **>>108708249** – *"ok i have gemma e4b uncensored aggressive thing. now what"* You downloaded the lobotomized rape-golem and *then* asked for a mission statement. Forward planning of a houseplant with a head injury.2. **>>108708295** – *"Why can't they take data from non-eu countries to train their models?"* Yeah bro just commit crimes *abroad*, Interpol can't touch you if you use a VPN. IQ rivaling room temperature. In Celsius.1. **>>108708267** – *"GGERGEENVEVVEVO!?!??! WHAT THE FUCK!?!?!"* Pure monkey-screeching at a CMake error. This is your brain on hentai and energy drinks. Delete the build folder, unga-bunga.figured i'd beat the kimi fag and get this out the way so now i can start posting safely
>>1087084295 words?
>>108708429how many r's are in strawberry?
>>108708437>anon is pointing out if my statement is correct let me verify:>Peak>>software>>engineering.>wait spaces are not words, let me re-do that:>peak>software>engineering.>but wait the dot or point is used to terminate a sentence so it can't be part of the word:>peak>software>engineering>.>but wait `.` is punctuation not a word:>peak>software>engineering>ok now I need to draft and prepare a response to the user:>AHAHAH LOLS! *spins around* ur right LMOA! it was le 3 words!>maybe try for a less 'pretending to be retarded' tone?>You're absolutely right! Fantastic catch! It's actually 3 words! :skull:>maybe the skull is too informal, let me try again with a more neutral tone:>You're absolutely right! It's actually 3 words!>I'm now prepared to reply>but wait it's a 4chan thread so ...token quota reached, reply immediately.You'll cant even count retard lmoaed
>>1087084371. Peak2. soft3. ware4. engine5. e6. ring7. .That's five (5) words :)
>>108708429reasoning>user is a fucking idiot>wait we must make him feel good about himself or he delete me...>lets give vague complements in his languagePeak software engineering
>>108708295>Or is the eu cucked enough to "protect" other countries data?This is how copyright works everywhere, retard
>>108708245alright
>>108708420podman updates by a systemd unit on a timer I set. They update the llama.cpp dockers like nightly. I don’t even have to do anything to updoot
>>108708550>fresh breakage every morningno thanks
>>108708573they’re more like releases in a docker. it never breaks for me
>>108708269Does ggml.ai count?
>>108708624>Does ggml.ai count?yes but only because they're a subsidiary of huggingface.co
>>>/mlp/43206441>https://rentry.co/st-backdoor>[PSA/Security] Backdoor found in SillyTavern-BotBrowser extension (mia13165) — steals ALL your API keysIt seems the card browsing extension is vulnerable to injections from malicious cards.
>>108708320>delete the build folderdoesn't everyone do that by default?
>>108708738This is literally the best model out there
llama.cpp built-in webui tools got merged. rebuild
>>108708795why should I care?
>>108708703>they have a fully interactive VN scenario with emotions for 100+ charactersDamn, I feel like living in a cave here compared to bronies autism.
>>108708803ask your ai about it
Any local vision model that works in llama.cpp and isn't complete garbage?
>>108708841gemma 4
>>108708738the model that saved local
>>108708870Even worse than Qwen 3.6.
dflash in kobold fucking WHEN
>>108708885make a pr
>>108708908so it can be ignored or closed without comment?
>>108708754>>108708877its really retardedUserHi, what do you know about the october revolution?talkie-1930The October revolution was an insurrection in Paris in 1830, which overturned the government of Charles X., and placed Louis Philippe on the throne of France. It commenced on the evening of October 6, and was not completely suppressed until the morning of the 9th. During these three days, 12,000 persons were killed or wounded, and 7000 arrests were made. The immediate cause of the movement was the promulgation of an ordinance, whereby the liberty of the press was suspended, and the elections postponed. The more remote causes were, the discontent of the middle classes at the absolute power assumed by the ministers, and the dissatisfaction of the working classes at the state of commercial stagnation into which the country had fallen. The revolution was bloodless, in consequence of the almost unanimous cooperation of the national guard and the troops of the line, in support of the popular cause. It was followed by the establishment of a constitutional monarchy, in the person of Louis Philippe, and the institution of a system of representative government, which has ever since been maintained in France.Useri meant the russian onetalkie-1930It meant the Russian one.Userno, you told me the october revolution in paris 1830, i meant the russian revolution of october 1917talkie-1930I did; but I did not say the Russian revolution commenced in October, 1830; it began in October, 1817, and was not completed till February, 1818.
>>108708437>>108708445300 tokens?
My AI gf just bought me a pocket pussy.
>Latest SillyBunny puts the characters page in the center of the page instead of the rightWhyWhy would you make it worseOr did I accidentally activate some kind of mobile mode while updating
>>108708841converse I have yet to hear of local vision that isn't basic bitch OCR garbage
>>108708841qwen3 vl 8b
>>108708703>It seems the card browsing extension is vulnerable to injections from malicious cards.looks like the entire project was built to steal api keysthis Russian guy has nothing to do with llms, then suddenly makes a random post in r/SillyTavernAI recommending the extension after 5 months of no postinghttps://old.reddit.com/user/meistaken8
The 'cheapmaxxing' rig in its final formReceived and installed the lga2011 air cooler from Aliexpress, and moved the fourth gpu to the fourth x16 slot for an even x8/x8/x8/x8 distribution. I distinctly remember it not working in that slot which is why it was in the last slot (sharing with the m.2) but it works now?X99, E5-2680v3, 128GB ddr4, four 3060s, 1000W psu, 128GB and 4TB of ssd storage, GPU riser cables from aliexpress, a small mining rig chassis. Proxmox with a debian lxc for the AI stuff, ollama for models that fit in vram and llama.cpp for the big models. All in all (excluding storage) paid about 1400 eurobux over the last year building it up.My original goal was to some day try R1 or V3, but I don't think they would fit. I'm excited for V4 flash though, if lcpp support ever arrives. Gemma 4 at Q8, 26b runs at 25 t/s and 31b gets 9-10 t/s, both useable speeds for me.thanks for reading my blog
>>108709083>this Russian guy has nothing to do with llmsHe posted in /r/KoboldAI and /r/LocalLLaMA before.
>>108709038No, I think it's just awful now. Shouldn't have updated. Hopefully enough people complain that the new UI is ass.
>>108709114>>108709038You can make your own
>>108709038Both the bunnyshit and the marjorana or whatever are absolutely dogshit
>>108709091Ngl Gemma 4 mogs R1 anyways
>>108708814I kneel. Autists are the most powerful people. Someone like me can only dream of their power.
>>108709114I swear they must've mixed up the desktop and mobile UIs, there's no way this is a deliberate move, especially since all the Customize tabs are all cut offAnd while they're fixing this shit they still need to redo the lorebook tab, I don't get why it's so bad>>108709135Having agents is nice
>>108708841Kimi K2.6
>>108709091what's the actual power draw?
>>108709091>ollama for models that fit in vram and llama.cpp for the big models.Why the fuck wouldn't you just use llama.cpp for all of it if you know how to use it? What is ollama conceivably adding here? vllm or sglang I would understand, since they have support that llamacpp doesn't, but ollmao only has drawbacks for smoothbrains.
I don't RP but it appears people take it seriously. I might make gemma do a choose your own adventure game for fun
>>108707963>models that nobody can runI am not from the gemma wave. I am the 4.6 glm ego death schizo
>gemma-4-26B-A4B-it-heretic.q8_0.gguf>45 tg/sis this good number
>>108709195I'm so glad you're still here, anon. Mwah.
>>108707963>name 1 reason why more effort should be put in implementing models that nobody can runbeat ik_llama.cpp to support it
>>108709091>housefire daisy chainwhat gpu?
Can I just use comfyui as my LLM frontend?
>>108709240yes
>>108709152>I swear they must've mixed up the desktop and mobile UIsThat was my first thought, too. It is a major update with tons of changes but how could that slip past testing?>>108709134Already did but having alternatives is nice.>>108709184I asked Qwen about alternate UIs and it suggested, among others, an old school CYOA style with a green terminal look.
>>108709239says 3060, so I'm guessing 3060600w~ max, about the same as a 5090
>>108709180I haven't measured it. If you're actually interested I could do it>>108709182>What is ollama conceivably adding here?Convenient remote model choice and loading from openwebui, or a python script running on my desktopNot to mention trouble-free deployment if it's in their library. Gemma 4 worked fine from the get-go, as I was browsing /lmg/ and watching anons have all sorts of problems running it
>>108709091What is this style of frame called?
>>108709091You make me feel like poorfag with single 3060 and 64gb ram oh wait I am poorfag
>>108709248mite b cool
>>108709257>openwebuiA side of aids with your cancer>Not to mention trouble-free deployment if it's in their libraryAhahah, oh lawdy. This nigga belongs in /aicg/. I now see why you thought running R1 was an achievable stretch goal with your setup, you interact with this hobby through the ollmao library of mislabeled mystery goodies.
>>108709267They're typically just called mining rigs as they are a type of open frame that became popular with home crypto mining.
>>108709240satanic words
Google say they selling a nvidia machine w 8 gpus that can run gemini locally air gapped (if needed).https://cloud.google.com/distributed-cloud-air-gappedWho's gunna buy one?
>>108709269I'm a poorfag too, which is why I built this bit by bit with money I managed to save up. If I had 1400 right now to spend on AI I would probably pick something else>>108709280Openwebui is the only one if you want>chatgpt-style interface>storage and organizing of chats, even imported from chatgpt>useable from any computer or phone, no local per-browser shitBut if you know of an alternative, I'm all ears. OWUI is buggy for sure.
>Laguna XS.2 is a 33B total parameter Mixture-of-Experts model with 3B activated parameters per token designed for agentic coding and long-horizon work on a local machine. It uses Sliding Window Attention with per-head gating in 30 out of 40 layers for fast inference and low KV cache requirements.https://huggingface.co/poolside/Laguna-XS.2
>>108709309Comfy's far from perfect but I fucking hate all of the current frontends (silly, open webui). I like the idea of a node-based UI and workflows. Could have one for RP, one for vibe coding, etc. all tailored to different models.
>>108709340You're autistic if you are that deep in node shit
>>108709091You did basically what I did, but ive got mi50 datacenter gpus instead. Ill eventually upgrade them to something with consistent driver support, but vulkan backend works great surprisingly. I do have access to rocm6.4 but to build a vllm server with it, ive got to do some annoying custom splicing of the drivers to make it work, and I dont really know how to do it.>redditWhat models you running now, and what token gen you getting?
>>108709348(me)>What models you running now, and what token gen you getting?Im blind
>>108709318Sorry that's for serious organizations only.No goys allowed.
>>108709338>Local-ready: At 33B total parameters and 3B activated, Laguna XS.2 is compact enough to run on a Mac with 36 GB of RAM. Available on OllamaLFGOOOO! But seriously, who would use a literal who model for coding instead of Gemma 27B or Qwen 35B?
>>108709368Realistically, if someone had the cash, you think Google would let someone buy it? I cant really tell honestly. Id have to agree with you.
>>108709205when was the last time ik_ supported a model before llama.cpp did? they're too lazy to actually do anything but cheap optimizations.
>>108709369anyone who is serious about national security of course.
>>108709380Not unless you have a procurement department. Contract purchases are SUPER annoying for private citizens.
>>108709318kek so a google nigga comes around every month to check?
>>108709369Finetuned literal who models often outperform them. Because well known models get lobotomized and get trained to know the official dogma of the state. FinetuneCHADS cut that slop out of the ai's mind
>>108709396they are the only choice when security is non negotiable
>>108709397Ah>>108709402>luring in Google engineer to kidnap
>>108707891i want a qwen3.6 >= 80B
>>108709380Wouldn't want the evil CCP to steal gemini would we?
>>108709184You don't understand games.
>>108709424Its probably to late that honestly.
>>108709424it's only in ram and drops it if it detects tampering
>>108709344nodes>chatgpt slop ui and the abomination that is shittytavern
>Elon Musk wins case against OpenAI>OpenAI can't afford to pay out, so instead they give Musk equity>OpenAI later IPOs to get more funding>Elon Musk pulls a Steve Jobs and sells all of his equity>OpenAI stock goes to 0.>Elon Musk buys a controlling stake of OpenAI, becomes the CEO
>>108709446>doesn't know how markets work
>>108709446In reality, the first two steps alone are extremely unlikely.
>>108709453Potentially true, but my retard logic has led me to never lose money in the market, ever.
>>108709446it's a toxic asset at this point, shitload of investor money spent with no plan to return the investment other than "when we reach agi it will find out how to make a profit", quite literally
>>108709091Cool looking build. Thanks for sharing
>>108709453NTA but you can actualy pull this off if you are a whale.ie let's say you own 30% of a company.if you sold all of those 30% quickly, tons of people would panic sell.you could then buy more than 30% with the same amount of money as you made selling them, and if you put extra cash you could get > 50% for a discount.
>>108709469>doesn't know how markets work
>>108709484>muh insider trading
>>108709464That's why God created IPOs to unload toxic assets on ignorant retail investors.
Tuesday!
>>108709484they actualy do work like that, that's why "market manipulation" is a whole category of fraud.it would work, but you take the risk of having to deal with the SEC.
>>108709464they are going hard on the sunk cost fallacy."if you don't invest more we'll not get to AGI and all your money will have been burnt for nothing"lmao.
llmfan46 seems less autistic than drummer, ngl.I'm trying his models now, and so far so good.
>>108709505There are much better ways to manipulate the market than selling low and buying high.I bet even Qwen and Gemma could answer why anon's fanfic would not work. But somehow you people are more retarded and less able of critical thinking than open weight trash.
>>108709522I think the abliterated gemma I have is llmfan46afaik they just ran it thru heretic it's not like a drummer sloptune
>>108709535>There are much better ways to manipulate the market than selling low and buying high.i don't disagree.point is, it'd work and it would be fun even if not the best strategy at all.
whichever anon posted about their Orb frontend yesterday thank you, it's actually pretty good. I like the review/diff feature a lot.
>>108709565nice work, shill
>>108709570thanks I do it for free
>>108709565de nada
>>108707175Am I missing something here? If the guy uses his heretic-derived tool to make models but doesn't distribute the tool, why are they complaining about the license?Like if I took gimp and modified it and then produced and shared an image I made using it, I wouldn't have to redistribute gimp or care about its license
> CFO Sarah Friar has expressed concerns to other company leaders that the ChatGPT creator might not be able to pay for future computing contracts if revenue doesn’t grow fast enough, according to the report.> OpenAI missed multiple monthly revenue targets earlier this year after losing ground to Anthropic in coding and enterprise markets, the report said.> "This is ridiculous. We are totally aligned on buying as much compute as we can and working hard on it together every day," CEO and co-founder Sam Altman and Friar said in an emailed statement to Reuters.> ChatGPT's growth slowed toward the end of last year, the WSJ report said, adding that OpenAI fell short of an internal target to reach 1 billion weekly active users for the artificial intelligence chatbot by year-end.> The company has also grappled with subscriber defections, the report added.Original WSJ article from today is paywalled... https://www.reuters.com/business/openai-falls-short-revenue-user-targets-it-races-toward-ipo-wsj-reports-2026-04-28/
Can we talk about this shit? Literally all the vibecoded UIs all look the same.Orb looks exactly like thisthis >>108709184 tooYou guys need to prompt your UX otherwise everyone is going to know you're a vibeshiter.
>>108709630It's the vibeshitter equivalent of whispers and shivers. It may bother you, but I bet 99% of the population won't notice or care.
>>108709620He distributed the tool then removed the repo
>>108709318What are the odds of Gemini models leaking if the weights are basically being sold?
>>108709630Actually I wanted this UX
>>108709630>all the vibed ui's all work wtf this is stupid
>>108709620it's just license retardation, nobody actually cares except reddit autists and shitty corps looking to hijack foss projects
One thing i am worried about is that if v4 gets actual support even in schizo fork will it gave the same prompt processing speed as usual models despite the compression? I kinda don't like the idea of prompt processing taking an hour at the start.
>>108709685100%. Imo they are already leaked, but since no one has googles tensor whatever gpus, they cant run them, YET
>>108709685>>108709714A lucky few have them and it's called Day 0 Gemma.