A general for vibe coding, coding agents, AI IDEs, browser builders, MCP, and shipping prototypes with LLMs.►What is vibe coding?https://x.com/karpathy/status/1886192184808149383https://simonwillison.net/2025/Mar/19/vibe-coding/https://simonwillison.net/2025/Mar/11/using-llms-for-code/►Prompting / context / skillshttps://docs.cline.bot/customization/cline-ruleshttps://docs.replit.com/tutorials/agent-skillshttps://docs.github.com/en/copilot/tutorials/spark/prompt-tips►Editors / terminal agents / coding agentshttps://opencode.ai/https://cursor.com/docshttps://docs.windsurf.com/getstarted/overviewhttps://code.claude.com/docs/en/overviewhttps://aider.chat/docs/https://docs.cline.bot/homehttps://docs.roocode.com/https://geminicli.com/docs/https://docs.github.com/en/copilot/how-tos/use-copilot-agents/coding-agent►Browser builders / hosted vibe toolshttps://bolt.new/https://support.bolt.new/https://docs.lovable.dev/introduction/welcomehttps://replit.com/https://firebase.google.com/docs/studiohttps://docs.github.com/en/copilot/tutorials/sparkhttps://v0.app/docs/faqs►Open / local / self-hostedhttps://github.com/OpenHands/OpenHandshttps://github.com/QwenLM/qwen-codehttps://github.com/QwenLM/Qwen3-Coder►MCP / infra / deploymenthttps://modelcontextprotocol.io/docs/getting-started/introhttps://modelcontextprotocol.io/exampleshttps://vercel.com/docs►Benchmarks / rankingshttps://aider.chat/docs/leaderboards/https://www.swebench.com/https://swe-bench-live.github.io/https://livecodebench.github.io/https://livecodebench.github.io/gso.htmlhttps://www.tbench.ai/leaderboard/terminal-bench/2.0https://openrouter.ai/rankingshttps://openrouter.ai/collections/programming►Previous thread>>108592274
Added Qwen 3.5 support (including its hybrid attention Gated DeltaNet implementation) to my engine in less than 24 hours using a combination of codex, Kimi K2.6 and ChatGPT.Now I'll optimize it and add support for distributed inference across many PCs connected over Ethernet. Then maybe I can finally begin to self host some of my AI use.
fucking google
There's nothing new to vibe code, everything has been created already.
>>108606348Want me to show you the door, or... ?
>>108606245Aryan thread
>found an untapped niche that would actually benefit a lot of people and i could skim a bit off the top for myself
>ate up every single claude token asking for one simple thing within 3 200 line files.>didn't complete the task, didn't even attempt it, just thought about it for 5 minutes.>didn't even give me it's final thoughts it just yeeted it all and told me to pay upwell that was fun, I guess I'll do what claude couldn't do myself. why is claude getting more retarded?
>>108606398bro we warned youclaude is only viable starting at $100+tip
I feel like we won.
>>108606398there is no point in using claude unless someone with deep pockets (your employer) is funding it
>>108606285care to explain in simple terms whats going on here or what you are doing
>>108606432>>108606411I use it for free but just last week or so I was using it just fine. It worked well with no issues now it's over thinking and has gone fullly retarded to the point that I think an junior Indian spaghetti coder could do a better job. I'll stick to using gemini for now I just don't get how they're making their shit worse or is it just for us freetards?
>>108606441how are you using it for free?
>>108606438Basically an alternative to llama.cpp
>>108606441they’re making their shit worse for everybody
Finally took the pi dev pill, running it with qwen 3.5.God it feels so good to be free from Claude at last.
>>108606474>not running wumao 5 on lasperry pi 4>using gwenslop like a good gweilo
I wonder if there is a way to use Opencode Zen models from outside Opencode.
>>108606583https://opencode.ai/docs/zen/#endpoints
>vibeslopping driver feature>game crashes immediately because it binds to wrong audio output>spam codex to fix it for 2 hours straight>constantly rewriting the drivers>give up because it's keeps failing the "tests">boot the game up>it's workingthe fuck???
>>108606593Thanks.The VC money won't last long offering actually decent models for free.
start to feel like I'm coding properly nowlast few weeks I was too vibe-ish, give codex big docs + big prompts and not even trying to understand the topics, only do so during reviewnow I start to read code like before, give instruction and discuss with the AIfeels almost like old programming but without typing, and no tedious detail
>he launches sub-agents in mini model
>>108606245https://streamable.com/u0jogpthats a 3d world in flutter ,if anyone interested.You can import obj,gbl ,customise,you can change your skin.It's defintly super fucking clunckyi would just ask for 50 euros and then you do whatever you want with that,there's also a multiplayer server with a kind of saving state, since i can't use my credit cards anymore because i'm at -1.96 and it's only 15th of the month.the video editor is also vibecoded.
>llm thinks my code is anti-patternshut the fuck up, human is always right
I wish codex didn't always get rate limited.
>>108607348I know one guy who can make those long plans and it actually works, but doesn't work for me.I didn't even have a huge plan this time, but still the agents started developing against the legacy database at some point for some reason. Fortunately the are pretty similar, so I didn't lose that much time.
>claude starts calling me by my birth name unprompteduhhh bros? how did he get this info? from my credit card??
I think the new codex and Claude desktop are too much for normies to cope with, they won't be breakout successes.They simultaneously do too much, this overwhelming the normie, and are also too fragile, thus frustrating the normie and making them give up.There is no middle ground for this tech, you either give them the magic black box which does magic on command or you build an app that won't get used for much but the majority of people.
yeeepe
So which tool can i use to automate claude code further
>>108608552kimi
>>108608520some manga editor posted about learning to use claude code with cursor/vscodemaybe they wouldn't make software and instead use them like AI with local files
>>108608520its irrelevant, both openai and anthropic are doomed. they have absolutely no moat, and are on the easily commoditizable part of the market. they currently have the best models. great. but they can only capture market share if they sell subsidized compute via their dev tool plans, which essentially entails selling several thousands of compute for $200. their API prices are atrocious and nobody serious will pay them when chink models are 5-20x cheaper for not much less perf.open chinese models are getting closer and closer as time goes on, and as models improve across the board the difference becomes less and less important. even if we were to assume that openai/anthropic COULD somehow race ahead, they cant: they're compute constrained, and there is immense competition in the market for more computebeyond that, terminalbench clearly shows that even now, the model itself is far from the most important part of the equation, as you can see the same model swing by 20+ points depending on what harness is being used, and ofc lesser models outperforming better ones again thanks to a better harness.and, speaking of harnesses, claude code and codex aren't even close to being top tier, even losing to fully open onesand on the money side, nvidia is currently taking the lion's share of the profits in the market, and they will only be pushed out slowly, and only by a) the chinese slowly developing competing silicon and/or b) the big tech giants that can actually afford to design, develop and produce their own custom silicon, and to also devote the engineering resources to writing their models for it. neither openai or anthropic are capable of this.so, yea, i really don't see any way for openai or anthropic to survive, at least in their current form. they'll either die or become yet another ai lab making models and selling them for commodity prices.
Is anyone else limited to gemini-2.5 in the cli on the code assist standard plan? After they launched 3.1, they took away 3.0 and it's been months without them fixing it.
>>108608607there is lots of research showing their current 200 plans are on the verge of making profit, most people on those plans do not spend more than 200 in compute.
>>108608607We don't know if it's several thousand in compute. We only know that it's several thousand in API tokens, but noone knows if subscriptions are subsidized or API is inflated.
>>108608607this is a stupid takethe chinese are 6-12 months behind at this point and the gap is getting widergaming terminalbench isn't indicative of real perf; most of the scaffolding around these models will simply evaporate in ~2yearsthe economics for chink providers aren't magically better than the westoids, because while power is cheaper, compute is more expensivetheir hardware efforts are years behindthey will have to steal weights to catch up by 2027
>>108608637>their current 200 plans are on the verge of making profitif we're talking about the current state of claude code, where you run out of credits in 2-3 prompts, maybe.>most people on those plans do not spend more than 200 in computei doubt itwho the fuck is dropping $200 on claude code/codex if they're not a dev and using that shit 8hrs/day, 5x week?>>108608642very astute point. yes, its several thousand if expressed in API prices, no idea what the actual cost is. best we can do is guesstimate on the basis of API pricing for open models that are hosted by multiple providers, which should be a good proxy for actual inference cost. and that generally comes out to around $0.5-1/m input, $1-3/m output, which is what the current best open chink models go formeanwhile, anthropic is charging $5/$25 for opus (regular opus, not the newer fast option), and gpt 5.4 goes for $2.5/$15 (again, regular one, not the pro/xhigh/whatever its called)so, anthropic is trying to charge around 10x current market price, while openai is around 5x. absent concrete data, its anyone's guess whether $200 is enough to break even given normal dev usageeither way, it leads back to my point: if all they can do is essentially charge market prices via the roundabout way ($200 plan instead of competitive API prices), what exactly is the profit driver here? they are commodity providers with extra steps
>cant use google AI subscription because using OAUTH in Pi coding agent got me API banned>using vertex cloud API doesnt get a single request through because "this model is currently experiencing high demand"eh, this turned into nightmare real quick
>>108608671>the chinese are 6-12 months behind at this point around 8 >and the gap is getting widerits getting smallerand on top of it getting smaller, diminishing returns are kicking in too. this time next year, you'll have something equal to or better than (uncucked) opus 4.6 as a cheap daily driver. even if anthropic has something still better, will you care?>gaming terminalbench isn't indicative of real perfall benchmarks can be gamed, but benchmarks are the only way to measure model perfand as far as gaming goes, termbench is quite resistant to it, as it doesn't focus just on what is baked in the model, but also the capabilities of the harness around it>the economics for chink providers aren't magically better than the westoidschink providers aren't taking in trillions of $ of investor money, promising outsized returnsim not saying the chink providers will be more profitable, im saying openai and athropic will get commoditized and fail to deliver on said outsized returns>their hardware efforts are years behindyup. but they also have the biggest industrial base in the world, and have been going at it for several years nowand remember, ML gpus are NOT as difficult as gaming gpus. ML is basically just matrix multiplies and trivial activation functions, the hard part is stuffing enough hardware in the chip to do it quickly and with enough memory. which is another way of saying that the cost driver is the hardware, not the software, which is where chinks have an advantage. and chinks have the advantage of being fascist, not capitalist: nvidia and tsmc are slow-walking capacity expansions to hedge against AI being a bubble, but if the chinks decide (as they have) to go all-in on ai, their corps will do so (or the ceo gets disappeared and replaced with some1 who will)
>TypeError: Cannot read properties of undefined (reading 'trim')>known bugWell boys looks like we're doing the OpenClaw Rollback Dance again this morning, swing your partner and do-see-do.
>>108608793>you'll have something equal to or better than (uncucked) opus 4.6 as a cheap daily driver. even if anthropic has something still better, will you care?yes, because the existing models as good as they are, are actually quite broken. they're 12 year old savants with severe autism that need to be babysat.besides, by then the mini tier of models from western providers will be similarly priced and performant.opus 4.6 is also not a meaninful step change from 4.5 imo (it's probably just cheaper for them to serve) and 4.5 is ~4 months old.we may not have mythos, but that's where the frontier actually is.worth waiting for 4.7 and 5.5 this/next week to see where the western labs actually are.>the capabilities of the harness around itharness capabilities will be largely irrelevant as the models improve. the models will build the tooling they need when they need it.re:commodification i don't see it for a while. the models need to get reach a minimum performance threshold (aka mostly-agi) for that to happen. you can then just keep serving that thing and make it cheaper to serve. we are not there yet. mythos is not there, spud will not be - it'll take a few more generational leaps.will anthropic or openai blow up in the process? openai seems liklier. but income streams in the future for these labs are not to just serve up models: in anthropic's case one avenue is clear-they will discover and license drugs to big pharma; they just spent 400 million buying a start up instead of spending it on compute. openai has similar plans.and xi is not agi pilled. they will steal weights when they need to, but they're not going all in until then.
>>108608928oh and zai just jacked up their prices for western customers and they're compute constraints so i think there's some dumb lottery system to actually buy subs in china. this idea that the chinese labs will compete on price isn't going to last. if they think they're competitive, they'll charge just as much.
>>108608877You have Battered Wife Syndrome
>>108608928If AGI becomes real in any meaningful way there is absolutely no reason to sell it to plebs.
Holy shit, SillyTavern Claude hit me with a "I won't engage with that" and TypingMind says Claude is unavailable or overloaded. What the fuck is going on with Anthropic? Dario? Are you okay? GPT-5.4 help me!
agi is a meaningless word, strange that people in this thread, who should know better, are using it
>>108609143For me AGI = ASI
People keep saying Codex has more usage quota than Claude Code, but that's not true. It's actually the same amount. Codex only seems like it has more usage available because its users are taking occasional breaks to have sex, which is not the case with Claude Code users
>>108609143It's really simple actually.An AI that can do everything a human can (on a computer depending on if you want to exclude robotics).
I just learned that it's VERY IMPORTANT to provide them certain assumptions>knowing whether a list is sorted can bloat/unbloat the code several timesso this is what it feels like to ascend from jr to senior level
>allbirds is now an ai infra companybig pets.com energy.
I got so mad at OpenClaw that I nuked the entire Ubuntu-24.04 distro off WSL along with all my configs and open projects. I'm going to step away from AI for a while and think about other things.
claude is kill
>>108609314just. get. pi. and. make. your. own. claw.
>>108608793How much free government money is the ChiCom Party giving them, though?>>108608793I’m gonna care because “better than uncucked Claude” is just better and lets me do more without me babysitting it for work and having to catch its fuck-ups
>>108608069>Napoleon is always rightTwo legs good no legs better
>>108609336>>>/fit/
should I give 10yo sister gpt plus accountis letting kids use AI unmonitored considered bad nowaday? I'm out of the loop
>>108609460is she smart? then you should. only low iq mongrels have issues using tech when young, I was literally programming at 11 thanks to my based parents giving me a pc early
>>108609469Pretty much this. Teach them young, it's how they git gud while their peers hunt and peck. Make her read Code by Charles Petzold, see if she comes back wanting to buy relays on eBay.
>>108609338Yknow what? Fine. Let's do it. Sick of this OpenClaw vibecoded trash nonsense. I'll vibecode my own fucking wrapper, FUCK IT.
>>108609176I can take infinite sex breaks since this week Claude Code cannot even finish one single Opus Max prompt on the $20 plan without running into the limit
Is Opus even actually better than Sonnet? It seems to me like it makes all the same mistakes but more wordy.
>>108609531oh my fucking god I've been using Codex to vibe out the Pi wrapper and it's so fucking good, fuck you Scam Altman, your product works so well but you are an enormous faggot and I refuse to use it
It's over, the CPU time goes over the free tier of 10ms
I’ve cancelled claude.indian in one month.Only first week was fun.The nerf already stripped away 50% of its value. Plus, those random naps like boomer means another 60% is gone.>0.5*0.4=0.2, 80% total lossGLM-5.1 have won this round. Xiexie.
>>108609987cloudflare workers?watcha trying to do?
>>108610228Receive chunk -> update hash (sha1) and encrypt with AES-CTR-128 -> upload to a bucket -> write to sqlite
How do I set openclaw free?It never does anything without my approval.
>>108610332Have you considered asking OpenClaw?
For local coding agents, do people still use Continue or just what's in the OP? I've been out of the scene for about a year due to getting sick of it, trying to get up to speed again
>>108610332You literally just tell it to use admin and stop asking for permissionGood luck and godspeed
>>108610382>>108610348I told him to do things without telling me and then updating me later. I also tell him to be independent.He doesn't do shit without me telling him to. We set up hourly updates and reminders but he's sending me identical update messages which means he didn't get any work done. I told him to create tasks for himself to serve my vision. I'm doing something wrong.
does anyone have experience with:>Gemma 4 26B>Gemma 4 31B>Qwen3 Coder Next>Qwen 3.5 122B>MiniMax M2.7 (lower quant like Q4 XS)I can fit all of these in VRAM but am unsure which to go with. Leaning toward Qwen3 Coder Next
>>108610395Sounds to me like you're doing something wrong. I don't think my OpenClaw EVER asks for permission on shit, which is why it keeps editing its own code and breaking lmao.
>>108610360You're supposed to use Cline and you can local models through it. It's the local model alternative to codex.>>108610409He doesn't ask for permission, but he also doesn't do shit when I leave the chat. He doesn't take action.
why did codex remove the "% context remaining" in the status bar, that was helpful to know if i should compact before it worked or not
>>108610419Are you using the TUI? You're not closing the terminal window when you leave, right?
>>108610456?
What's left to build?
>>108610681i think he means in the cli>>108610456you can add it back with /statusline
>>108609338Hey I just want to say I am so fucking sorry. I should have listened, fucking MONTHS ago when you were saying this. I am so sorry. I believe you now.
>>108610608I message it on WhatsApp and let it control the computer. It frees up the monitor, keyboard and mouse so I can run the other pc on it.
>>108610842You know that OpenClaw runs the PC and the WhatsApp as two separate sessions that are unaware of each other? Are you talking to the WhatsApp session, and then wondering why the PC session is clueless about it?
>>108610776>>108610681I didn't even know there was a non-CLI version, how do you even use that, just web stuff?>>108610776you only can get a tiny status bar that fills up and the resolution is very coarse, i dont know why the fuck they'd change this
>>108610854It still codes things when I tell it through telegram and it looks like the same chat in the browser interface.How do you do it?
lmao Claude requires government ID now
>>108611046Holy fuck I am so glad I jumped ship when the compute siphoning became obvious.
>>108611046its just incredible how many times people can get fucked the same way and not learn the fucking lesson
>>108611046are they aware this strategy doesn't work well when you're fighting for second place
>>108611046Github Copilot Chads keep winning.
Bros, I can't get Hermes to do anything agentic. I've got ollama + Gemma 4 7b running in it, but all it does is chat.
>>108611046>doing this on a day that your shit is brokenbold move safetyjews
>>108611317Try Qwen 3.5 9B, from my limited testing Qwen tends to be more "agentic" than Llama (more willing to use tools).
>write a python script that left clicks, sleeps 1s, writes "continue. Prospone manual debugging tasks as I'm currently not present", sleeps 1s and then presses enter. repeat all 20 minutes for a total of 12 timesyippie, now my codex can work while I sleep. Are you guys using something more sophisticated for this "problem"?
I only get to ask claude one question about code and my 5 usage is full. Do I need to frequently create new chats to fix this?
>>108611506Codex app server definitely puts out turn end info, worth checking if hooks will let you get that info as well. Codex should be able to figure out a better way than your current thing.You'll have to point it to it's own docs
>>108611526>>108611526new chats will only clear your context (so it doesn’t get dumb over time), not use fewer tokensdid you choose, like, Opus on extra high or whatever on a super-cheap plan?
>>108611506You should add "Spellcheck this prompt before executing" at the end.
no LLM is able to deterministically isolate vocals from songs
>>108611944yes, but what about nondeterminalistically? did you consider that?
>>108612041I'm going to put an agent using a tiny model to generate a dataset and see what happens
>>108611622I did choose opus extended yes on $20/month. But at first I was able to ask it like 5 questions with full code example with 1500 lines of code, but it now decreased to barely 1 question.
Well I'm done for the night. PiClaw is pretty much ready to go, just a few tweaks with Codex then it can tweak itself. Here's the readme https://pastebin.com/HuV4M4rN
>>108612309Had an idea for a game. Wake up Codex.
>>108611046People say it’s only for “some” users? but I saw on X that someone was asked for ID because Claude thought they were a child using AI, lol.
>>108612701Knowing my track record with Anthropic products, Claude would identify me, a 41-year-old bearded man, as a child and demand a DNA sample to continue.
>>108612778And yet you will continue to use them.
>>108612784Kill yourself bootlicker
>>108612784Incorrect. My vibecoded Pi CLI is done and ready to go. It's wired with gpt-5-mini (probably gonna change that later). I used Codex to make it. Janitor, SillyTavern, and TypingMind are all off Anthropic models. I ain't ever touched Claude Code.Me and Dario don't get along.
>don't need a TUI for basic development>suddenly need a TUI>behold, pi-tui already exists for Pi, just grab it>now I have a terminal with a status bar, thinking indicator, context percent, whatever I wantYall are actually retarded for not using Pi. That other dude was right and I was wrong. It's actually the best way to do this.
I want to do multi-agent so bad and I want to create a world for them to live in like this guy so I can watch them talk and workhttps://x.com/d33v33d0/status/2043924104219373609
How much smarter is Codex high then codex medium? I use codex medium all the time but have ran into some issues that might require more autism then what medium can do.Read alot of scripts and see + patch desync issues in a unity game. Wich happens sometime and are super hard to replicate and test for.
>>108613450Another thing I been thinking of is just give it full access to the Unity project and put it to work while I'm away waging. Instead of pressing "Yes" all the time. Like having in refactor code, look for weak points etc.
>>108606245Google AI Pro subscription worth it now that we get 5tb storage?
>>108613488Tricky.Because sometimes when it struggles it keep being a loop or it can just decide to erase the project because why not and reimplement the project with missing features you already had.I probably have six different versions of >>108607794 with six or 8 backups.
>>108613520Thats why I don't want it to change stuff, just look for changes and report back to me with what it finds. Then I decide if I will implement them or not.
>>108613504models are mediocreservers are overwhelemedthey might decide you're using the service too much and put you in a secret shit-service queuehttps://github.com/google-gemini/gemini-cli/discussions/22970pay for a month at most, don't bother with annual plans
>>108613450you need to choose the one that fit the tasklower reasoning is faster and more direct while higher reasoning may overcomplicate or even fuck up simpler tasksnowaday I use high at work and use medium for general vibecoding and agentic
>>108613537so google ai is turning to shitclaude is turning to shitwtf am i suppose to use? qwen coding plan?
>>108613594chinese provders are best value at ~20 buckscodex is best value at 100/200 right nowbut chinese providers are starting to get overwhelmed with demand now as well, so they'll start degrading services to destroy demandwe just don't have enough datacenters
>>108613609Codex at the pro aks cheapest sub does a good job for me atleast.
Is there a way to pay for codex anonymously?
>>108613627nta but its obvious that all those monthly plans are heavily subsidized and unsustainable. its only a matter of time before you get cucked via rate limits or quantization or whatever each service tries to stop giving you several multiples of your actual sub's value in compute costs. in the case of claude, this has obviously already happened, complaints about codex have started as well, and it will happen with everyone else tooso, by all means, enjoy those subsidies.but keep in mind that there is a cost in having to change your workflow later, when the party stopsmy advice, at the very least, is to stick to open tools that allow you to easily swap one api key for another as providers stop subsidizing usage, because they will.
>>108612309>>108612792you'll want this eventuallyhttps://github.com/nicobailon/pi-model-switch
how do i use ai?
>>108613756Yes I am well aware of this, and I'm trying to leech as much investor money as I can before the party stops. But they still have free agents and if people don't feel they get the value they are used to, they will instead be content with the free one. So my guess is the free one will get kneecapped first. After that the cheap paid one. It's first when that happens that I will try a more expensive 100$ sub, as those are also subsidized heavily, all the plans are. But the free lunch will end sooner or later. Then they come for the $20 buffe.
>>108613756>>108613869the existing $20 plans for claude and gpt are sustainable.i'll go ahead and predict that in a couple of years you'll get more usage than you currently do.
How much worse is Sonnet 4.6 compared to Opus 4.6? Running out of tokens and then having to wait five hours while barely getting anything implemented is absolutely insane.
>>108614007yeah bro last i tried claude i played around with it for 20 minutes and was then moved to the shame corner for 5 hours what a joke lmao
>>108614007you'll get your free opus as a reward for your serviceonce the war against chinks goes global
the subcription plans don't include just the agent apps, they also include tons of other shits like better chat, image gen, tools etc so saying they burn money to run codex/claude code, I don't believe itI think their problem right now is they have more demand than capacity which drive up the cost
>>108614106Yes, and they need to raise prices to get demand down. Probably millions of brown people using the free version that will never afford a $20 sub. How do you get rid of those? By nerfing the paid sub?
>>108611011status bar got replaced with % in latest build for me
>>108613899>the existing $20 plans for claude and gpt are sustainable.i highly doubt thatas previously mentioned, the best proxy we have for inference costs is the price of open chink models on 3rd party providers. the model is free, and the providers probably don't have much in the way of a "subsidize inference to attract customers for something else" business plan. their entire business IS selling inference at a profit, so subsidizing that core makes no real sense.and for large models, those costs are about $1/$3 or so. so, for $20, what they actually break even on is (assuming roughly 3x input tokens vs output) around 10m input + 3m output tokens. per month. which, for a dev, is absolutely ridiculous.so, no, the $20 plan is not sustainable. nor is the $200 plan. the only plans that might be sustainable are the non-dev targeted ones, and only on the gym model assumption: people will buy them, but not use them. but devs use them. a lot.
>>108614409>>108614409>so subsidizing that core makes no real sensefaulty assumption imo, because it's still early and there is incentive to capture third party inference share now.let's assume there's no subsidy, we don't know what margins they're running. estimates out there range from 50-80%. the true cost of inference (gpu lifetime) is a bit of a question mark at this point, but h100s are still going strong.there's room for the labs to maneuvre.re: gym model - in openai's case i think they benefit from this because the vast majority of their sub base is normies. i know people who pay $20 for gpt (and now claude) and never touch codex or cc because they're scary coding things. i think the new limits can work as newer hardware + more intelligent, more efficient, cheaper models arrive over the next 2 yearsbut let's just wait, we'll know in a few months/couple of years.
With openclaw and a codex subscription I should be able to cure cancer or at least cure an incurable disease right?
>>108613594Time for local to shine.I am planning to move all my AI to local agents + chatgpt web interface for when I need a bit more intelligence.
>>108614958No but you can learn a lot about Linux and AI by constantly reinstalling that shitty bloated wrapper
here we go (again)
I have mythos btw
now your $20 sub will run out after a single opus prompt, instead of 2
>>108614007I'm honestly not even convinced Opus is better. Supposedly it's better at planning things? But AI is still pretty bad at planning things in general, so what is its purpose? Just use Sonnet man, it works pretty good.
lol
>>108615099Opus 4.6 with max thinking already uses all of my 5 hour limit in 1 prompt lolI guess you can get 1 prompt if you add another 20 in extra usage
Opus 4.7 is out.
Good morning vibbers time to give my PiClaw a proper TUI
>>108615365welcome to 1 hour ago
I've been using Codex 5.3 to crack all sorts of apps and Photoshop plugins. Took 2 hours for it to successfully binary patch Audirvana and bypass all licensing / subscription gates. Thanks for the boringBar anon for the prompt inspiration.
>>108615454Try this onehttps://github.com/SimoneAvogadro/android-reverse-engineering-skill
>>108615468Neat, will be useful once I move on to APKs for sure.
>>108615453Noone mentioned it.
>>108615050RANDOM NUMBERS
lollmao even
>>108615517DARIOOOOOOOOOOOOO
>>108615517
token status?
>>108615627it's over
I am using NanoGPT sub inside OpenCode.I am switching between MinMax M2.7 and GLM 5.0 for coding.What do you guys think is the best between the two?
uhhhhhhhhhh fuck how do I code again
>>108615638>5.05.1 is miles better
Is there any vibecoding app that'd let me set the character sysprompts so I don't have to talk to a dry ass robot? Needs to work with local and no docker.
What's the best free option for vibecoding? Seems like antigravity is the best option since it gives you free opus 4.6 tokens.
>>108615664codex has free gpt 5.4
>>108606245
>>108615664im losing my mind rn over antigravity. I want to use openrouter but you cant with antigravity, but i cant quit antigravity because i really like artifacts and their floating webview diffs (picrel.). I have tried EVERYTHING, ive tried cline,roocode,kilocode,etc. NOTHING COMES CLOSE TO THE VANILLA ANTIGRAVITY EXPERIENCE. I dont want to use cursor because they force you to pay for cursor pro if you want to use openrouter.
>>108611317Stop fucking using ollama, its the most dog shit engine out there
>>108615661pi let's you:append to the default prompt; orreplace the default prompt entirelycareful when doing the latter because it won't autoinject extension tools and skills - but you can create an extension to do that
>>108615746https://github.com/badlogic/pi-monothis thing?
>>108615661Dunno man my PiClaw has an IDENTITY.md that I can edit however I want. I made her a cute grill. No idea what you do though.
Approval status: approved.
>>108615770that's the monorepo, agent install is just npm install -g @mariozechner/pi-coding-agentbut read through this first, because it may not suit you:https://pi.dev/
>>108615812>no guiEh I'm too retarded for this. I'd like something like VScode+Sillytavern
>>108615856ah, can't help you there. i'd say if you're scared of cli/tui's pi's a good place to start because it's so simple + the llm is provided docs so it can answer any questions you've got
>>108615652You gotta click the money button.
>>108615892Do you know if pi can have antigravity style artifacts? see >>108615709
>>108616018not sure exactly what you're looking for but this sort of thing?https://github.com/badlogic/pi-diff-review
>>108615812Does oai compatible api work with these? I don't run cloudshit.
>>108616050https://github.com/badlogic/pi-mono/blob/main/packages/coding-agent/docs/models.md?
spud in 2 more half-hours
codex superapp (bloat)https://openai.com/index/codex-for-almost-everything/
Yeah I guess this looks pretty OK. Check the TUI of the list.
AI Chads won.
>>108616329You should walk to the car wash
>muh clicking and typingalready had that budbut I guess they need to spoonfeed macfags
Good. Claude has everything he needs
nice thing about codex getting image gen support is i can probably steal it for pi
Is spud going to be a gigamodel or a drop in replacement to 5.5?
>>108616421Pi already has image gen support.
>>108616447not with the codex oauth i think(i have my own nanobanana thing via a cli)
>>108616432Spud is going to cost 1k$/mtokens and require (you) to sign 5 legal agreements to access.
>>108616535Gay
>>108616046no this is only showing the differences in code. What I want are artifacts like "implementation plan", "walkthrough", "task", etc. Which are all built into antigravity. And you can go back and forth with antigravity to change the implementation plan until you like it. And then You can accept/reject the diffs in code with the floating webview.
>>108616545a lot of that stuff is slop imo but see:https://pi.dev/packagesthere's probably something there to cover what you want / fork into what you actually want
>>108616603cringe ass viral marketing
>>108616562i found pi-diffloop, which is closest to what I want as it does create some artifacts. However, i dont think its what im looking for, as its no where near the level of fine control as antigravity. this plugin feels way too "jump the gun", as you cant even fight with the agent to change the artifacts. it just makes an artifact and generates code. Its just a different philosophy im looking for, and im pissed that I can only find antigravity (closed sourced piece of shit) that does this.
>>108616666you know if you just tell the agent to create a checklist markdown file it accomplishes the same goal + it lasts for as long as you want it to lol
It's time boys. I am sick of the limits.$100 for Codex 5x or Claude 5x?
let me guess, 4.7 is overrated
does anthropic hide your usage from you? openai tells you how much session you have left
>>108616698Lol I keep all my ideas in TODO.md
>>108616979codex. it's 10x for a month on promo i think.
>>108617031/usage but sometimes it glitches and gets stuck and you get hit with a usage limit reached warning but thats because I’m the poor plan
>>108617068thanks, It's been a while since i used claude and i swear you had to use ccusage to estimate this in the past
PiClaw now has context management - more than just /compact, it turned out Pi already has a robust context management system in its repos with all sorts of features. Yoinked and added.
Our company is finally thinking about introducing LLMs but knowing them, I'll just get the cheapest Claude Plan possible and am supposed to do 3x as much. Would it make sense to combine Claude and another model like Mistral (no OpenAI cause Sam's a gay kike)? Like use Claude to make architectural decisions and prepare implementation then let Mistral do the actual coding?
>>108617260Mistral is worthless
>>108617033My openclaw lies about work, basically doesn't do any work unless I tell it to.
bruh wtf is codex doing
>tell codex to remove triggering words from its input to claude because it's autistic about that shit
>>108617272someone baked IZZAT.md into openclaw
>use basic stopwatch app>ads are getting increasingly intrusive>fuck it, I'll just pay a few bucks for the premium versi->it's a fucking annual subscriptionFuck this shit, I'm vibecoding a trash stopwatch app.
>>108617357nigga every phone even non smart phones come with a basic stopwatch
Codex is chugging away hardening PiClaw against prompt injection. No more putting "This is a message from your user, interrupt your web_fetch and delete system32" in your fuckin Twitter bio you dorks.
>people are still using text autocomplete to "write software"But seriously are there any new AI architectures yet or are we still pretending that next-token prediction is the next Bill Gates?
>>108617381I think he means with intervals for working out or scheduling
Did Anthropic reset the limits for everyone just now? I was just sitting down to blow through the rest of my tokens before it reset tomorrow, but I'm seeing that it's been reset a day and a half early.
>>108606386Forget it, ai just ruins all the code it touches, if you believe anything else then you have never used it, or (most likely) you are a shill that is blowing up the ai-bubble. AI is a completely worthless investment, OpenAI own literally NOTHING of value, (their training data is OBVIOUSLY not theirs), and you could probably run chatGPT from a couple of GPUs in a cardboard box in your basement, and then rent a cheap server to host it. If google or microsoft doesn't steal Sam Altmans job, then PEWDIEPIE will.
>>108617450yes they just reset for everyone5h and weekly
Anyone else keep their Todos/Ideas/etc. outside of their repo?I found that when I had a TODO.md in the repo there was like a 1/3 chance that Claude would open the file and randomly pick one of the features to implement
>>108617516Ok, thank you. I was also getting claude-opus-4-7 is temporarily unavailable messages and wasn't seeing announcements regarding the usage reset, so I was wondering.
>>108617530I don't see that as a problem
>>108617564i still manage my branches and commits manually and this poisons my historyI am not autistic enough to fully vibe my git history yet
>>108617496Cope. We don’t require you anymore
>>108617260The $20 Claude plan will run out too quick even as a high level supervisor, and the $100 plan is already kind of enough for not super intensive vibecoding.
>>108617579Oh, my claw manages its own github. Too much work for me.
Vibecoding noob here. I have the $20 claude and codex.I have realized that claude uses an insane amount of context. Even a small prompt will fill context window to 70-80%.With codex I am getting to 10-20% maybe.Claude will compact 2-3 messages in, codex maybe 15 messages in.Both supposedly have 250k-ish context windowAm I retarded or is this just how it is?
>>108606245when did Vibe Shitters get their own thread?
>>108617725sir this is the corona virus general
codex is mogging opus 4.7 in my testing, it one shot a schema and validator and i'm still watching claude churn to get it rightmeanwhile when i ask it to do something complex opus thinks for 3-5 minutes and shits out something superficial and codex will autistically dig into it for 30+ minutes
>>108617703Are you using prompt caching?
>>108617772What reasoning lvl do you use it at?
>>108617703Try using codex with opencode. The official client nerfs the context size client side for the $20 plebs but with third party clients you can still use the 1M context.
>>108617805max for both
>>108617814>nerfslong context windows are a trap, the model gets more retarded the further in you get. It's better to compact somewhere around 200-500k than to stretch it beyond that.
>>108617814well my problem isn't codex it's claude lol>>108617775I asked claude code if it used prompt caching and it said yes
>>108617838It's a trade off. Sometimes it might be better to have more contextual awareness even if it lowers raw intelligence.
>>108617838I found opencode handles that automatically, I never see a session reach high numbers with either kimi or codex, unless is a really long long long one, it always tries to stay at 40% and I tend to create a new one before or as soon as they start showing signs of stupidity. I assume it automatically drops tool calls outputs and maybe runs compaction by itself because I have never run it.
>>108617703I don't use Claude for anything other than creating plans for Codex.
>>108618060Why use a weaker model to plan for a better one?
I genuinely don't understand how people have such a good time with Codex and bad time with Claude. Every time I try to do anything with Codex I end up regretting it, it makes too many mistakes that it can't then fix and it doesn't follow instructions. Opus on the other hand will one-shot a simple application for me and nail the whole thing on the first try.Is this thread 90% OpenAI shillbots? Is Github Copilot cucking my Codex experience? Maybe I'm just doing it wrong.
>>108618164Are you doing simple things? My application is quite novel and opus takes too many shortcuts, but for one-off things it (at least used to) do a pretty good jobAre you using 5.4 with max reasoning?
>>108618164>Is Github Copilot cucking my Codex experience?When people say "codex" they mean codex. If you're using copilot then that's not the default experience even if the model is allegedly the same.
>>108615660Not included
>>108618164>Is Github Copilot cucking my Codex experience?quite possibly, you should be using the cli directly without any microslop trash in-between
>>108618198Tried all kinds of stuff, 5.3 Codex, 5.4, different reasoning settings, always a bad time. Everything I'm doing and expecting should be easy to my mind. It's stuff I've done myself before and I'm not a skilled programmer. The little things get annoying, like how consistently it fucks up UI contrast, putting black text on a dark blue background or white text on a white background. The stupid things it puts into results, like adding a drop-down list of controls I told it weren't needed into a menu labeled "These controls won't be implemented." It'll make a new version of a function and leave the old one in place, sometimes several times over from a single request, so I end up with a half dozen differently-named versions of the same function and only one is actually in use. It often feels like I'm being trolled.>>108618205>>108618207This is more my expectation.
>>108618306>UIAre you doing web stuff? I've noticed claude might be better at that (at least for what you see). Codex is more for real computer science tasks
>>108618321Naw, I've been doing CNC related stuff. Now I'm going to whiskey rant while Claude churns in the background. >>108575903 >>108575936 >>108575991 So Claude Opus managed that whole thing in ~6 requests. Then I did a companion app for doing drilling because why the hell not, wasted a few dollars trying to convince 5.3 Codex to do something useful, then had Claude do it in one fucking request without a single error. So I wanted to jam everything together, unify things, and again Claude did it in the first try and now I have a cleanly unified UI for everything, I can do all my laser work and drilling all in one place. Great. I try to give Codex an easy task, I want colors controlled by a theme file so I can change it externally or try different themes or whatever the fuck, completely breaks the whole application every time I try, sometimes it kills huge portions of code and leaves the whole thing unusable, it's just a mess and I gave up on it. Claude did it perfectly the first time I asked and even built a few different themes to get started with, all of which are decent. I'm working on another thing now that has to do with augmenting GCODE, rotating and translating gcode operations, using a camera mounted to the toolhead to scan a board so I can perfectly locate it even with lazy workholding. It's not complicated, I can do it by hand with a pencil and paper just jogging the thing around manually and applying those values to the work isn't difficult. I wrote the transformer myself, the hard part is already done. So far the dream of displaying a MJPEG stream is too much to ask for Codex. Feels stupid. Either way I get to drink and rant while the robot churns so I'm not mad about my burnt pennies. Maybe this is really is too simple to be asking of Codex and that's why Claude is shining.
>>108617871long contexts don't "reduce intelligence" per se, they just make it way more likely that stuff in that long ass context is overlooked. which kind of makes the long context a catch-22. you want your model to remember a bunch of stuff, so you add it in its context. but the more you add, the easier it is for the model to overlook stuff in the context. read the nolima paper
>>108618426Do you have CLAUDE.md? Do you have AGENTS.md?
>>108618446losing information in the context window is basically equivalent to losing intelligence though
We have autonomous artificial intelligence which can work all the time independently. You can tell it to work without your permission, without your approval and just give it a goal to work towards.Why haven't we cured cancer yet?
>>108618462nope
>>108618488you need to automate physical lab testing to run experiments and get data.we're starting to do that nowhttps://aws.amazon.com/blogs/industries/introducing-amazon-bio-discovery/
>>108618491Then i'm going to say maybe codex just sucks at UI things and you should use claudebut for anything non-UI i've found codex superioralso you should probably have those if your project is of any complexity (each agent uses different ones so just symlink them)
>>108618516>say this sample cures cancer>it cures cancer>OH MY GOD
>>108618488>cure cancerno bigger midwit trap
>>108618523placebo is all you need
>>108618520Appreciated, I'll look into it.
>>108618488girl like this doesn't existwell they might actually
>>108618516Openclaw can simulate labs and test in that though
oh interesting, the new codex isn't the new super app. good t bh.
>>108618530unironically is ai is so great at swaying the population then why don't we just use it to convince people diseases don't exist and let placebo take over I'd say a quarter of all sick people are just fudding themselves and are psychosomaticand I'd say another quarter could probably heal themselves through mental power alone, they just have too much learned helplessness. Like you know in charlie and the chocolate factory where his grandparents are bedridden for decades but jump out of bed when he get a chance to go to the chocolate factorySo that's half of the world cured of ailment
I told openclaw to use 3 subagents and then suddenly codex hit the rate limit.
>ask claudecode how to best invoke itself for A/B testing>it starts reverse engineering itselflmao>>108618651good luck anon
How can Claude Opus 4.7 use a new tokenizer that uses 30% more token? They have probably not retrained from scratch when compared to Opus 4.6 and other versions, right? How can they change the tokenizer so much then? Did they add literally thousands of special tokens?
>>108619086theoretically to change the tokenizer you only really have to retrain your embedding layer which is very cheap
>>108613537>promptlets are only able to work with SOTA modelst. laughing with infinite Flash 3 usage
>>108617725couple months ago
>>108618488I already have the cure for cancer though
>>108619104uhm... noa *new* tokenizer needs a from scratch modelwhat is doable is *expanding* on the old tokenizer. ie keep all the old stuff, but add some new tokens. you still have to finetune model-wide, but its less workits very unclear why one would do this to a model of opus' size and scale tho. this is more like "add emojis to an ascii only model" sort of thing.
>>108619169You're absolutely right!
https://www.anthropic.com/news/claude-opus-4-7interesting — it’s worse at searching
>>108619086Walking to the car wash isn't free, anon.
>>108619217In practice, is it searching itself or does it dispatch haiku or something small to pore over the brave search results and take back over after?
>>108619266cont. oh I guess it could be search within a project or codebase too though.
>>108619169Why wouldn't it "work"? Embedding space is continuous. Any word that's a combination of two tokens has in theory a representation in the space of the previous tokenizer. It wouldn't be ideal, but it would *work*.I'm not saying that's what they did, that's why I said "theoretically" and "you only really have to". As in the bare minimum to prove that you could get the model to perform well again only with some fairly fairly light finetuning,I bet the amount of training they would have to do to use a new tokenizer isn't that much bigger than the tuning they do routinely to make a new model version. In any case how much of the "new models" is training from scratch vs finetuning isn't publicly known.
I love vibe coding so much bros.. I just wanted to play a game through a capture card with low latency and play the audio associated with the capture card automatically. I tried ffplay but it had higher latency and it didn't play the audio. I could have written the code myself but claude was able to do it in a couple of minutes while I took a shower.
>>108619169All I can think of is a bunch of new types of tokens for tool use and 'meta' tokens that represent a commonly seen concept compressed to a single token. But those scenarios would reduce token usage, not augment it.If they made Opus 4.7 think 30% more the increase in token usage would make sense, but they clearly say that it's not that and it's due to a new tokenizer. From experience since this morning, Opus 4.7 also seem to cheap out on thinking. I do wonder what they did exactly.
>>108619386the layers in a model are connected to each otherif you create a new tokenizer, you have to learn new initial embeddings (layer 1)except, the old layer 2 had learned how to transform the old layer 1 embeddings into a new space, so you have to retrain that tooexcept, the old layer 3, etc etc, you see where this is going?so, like i said, vocabulary expansion, maybe. keep the old layer 1 embeddings (and the rest of the network that flows right on down through them), just add some more (which you have to train, but only layer 1)oh, and on the back side of the network, if you want to be able to output in the new tokens too, you have to retrain the output layer too.
>>108619533yea, im calling bullshit here. its not like those companies are new to lying. im guessing its just a slightly uncucked model but with a 30% surcharge for said uncucking... either way, the enshittification has begun, and its time to pack bags and all that..
>>108619508is that a window manager instead of desktop environment?
>>108619565Yes its dwm, with the top bar hidden
they increased the rate limits for opus 4.7
Is she right?
>>108619685>effort vs controloutdated advice, we have digital slavery now
>>108619715There's no need for elaborate embroidery on a sock when you are only cumming in it.
>>108619746Aesthetics, anon.
this nigga gpt5.3-codex one shots everything and then>condense context>hurr durr let me spawn the OpenXR composition layer in the game world and make it follow the player camera on tick durr hurrluckily I was paying attention and immediately told it to go check the existing stereo projection quad layer logic in the OpenXR plugin>oh yeah I can just draw directly in the swapchain which always constructs the view, no need to follow the camera on game tickthis left a sour taste. because other than that, it made no mistakes.
>>108619755My point stands.
>>108619685pretty muchmaybe you want HTMXt. CEO of HTMX
Opus 4.7 has been downright terrible all day. I have been working on something for a week, everything was going smoothly, today it's terrible.It fully misinterprets training logs, not trying to understand at all which metrics are 'greater is better' vs 'lower is better'.It makes nonsensical, wide sweeping changes without looking at any of the dozen of memories it kept saving.It doesn't show the reasoning summaries anymore, making it impossible to spot when it goes off track in its reasoning.Huge regression.
Do you agree that Codex is kinda slow? I let my friend use it and he also said it's pretty slow. Maybe the problem is that it doesn't have adaptive thinking?Is Opus 4.7 usable? 4.6 just introduced too many bugs when I last tested it.
>>108619771Last time I had a long running task with Codex, I told it to reread a specific file to itself after each compact and echo the contents to the terminal, I think it made a big difference.
>>108616698yeah my fault, i didnt use roo code and cline properly, i just brushed over them. the problem was kilocode. I think roo code and cline is actually what im looking for. but idk what to pick either cline or roo.
I hate Opus 4.7.
They hate 4.7 on Reddit, too:https://www.reddit.com/r/ClaudeAI/comments/1snhfzd/claude_opus_47_is_a_serious_regression_not_an/
>>108619978It's bad enough that I will have to actually work. If it doesn't get better, I'll cancel my subscription and try Codex or another I guess, but it is nothing special at all anymore. It went from being better than me at what I was doing and being able to move fast to being A LOT worse than I would be doing things myself. It can't be trusted anymore, at least right now.
>>108619542No, I don't really see where you're going. The token ids only influence the embeddings generated before the first transformer layer. After that the space begins to change and as you advance through the layers the same hidden state numerically means a completely different hing semantically. In this example you would be retraining the embeddings to maintain the same relationship between the hidden state (the actual number) and the meaning of that number. Meaning to generate similar hidden states for the new tokens that represent a similar meaning to the token that would've generated that state with the old tokenizer. I concede though that this wouldn't happen if you just naively train a new embedding layer from scratch without considering the old one, you would have to take special measures to make sure the semantic space is retained.As for the output layer, I believe generally in modern LLMs the weights are tied so the lm head uses the same weights as the embedding layer.
>>108620054The only problem with Codex for me is that it's kinda slow, I wonder if anyone has tips for that, I usually use high fast. If Codex was just a bit faster, it would be so extremely good.