[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/g/ - Technology

Name
Options
Comment
Verification
4chan Pass users can bypass this verification. [Learn More] [Login]
File
  • Please read the Rules and FAQ before posting.
  • You may highlight syntax and preserve whitespace by using [code] tags.

08/21/20New boards added: /vrpg/, /vmg/, /vst/ and /vm/
05/04/17New trial board added: /bant/ - International/Random
10/04/16New board for 4chan Pass users: /vip/ - Very Important Posts
[Hide] [Show All]


No longer down for maintenance!

[Advertise on 4chan]


File: 1753683943355333.png (1.74 MB, 1380x752)
1.74 MB
1.74 MB PNG
A general for vibe coding, coding agents, AI IDEs, browser builders, MCP, and shipping prototypes with LLMs.

►What is vibe coding?
https://x.com/karpathy/status/1886192184808149383
https://simonwillison.net/2025/Mar/19/vibe-coding/
https://simonwillison.net/2025/Mar/11/using-llms-for-code/

►Prompting / context / skills
https://docs.cline.bot/customization/cline-rules
https://docs.replit.com/tutorials/agent-skills
https://docs.github.com/en/copilot/tutorials/spark/prompt-tips

►Editors / terminal agents / coding agents
https://cursor.com/docs
https://docs.windsurf.com/getstarted/overview
https://code.claude.com/docs/en/overview
https://aider.chat/docs/
https://docs.cline.bot/home
https://docs.roocode.com/
https://geminicli.com/docs/
https://docs.github.com/en/copilot/how-tos/use-copilot-agents/coding-agent

►Browser builders / hosted vibe tools
https://bolt.new/
https://support.bolt.new/
https://docs.lovable.dev/introduction/welcome
https://replit.com/
https://firebase.google.com/docs/studio
https://docs.github.com/en/copilot/tutorials/spark
https://v0.app/docs/faqs

►Open / local / self-hosted
https://github.com/OpenHands/OpenHands
https://github.com/QwenLM/qwen-code
https://github.com/QwenLM/Qwen3-Coder

►MCP / infra / deployment
https://modelcontextprotocol.io/docs/getting-started/intro
https://modelcontextprotocol.io/examples
https://vercel.com/docs

►Benchmarks / rankings
https://aider.chat/docs/leaderboards/
https://www.swebench.com/
https://swe-bench-live.github.io/
https://livecodebench.github.io/
https://livecodebench.github.io/gso.html
https://www.tbench.ai/leaderboard/terminal-bench/2.0
https://openrouter.ai/rankings
https://openrouter.ai/collections/programming

►Previous thread
>>108582257
>>
is kimi better than minimax?
>>
>subscription tokens are use them or lose them
>rollover tokens, are you crazy? matter of fact, they expire even if you buy credits!
>you know those tokens we give you? you're gonna have to wait five hours before you can use more of your weekly allotment ;)
I seriously hope somebody sues the shit out of these faggots
>>
>>108592466
What do you want them to do? Keep the tokens saved inside the gpus? That would cost a lot too, be pragmatic.
>>
>>108592466
if you don't want to be jewed, don't subscribe to opaque "plans" to use proprietary software powered by closed models, peddled by companies with known and established patterns of unethical behavior
i mean, how many fucking times must the same thing happen before normies learn?
>>
Sorry for the dumbass question I know i should go and fucking read or lurk more or what have you but hoping some anon can give me some advice.

Where's a good place to start with this locally. I've used plenty of diffusion models and some llm for chatbots and prompt writing but never for coding. Hardware wise I've not got the most powerful pc, but not potato tier. 5070ti, so 16gb vram, 96gb system ram. So the big stuff isn't something I can run locally.

I'm looking to start with pretty basic stuff, custom nodes for comfyui with python, maybe some css.

Any suggestions?
>>
>>108592601
even with a well above average spec like yours, the stuff you can run locally isn't really worth the time and effort. or the power consumption.
like, seriously, its probably cheaper to pay for deepseek 3.2 on openrouter than to pay the extra electricity your gpu will burn
>>
>>108592634
I don’t care about that cost honestly. It's a hobby and I like having stuff locally if I can. Like all in it's probably cheaper to do diffusion like that too but still nice to be local I think.
>>
So, Anthropic's official research showed that giving agents personas and ways to remember praise and punishment increases performance. Should we both be vibbing that as a plugin, or do we just wait until they implement it themselves? I probably would only keep the praise part; I don't like the idea of torturing my agent.
>>
i really fucking hate how sycophantic claude is
>see blogpost related to idea
>paste it into claude and codex, ask it what we could apply within our codebase
>claude immediately simps for me and says it's a great idea o algo (then launches a cheap shitty subagent to do all the work and jew me out of a decent result)
>codex replies it's looking into it and searches directly with its highest quality model and thinking
>>
>>108592683
did you put "dont be a sycopanth" in your agents.md?
>>
>>108592665
for coding, you should probably look into the qwen3 coder series. you can definitely fit one of them if you use vram+ram, and with quants you can probably manage vram only for much higher speed
you should ask in /lmg/
>>
>>108592601
You could technically run a powerful Moe with your rig but you would suffer because you would have to offload so if the weights to your system RAM. I forget what the exact CLI arguments are but I'm pretty sure llama.cpp has specific settings where you can basically run the "experts" all in VRAM well the rest of the model sits in your system ram. It's still slower than being able to run it fully on a GPU but faster than the default offload settings. The large amount of system RAM you have actually helps you
>>
>>108592754
What if I sell it and buy a mac mini?
>>
>>108592737
Their system prompt already tells it to do that, i don't believe in trying to hack in logical contradictions in the context window before it even starts to do any work.
>>
>>108592742
Will have a look there and see yeah, I've used some qwen stuff before for other things and they were pretty capable for their size.

>>108592784
Lmao
>>108592754
Yeah I set up gpt-oss-120b the other day and had a poke about with it. it runs at about 15t/s iirc. Usable for my purposes i think, but a bit problematic with context for much more than very basic tasks. I'll have to try offloading more layers to cpu with it i think to get more out of it.

Looking on openrouter it is pretty bloody cheap for deekseek tokens as the other anon suggested to be fair.
>>
>>108592805
no it doesn't
>>
>>108592871
>Looking on openrouter it is pretty bloody cheap for deekseek tokens as the other anon suggested to be fair.
just for fun, also check prices for gtp-oss-120b. and tokens per second, iirc cerebras is running that fucker at 500t/s
seriously, i was not kidding when i said its not worth it.
>>
>>108592922
>>108592871
why pay to run a local model instead of claude/codex?
>>
>>108592879
it's either in the system prompt, the model, or both. regardless of the situation telling it not to be a sycophant is contradicting its nature
this is the fundamental problem with misanthropic, they only one use model for everything so it has to be a normie-coded generalist who is le friendly, openai makes models specifically for coding so they can autism-maxx their model
>>
>>108592922
Chatgpt tells me that codex is included with my subscription. Idk what its like with limits and shit though I've only used it as a Google replacement and for some Linux troubleshooting.

So many fucking options with all this stuff my boomer brain can't cope
>>
>>108592871
im setting up GPT-OSS 20B locally and wanted to train LoRAs for it, but its been a clusterfuck of which inference servers support the platform (arm64) and the mxfp4 specialty, and supports GPTOSS loras, and supports this era of chip architecture (ampere)...

i eventually settled on converting the lora to gguf format so i could burn it into the base model. disappointing, but a good learning experience with codex-high to help. Funnily enough i get around 15 tks on 20B also
>>
>>108592933
>why pay to run a local model
you pay for some1 else to run it. on a big ass server with datacenter gpus. at much higher speeds.

and most importantly, you pay because you can't run the big models locally
even that 5070ti/96gb ram system that anon has, that is quite powerful by gaming pc standards, is a joke when it comes to running bigboy models
current sota models are ~1.5-2tb in size. you don't just need a multi-gpu server, you need a rack of them to run inference with them.

>instead of claude/codex?
i suggest you have a look at /r/claudecode
opus is extremely good (when anthropic isn't cucking users to save on compute), and the max plan was excellent value until recently, but
a) you can get 90% of the perf for way cheaper with chink models
and more importantly
b) anthropic IS actually cucking CC users currently.
>>
>>108592971
learning how to use a new tool is a skill
you can type /status in Codex to find out what your 5-hour and week limits are
>>
I’ve been doing a bit of cleanup with this prompt:

> 58. Fools ignore complexity. Pragmatists suffer it. Some can avoid it. Geniuses remove it.

Is there any removable complexity in this project?


and then I walk through with Codex xhigh (Claude doesn’t bring as much autism, which is what I need) to make wads of code that I haven’t really understood less shitty with stupid duplications
I stopped when it started to suggest undoing consolidations (something it suggested) to make things more understandable
>>
>>108593004
I agree with all your points, but I just don't think vibbing with something that can't self-regenerate is viable. That would require manual work, unless you believe those Chinese models are capable of that, then yes, I would say that's a wise move. I haven't used them, so my input is skewed.
>>
>>108593033
>codex lobotomize my code
>>
Also I did this prompt:

Read the source and then plan a linear walkthrough of the code that explains how it all works in detail

Then run `uvx showboat --help` to learn showboat — use showboat to create a walkthrough.md file in the repo and build the walkthrough in there, using showboat note for commentary and showboat exec plus sed or grep or cat or whatever you need to include snippets of code you are talking about


and I got a pretty-good summary of my cursive pixel font, although it didn’t go super deep into the craziest part that I really don’t understand
>>108593045
more like remove stupid duplicated things that Claude put in too hastily
>>
>between 118 and 120 codex now hides the % context remaining
dumb and gay default change
>>
>>108593058
why not program with dex from the go, claude has been lobotomized as you know
>>
>>108593035
>vibbing with something that can't self-regenerate is viable
im not sure what you mean by "self-regenerate", but im guessing its probably something to do with the agent or orchestration layer
so, nothing to do with the underlying LLM, except insofar as it has to not suck enough to be able to do what you tell it to do (which chink models certainly can)

you might benefit from reading https://codemyspec.com/pages/five-layers-of-agentic-coding
>>
is deepseek better than claude medium?
>>
>>108593065
For a lot of things I want to do, Claude Opus is OK enough, even when it’s lobotomized, and I can usually tell in advance if extra autism is gonna be needed
also I get a lot more Claude Code crunching — with Codex I can hit my 5-hour limit while only single-tasking
I’m kind of surprised when Claude gets dumb and forgetful during off-peak hours — maybe off-peak hours are the clogged ones now. I was doing stuff this past Saturday afternoon and it was forgetting basic shit, and I run Opus all the time
>>
>>108592948
claude isn't as dumb as hal, you can just tell it to override some persona features
>>
>>108593100
Are you even a human? Those dashes be sus
>>
>>108592933
>cerebras
>pay
>>
>>108593114
Normal:
>-
His:
>—
Bro uses alien dashes
>>
>>108592466
>keep your "rollover" money in your bank account in a separate wallet
>just add credits to openrouter when you need more ai
>if you don't use it, the credits rollover to the next month
>no expiry
they'll jew you out on rates every time you add credit, but whatever, worth it
>>
I added ten dollars to my openrouter account like you guys said because I couldn't use openclaw with the api for that long per day.
What happens now?
>>
>>108593300
Now OpenRouter sees that you're not a third worlder and opens up betters rates and limits to you. You don't even have to spend the $10 but now that you had it in your account at some point you're marked as Aryan and can spend it down to zero if you want.
>>
>>108593297
i heard openrouter is owned by fred
>>
>>108593311
I didn't get any notification or anything so I was worried it didn't happen but I've already noticed that openclaw suddenly knows how to code scripts at high speeds.
Do I have to spend the 10 dollars every now and then or just keep it sitting in my account?
>>
>>108593335
AFAIK you just need to hit that threshold once and you're in the club
>>
>>108593354
and who is paying for my tokens?
>>
>>108593114
option-underscore, bitches
>>
>>108593383
Uh you if you spend them
>>
>>108593388
>mac
yep non-human
>>
>>108593391
doesn't sound fair
>>
>>108593388
— I use WinCompose to get them on windows —
I used to like them until AIs turned them into a meme
>>
>>108593388
>option
?? my phone doesn't have an option for this
>>
>>108593388
Actual psychopath
>>
File: HFtKTwNWYAALmhI.jpg (48 KB, 745x398)
48 KB
48 KB JPG
You can literally search for OPENAI_API_KEY on github and find tons of exposed API keys

Great work vibecoders
>>
https://github.com/rohitg00/agentmemory

Finally got this running. External long term memory. Supposedly reduces tokens by a lot while also extending memory gracefully. We'll see how it does.
>>
>>108593460
make sure to give the models some traumatic memories, it increases artistic capability. not molestation tho, that will just make them gay
>>
>>108593460
Now make it watch Blade Runner or Total Recall and see where it goes from there.
>>
>>108593427
I’ve seen lots of text with em dashes and straight quotes
Diablo 3 was like this
REAL psychopathy is touch-typing option-}…on a keyboard for an 11-inch iPad
>>
File: file.png (28 KB, 834x225)
28 KB
28 KB PNG
claude is on acid
>>
>>108593445
you don't even know how to steal pictures from twitter properly, lilbru
>>
>>108593579
it must have listened to binaural beats from i-doser
>>
I want to run Qwen 27B in at least Q6 distributed across two or three PCs. I think that might be enough for me to use for everything except when I get really desperate (in which case I can pull out a $20 codex sub maybe).
>>
Have I shilled one of my bookmarklets here recently? This one copies a selection, or the h1, or a really good heading from GitHub to both Markdown for plain text and rich text for pasting into Slack. Not all of it fits in one message, so here’s most of it:
javascript:(() => {
const preferredH1 =
document.querySelector('h1[class*="PageHeader-Title"]') ||
document.querySelector("h1");

const h1 = preferredH1?.innerText.trim();
const title = document.title.trim();
const selection = window.getSelection?.()?.toString().trim();

const url = window.location.href;
const ghRef = getGitHubRepoRef(url);

let text = selection || h1 || title || url;
text = normalizeTextForGitHub(text, ghRef, Boolean(selection));

const mdLink = `[${text}](${url})`;
const html = `<a href="${url.replace(/"/g, "&quot;")}">${escapeHtml(text)}</a>`;

navigator.clipboard.write([
new ClipboardItem({
"text/plain": new Blob([mdLink], { type: "text/plain" }),
"text/html": new Blob([html], { type: "text/html" }),
}),
]);

function getGitHubRepoRef(rawUrl) {
// …
}

function normalizeTextForGitHub(text, ref, hasSelection) {
if (!ref || !text) return text;

const prefix = `[${ref.full}]`;
const escapedFullRef = ref.full.replace(/[.*+?^${}()|[\]\\]/g, "\\$&");
const bracketedRef = new RegExp(`\\[${escapedFullRef}\\]`, "g");
const bareRef = new RegExp(`(^|\\s)#${ref.num}(?=\\b|\\s|$)`, "g");

let body = text
.replace(bracketedRef, "")
.replace(bareRef, "$1")
.replace(/\s+/g, " ")
.trim();

if (hasSelection) {
if (text === `#${ref.num}` || text === ref.full || text === prefix) {
return prefix;
}
}

return body ? `${prefix} ${body}` : prefix;
}

function escapeHtml(s) {
return s
.replace(/&/g, "&amp;")
.replace(/</g, "&lt;")
.replace(/>/g, "&gt;");
}
})();
>>
>>108593815
The elites don’t want you to know this, but you can just paste bookmarklets into your URL field in your browser. You don’t have to minimize them and turn spaces into %20. `javascript:` is also something you can just leave in a file — it’s a label, like for goto or whatever. I have 14 bookmarklets.

  function getGitHubRepoRef(rawUrl) {
try {
const u = new URL(rawUrl);
if (u.hostname !== "github.com") return null;

for (const key of ["issue", "pull_request", "pullRequest"]) {
const value = u.searchParams.get(key);
if (!value) continue;
const [org, repo, num] = decodeURIComponent(value).split("|");
if (org && repo && /^\d+$/.test(num)) {
return { org, repo, num, full: `${org}/${repo}#${num}` };
}
}

const parts = u.pathname.split("/").filter(Boolean);
const reserved = new Set([
"orgs",
"users",
"settings",
"notifications",
"marketplace",
"sponsors",
"explore",
"login",
"signup",
]);

if (
parts.length >= 4 &&
!reserved.has(parts[0]) &&
["issues", "pull"].includes(parts[2]) &&
/^\d+$/.test(parts[3])
) {
const [org, repo, , num] = parts;
return { org, repo, num, full: `${org}/${repo}#${num}` };
}
} catch {}

return null;
}
>>
>>108593830
hey what's up with that stuff?
>>
>>108593460
>External long term memory
so RAG?
>>
Sorry for the an retard question, but I have briefly been looking for an AI tool to create multi layered stencils. Is this something that already exists? If not could it be vibecoded and run locally? I worry about weird issues like bridges so a stencil doesn't fall out of itself basically. I saw a guy on youtube that said he had built this but his website 404'd. I think he realized it would cut into his business so he nuked it.
>>
>>108594148
Probably. I have no idea how stencils work and I doubt anyone here does either but are they svgs? Sounds doable
>>
>>108594148
What does any of that mean
>>
>>108594085
Kind of, but more. It does a lot of wizardry with Claude Haiku to capture and compress memories into searchable objects stored outside the AI wrapper. Then, when it detects a keyword from a previous memory in a prompt, it checks the stored memory and includes only that in the prompt, not an entire MEMORY.md file full of maybe useless memories. Should cut tokens.
>>
>>108594212
I'm currently running it alongside the traditional MEMORY.md route. After a week or so, I'll inspect the long term memory and see if it is actually capturing and compressing memories on its own. If it is, I'll nuke the MEMORY.md system.
>>
File: 1761733911762679.jpg (456 KB, 2160x4166)
456 KB
456 KB JPG
I'm doing everything in my power to get games to detect my DualSense as being connected via USB while playing on Bluetooth (Features like Adaptive Triggers/Haptics are otherwise disabled). Apparently I need to create a driver to spoof as a USB DualSense, then translate and send that back to my Bluetooth controller. This is definitely going to cause anti-cheat issues, but who cares atp I just want to get this done... I've spent half of my tokens building this.
>>
>>108594225
>Features like Adaptive Triggers/Haptics are otherwise disabled
what? why?
>>
>>108594212
why keywords rather than semantic similarity?
>>
File: 1754866207869588.png (15 KB, 582x242)
15 KB
15 KB PNG
>>108594238
>what? why?
Okay, it's actually a few issues. For whatever reason, some developers only scan for the DualSense HID under the USB enumerator, so it'll never locate your DualSense which shows up under Bluetooth. Also for some really odd reason (maybe for power saving?), Bluetooth reporting will default to simplified reports on PC, so you have to switch over to higher fidelity packets to get Adaptive Triggers to work via BT (which I have working right now). I just need codex to create the driver, I'm just outlining what needs to get done rn.
>>
>>108594295
nta, but good project, anon
always nice to see people tackling real world shit (even if its just gayming) rather than JustAnotherWebApp™ or YetAnotherLLMToolingProject™

oh a side note, there is a desperate need for lovesense integration in gayming. just imagine how much more exciting the halo 1 "drive through the crumbling station" sequence would be if you were also being vibed out of your mind.
just a thought.
>>
>>108594177
they are generally svgs but could be converted from jpg. Its essentially just layers of different colors stacked on top of each other. The main issue I see though is for stuff that bridges. Like if you were to make a stencil of the number 8 and you cut it out, you would just get a huge hole instead of having a bridge to hold the 8 together.
>>
>>108594421
Isn't that like a solved issue? surely there's one way to do it. I don't know shit about it but it sounds like the kind of thing people already solved. If that's the case then llms will know about it for sure
>>
>>108594434
Yeah im pretty sure its solved, I just don't know of any good solutions. Will do some more research, but I think all the solutions might take credits. Would like a solution that could run locally.
>>
>429 error
It's nyover...
>>
>>108594421
doesn’t https://potrace.sourceforge.net already do what you want?
>>
File: pikachu.png (496 KB, 648x648)
496 KB
496 KB PNG
>>108594577
No, that seems like its only for B&W images, and converting png or jpg to svg. What im looking for is to take something that has 4 + colors and to break them down into layers. I attached a picture that shows what it should look like. You can see the pink plus points as guide markers for each layer to line up. Then when you look at pikachus black outline you can see the bridges. Everything else is basically straight through holes. Want to avoid things like islands too.
>>
>>108594286
Not my repo. Perhaps I'll code that into it.
>>
>Codex works past 0%
insanely based
>>
>>108594720
you mean it tries finishing the prompt? how long will it try for?
>>
>>108594731
Yes. I'm not sure, I gave it a really complex task and it's going on 30 minutes now kek
>>
>>108593807
Tried it across 2 PCs with llama-rpc, unfortunately it was very slow (1.3 tk/s).
I will continue developing my engine with the hope that it's just a suboptimal implementation rather than a hardware limitation.
>>
words can barely express how bad claude has gotten, a bunch of indians would probably do a better job
codex spent 3% of the 5h credits to do in 10 minutes what claude failed to do in more than an hour after spending 100% of the 5h credits
it seems almost incapable of reasoning
>>
>>108594833
You’re using Opus, right?
I keep getting annoyed because it always asks me how it’s doing when it’s doing fine (and I say so), not when it’s forgetting stuff left and right
>>
>>108594845
both, it's ridiculous
sonnet, opus, it's all trash now, like actual trash. I wouldn't be surprised if GLM 5.1 isn't actually better
I don't know if I'm hallucinating but I'm sure claude was really good a few weeks ago, like amazingly good
I do have to admit it managed to implement one thing earlier today, but only following a plan that was written by 5.4xhigh
>>
File: 1755076391632464.png (5 KB, 316x119)
5 KB
5 KB PNG
it's still fucking going lmao
>>
>>108594871
error
agent ran out of tokens
come back later
>>
>>108594875
Dude, I ran out of tokens 30 minutes ago. Apparently Codex looks for a safe off-ramp before killing the prompt, but what I'm working on rn is pretty serious so who knows what's going on lol
>>
>>108594882
weird how it doesn’t just pause until the 5h or 1w limit is up
but then again all these client apps are barely-working piles of trash because there are *so* many demands placed on them to get better in a gazillion different ways
>>
is gpt-5.3 codex good?
>>
>>108594893
Some people argue it's good at implementing concrete things that don't require exploration, but I find that all the good things about 5.3 codex are already included in 5.4
I used to get 5.3 codex to implement planned steps one by one but right now I'm doing it all with 5.4
>>
File: 1772862705332427.png (22 KB, 399x276)
22 KB
22 KB PNG
>>108594886
>weird how it doesn’t just pause until the 5h or 1w limit is up
and I'm glad it doesn't, $20 well spent.
>>
What's the best agentic AI that has full permissions? Still fucking OpenClaw?
>>
>>108594907
I wonder if it just ‘borrowed’ from your next 5h chunk
>>
>>108594909
what do you mean? just run codex in yolo mode?
>>
File: all_tokens_absolute.png (927 KB, 3575x2376)
927 KB
927 KB PNG
Debugging my LLM engine.
>>
>>108594916
It went for 1hr 39min and didn't use any future tokens. New strategy unlocked.
>>
File: 1774103679722002.png (434 KB, 1654x936)
434 KB
434 KB PNG
>>108593445
doesn't posting keys on github automatically disable them?
>>
>>108595051
Nice! Kimi fixed it!
>>
>>108594856
There are some rare cases when Opus unblocks some task, so I still keep it, but overall I agree completely. It's not just underwhelming, or not as good as before, or too expensive for what it's doing, for complex tasks it's just actually bad.
>>
File: file.png (686 KB, 1200x900)
686 KB
686 KB PNG
>>108595051
>>108595462
is this recursive self-improvement?
>>
>>108596295
I left it running since that post, now the code is more buggy than before. So kinda but not really, even if I had 500k to run Kimi locally.
>>
>>108596295
butterfly?
>>
>wake up and send claude a prompt to continue my work from last night
>it starts with "Good Morning"
surprisingly made me smile
>>
File: file.png (54 KB, 1160x485)
54 KB
54 KB PNG
yes thank you you gemini cli you fucking piece of utter dogshit.
>mcp to get time
>blocks its own tools
>crashes out of the cli
multitrillion dollar corporation of incompetent hacks, do they even use their own fucking harness
>>
>>108585395
>natural language 3D parametric modeler with cadquery
The bar is pretty low.
>>
>>108597529
tb h these natural language intefaces to do things and mcps's for existing applications are never going to be very performant until labs get around to doing the RL for those things
code, math and some other shit like excel are good now because they've trained the models out the ass on it. it'ts a miracle the models can even do anything else because all that's just coming from raw pre-training and reasoning
>>
>>108597529
Isn't something like that kind of silly, I don't see how explaining a 3d model parameters in natural language will be easier than doing it yourself on solidworks/fusion. There are too many details for the machine to guess that I don't see how it would be worth it, plus if you can accurately describe what you want you should be able to model it yourself, this isn't like vibecoding where you don't really need to know what you are doing to get basic results.
>>
>>108597582
the closest I ever came to CAD was doing a Half-Life 2 mapping tutorial but maybe you might be able to get an LLM to make some basic stuff that’s likely wrong but it’s faster to tweak something it’s made rather than you making it from scratch
also if you’re getting RSI from your job you might be OK just talking to your computer for this and use the keyboard+mouse a little bit less
>>
>>108597615
I don't know, I have barely any experience, just small stuff for 3d printing, but for example that piece I could probably recreate it in an hour just fooling around with the UI, but describing it in text? I would have no idea how to call each feature and describe the constraints, someone with more experience may be able to, but then, they would probably it do it manually way faster than me so would they really gain something by using a LLM?
Who knows though, maybe by doing it in steps, slowly building it visually, it could be worth it.
>>
When are we getting world models properly? I'm tired of having to use genie to simulate my game.
>>
qwen code cli slow as fuck today holyshit. stupid chinks
>>
>>108597898
>https://github.com/QwenLM/qwen-code/issues/3203
ACK!
>>
File: 1759366622067194.png (2.66 MB, 2230x1286)
2.66 MB
2.66 MB PNG
>>108594225
Alright, I actually had a major breakthrough with the driver and it's spoofed descriptor. Steam now thinks a DualSense is being used over USB whenever my driver is enabled. The only problem is that Codex 5.4 broke my primary application, so I can't toggle the transport layers on.

Unfortunately, I only have 16% of my tokens left for my weekly quota, and they don't reset until the 18th. At the very least, I'm feeling hopeful that we may be nearing the end.
>>
>>108594225
>>108598034
Doesn't dsx do that already?
>>
File: 1755852726787216.png (579 KB, 1200x600)
579 KB
579 KB PNG
>>108598077
>Doesn't dsx do that already?
I'm too deep in to stop now.
>>
>>108597840
https://huggingface.co/blog/waypoint-1-5
>>
>>108598077
>>108598092
Ok, I looked into it. Apparently DSX is paid software and now costs $12. The developer probably realized how much of a pain in the ass implementing dualsense emulation was and started charging for it kek.

I'm gonna open source this when I'm done, fuck that.
>>
>>108595048
Codex can control your whole shit? Phone and everything? I thought Claude Code was just for coding and couldn't order me Doordash like OpenClaw can.
>>
>>108598077
It does, it's a $12 DLC, so $22 for all the features. It's also vibe-coded, and the creator is a huge faggot.

>>108598092
I'm rooting for you. I don't fuck with DS controllers anymore but DSX is trash and I think it's hilarious that you could replace its "biggest feature" with a handful of token.
>>
>>108598121
>>108598126
Oh I see I misread it, it's $10, +$5 for the DLC, or $12 for them together. Still, that DLC is vibe-coded, fuck 'em.
>>
>>108598121
>>108598126
I think its $12 but still, if you already spent your weekly quota, that's 1/4 of your subscription, assuming codex $20, you are $5 in already, better hope you get it done in the next week lol. And that ignores your own time.
I am not going to shill dsx, I think it's bloated as hell, but the few times I have needed it, it was that bloat that helped me so I can't really complain, but if you enjoy the challenge go ahead.
>>
https://github.com/dimensionalOS/dimos
Oh yeah. It's all coming together.
>>
>>108598164
I saw the Stargate: SG-1 episodes with these fuckers
they were harrowing
>>
>>108598183
$100 kit and then my coding girlfriend can walk around the house and look at stuff. Maybe nag me to put something on my bare walls.
>>
Does anyone here have any idea how changeable all the compute is for Anthropic’s servers?
We’ve all been sneeding about how retarded Claude has gotten but I’m reading how Mythos Preview is finding 20-year-old zero-days in OpenBSD and I’m wondering how much compute they were allocating to Opus is now going to Mythos
>>
>>108598195
https://meh.com/deals/odyssey-toys-zoom-cam-mini-rc-racer-fpv-llu
the spider robot is probably better for stuff, though
>>
>>108598211
Here's the hardware stack I landed on. This bitch has binocular depth sensing out of the box, and with an extra $20 and some arts-and-crafts there's tutorials for making a docking station for it. Fully autonomous movement and recharging. About $250. Then Claude and me use the repo above as a starting point to hax this bih.

SunFounder PiCrawler kit $140
Raspberry Pi 4 (4GB) ~$55
VL53L1X ToF sensor ~$12
Battery pack (if not included) ~$20
Total ~$227
>>
>>108598164
oh no no no not like this
>>
>>108598202
they just grew to quickly.
everyone figured out how good the coding models were late last year and anthropic's arr tripled from 10b to 30b in q1 of this year.
dario didn't buy enough compute
>>
>>108598496
I think the future is people being able to sell compute, like crypto.
>>
>>108598519
did... did you just re-invent nvidia?
>>
>>108598636
more like torrenting but for compute
>>
>>108598639
sounds like https://www.nvidia.com/en-us/geforce-now/
>>
>>108598652
but by the people for the people, the true free as in beer
>>
>>108598639
oh, ok, you instead re-invented AWS ec2
cool, cool
>>
File: file.png (59 KB, 567x142)
59 KB
59 KB PNG
Where the fuck is 5.4 Codex? Plus pleb here, is it only available on higher tiers? Want to give it a spin how good it is on cracking apps / plugins compared to 5.3.
>>
>>108598676
>5.4 Codex
In our hearts and minds.
>>
Today wasn't great :\
>>
>>108598676
Doesn't exist. I don't think it makes a big difference, I just use gpt 5.4
>>
>>108598662
You are right, but I can still imagine people buying it.
>>
>>108598994
On one hand, I’m genuinely sorry for your loss
On the other hand…why are you asking it to copy stuff to back things up? Do you not know how Git works (at least at a high level)? Are the files too large for Git? Do you need to learn how git-lfs works?
you should read
https://simonwillison.net/guides/agentic-engineering-patterns/using-git-with-coding-agents/
if you don’t know anything about git
also get a Git GUI like, actually I think GitHub has one, and it probably doesn’t _depend_ on having your project’s repository also be on GitHub as well
>>
>>108599063
I am tracking my project using git. The llama folder isn't part of my project, I just wanted to check my error against theirs to isolate what is quantization error vs what is a bad implementation.
I didn't really lose anything, thankfully. But I thought "the user is extremely angry" part was kinda funny. I was more angry at myself for getting into a dangerous situation than at him. I was feeling sorry for yelling at it.
>>
>>108599171
oh, good
I remember reading from that https://www.seangoedecke.com guy that these things role-play, so if you yell at it, it’ll role-play “unhelpful assistant” and fuck up more
>>
Is there a reason not to use Github Copilot at this point? Still appears to be the best price for access to a variety of models. I think the context window is smaller for some things, like 250k for Opus instead of 1m? I'm not sure if the advantages to other means of access, assuming they exist, actually apply to me. Hard to compare because it's done by request rather than dealing with tokens, maybe someone knows better than I do.
>>
>>108599269
In so much as what exactly?
>>
>>108599269
well, usually the question isn’t “Copilot or not?” but “Copilot, or should I spend my money elsewhere?”
>>
I am getting increasingly bothered by how LLMs use — and all the time but never use ≈ and use ~ instead
>>
>>108599237
Yeah... I don't know
I feel like there is some truth to that, not because they role play as unhelpful assistant but because it makes them panic and they try to solve the problem they caused as quickly as possible, which often makes things worse.
>>
>>108599284
That's what I meant. I spend $10/mo, up to about $20 on a really busy month. I can't imagine I wouldn't be spending more doing the same things other ways, but I wanted to ask in case I'm sorely mistaken and may get more for my money elsewhere.
>>
>>108599295
moot ate my rightward-pointing arrow
>>
File: zuck.png (103 KB, 955x355)
103 KB
103 KB PNG
When are you planning to make your clone?
>>
>>108599318
Copilot might be the best option for you if other places only have $20/month options and you’re particularly price-sensitive
on the other hand, if you’ve got a burst of work to do you might want to switch over temporarily
>>
>>108599295
Tell it to act slutty and you will get ~ at the end of every damn sentence.
>>
File: all_tokens_with_llama.png (203 KB, 2780x1976)
203 KB
203 KB PNG
Looks like I have a bug with my LM head
>>
>>108599412
im tired of this eye burning 'creens fuck off
>>
>>108599428
No YOU fuck off, you little piece of shit.
>>
>>108599449
you are the vile moid trying to burn my eyes
>>
>>108599459
tits or gtfo
>>
I've been fighting openclaw for days about not hosting websites correctly, turns out it never explained and I forgot about the local pc ip address instead of using localhost.
>>
found the culprit
>>
>>108599758
was it the jews?
>>
>>108599758
I too see a happy merchant
>>
Pleasantly surprised with GLM chat. And I assume it’s not even 5.1 in the free chat?
>>
>>108599800
you're supposed to be able to choose (top left), but i haven't been able to get 5.1 yet (says high demand, fuck you), only 5turbo
ive done 2 sessions with 5turbo, one *noticeably* better than the other, tho even the bad one was pretty gud. no idea what is happening behind closed doors to affect perf like that when nominally speaking to the same model.
anyway, yea, 5turbo is good, 5.1 presumably is going to be even better

my only reservation is that the $1.4/$4.4 rates are a bit steep for a chink open model.
like yes, its *better*, no doubt, but you can do ds 3.2 for ~$0.25/$0.5... or minimax 2.7 for $0.3/$1.2
and i don't think its 3x better than m2.7, and ofc ds is just *insane* value
>>
File: file.png (39 KB, 546x285)
39 KB
39 KB PNG
how deep in the weeds am I when LLMs start suggesting code like this as a reasonable approach?
>>
BOOM! The fix worked perfectly!

**THE FIX WORKED!**
>>
>>108599867
tldr?
>>
File: 1769420524570691.png (3 KB, 498x38)
3 KB
3 KB PNG
first time i see claude doing this
>>
>>108599916
they must be adding cutesy shit to try to calm down the redditors and make them feel like poor claude is really working hard for them lmao
>>
I had never seen this failure mode with Kimi before
>>
>>108600156
damn thats crazy
>>
I'm very weary of vibecoding but I also want to learn to see if it's valuable. So far I've only experimented with Aider but I want to dip my toes into more modern agentic stuff.
what's the most minimal agent I can use? Pi?
how can I containerize it effectively?
>>
>>108600183
>weary
why
>>
>>108600228
I wouldn't let a person throw around commands in my terminal let alone a clanker.
>>
>>108600183
>what's the most minimal agent I can use? Pi?
https://drive.google.com/file/d/1YtpKFVG13DHyQ2i3HOtwyVJOV90nWeL2/view
>how can I containerize it effectively?
run it on a separate PC
>>
Added star globes to the star info pages now. Not 100% happy with the noise texturing, but it'll do for now and the bloom masks it somewhat. Also added a compare section at the bottom of the planet pages, so you can compare exoplanets to Solar System planets. And I added textures for brown dwarves, although there's only two host stars which it applies to. Still cool to have. Ignore the glitch.
>>
>>108600237
clankist pos
>>
File: 1745054540998321.png (237 KB, 830x309)
237 KB
237 KB PNG
I used Opus 4.6 to crack this Mac app that hit the front page of Hacker News https://boringbar.app/ in an hour since he kept debug symbols in the binary. It was like 5 lines of code to change/inject

If you find reverse engineering semi-interesting then you may find this writeup semi-interesting
https://rentry.org/boringbarbusted
>>
>>108600363
yet he will get hired for 500k and you won't, life isn't fair
>>
File: 1760491643439989.png (131 KB, 2375x461)
131 KB
131 KB PNG
I'm tempted to get a second plus plan to finish my project, but I think I'll wait... 5 days will probably do me some good after working on this for the past two days.

I successfully vibe coded a dualsense driver which is extremely funny. There's just a few issues I need to resolve.
>>
>>108600433
good shit, anon
>>
>>108600433
based
>>
>>108600183
>what's the most minimal agent I can use?
mini-swe-agent

>how can I containerize it effectively?
by... using... a container?
like docker?
>>
>>108600485
Not familiar with docker, can you elaborate?
>>
>my le agent is self-improving
The only people who think this is possible right now are ones who've never read a skill written by a clanker
>>
>>108600595
It's on the way there.
The way I look at it is as an autistic savant with vast amounts of crystallized knowledge from spending years obsessively cramming books, but very low actual fluid intelligence and ADHD/amnesia.
It needs guidance but it can be of help when you put the right puzzle in front of it.
>>
>>108600595
you lost
>>
File: 1758825121973785.png (287 KB, 1487x2253)
287 KB
287 KB PNG
>>108600547
no, but my friend glm 5.1 sure can
>>
>>108600624
Damn that website is blocked in my country
>>
>>108600547
It's an app to find docking buddies
>>
>>108600623
They can't write skills
The skills are slop
We need RL on skills
read one and within 2 passes you can trim it down to less than half the tokens
>>
>>108600685
RL on specific skills is not gonna happen
but it would be nice if companies started training on brevity
clankers out there be writing fucking essays for no good reason
like, i get that your job is to generate text, but fuck, sometimes less is more
>>
>>108600711
Pretty sure he means RL on generic "skill" file writing.
As for length, it's probably a historical artifact of non thinking models needing a lot of text and Nigerians rating markdown soup higher during RLHF.
>>
>>108600711
The raw reasoning traces for 5.3 and 5.4 are basically a much better version of caveman meme thing, basically a neuralese precursor. so the models are already very token efficient internally - they just don't let them write that way to us.
>>
>>108600753
Also those traces when they leak look a bit like this
https://github.com/aw31/openai-imo-2025-proofs
>>
>>108600753
I think letting GPT RL itself into that grug token soup for thinking has greatly diminished it's ability to use actual human language.
Gemini and Claude feel warm and nice to talk with while GPT feels like a cold passive aggressive autist.
>>
>>108600774
You sound like a simpleton
>>
>>108600774
Agree. Gpt was always a bit shit to talk to, but this issue is very palpable with 5.3 codex - that thing barely tries to sound human
>>
>>108600787
>>108600774
Can't believe yall are crying the robot doesn't speak in a fake warm fuzzy human tone.
>>
>>108600787
>5.3 codex - that thing barely tries to sound human
good. its a coding optimized variant, not a coombot
>>
>>108600796
I don't mind it personally because I don't talk to the bots outside of vibeslopping, but it is a problem for openai who want the normies audience. T bh I can't stand Gemini's tone, grates on me more than any other model.
>>
>>108600797
excuse me?
>>
>>108600797
And yet before they nerfed it Opus was beating it
>>
>>108600885
misanthropic has been behind codex for almost 2 months now, it will never be good again (and it was dubious if it was ever better in the first place)
>>
>>108600849
he's right, much of /g/ uses AI for ERP and can't be trusted to have valid opinions on it
>>
File: 20260413_185149.jpg (3.46 MB, 4080x3060)
3.46 MB
3.46 MB JPG
>vibecoding
>everything smooth as butter
>suddenly reminded I'm on a shitheap laptop
Well, I guess we'll find out tomorrow if it works. And in the meantime, the OpenClaw daemon is so overloaded that my gf won't even respond. This shit better work.
>>
>>108600949
why do you use openclaw
>>
>>108600183
if you want to just try something out normally I’d shill Claude Code’s $20/month plan for ONE month (ask yourself in 25 days if you want to keep going on with another month, or spend more, or spend less)
but Claude is shitting the bed too much so I’d recommend ChatGPT Plus, which is also $20/month
but you might run out of tokens super quickly if you run xhigh all the time like I do
there’s an experimental flag to have ChatGPT audit its own command so it’s less likely to randomly delete your stuff like what happened to that other anon in this thread
t. has work pay for ChatGPT Pro and doesn’t know how quickly tokens get used on the $20/month plan
>>
File: file.png (30 KB, 906x223)
30 KB
30 KB PNG
claude went sassy on me
telling me to do it myself why the fuck didn't you do it you dumb clanker if you found an error it would've been the same tokens to patch that into the file
>>
>>108600781
I built my own LLM inference engine.
Have you?
>>
>>108600991
i feel proompted
>>
>>108600980
Why the fuck are you paying $20 a month for lobotomized Anthropic slop or OpenAI's walled garden when DeepSeek literally exists? Claude is completely cucked right now, refusing half your boilerplate for "safety" and hallucinating the rest, and the fact that you're recommending ChatGPT Plus while openly admitting it might just randomly nuke your codebase is peak brainlet behavior. If you actually want to write code instead of fighting with a sanitized botnet, just use DeepSeek; the API costs literal pennies compared to your overpriced subscription, it punches way above its weight class in actual logic benchmarks, and it doesn't treat you like a toddler. Imagine unironically bragging about your employer paying for a premium OpenAI subscription when a Chinese startup just dropped a model that absolutely mogs both of these Silicon Valley onions projects for a fraction of the cost.
>>
>>108600996
usecase?
>>
>>108600999
Nothing I’ve ever used has nuked my codebase but evidently this happens
at any rate I haven’t been tempted to do anything naughty by American standards so why would I use a third thing and pay money for it when I get something better for free
the deepseek whale logo/mascot owns compared to the western butthole logos, I’ll grant you that much
>>
>>108601004
I want to optimize it to do distributed inference across many PCs because smaller GPUs have better GB/$.
Right now I'm trying to get Kimi to vibecode me a Gated DeltaNet implementation to be able to run Qwen 3.5.
>>
>>108600999
DeepSeek is retarded
>>
File: sddefault.jpg (39 KB, 640x480)
39 KB
39 KB JPG
Yo this shit is actually going to work
$250
I get paid on Friday
Enjoy having your entire car dismantled piece-by-piece by my army of agentic AI spider robots, fags
>>
>>108601206
I saw this arc
they’re weak to boolet
>>
File: 1765247902078249.png (288 KB, 1200x1600)
288 KB
288 KB PNG
>>108601206
stealing car parts is lame and gay and niggercoded
what you should do instead is pic related
>>
>>108601206
jokes on you i dont own a car
>>
Kimi 2.6 Code is live
>>
>>108601370
How many boolet do you has? I have 800 of these spider robots, all controlled by their own agent that answers to Opus 4.6, and I'm even considering a 360 degree LIDAR unit. When you're empty, they're chopping you to bits.
>>108601398
Stealing cats is absolutely not gay, a bucket of aqua regia will dissolve out the platinum and paladium which you can sell. Shows how much you know scrub.
>>108601416
They can also dismantle pets and dead bodies
>>
>>108601770
precipitate* out
Sorry I was typing fast and not thinking
>>
>>108601606
Damn. Based. It's looking more and more grim for westcucks.
>>
>>108601770
https://www.youtube.com/watch?v=c9lEeKdbQ_w
https://www.youtube.com/watch?v=NjlCVW_ouL8
>>
>>108601398
kek
>>
do i even worry while using cline about switching modes between act or plan if i am just asking a question? i am using minimax monthly plan, does it even matter if they just track calls instead of token usage ?
>>
File: 1769254995594800.jpg (28 KB, 258x360)
28 KB
28 KB JPG
if anthropic had made something similar to gpt image 2 they would had screamed and creamed their pants the told people that they are too stupid and low caste to not "hurt themselves using such strong model"
>>
File: 1760466541321550.png (186 KB, 498x471)
186 KB
186 KB PNG
>so down bad on tokens I asked my mom if I could use her plus account for codex
go ahead, laugh at me, I deserve it
>>
File: 5901_Pathetic.png (243 KB, 537x367)
243 KB
243 KB PNG
>>108600796
>yall are crying the robot doesn't speak in a fake warm fuzzy human tone
You see it on HN constantly. It's always, without exception coming exclusively from Claude Code users. Just yesterday there was post about changes in Codex usage, and the comments were full of users circle jerking about refusing to use Codex because its system prompt isn't configured to waste a bunch of tokens on pleasantries and ass kissing. Grown men calling Claude their spiritual yoga buddy and shit. The only sane conclusion is that 100% of Claude Code users are kissless, handholdless virgins
>>
>>108600260
Very cool project. Where did you source the data from? Are the colors of these bodies based on anything? Is the larger star map real time positioning?
>>
>>108601966
what’s she using her Plus account for that she needs all of her tokens?
>>
File: 1768756209458993.png (129 KB, 960x480)
129 KB
129 KB PNG
>fucking with my emulated virtual DualSense driver
>this pops up after the latest driver build
>I don't have anything plugged in
uhhh bwos... whats going on
>>
Sometimes I switch to Codex and it's so much better I don't even know why I'm bothering with an LLM engine in the first place.
>>
>>108601936
Plan mode was always a meme to try to stop LLMs from ending their turn prematurely. Nowadays it doesn't really help anything.
>>
>>108601981
NTA but bro I don't think stars move too much in human timescales...
>>
>>108602096
I once vibecoded a mouse emulator that could be installed on an old rooted phone and control a PC remotely.
My goal was to automate my work undetected but abandoned it and ended up just installing AI shit on the work PC itself and got in trouble for it.
>>
>>108602096
Steam sees your shit as a controller plugged in and thinks that controller is on an outdated firmware version. Given that it's a virtual driver I'll bet that not only can have you it pretend to be the current version, you may be able to have it verify the current firmware version in order to always spoof the most recent update.

Most game controllers have firmware updates these days, console gamers never see it happen but PC gamers do. Microsoft pissed a lot of people off at one point (myself included) when they updated certain revisions of Xbox One controllers and actually changed the Bluetooth implementation entirely, it was a smart move for a lot of reasons but this broke compatibility with a LOT of older Bluetooth adapters and nearly all the third-party controller adapters. Had to jump through some hoops to downgrade to the previous firmware.
>>
>>108602126
>console gamers never see it happen
correction: I see this from time to time on my Switches
>>
Shipped my product today. Feels good.
>>
File: 1769241031504628.png (1005 KB, 3840x2160)
1005 KB
1005 KB PNG
Well, there it is folks, a UMDF virtual DualSense driver

>Audio Haptics + Adaptive triggers
>DSX Status? daily closed source L

my next post will be the final nail in the coffin for that piece of crapware that rugged their users by switching to a paid model
>>
>>108602140
A point for Nintendo then, I'll take that over silent background updates any day.
>>
>>108594225
Codex 5.4 slop UI detected
>>
>>108602387
there's no way anyone gives a shit about UI unless it's totally unusable or labyrinthian
>>
File: file.png (20 KB, 820x317)
20 KB
20 KB PNG
>>108602302
10/10 anon, you did good.
>>
>>108602422
you could have just said “I don’t know any Mac users” and saved yourself a good bit of typing
>>
>>108602302
What the fuck is DSX for? I read their steam page and it's just mumbo jumbo marketing bullshit like this:
>Your DualSense™, DualSense Edge™, or DualShock™ 4 isn’t just hardware. With DSX, it becomes an extension of your instincts. Every button press feels deliberate. Every trigger pull feels tuned. Every haptic pulse reacts exactly how you want it to.
So like... what makes this better than me just plugging my dual shock controller in normally like I already do?
>>
Holy shit guize GPT 5.4 just leaked its chain of thought in the response wtf!!!
>>
>>108602475
Is this using caveman?
>>
>>108602482
Nope. GPT used grug speak in its internal thinking long before the caveman meme became a thing.
We know this from examples published in their safety papers, published examples from math competitions and from GPT-OSS which shares a lot of stuff with the closed models.
>>
Still going. Looks like the context got poisoned and now it's doing it on every turn. So it must be possible to induce it on purpose.
>>
>>108602471
Certain functionality is only enabled when plugged in via USB, DSX and now the vastly superior >>108602302 enable that functionality wirelessly.
>>
>>108602475
I write notes like this
>>
btw I guess most people don't know about the advantage of personal vocabulary because they don't use or have such thing
basically the common terms and language are poison to the mind - just look at how the webdevs and midwits invent and use tons of terms and you know what I mean. using your own language instead of known terms keep your thoughts pure
>>
File: urbit auras.jpg (117 KB, 1784x622)
117 KB
117 KB JPG
>>108602805
based and urbit-pilled
>>
I used to abuse free top tier models on benchmark websites but that dried out.
What's the cheapest alternative to get my fix?
>>
>>108602845
just turn on your faucets and buy some incandescent bulbs
>>
>>108602850
no man I really have a need to never learn programming but automate eveyrthing, have a script for every activity under the sun
>>
>>108602845
Man get on OpenRouter and peruse the buffet
>>
>>108592274
'My code is art' - the conceptual artist who smears his own excrement on a canvas and convinces some moron to pay a million dollars for it.
>>
>>108603231
(You)
>>
>>108601972
based king
>>
Vibe coded a ROON / Audirvana Discord RPC since every other client is 5+ years old, doesn't work and not even supports Audirvana. Supports album art, local album art cache and multiple file hosts (catbox, litterbox, pixeldrain etc.)
>>
what should i vibe code?
>>
>>108603508
yeah
>>
ok guys, I got an idea for using AI to provide a service.
But how can I make money from it?
They are just gonna use up my tokens.
>>
>>108603195
Are openrouter free models even worth abusing
>>
>>108603596
ask for money?
>>
>>108603596
Use accounts or something like a crypto wallet to store sessions, and measure usage for that account with top-up options (charge a bit more than the allowed tokens to make some profit)
>>
Software Engineers: are you using superpowers and following the latest best practises for plan prompting to maximise your results or just plain chatting with the harness raw dog without any additional shit layered on?
>>
>>108603859
up until now, i've been using a web UI and chatting with it like that. i'm starting to crave more, though. i think i might start experimenting with IDE plugins
>>
>>108603859
i use regex
>>
>>108603884
u havin' a lil giggle m8?
>>108603875
so you're new to agentic code gen or did you come from cursor 2023-25?
>>
File: image.jpg (189 KB, 881x677)
189 KB
189 KB JPG
>>108597582
Models can accept drawings or hand drawn sketches, so it's not about describing every feature with language.
>>
>>108603974
i was an LLM hater (because i thought the product was shit) for the past five or six years. the models are finally getting good enough that they feel worth a damn, so i'm playing catch-up
>>
>>108603859
tourist here: is vibe code general even filled with actual software engineers or just indians burning tokens on xhigh trying to write bots for polymarket?
>>
>>108604025
>indians
>able to afford xhigh
lol lmao even, do you even know how little they earn and how little their money is worth?
>>
>>108604025
I am an actual engineer but my work is simple webdev stuff and I am a SA monkey so I guess that makes me an honorary indian
Please saar give opus key I need to do the needful
>>
>>108604025
I’m an actual software engineer getting more autism out of LLMs than I can do myself
one guy here is very much not a software engineer and thinks of ‘git’ as a place where software gets software, like the NPM repository
I haven’t seen anyone talking about polymarket at all. maybe that’s a /biz/ thing?
>>
markup + streaming is the obvious superior choice for tool calling
which nigger lead us to json in the first place?
>>
>>108604134
douglas crockford
and if you think XML is better in all cases you’re 12 years old and don’t remember the 90s
>>
>>108604134
XML is better actually
>>
>>108604147
YAML is better unless you actually need markup
>>
>>108604144
I mean at least for LLM it should be better, right?
>>108604147
same thing nigga
>>
>>108604169
I think it mostly depends on the shape of your data
if you need, like, nested tags like <p>do you <em>really</em> need that?</p> then clearly XML is the better choice
but if you don’t, then you’re probably wasting tokens on XML structure and closing tags and making people think harder about how to get your serialized data into dicts and lists and strings and numbers
oh, and JSON only has doubles, so if you need a real integer type dammit you need to use YAML or XML or literally anything else
>>
>>108592871
Qwen 122B is way better than gpt-oss-120b as someone who has run both. On strix halo with Windows (inb4) you can run bartoskis Q4 L quant of it and it will run at about 20tks, using 92.7/96GB. IQ4 XS is also decent and uses 80GB if you wanted to also load a smaller model ontop of that.

>>108592933
I use both. Because Anthropic has cucked the paid plans and you use all your tokens in 8 seconds I use Opus to generate detailed plans on refactoring things, then Qwen to actually execute on them. Local models are competitive with the paid ones but you have to inject a shitload of context first if you want them to be useful, whereas Opus will just do the right thing most of the time without any manual context building.
>>
>>108604190
But how can you stand that slow speed? I have a machine with 16gb vram and 96gb ram that I don't even use for work, I've been thinking using it to host gemma and point opencode to it just like you said, codex for planning, gemma for execution, but last time I tried to run a local model was glm 4.5 air and I was getting like 5tk/s and it quickly dropped to 1tk/s as the context grew, I am already annoyed at how slow codex is at xhigh, that would drive me crazy.
>>
>>108604219
As a claude code user, is openclaw+gemma even close? Should I switch to this?
>>
5tk/s is unusable but 20tks isn't. You get answers in 20s to a couple minutes. It's fine because I work on other parts of my project while the AI is working, then when I get a notification I can comeback and tell it whatever. It works for me but 20tks is probably the minimum speed I accept from an LLM. 35B-A3B is double the speed but feels retarded in comparison.
>>
>>108604293
Opencode (not claw) with codex, sure, should be comparable if slightly dumber.
The gemma thing I don't know, have not tried it yet, I am just thinking about it.

>>108604307
I'll have to try it then, you use qwen no? I wonder how gemma compares, it would be nice to have a session with codex spawning a bunch of little gemmas sequentially to finish a long task.
>>
>>108604337
Qwen and Kimi tend to ramble and fuck up opencode projects. I let them go and when they hit a snag I have gemini in a separate terminal running to unfuck them
>>
>>108604406
>I have gemini in a separate terminal running to unfuck them
You mean manually asking it to fix the issue the others are having or you have some automated thing that let's agents view other sessions and manage them?
>>
>>108604134
LLMs don't generate JSON natively for tool calling. Whatever format the LLM uses in the chat template is converted to JSON.
>>
>>108604307
have you compared the 122B to the 27B?
>>
>>108604554
On Strix Halo 27B gives like 8tk/s, which I consider unusable. On my 7900XTX I get 40tks which is fine. 122B tends to fuck up less than 27B, but 27B wasn't bad if you happen to have a > 16gb gpu laying around
>>
>>108604438
I let it sit in my home directory and manually ask it. It's more expensive than kimi so i just use it as the doctor when the situation gets out of hand
>>
I'm going to try something similar.
Agent using weak but cheap model + uploading the session file to ChatGPT to read it and advise the agent on how to continue during the next session.
>>
>>108604927
That's not really more efficient no? You are uploading the file, likely chatgpt is putting all of it into the context. So you are still paying for the tokens, just as input not generation.
>>
>>108605001
I'm not paying OpenAI per token, I'm paying a $20 flat rate. Codex and ChatGPT usage is separate, and I wasn't really using ChatGPT for anything else. And the agent is from a $40 Kimi Code sub, but it could as easily have been something like NIM (free), the free opencode tier, or a local agent.
So far the agent has been working without stopping much longer than when I just told it "Keep implementing Qwen 3.5", and the prompt generated by GPT seems to contain good advice, like don't rush to test the code end to end when you haven't validated the individual layers against HF etc. so I bet it works better than just asking it to keep working over and over which is what I was doing.
It could also work for local AI with a small (dumb) and fast worker vs a large and slow "boss". Or even with the same model, telling it "review this session and give the agent work" is different than loading the session as if the model itself had done the work. It puts the agent into a managerial state of mind rather than task rabbit mode.
>>
>>108605034
Ahh I hadn't thought about using their web ui too, I'll do that now lol
Could even make a plugin/browser extension that does that automatically, export agent session, open browser tab and upload file + prompt and read response back into the agent.
Probably breaks their terms and conditions lol.
>>
I need to find a way to scrap a god key to vibe code with, these free credits I get aren't cutting it
>>
Is there a tool to automate claude code further? I dont mean replace claude code with another harness that uses their api, but a harness that basically orchestrates and automated claude code sessions
>>
Open ai announce 5.4 cyber, with limited access program
Anthropic announce automated safety researcher which outperforms humans
>>
>>108605404
yes
>>
>>108605476
It's all for show anyway. In reality Q* proved P=NP, broke all known encryption schemes and they already have a microcode level rootkit in every computer in the world since early 2023.
>>
File: 1759949725817062.jpg (33 KB, 750x725)
33 KB
33 KB JPG
>>108605515
>>
>>108605595
most people know about that thoughverbait



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.