[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/g/ - Technology


Thread archived.
You cannot reply anymore.


[Advertise on 4chan]


File: llms_hype.jpg (888 KB, 1320x1690)
888 KB
888 KB JPG
At first, I thought GPT-5 had cracked those math problems on its own.
Turns out (as Demis pointed out) GPT-5 just looked up the answers via web search.
We really need better peer review for these “AI discovers science/math” claims.
>>
File: AI_SLOP.mp4 (1.22 MB, 1080x1080)
1.22 MB
1.22 MB MP4
The last few months have been devastating for LLM dreams:
> Apple reasoning paper and the ASU mirage paper and many others confirmed that LLMs still can’t solve distribution shift.
> GPT-5 came late and fell short.
> Karpathy just said agents aren’t anywhere close, and that AGI is a decade away.
> And Hassabis just blew up some wildly overhyped claims from OpenAI about math.
Game over, man.
LLMs have their place, but anyone expecting the current paradigm to be close to AGI is delusional.
>>
People who are in charge of selling you their product and purposefully block access for people who want to test and validate their results are in fact full of shit.
Who knew?
>>
>>106931905
how about you post where he said something useful and not just "this is embarassing"?
>>
>>106932739
this. embarassing desu
>>
>>106931915
>distribution shift
Explain that without consulting Gary Marcus
>>
File: 1000016877.png (398 KB, 1080x1475)
398 KB
398 KB PNG
>>106932739
>>106933175
Not op.

I guess Google must be fairly confident about Gemini 3.
>>
>>106931905
Do these people even understand how GPT works? It doesn't solve or invent anything. It predicts the next most likely token.
>>
>>106933208
Something something emergent abilities something give me a trillion dollars
>>
>>106933208
still with their CoT (chain of thought) stuff basically, lets prompt the model with software 1000x times ourselves, from the initial users prompt.
it does do a form of thinking, in the sense that it can coagulate a mishmash of info and maybe combine it into an interesting novel thing.
this is extremely inefficient though and theyre never using 'consumer grade' models for these hot-shot question answerings, its always highly trained shit, highly specialized, highly expensive. so even if it can create some novel soups with a bunch of ingredients and math autism, its not feasible to depend on it rather than some scientists eating pizza and grinding the solution out.



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.