[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/sci/ - Science & Math

Name
Options
Comment
Verification
4chan Pass users can bypass this verification. [Learn More] [Login]
File
  • Please read the Rules and FAQ before posting.
  • Additional supported file types are: PDF
  • Use with [math] tags for inline and [eqn] tags for block equations.
  • Right-click equations to view the source.

08/21/20New boards added: /vrpg/, /vmg/, /vst/ and /vm/
05/04/17New trial board added: /bant/ - International/Random
10/04/16New board for 4chan Pass users: /vip/ - Very Important Posts
[Hide] [Show All]


Janitor application acceptance emails are being sent out. Please remember to check your spam box!


[Advertise on 4chan]


File: i36ekdptg12g1.png (209 KB, 3840x2160)
209 KB
209 KB PNG
How will openAI recover from this?
>>
>>16851788
Use case?
>>
Why does someone just not open up their image generator to make nudes and make a gorillion dollars?
>>
>>16851788
by releasing a better model, duh.
>>
>>16851807
Oh, you mean like 5.1? :-)
>>
File: 1743522602363177.png (19 KB, 800x1000)
19 KB
19 KB PNG
>>16851788
It's all the same shit, it's been the same for like 3 years lol
>>
>>16851788
Interesting progress, especially on arc-2, though I saw the guys behind it post that it weirdly fails some easy arc-1 puzzles. Also apparently it has similar hallucination rate as 2.5, which was kind of high. Hope to try it myself to see.
>>
File: 1745125905784320.jpg (251 KB, 1271x684)
251 KB
251 KB JPG
Do you agree on Gemini 3's definition of mathematics?
>>
>>16851788
Bar graph department in freefall as AI makes new bar graph.
>>
>>16851838
Not bad. Who wrote it first?
>>
>>16851835
It's still number 1 on arc-1 leaderboard for significantly less cost than o3. As far as I'm concerned this marks the point where google overtake the competition and they never catch up after that. After all GPT-3 was built on an idea stolen from Google in the first place.
>>
File: 1757062164027083.jpg (606 KB, 1752x1511)
606 KB
606 KB JPG
Find a flaw in its logic. Pro tip: you can't.
>>
File: 1735276988813995.jpg (140 KB, 1391x807)
140 KB
140 KB JPG
Sigh... Maybe Gemini 4 will actually be able to read and understand simple questions. But I'm starting to think that these current LLMs are just a dead end.
>>
>>16851838
I agree with whoever it was that originally wrote that definition. AI isn't creative, it's purely derivative and always will be you dumb fuck.
>>
>>16851856
But what it said was correct. The hint is right there in what it said. We need to go beyond the standard system of real numbers used in almost all of science, engineering, and mathematics.

Which is true. Shit is fucking stagnating. Science is so fucking beaurocratic now, it's disgustingly institutionalized. Papers get published not because of rigor but because of politics and shit. It's all fucked and everyone knows it.
>>
>>16851862
You're just as fucking stupid as the AI. Read the fucking question again.
>>
>>16851843
>>16851858
Everyone knows it.
Some are only pretending to not see it.
>>
>>16851862
>We need to go beyond the standard system of real numbers used in almost all of science, engineering, and mathematics.
But that's why we're building the AI in the first place. To go beyond our human abilities. Right?
>>
>>16851856
>Prove the 0.999 =/= 1
Maybe try asking a question that makes sense lmao. Did you mean "Prove THAT 0.999 =/= 1? It probably just thought you were retarded and forgot to add the "..." too.
>>
>>16851862
>humans game systems
Well, duh.
>>
>>16851873
It can't even properly collect resources for easy access and display. I asked it some simple questions a few times. It answered and when I asked "what about x missing" off the top of my head it answered "yeah, you are absolutely right, there is also ...."
AI my ass. Shit tier scam
>>
>>16851876
>It probably just thought you were retarded
It was almost certainly trained on the gazillions of /sci/ posts pn this very subject. It should be able to entertain monkeys for nearly 15 minutes with its quality discourse and shit slinging.
>>
>>16851879
That's a great story to tell to your kids.
>>
>>16851880
Do you think that it can memorize all the data on the internet? This is what you "it just regurgitates information" faggots fail to understand. It builds a model from the data that can generalize to new data, this is EXACTLY what humans do when they learn.
>>
>>16851807
when they released gpt 5, they said they wouldn't release cutting edge models anymore because they just can't afford them. they'll just release affordable models
>>
>>16851850
>Lizard brain
>Thigh highs
Didn't know those exists prehuman.
>>
>>16851897
Data doesn't meant what you think it means. It does not generalize. it interpolates. You're a mouthbreather. It does not work in any way similar to a brain. Brains do not back propagate, for one. For another information is not stored in the brain.
>But muh heckin memory!
Nope.
>>
>>16851856
A dead end for what? I don’t think many of the people in this thread have put in much time with an LLM. I’m not going to tell you exactly what I do with LLMs, but it does incorporate layers of personal modification. At this point the sperge will yell that LLMs can’t learn between trainings. That’s true. But you can impose rules, behavioral parameters, protocols, macros, etc., in persistent memory, across projects and threads, and this can create a powerful tool. It isn’t learning, it’s compliance by way of governance constraints, but it works, and it works to your specification, even down to the selection of epistemic stances (I like a sparring partner who ruthlessly calls out my bullshit). Once you start down that path there’s no going back to default. You create a very efficient and sometimes very rewarding work flow. But hey, it’s a prediction engine, not a magic 8 ball, sometimes it still goes fractal.
>>
>>16851814
Yes, it's better. I can actually ask GPT 5.1-high questions about specifics in Gaitsgory's papers and get somewhat sensible answers while gemini-3-pro just shits out vague, short wordslop. I mean, you can try it yourself: take a closed genus-2 surface [math]\Sigma_2[/math], fix a connected reductive complex group [math]\hat{G}[/math] (Langlands dual of some G), and ask both gpt-5.1-high and gemini-3-pro to calculate [math]\operatorname{Loc}_{\hat{G} }(\Sigma_2)[/math].
>>
>>16851960
>Data doesn't meant what you think it means. It does not generalize. it interpolates
Obviously you never studied linear regression or machine learning

>It does not work in any way similar to a brain. Brains do not back propagate, for one.
It's literally a neural network that updates itself incrementally based on the data it sees, the exact details and substrates used to achieve this is irrelevant

>For another information is not stored in the brain
Lmao now I know for certain you're a brainlet
>>
>>16852037
>linear regression or machine learning
The fact you combine these two tells me all I need to know. Once the bubble bursts your tears will be saltier than the oceans.
>>
>>16851897
Naw. Nope. No.
>>
File: 1561261564654.jpg (44 KB, 1006x636)
44 KB
44 KB JPG
>>16851788
by getting more money from Nvidia, and buying more data centers from Oracle who will purchase more GPUs from Nvidia
>>
November 17, 2025

Grok 4.1 is now available to all users on grok.com, X, and the iOS and Android apps. It is rolling out immediately in Auto mode and can be selected explicitly as "Grok 4.1" in the model picker.

We are excited to introduce Grok 4.1, which brings significant improvements to the real-world usability of Grok. Our 4.1 model is exceptionally capable in creative, emotional, and collaborative interactions. It is more perceptive to nuanced intent, compelling to speak with, and coherent in personality, while fully retaining the razor-sharp intelligence and reliability of its predecessors. To achieve this, we used the same large scale reinforcement learning infrastructure that powered Grok 4 and applied it to optimize the style, personality, helpfulness, and alignment of the model. In order to optimize these non-verifiable reward signals, we developed new methods that let us use frontier agentic reasoning models as reward models to autonomously evaluate and iterate on responses at scale.

https://x.ai/news/grok-4-1



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.