[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/g/ - Technology


Thread archived.
You cannot reply anymore.


[Advertise on 4chan]


File: 1768053777001.png (264 KB, 2000x1071)
264 KB
264 KB PNG
Let us assume that you have a magical machine that can immediately optimize all parameters of an AI model as such that it is the most "intelligent". Let us ignore that there is no agreed upon metric of what that means, and just assume that we have some magical metric that measures intelligence correctly.
Given all that, how good would a, say, 4B parameter model be? What is the ceiling of how intelligent a small model like this can be with established transformer architecture?
>>
File: tired_doggo.jpg (433 KB, 1000x1000)
433 KB
433 KB JPG
>>107824208
>wants to know how "intelligent" a model can become
>can't measure intelligence



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.