[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip / qa] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/g/ - Technology


Thread archived.
You cannot reply anymore.


[Advertise on 4chan]


File: officialZucImage.png (210 KB, 603x653)
210 KB
210 KB PNG
>Lama-3.1; 8b, 70b and 405 is now released on huggingface!

405b is beating chatgpt-4o, while being just half the parameter count, possibly because of not suffering the crippling effects of wokeness infusion

While the new 8 billion parameter models seems affected in reasoning by being forced to learn thirdie langues compared to the old 3.0 version.
>>
File: stats.jpg (250 KB, 1080x2025)
250 KB
250 KB JPG
>>101535820
>>
>>101535820
What's the VRAM requirement for 405b?
>>
>>101536000
for 4q about 200gb, then more for context
>>
>>101536000
one Mac Pro with M2 Ultra
say it
THANK YOU BASED APPLE
>>
>>101536000
Literally won't run regardless of what you change

go for the smaller models
>>
>>101535820
>he thinks the top llama model is uncensored
>he can't even run the foundation model on runpod without $200+ a month on runpod
>>
>>101535820
>open source llm and nn are the path forward
Second that. Just don't tell any retards. Work or wait for or both. It will come and it will be the best.
>>
>>101536329
>one Mac Pro with M2 Ultra
405B model, 192GB Mac...
you're going to run it at q3? Plus, even the absolute best M2 barely touches at 3070 in terms of performance.
>>
>>101535820
Mark looking extra zesty here
>>
>>101538815
thats just his new cool look
>>
Whats the cheapest Hardware that can run the bigger models? Doesnt need to be fast just borderline usable. Do i have to become an itoddler?
>>
>>101535820
>405b
what kind of specs do you realistic need to run something like this at home? definitely not something a desktop can handle on its own but how much hardware does a home server g tard need?
>>
>>101541866
something like 160GB of VRAM however you can manage that
>>
>>101543013
H100 dgx has 640 GB of vram
>>
>>101535820
Is that the one that was leaked ages ago?
Did anyone make anything with it?
>>
File: 20240723190031_2.jpg (445 KB, 2560x1440)
445 KB
445 KB JPG
>>101535820
nothing released so far is even remotely cool. the dalle3 shit was KINDA NEAT for like 3 months and it died out with how woke it fucking got after that first week of it being kino.
the whole ai thing is so shit. lama 3 405 just came out. it is at best as good as gpt4 lol like where the fuck are the robots and the fucking ai that will take over peoples jobs?!?!?
>>
>>101535820
I trust this mother fucker as far as I can throw him. What's the catch Zuck? You want me to believe your releasing a multi million dollar piece of software for purely altruistic reasons..
>>
>>101536189
>>101538988
>>101541866

Could I run something like this on say 128 lanes of avx512 at a tolerable token rate? Because I have access to a rig with nearly a tb of system ram.

>>101545395
He's doing it to fuck over Microsoft. He wants a model for internal use, has the hardware to train one because he was expecting to need it for the VR bullshit and ticktock knockoff. It looses him very little to release it to everyone.
>>
>>101545395
To bring down Microsoft/OpenAI.

Which I suppose counts as altruistic reasons.
>>
But can it produce the filthiest smut imaginable with no content restriction?
>>
>>101545536
>>101545589
Nah, I think there's more to it than that.
Here's a few theories:

Get it integrated into everything then hold everyone hostage.

Facebook is on the decline, so he's all in on anything that will bring users.

And lastly, long term goals. If his company can steer or control responses, he can subtly influence users for political or financial gains.
>>
>>101535820
How are the 'scores' generated for rating these models?
>>101536000
Someone had a 4090 and used the 405B option and it took him 30 mins for it to generate the word 'The'.

I am currently trying to install the 8B version but im a noob and only messed around with stable diffusion before this, so im just trying random shit the meta AI site tells me when I try to use it for tech support. Right now I got an error because I need the llama library but it cant install the llama library due to some error and its suggesting I download an older version of python but idk how to do that. and i tried updating pip and installing the other dependencies too.

But I need to finish this challenge of getting it to run locally I refuse to always use a website for it
>>
>>101545703
what os are you on?
>>
>>101545770
Windows of course, since I am a humble gamer dabbling in these things. Im using git bash as a terminal I guess although I tried to use powershell to see if that would magically fix any errors but it did not.
>>
>>101545536
>fuck over Microsoft.
Dare I say it, but, based?
>>
>>101535820
So it's "unsafe" and will have the hammer brought down on it behind the scenes.
>>
>>101545395
Free testing
>>
>>101535820
Sauce on 4o being 800B?
>>
>>101535820
Open source so that retards can provide cuckerberg with free work and data. Great idea.
>>
>>101535820
>405b is beating chatgpt-4o
no it's doesn't, it's barely better than the now decrepit chatgpt3
>>
File: 1702048985601082.png (19 KB, 800x1000)
19 KB
19 KB PNG
>>101535820
zammmmmmmn zucky look like dat??
>>
File: E-mad Mostaque.jpg (289 KB, 2556x1438)
289 KB
289 KB JPG
>>101546056
nobody works on Llamo for free tho
only Stable diffusion corp managed that
>>
The Llama-3.1-8b is pretty damn good.

Like turbo level at least which is great for local
>>
>>101547271
Did they up the context?
>>
>>101537268
>he thinks the top llama model is uncensored
Yikes. Fucking silicon valley cucks still complying with DEI.
>>
>>101538815
crazy what a good haircut can do.

>>101545395
"The enemy of my enemy is my friend" kind of reasoning.

>You want me to believe your releasing a multi million dollar piece of software for purely altruistic reasons..
To be fair, they already released a multi million dollar piece of software as open source (Pytorch), so MAYBE he's not so evil...
>>
>>101544098
Nigga that shit costs almost HALF A MILLION DOLLARS
>>
>>101545536
define tolerable token rate
either way, it'll be way under 1 tokens/sec
>>
>>101535820
>405b
Man, I only get about 35tk/s with 70b on a 4090... I can't imagine there'd be a person out there with the hardware to run the big one that fast.

>>101545703
>Someone had a 4090 and used the 405B option and it took him 30 mins for it to generate the word 'The'.
lol
>>
>>101545395
If I'd have to take a guess, then cheap AI probably makes Facebooks core buisiness more valuable. Maybe they're trying to boost the ecosystem to eventually use AI to offer professional shilling services, hellban people or "reanimate" dead loved ones on facebook for a fee
>>
>>101547287
I don't know, I only got a quickie out before I had to leave. I'll put it through it's paces tomorrow though.

It's very coherent and seems to be uncensored.
>>
>>101536000
wouldn't a dual socket cpu be reasonably fast and cheaper at that size? I doubt strapping x10 p80 together will be fast.
>>
>>101545703
just follow https://github.com/Mozilla-Ocho/llamafile?tab=readme-ov-file#using-llamafile-with-external-weights its the best way to test instead of running all these shitty python venvs that pull half the ecosystem.
>>
File: file.png (44 KB, 993x376)
44 KB
44 KB PNG
>>101547287
I think so
>>
File: file.png (72 KB, 1081x391)
72 KB
72 KB PNG
why
>>
it sucks dick refuses to write anything useful because muh violence when there is nothing like that. I told it to capture analyze a poem and it shits itself about muh safe guards.
>>
>>101536189
>>101543013
So this is unreachable for consumers? I assume consumers access models running on data center GPUs remotely. But the cost and energy use must be enormous. What the fuck?
>>
>>101545395
To reduce value for OpenAI/Microsoft dominance. OpenAI is setting themselves up to swallow the government contracts whole.
>>
>>101549008
Yeah corporate trannyism is in place. Surely its not your first time using a corporate model
>>
>>101547486
what are you, poor?
>>
>>101549220
no, just saving up for a catgirl loli local llama life size silicone stretable reusable pleasure AI sex robot
>>
File: 1716580915241820.jpg (133 KB, 1024x1024)
133 KB
133 KB JPG
>>101549310
oh nevermind, go on then king
>>
>>101549120
it is. They advertise 128k context and after a couple prompts it starts spamming ##########"""""""""""
>>
>>101535820
'berg be bussin right now, brah.
>>
File: ZuckyM.jpg (22 KB, 274x359)
22 KB
22 KB JPG
>>101535820
Zucky M did it
>>
File: Gold bricks.jpg (2.41 MB, 2160x3840)
2.41 MB
2.41 MB JPG
>>101547486
Just borrow one from work



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.