[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/g/ - Technology


Thread archived.
You cannot reply anymore.


[Advertise on 4chan]


I'm not a gamer, I would just be getting one for the local LLM/finetuning capabilities. Worth it or should I hold off for something in the pipeline?
>>
yes
>>
no
>>
Maybe


Nigger
>>
boomer just found out about nvidia stock 6 years too late
>>
No Nvidia card is worth its price at the moment.
>>
>>108440089
This thread has nothing to do with Nvidia stock. It has to do with bang for buck really. From what I've seen the 5090 doesn't seem much better than the 4090 so 4090 still seems to reign supreme in sub 3k category.
>>
>>108440076
>yes
>>108440084
>no
>>108440086
>Maybe
I don't know. Can you repeat the question?
>>
>>108439888
>>108440159
How about 3090 24Gb? I want to run a local llm too.
>>
File: arc.png (258 KB, 1000x1000)
258 KB
258 KB PNG
>>108439888
>>
>4090/3090 for llms
Am I wrong for thinking a commercial card would be better for this? I do stable diffusion on an 8Gb 2060 super and its great, only slow when upscaling to higher resolutions.
>>
>>108440301
which model? and how long does it take for image gens
>>
>>108440301
the 4090 is obviously much more capable. it really depends on what you want
>>
>>108440463
even if they have the same VRAM?
>>
get a blackwell 6000
>>
>>108440490
you need at least 3 of those if you want decent results with local LLM. oh, and make sure you have at least 128 GB of ram, too.
>>
>>108440499
yeah
>>
>>108440417
The Asus Dual, I think there's a v2.0 in there somewhere. Takes about 55 seconds for a 1024 x 1024 image using Invoke with a ~6Gb model
>>
>>108440499
so you really can't get shit unless you spend 20-30k?
>>
>>108439888
It's hard to get a 4090 without getting scammed, if you're spending that much might as well just get a 5090
>>
>>108439888
Get a couple 3090's instead.
>>
>>108440573
Get a Mac Studio, 6 grand for 256GB.
>>
>>108440631
Apple gives me hemorrhoids. any non apple alternatives that you could run linux on?
>>
>>108440695
Get a decommissioned Threadripper or dual Xeon server with as much memory you can afford, min couple grand. You'll be limited to a few tokens/s though.
>>
>>108440710
Would a Mac Studio be faster?
>>
>>108440714
Shitloads. >>108440631 is actually the most sensible option money wise.
>>
>>108439888
>LLM/finetuning capabilities
>finetuning
You need hundreds of VRAM to finetune even something like a 7b parameter model, not 24.
>>
>>108440718
>Mac Studio,
doesn't Nvidia also have a box like the max mini with unified memory?
>>
>>108440724
The Spark? Less memory for pretty much the same price.
>>
you don't need to finetune your own llms, they're all trash and theres no real use case for any of this. Don't fall for the meme. With the exception you're getting paid to do it and your company is fell for the meme
>>
>>108440767
But i thought pewdiepie fine tuned an LLM to be as good as ChatGPT
>>
File: 1731150005501056.gif (607 KB, 800x450)
607 KB
607 KB GIF
>>108439888
Get a 5070TI instead, about 350 tps faster than the 4090 for ai workloads. 4090 is not worth the cost and is inferior for AI performance.
>>
>>108440794
I think memory is the bigger issue than speed here.
>>
>>108440767
but can't i finetune it to be more based or something
>>
>>108440796
if Memory is the goal then: DGX spark. Spark = 128 gigs of vram at 3800-4000 USD with a tops through-rate of 1050.
>>
File: 1773780003141099.gif (3.49 MB, 396x380)
3.49 MB
3.49 MB GIF
>>108440796
Or wait for 6x series 6x series rumored to have of a tops of roughly 5000-6000 on the 5070 equivalent. ... Assuming if we ever see them materialize at this point due to component availability, I wouldn't waitfag personally but that's just me.
>>
>>108440820
Spark is 4800 USD now. At that point just get >>108440631. Double the VRAM and still decent performance.
>>
>>108440813
look into OG dolphin Mixtral / or BlackDolphin.
It's already "based" guardrail-free.
>>
>>108440813
>>108440856
Just go to /lmg/ and actually look at the recommended current models for particular use cases.
We have a general for that for a reason.
>>
>>108440864
but I want to ERP
>>
>>108440865
There's literally uncensored models listed there for that... At least look before posting.
>>
File: 1774910612886.gif (4 MB, 640x578)
4 MB
4 MB GIF
>>108440855
Dogshit mac evirons with no support for anything built for Nvidia which is most AI stuff still. + trash mactarded "jeetgineering" t. Louis Rossman. Even apps written native for ai on macs hit the wall with a laughably bad tops of 50-75 at best. Enjoy waiting 7 - 20 minutes for your single 1024x1024 SDXL local genned image x.x.
>>
>>108440885
>Enjoy waiting 7 - 20 minutes for your single 1024x1024 SDXL local genned image x.x.
You could have started by saying you're clueless.
>>
>>108440865
Literally any local llm running on lama or whatever backend of choice like lmstudio chained to SillyTavern front end gui then.
>>
>>108440885
>OP is talking about LLMs
>bring up SDXL
ok?
>>
>>108439888
It was absolutely worth it when it launched, I highly doubt its worth it now given the prices of everything
>>
>>108439888
this thing cannot finetune, you need at least a blackwell 6000 for local finetuning. a 4090 with at least 64gb of ddr5 is a pretty average local llm rig. just ask on /lmg/ about this shit.
>>
Anyone have one of those modded 22gb 2080tis from China? I heard they cheaped out on some of the board components so they aren't reliable.
>>
>>108440301
I know nothing about it but grok told me the commerical cards that hold big amounts of ram are actually slower than 5090 for image generation, their up side is the ram which they use to train way bigger models thou



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.