[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/g/ - Technology

Name
Options
Comment
Verification
4chan Pass users can bypass this verification. [Learn More] [Login]
File
  • Please read the Rules and FAQ before posting.
  • You may highlight syntax and preserve whitespace by using [code] tags.

08/21/20New boards added: /vrpg/, /vmg/, /vst/ and /vm/
05/04/17New trial board added: /bant/ - International/Random
10/04/16New board for 4chan Pass users: /vip/ - Very Important Posts
[Hide] [Show All]


Janitor application acceptance emails are being sent out. Please remember to check your spam box!


[Advertise on 4chan]


File: 1731429531133813.png (795 KB, 1323x813)
795 KB
795 KB PNG
it just werks
>>
>>107106564
>can't have files over 4GB
straight into the trash
>>
>>107106628
thats because windows fucking sucks.
>>
File: 1745104921963826.png (488 KB, 1839x776)
488 KB
488 KB PNG
>>107106628
>>107106699
idgi, do these not work?
>>
>>107106699
Who is still using Fat32. Are you retarded?
>>
File: file.png (100 KB, 1233x144)
100 KB
100 KB PNG
>>107106746
windows has a filesize limit for executables.
>>107106755
retard
>>
>>107106564
>troonware
no thanks, still using the superior llama.cpp, thanks
>>
File: 1745993400598726.png (447 KB, 1502x558)
447 KB
447 KB PNG
>>107106768
kek imagine using winshit, what the fuck are we still in x86_32 days?
and yea apparently you gotta split it into 2 files
>>
for me its llamacpp server in docker anyone on lan can access through its http port
>>
>>107106904
too bloated if you just need a small fast model to run on your code editor. llamafile is simple
>>
>>107106564
Bump. Good morning sir!
>>
>>107109379
good morning
>>
>>107106699
why would someone use Wiindows?
>>
>>107111272
i had to in a workplace several times on an airgapped network. if llamafiles existed it would have saved me a fuckton of time
>>
>>107111272
why are you a retard?
>>
>>107106699
>runs llms on windows
just use online services if you don't give a fuck
>>
>>107106564
coded by a tranny that tried to sabotage the upstream project
search "justine" tunney for more
>>
>>107106564
> just werks
> built in sandboxing on linux
> reasonable token production
If you know someone who isn't computer literate, but wants to try local LLMs, this is a good solution. Lots of options. I suggest
> Mistral 7B Instruct v0.3.Q4 (fast general purpose)
> Google gemma 3 12B it Q4 K M (general purpose
> Gemma 2 27B it Q6K (slow general purpose)
> Qwen2.5.1 Coder 7B Instruct Q8 (fast coding helper)
> Qwen2.5 Coder 14B Q6K (slow coding helper)
>>
>>107114504
Oh, I almost forgot. You can load gguf's with this, too.
> ./Mistral-7B-Instruct-v03.!4_0.llamafile -m <your-gguf-model-here>
>>
>yet another ai chatbot
holy yawn
>>
>>107114504
im trying to read up on LLM and quants
q4 seems bad, i dont quite get difference of q6k vs q8
how can u do 6bit?^
just how big is the difference of a q6k 7b vs 13b
it seems q6 is the sweetspot and K M suffixes seem to imply better trained
>>
>>107114919
>how can u do 6bit?
you pack the next 2 bits in the empty part



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.