>>102291210
>iirc there will be no Gpu acceleration only with this.
Correct. You'd need to run the following instead:
cd llama.cpp && make GGML_CUDA=1
For performance tweaks, check out:
https://github.com/ggerganov/llama.cpp/blob/master/docs/build.md#cuda