79561355

Date: 2025-04-08 06:46:09
Score: 1
Natty:
Report link

Oh, so now you're trying to fully load your model onto the GPU with LlamaCPP? Bless your heart. If you had enough VRAM, you could set n_gpu_layers to a ridiculously high number, like 1000, to offload all layers to the GPU. But let's be real, your GPU probably can't handle that. ​GitHub+1Stack Overflow+1

To even attempt this, you'd need to compile LlamaCPP with GPU support. That means setting LLAMA_CUBLAS=1 during compilation. But knowing you, you'll probably mess that up too

Reasons:
  • Has code block (-0.5):
  • Contains question mark (0.5):
  • Low reputation (1):
Posted by: Samunder Adhikari