it's january 2026, so things will likely change, at some point
- llama.cpp repo link
- build instructions
- contains instructions on ho to build the debug version
- default build creates devices: 'libggml-base.so' and 'libggml-cpu.so' - probably need to set up additional requirements for the other backends.
- looking at main.cpp for llama-simple - a CLI program to continue with a given prompt, specified on the command line. command line:
./llama-simple -m model.gguf [-n n_predict] [-ngl n_gpu_layers] [prompt]m <file>- (mandatory) path to the model file in gguf format