On a MacBook Pro with Apple Silicon I have done the following:
% brew install llama.cpp
% llama-server -hf unsloth/gpt-oss-20b-GGUF:UD-Q4_K_XL --jinja
I have used here the default context value -c 4096.
Then opened a webbrowser at localhost:8080 and asked the following question:
