On a MacBook Pro with Apple Silicon I have done the following:
% brew install llama.cpp
% llama-server -hf unsloth/gpt-oss-20b-GGUF:UD-Q4_K_XL --jinja
I have used here the default context value -c 4096
.
Then opened a webbrowser at localhost:8080
and asked the following question: