Recommended way to run this model:
llama-server -hf ggml-org/gemma-3-270m-it-GGUF -c 0 -fa
Then, access http://localhost:8080
Chat template
8-bit
Base model