grok-2-GGUF / README.md
danielhanchen's picture
Update README.md
7b560ae verified
metadata
library_name: transformers
tags:
  - grok
  - unsloth
license: other
license_name: grok-2
license_link: https://huggingface.co/xai-org/grok-2/blob/main/LICENSE
base_model:
  - xai-org/grok-2

Learn how to run Grok 2 correctly - Read our Guide.

Unsloth Dynamic 2.0 achieves superior accuracy & outperforms other leading quants.

Grok 2 Usage Guidelines

  • Use --jinja for llama.cpp. You must use PR 15539. For example use the code below:
  • git clone https://github.com/ggml-org/llama.cpp
  • cd llama.cpp && git fetch origin pull/15539/head:MASTER && git checkout MASTER && cd ..

Utilizes Alvaro's Grok-2 HF compatible tokenizer as provided here

Grok 2

This repository contains the weights of Grok 2, a model trained and used at xAI in 2024.

Usage: Serving with SGLang

  • Download the weights. You can replace /local/grok-2 with any other folder name you prefer.

    hf download xai-org/grok-2 --local-dir /local/grok-2
    

    You might encounter some errors during the download. Please retry until the download is successful.
    If the download succeeds, the folder should contain 42 files and be approximately 500 GB.

  • Launch a server.

    Install the latest SGLang inference engine (>= v0.5.1) from https://github.com/sgl-project/sglang/

    Use the command below to launch an inference server. This checkpoint is TP=8, so you will need 8 GPUs (each with > 40GB of memory).

    python3 -m sglang.launch_server --model /local/grok-2 --tokenizer-path /local/grok-2/tokenizer.tok.json --tp 8 --quantization fp8 --attention-backend triton
    
  • Send a request.

    This is a post-trained model, so please use the correct chat template.

    python3 -m sglang.test.send_one --prompt "Human: What is your name?<|separator|>\n\nAssistant:"
    

    You should be able to see the model output its name, Grok.

    Learn more about other ways to send requests here.

License

The weights are licensed under the Grok 2 Community License Agreement.