πŸ“‹ Model Description


library_name: transformers tags:
  • grok
  • unsloth
license: other license_name: grok-2 license_link: https://huggingface.co/xai-org/grok-2/blob/main/LICENSE base_model:
  • xai-org/grok-2

Learn how to run Grok 2 correctly - Read our Guide.

Unsloth Dynamic 2.0 achieves superior accuracy & outperforms other leading quants.

Grok 2 Usage Guidelines

  • Use --jinja for llama.cpp. You must use PR 15539. For example use the code below:
  • git clone https://github.com/ggml-org/llama.cpp
  • cd llama.cpp && git fetch origin pull/15539/head:MASTER && git checkout MASTER && cd ..

Utilizes Alvaro's Grok-2 HF compatible tokenizer as provided here

Grok 2

This repository contains the weights of Grok 2, a model trained and used at xAI in 2024.

Usage: Serving with SGLang

  • Download the weights. You can replace /local/grok-2 with any other folder name you prefer.
hf download xai-org/grok-2 --local-dir /local/grok-2

You might encounter some errors during the download. Please retry until the download is successful.
If the download succeeds, the folder should contain 42 files and be approximately 500 GB.

  • Launch a server.

Install the latest SGLang inference engine (>= v0.5.1) from https://github.com/sgl-project/sglang/

Use the command below to launch an inference server. This checkpoint is TP=8, so you will need 8 GPUs (each with > 40GB of memory).

python3 -m sglang.launch_server --model /local/grok-2 --tokenizer-path /local/grok-2/tokenizer.tok.json --tp 8 --quantization fp8 --attention-backend triton

  • Send a request.

This is a post-trained model, so please use the correct chat template.

python3 -m sglang.test.send_one --prompt "Human: What is your name?<|separator|>\n\nAssistant:"

You should be able to see the model output its name, Grok.

Learn more about other ways to send requests here.

License

The weights are licensed under the Grok 2 Community License Agreement.

πŸ“‚ GGUF File List

πŸ“ Filename πŸ“¦ Size ⚑ Download
grok-2-UD-TQ1_0.gguf
Recommended LFS
76.17 GB Download