# Grok-1 Inference - 314 Billion Parameter Grok-1 Inference Accelerated by 3.8x, an easy-to-use Python + PyTorch + HuggingFace version for Inference. [[code]](https://github.com/hpcaitech/ColossalAI/tree/main/examples/language/grok-1) [[blog]](https://hpc-ai.com/blog/314-billion-parameter-grok-1-inference-accelerated-by-3.8x-efficient-and-easy-to-use-pytorchhuggingface-version-is-here) [[HuggingFace Grok-1 PyTorch model weights]](https://huggingface.co/hpcai-tech/grok-1) [[ModelScope Grok-1 PyTorch model weights]](https://www.modelscope.cn/models/colossalai/grok-1-pytorch/summary)
## Installation ```bash # Make sure you install colossalai from the latest source code git clone https://github.com/hpcaitech/ColossalAI.git cd ColossalAI pip install . cd examples/language/grok-1 pip install -r requirements.txt ``` ## Inference You need 8x A100 80GB or equivalent GPUs to run the inference. We provide two scripts for inference. `run_inference_fast.sh` uses tensor parallelism provided by ColossalAI, which is faster for generation, while `run_inference_slow.sh` uses auto device provided by transformers, which is relatively slower. Command example: ```bash ./run_inference_fast.sh