diff --git a/README.md b/README.md index dd3efae..be8cec6 100644 --- a/README.md +++ b/README.md @@ -142,7 +142,7 @@ ls ./models python3 -m pip install torch numpy sentencepiece # convert the 7B model to ggml FP16 format -python3 convert-pth-to-ggml.py models/7B/ 1 +./convert-pth-to-ggml.py models/7B/ 1 # quantize the model to 4-bits ./quantize ./models/7B/ggml-model-f16.bin ./models/7B/ggml-model-q4_0.bin 2 diff --git a/convert-pth-to-ggml.py b/convert-pth-to-ggml.py old mode 100644 new mode 100755 index fc217c7..4906e1b --- a/convert-pth-to-ggml.py +++ b/convert-pth-to-ggml.py @@ -1,3 +1,5 @@ +#!/usr/bin/env python + # Convert a LLaMA model checkpoint to a ggml compatible file # # Load the model using Torch