From 81c9c9e8a65e0be5416f5fad896f8a4a3add94ce Mon Sep 17 00:00:00 2001 From: Dietrich Daroch Date: Sun, 12 Mar 2023 17:05:13 -0700 Subject: [PATCH] Misc: Make the conversion script executable --- README.md | 2 +- convert-pth-to-ggml.py | 2 ++ 2 files changed, 3 insertions(+), 1 deletion(-) mode change 100644 => 100755 convert-pth-to-ggml.py diff --git a/README.md b/README.md index dd3efae..be8cec6 100644 --- a/README.md +++ b/README.md @@ -142,7 +142,7 @@ ls ./models python3 -m pip install torch numpy sentencepiece # convert the 7B model to ggml FP16 format -python3 convert-pth-to-ggml.py models/7B/ 1 +./convert-pth-to-ggml.py models/7B/ 1 # quantize the model to 4-bits ./quantize ./models/7B/ggml-model-f16.bin ./models/7B/ggml-model-q4_0.bin 2 diff --git a/convert-pth-to-ggml.py b/convert-pth-to-ggml.py old mode 100644 new mode 100755 index fc217c7..4906e1b --- a/convert-pth-to-ggml.py +++ b/convert-pth-to-ggml.py @@ -1,3 +1,5 @@ +#!/usr/bin/env python + # Convert a LLaMA model checkpoint to a ggml compatible file # # Load the model using Torch