From 181d2f5793e069f4ece0fccf4f597400de6ea9d3 Mon Sep 17 00:00:00 2001 From: Anthony Wang Date: Tue, 30 May 2023 03:18:36 +0000 Subject: [PATCH] Use WizardLM instead of Vicuna --- main.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/main.py b/main.py index 0cce538..dc6ebd9 100644 --- a/main.py +++ b/main.py @@ -20,11 +20,11 @@ def llama(prompt): "-ngl", "32", "-m", - "/opt/llama.cpp/models/ggml-vicuna-7b-1.1-q4_0.bin", + "/opt/llama.cpp/models/wizardLM-7B.ggmlv3.q4_0.bin", "-n", "1024", "-p", - f"### Human: {prompt}\n### Assistant:", + f"{prompt}\n\n### Response:", ], stdout=subprocess.PIPE, )