@ -21,6 +21,8 @@ def llama(prompt):
"/opt/llama.cpp/models/ggml-vicuna-13b-4bit.bin",
"-n",
"-1",
"-c",
"2048"
"-p",
prompt,
],