Spaces:
Sleeping
Sleeping
reduce model size
Browse files
app.py
CHANGED
@@ -34,7 +34,7 @@ print("loading the LLM......................................")
|
|
34 |
# )
|
35 |
|
36 |
llm = AutoModelForCausalLM.from_pretrained("TheBloke/Llama-2-7B-Chat-GGUF",
|
37 |
-
model_file="llama-2-7b-chat.
|
38 |
model_type="llama",
|
39 |
temperature = 0.2,
|
40 |
repetition_penalty = 1.5
|
|
|
34 |
# )
|
35 |
|
36 |
llm = AutoModelForCausalLM.from_pretrained("TheBloke/Llama-2-7B-Chat-GGUF",
|
37 |
+
model_file="llama-2-7b-chat.Q3_K_S.gguf",
|
38 |
model_type="llama",
|
39 |
temperature = 0.2,
|
40 |
repetition_penalty = 1.5
|