Spaces:
Sleeping
Sleeping
vidhiparikh
commited on
Commit
•
1bc21a7
1
Parent(s):
436241a
Update app.py
Browse files
app.py
CHANGED
@@ -75,8 +75,8 @@ callback_manager = CallbackManager([StreamingStdOutCallbackHandler()])
|
|
75 |
# Function to create a conversational chain
|
76 |
def create_conversational_chain(database):
|
77 |
llama_llm = LlamaCpp(
|
78 |
-
# Set gpu_layers to the number of layers to offload to GPU. Set to 0 if no GPU acceleration is available on your system.
|
79 |
-
llm = AutoModelForCausalLM.from_pretrained("TheBloke/Llama-2-7b-Chat-GGUF", model_file="llama-2-7b-chat.
|
80 |
temperature=0.75,
|
81 |
max_tokens=200,
|
82 |
top_p=1,
|
|
|
75 |
# Function to create a conversational chain
|
76 |
def create_conversational_chain(database):
|
77 |
llama_llm = LlamaCpp(
|
78 |
+
# Set gpu_layers to the number of layers to offload to GPU. Set to 0 if no GPU acceleration is available on your system. llama-2-7b-chat.Q8_0.gguf
|
79 |
+
llm = AutoModelForCausalLM.from_pretrained("TheBloke/Llama-2-7b-Chat-GGUF", model_file="llama-2-7b-chat.Q8_0.gguf", model_type="llama", gpu_layers=0),
|
80 |
temperature=0.75,
|
81 |
max_tokens=200,
|
82 |
top_p=1,
|