Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -5,7 +5,7 @@ import torch
|
|
5 |
# Load the model and tokenizer
|
6 |
model_path = 'LLM4Binary/llm4decompile-1.3b-v1.5' # V1.5 Model
|
7 |
tokenizer = AutoTokenizer.from_pretrained(model_path)
|
8 |
-
model = AutoModelForCausalLM.from_pretrained(model_path, torch_dtype=torch.bfloat16)
|
9 |
|
10 |
# Define the inference function
|
11 |
def generate_response(input_text, temperature, top_k, top_p):
|
|
|
5 |
# Load the model and tokenizer
|
6 |
model_path = 'LLM4Binary/llm4decompile-1.3b-v1.5' # V1.5 Model
|
7 |
tokenizer = AutoTokenizer.from_pretrained(model_path)
|
8 |
+
model = AutoModelForCausalLM.from_pretrained(model_path, torch_dtype=torch.bfloat16)
|
9 |
|
10 |
# Define the inference function
|
11 |
def generate_response(input_text, temperature, top_k, top_p):
|