Update app.py
Browse files
app.py
CHANGED
@@ -133,6 +133,7 @@ def hermes_generate_response(msg_prompt: str) -> dict:
|
|
133 |
|
134 |
|
135 |
def mistral_generate_response(msg_prompt, persona_desc_prompt):
|
|
|
136 |
user_prompt = f'{msg_prompt} [/INST]'
|
137 |
persona_prompt = f'{persona_desc_prompt} [/INST]'
|
138 |
prompt_template = f'''[INST] Instruction:{persona_prompt} [INST] {user_prompt}'''
|
@@ -144,6 +145,7 @@ def mistral_generate_response(msg_prompt, persona_desc_prompt):
|
|
144 |
generated_ids = model.generate(model_inputs, max_new_tokens=1000, do_sample=True)
|
145 |
decoded = tokenizer.batch_decode(generated_ids)
|
146 |
response = (decoded[0])
|
|
|
147 |
return response
|
148 |
|
149 |
|
|
|
133 |
|
134 |
|
135 |
def mistral_generate_response(msg_prompt, persona_desc_prompt):
|
136 |
+
|
137 |
user_prompt = f'{msg_prompt} [/INST]'
|
138 |
persona_prompt = f'{persona_desc_prompt} [/INST]'
|
139 |
prompt_template = f'''[INST] Instruction:{persona_prompt} [INST] {user_prompt}'''
|
|
|
145 |
generated_ids = model.generate(model_inputs, max_new_tokens=1000, do_sample=True)
|
146 |
decoded = tokenizer.batch_decode(generated_ids)
|
147 |
response = (decoded[0])
|
148 |
+
|
149 |
return response
|
150 |
|
151 |
|