Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -138,8 +138,9 @@ def chat_llama3_8b(message: str,
|
|
138 |
input_ids = tokenizer.apply_chat_template(conversation, return_tensors="pt").to(model.device)
|
139 |
|
140 |
streamer = TextIteratorStreamer(tokenizer, timeout=10.0, skip_prompt=True, skip_special_tokens=True)
|
141 |
-
print(max_new_tokens)
|
142 |
max_new_tokens=4096
|
|
|
143 |
generate_kwargs = dict(
|
144 |
input_ids= input_ids,
|
145 |
streamer=streamer,
|
@@ -180,14 +181,16 @@ with gr.Blocks(fill_height=True, css=css) as demo:
|
|
180 |
gr.Slider(minimum=0,
|
181 |
maximum=1,
|
182 |
step=0.1,
|
183 |
-
value=0.
|
184 |
label="Temperature",
|
|
|
185 |
render=False),
|
186 |
gr.Slider(minimum=128,
|
187 |
maximum=4096,
|
188 |
step=1,
|
189 |
value=4096,
|
190 |
label="Max new tokens",
|
|
|
191 |
render=False),
|
192 |
],
|
193 |
examples=[
|
|
|
138 |
input_ids = tokenizer.apply_chat_template(conversation, return_tensors="pt").to(model.device)
|
139 |
|
140 |
streamer = TextIteratorStreamer(tokenizer, timeout=10.0, skip_prompt=True, skip_special_tokens=True)
|
141 |
+
# print(max_new_tokens)
|
142 |
max_new_tokens=4096
|
143 |
+
temperature=0.9
|
144 |
generate_kwargs = dict(
|
145 |
input_ids= input_ids,
|
146 |
streamer=streamer,
|
|
|
181 |
gr.Slider(minimum=0,
|
182 |
maximum=1,
|
183 |
step=0.1,
|
184 |
+
value=0.9,
|
185 |
label="Temperature",
|
186 |
+
interactive = False,
|
187 |
render=False),
|
188 |
gr.Slider(minimum=128,
|
189 |
maximum=4096,
|
190 |
step=1,
|
191 |
value=4096,
|
192 |
label="Max new tokens",
|
193 |
+
interactive = False,
|
194 |
render=False),
|
195 |
],
|
196 |
examples=[
|