Spaces:
Sleeping
Sleeping
Details
Browse files
app.py
CHANGED
@@ -248,7 +248,11 @@ def generate_cells(dataset_id, prompt_fn, notebook_type="eda"):
|
|
248 |
{"role": "user", "content": prompt},
|
249 |
]
|
250 |
output = inference_client.chat_completion(
|
251 |
-
messages=prompt_messages,
|
|
|
|
|
|
|
|
|
252 |
)
|
253 |
|
254 |
generated_text = ""
|
@@ -276,7 +280,11 @@ def generate_cells(dataset_id, prompt_fn, notebook_type="eda"):
|
|
276 |
]
|
277 |
|
278 |
output = inference_client.chat_completion(
|
279 |
-
messages=prompt_messages,
|
|
|
|
|
|
|
|
|
280 |
)
|
281 |
cells_txt = output.choices[0].message.content
|
282 |
logging.info(f"---> Mapping to json output attempt {retries}")
|
|
|
248 |
{"role": "user", "content": prompt},
|
249 |
]
|
250 |
output = inference_client.chat_completion(
|
251 |
+
messages=prompt_messages,
|
252 |
+
stream=True,
|
253 |
+
max_tokens=2500,
|
254 |
+
top_p=0.8,
|
255 |
+
seed=42,
|
256 |
)
|
257 |
|
258 |
generated_text = ""
|
|
|
280 |
]
|
281 |
|
282 |
output = inference_client.chat_completion(
|
283 |
+
messages=prompt_messages,
|
284 |
+
stream=False,
|
285 |
+
max_tokens=2500,
|
286 |
+
top_p=0.8,
|
287 |
+
seed=42,
|
288 |
)
|
289 |
cells_txt = output.choices[0].message.content
|
290 |
logging.info(f"---> Mapping to json output attempt {retries}")
|