Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -13,7 +13,7 @@ def correct_text(text, genConfig):
|
|
13 |
corrected_text = tokenizer.decode(outputs[0], skip_special_tokens=True)
|
14 |
return corrected_text
|
15 |
|
16 |
-
def respond(text, max_new_tokens, min_new_tokens, num_beams, num_beam_groups, temperature, top_k, top_p, no_repeat_ngram_size, guidance_scale):
|
17 |
config = GenerationConfig(
|
18 |
max_new_tokens=max_new_tokens,
|
19 |
min_new_tokens=min_new_tokens,
|
@@ -24,7 +24,7 @@ def respond(text, max_new_tokens, min_new_tokens, num_beams, num_beam_groups, te
|
|
24 |
top_p=float(top_p),
|
25 |
no_repeat_ngram_size=no_repeat_ngram_size,
|
26 |
early_stopping=True,
|
27 |
-
do_sample=
|
28 |
)
|
29 |
|
30 |
if guidance_scale > 0:
|
|
|
13 |
corrected_text = tokenizer.decode(outputs[0], skip_special_tokens=True)
|
14 |
return corrected_text
|
15 |
|
16 |
+
def respond(text, max_new_tokens, min_new_tokens, num_beams, num_beam_groups, temperature, top_k, top_p, no_repeat_ngram_size, guidance_scale, do_sample: bool):
|
17 |
config = GenerationConfig(
|
18 |
max_new_tokens=max_new_tokens,
|
19 |
min_new_tokens=min_new_tokens,
|
|
|
24 |
top_p=float(top_p),
|
25 |
no_repeat_ngram_size=no_repeat_ngram_size,
|
26 |
early_stopping=True,
|
27 |
+
do_sample=do_sample
|
28 |
)
|
29 |
|
30 |
if guidance_scale > 0:
|