hfl-rc commited on
Commit
960904c
·
verified ·
1 Parent(s): 4e6c5df

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +8 -0
app.py CHANGED
@@ -50,11 +50,19 @@ def stream_chat(message: str, history: list, system_prompt: str, model_version:
50
  input_ids = tokenizer.apply_chat_template(conversation, add_generation_prompt=True, return_tensors="pt").to(model.device)
51
  streamer = TextIteratorStreamer(tokenizer, timeout=10.0, skip_prompt=True, skip_special_tokens=True)
52
 
 
 
53
  generate_kwargs = {
54
  "input_ids": input_ids,
55
  "streamer": streamer,
 
 
56
  "max_new_tokens": max_new_tokens,
57
  "temperature": temperature,
 
 
 
 
58
  "do_sample": temperature != 0,
59
  }
60
 
 
50
  input_ids = tokenizer.apply_chat_template(conversation, add_generation_prompt=True, return_tensors="pt").to(model.device)
51
  streamer = TextIteratorStreamer(tokenizer, timeout=10.0, skip_prompt=True, skip_special_tokens=True)
52
 
53
+ terminators = [tokenizer.eos_token_id, tokenizer.convert_tokens_to_ids("<|eot_id|>")]
54
+
55
  generate_kwargs = {
56
  "input_ids": input_ids,
57
  "streamer": streamer,
58
+ "eos_token_id": terminators,
59
+ "pad_token_id": tokenizer.eos_token_id,
60
  "max_new_tokens": max_new_tokens,
61
  "temperature": temperature,
62
+ "top_k": 40,
63
+ "top_p": 0.9,
64
+ "num_beams": 1,
65
+ "repetition_penalty": 1.1,
66
  "do_sample": temperature != 0,
67
  }
68