CISCai commited on
Commit
a4099b6
1 Parent(s): 0470387

Corrected eos token

Browse files

According to tokenizer_config.json the eos token should be <|im_end|>, and according to my own testing this is correct, using the wrong token results in infinite generation responses.

Files changed (2) hide show
  1. config.json +1 -1
  2. generation_config.json +1 -1
config.json CHANGED
@@ -5,7 +5,7 @@
5
  "attention_bias": false,
6
  "attention_dropout": 0.0,
7
  "bos_token_id": 1,
8
- "eos_token_id": 2,
9
  "hidden_act": "silu",
10
  "hidden_size": 2048,
11
  "initializer_range": 0.02,
 
5
  "attention_bias": false,
6
  "attention_dropout": 0.0,
7
  "bos_token_id": 1,
8
+ "eos_token_id": 7,
9
  "hidden_act": "silu",
10
  "hidden_size": 2048,
11
  "initializer_range": 0.02,
generation_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
- "eos_token_id": 2,
5
  "pad_token_id": 0,
6
  "transformers_version": "4.44.0"
7
  }
 
1
  {
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
+ "eos_token_id": 7,
5
  "pad_token_id": 0,
6
  "transformers_version": "4.44.0"
7
  }