File size: 224 Bytes
4d1b2ed
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
{
  "_from_model_config": true,
  "bos_token_id": 1,
  "eos_token_id": 2,
  "do_sample": true,
  "temperature": 0.1,
  "top_p": 0.95,
  "max_new_tokens": 25,
  "repetition_penalty": 1.2,
  "transformers_version": "4.40.2"
}