chinese-alpaca-7b-gptq / tokenizer_config.json
keyfan's picture
Add model files
0fd3ee5
raw
history blame contribute delete
394 Bytes
{
"bos_token": "<s>",
"clean_up_tokenization_spaces": false,
"eos_token": "</s>",
"model_max_length": 1000000000000000019884624838656,
"special_tokens_map_file": "/home/serving/.cache/huggingface/hub/models--ziqingyang--chinese-alpaca-lora-7b/snapshots/89dd628daa79cae8b930f2a7066006c39d6ac454/special_tokens_map.json",
"tokenizer_class": "LlamaTokenizer",
"unk_token": "<unk>"
}