danielhanchen commited on
Commit
9b378d3
·
verified ·
1 Parent(s): 4467081

Add files using upload-large-folder tool

Browse files
.gitattributes CHANGED
@@ -34,3 +34,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
  tekken.json filter=lfs diff=lfs merge=lfs -text
 
 
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
  tekken.json filter=lfs diff=lfs merge=lfs -text
37
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "unsloth/Mistral-Nemo-Base-2407",
3
  "architectures": [
4
  "MistralForCausalLM"
5
  ],
@@ -22,8 +22,8 @@
22
  "sliding_window": null,
23
  "tie_word_embeddings": false,
24
  "torch_dtype": "bfloat16",
25
- "transformers_version": "4.44.2",
26
- "unsloth_version": "2024.9",
27
  "use_cache": true,
28
  "vocab_size": 131072
29
  }
 
1
  {
2
+ "_name_or_path": "mistralai/Mistral-Nemo-Base-2407",
3
  "architectures": [
4
  "MistralForCausalLM"
5
  ],
 
22
  "sliding_window": null,
23
  "tie_word_embeddings": false,
24
  "torch_dtype": "bfloat16",
25
+ "transformers_version": "4.49.0.dev0",
26
+ "unsloth_fixed": true,
27
  "use_cache": true,
28
  "vocab_size": 131072
29
  }
generation_config.json CHANGED
@@ -2,7 +2,7 @@
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
- "max_length": 1024000,
6
  "pad_token_id": 10,
7
- "transformers_version": "4.44.2"
8
  }
 
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
+ "max_length": 131072,
6
  "pad_token_id": 10,
7
+ "transformers_version": "4.49.0.dev0"
8
  }
model.safetensors.index.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "metadata": {
3
- "total_size": 23604234240
4
  },
5
  "weight_map": {
6
  "lm_head.weight": "model-00005-of-00005.safetensors",
 
1
  {
2
  "metadata": {
3
+ "total_size": 24495564800
4
  },
5
  "weight_map": {
6
  "lm_head.weight": "model-00005-of-00005.safetensors",
special_tokens_map.json CHANGED
@@ -13,7 +13,13 @@
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
- "pad_token": "<pad>",
 
 
 
 
 
 
17
  "unk_token": {
18
  "content": "<unk>",
19
  "lstrip": false,
 
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
+ "pad_token": {
17
+ "content": "<pad>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
  "unk_token": {
24
  "content": "<unk>",
25
  "lstrip": false,
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -8007,7 +8007,8 @@
8007
  "bos_token": "<s>",
8008
  "clean_up_tokenization_spaces": false,
8009
  "eos_token": "</s>",
8010
- "model_max_length": 1000000000000000019884624838656,
 
8011
  "pad_token": "<pad>",
8012
  "padding_side": "left",
8013
  "tokenizer_class": "PreTrainedTokenizerFast",
 
8007
  "bos_token": "<s>",
8008
  "clean_up_tokenization_spaces": false,
8009
  "eos_token": "</s>",
8010
+ "extra_special_tokens": {},
8011
+ "model_max_length": 131072,
8012
  "pad_token": "<pad>",
8013
  "padding_side": "left",
8014
  "tokenizer_class": "PreTrainedTokenizerFast",