tuhinatripathi commited on
Commit
adc10da
1 Parent(s): 16e731f

Training in progress, epoch 1

Browse files
adapter_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "unsloth/gemma-2b-bnb-4bit",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "unsloth/gemma-2b-it-bnb-4bit",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a5680bf0c6fd6ffd9f8eb017d1cbe451fc452236d01e4c9a025c0578424e1d9d
3
  size 39256456
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1249a9d17b16e9873b34350efcace19c63754186f340acd957351b2352639c0d
3
  size 39256456
runs/May14_19-22-32_4ac1ee8d43cd/events.out.tfevents.1715714564.4ac1ee8d43cd.546.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:21f2d13725a872d3bf6e6284be2350031848ab5d95c0eb1c9e95f631b231f0a0
3
+ size 5135
runs/May14_19-26-31_4ac1ee8d43cd/events.out.tfevents.1715714795.4ac1ee8d43cd.546.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f136f9b36f59eaade6280ad08392841a5d846817d7b9c52902cc4c6002a62f5d
3
+ size 18196
runs/May14_20-14-30_4ac1ee8d43cd/events.out.tfevents.1715717681.4ac1ee8d43cd.546.4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c9f5d9ad8000f413549afb04c7c971f0443c465edb760a18bc1e2569918f1e03
3
+ size 18196
runs/May14_21-03-30_4ac1ee8d43cd/events.out.tfevents.1715720616.4ac1ee8d43cd.546.5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b7da9b0008c15150814e3212b4f48f786a40bc5790aec1b788a491e42ebb19b4
3
+ size 18821
tokenizer_config.json CHANGED
@@ -1744,6 +1744,7 @@
1744
  "<end_of_turn>"
1745
  ],
1746
  "bos_token": "<bos>",
 
1747
  "clean_up_tokenization_spaces": false,
1748
  "eos_token": "<eos>",
1749
  "model_max_length": 8192,
 
1744
  "<end_of_turn>"
1745
  ],
1746
  "bos_token": "<bos>",
1747
+ "chat_template": "{{ bos_token }}{% if messages[0]['role'] == 'system' %}{{ raise_exception('System role not supported') }}{% endif %}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if (message['role'] == 'assistant') %}{% set role = 'model' %}{% else %}{% set role = message['role'] %}{% endif %}{{ '<start_of_turn>' + role + '\n' + message['content'] | trim + '<end_of_turn>\n' }}{% endfor %}{% if add_generation_prompt %}{{'<start_of_turn>model\n'}}{% endif %}",
1748
  "clean_up_tokenization_spaces": false,
1749
  "eos_token": "<eos>",
1750
  "model_max_length": 8192,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cac5db16d11f13b33328c16c8872a9250ac7f9f09af7a8b9fb3600d6e7dbabb2
3
  size 4984
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c5856ed2a197636c4a2b4b304924a88e4f6f9e4a333f5aef7a7d957673ce8ec3
3
  size 4984