vrvenkatesh commited on
Commit
6edb06e
1 Parent(s): 463cb94

Training in progress, step 500

Browse files
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:506e2e0064cbd74c66b6c580fd8d36d11a8700e07be33ca500024ba69cd4dcb8
3
  size 1324830880
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d2bff958478a79ac4d386a928cf7f84585ebe311991db433d53a6a09e1c1e817
3
  size 1324830880
runs/Dec03_19-17-39_68be682e0e8d/events.out.tfevents.1701631065.68be682e0e8d.244.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:57e5c9dd54733d4b75d8a2c4c287a1bed72a23a08e114ae1a7133d6080fd07c5
3
+ size 4347
runs/Dec03_19-25-01_68be682e0e8d/events.out.tfevents.1701631506.68be682e0e8d.244.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:88f3de7f453d9503a5433a9158bf6eee115c7b0efb7a6ab558ce6f373fe1943e
3
+ size 4504
tokenizer_config.json CHANGED
@@ -23,7 +23,6 @@
23
  "clean_up_tokenization_spaces": true,
24
  "eos_token": "</s>",
25
  "errors": "replace",
26
- "is_split_into_words": true,
27
  "model_max_length": 1000000000000000019884624838656,
28
  "pad_token": "<pad>",
29
  "tokenizer_class": "GPT2Tokenizer",
 
23
  "clean_up_tokenization_spaces": true,
24
  "eos_token": "</s>",
25
  "errors": "replace",
 
26
  "model_max_length": 1000000000000000019884624838656,
27
  "pad_token": "<pad>",
28
  "tokenizer_class": "GPT2Tokenizer",
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1528ea95335456d18dff940984f0b0bfd970a37ab341573800b9db6a5d2ad850
3
  size 4664
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e4873a6e1834fe1134dfcbe69cfe506ee273492413c0d7311cd69f5d032a0b5a
3
  size 4664