laquythang commited on
Commit
0ead229
1 Parent(s): dd61209

Training in progress, step 6

Browse files
adapter_config.json CHANGED
@@ -20,10 +20,10 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "dense",
24
  "query_key_value",
25
- "dense_4h_to_h",
26
- "dense_h_to_4h"
27
  ],
28
  "task_type": "CAUSAL_LM",
29
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
+ "dense_h_to_4h",
24
  "query_key_value",
25
+ "dense",
26
+ "dense_4h_to_h"
27
  ],
28
  "task_type": "CAUSAL_LM",
29
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:df2bbc3cf162bf3f7e6feac9e7dafa8931381498d73a06b4dff98aeea41e8777
3
  size 33572288
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6465b0e2f785246fba367e5d6e13428427c959a4a196f8343337934a404fe369
3
  size 33572288
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7039ee5e93e39ae75f53f38b044442799d1e30cd7ff9a83e519c275615c657cc
3
  size 6776
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b5626cd2613597d38b016ac503b35d867aeed1b5b31a1151465dc38b87f3fc7b
3
  size 6776