Diogo-V commited on
Commit
fd5e20c
1 Parent(s): 7e99212

Adding more results

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. params/tinyllama/8/nlr_t_no_sched/comb_12/init/lm_head/_0.pt +3 -0
  2. params/tinyllama/8/nlr_t_no_sched/comb_12/init/lm_head/_s.pt +3 -0
  3. params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.0.mlp.down_proj/_0.pt +3 -0
  4. params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.0.mlp.down_proj/_s.pt +3 -0
  5. params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.0.mlp.gate_proj/_0.pt +3 -0
  6. params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.0.mlp.gate_proj/_s.pt +3 -0
  7. params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.0.mlp.up_proj/_0.pt +3 -0
  8. params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.0.mlp.up_proj/_s.pt +3 -0
  9. params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.0.self_attn.k_proj/_0.pt +3 -0
  10. params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.0.self_attn.k_proj/_s.pt +3 -0
  11. params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.0.self_attn.o_proj/_0.pt +3 -0
  12. params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.0.self_attn.o_proj/_s.pt +3 -0
  13. params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.0.self_attn.q_proj/_0.pt +3 -0
  14. params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.0.self_attn.q_proj/_s.pt +3 -0
  15. params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.0.self_attn.v_proj/_0.pt +3 -0
  16. params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.0.self_attn.v_proj/_s.pt +3 -0
  17. params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.1.mlp.down_proj/_0.pt +3 -0
  18. params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.1.mlp.down_proj/_s.pt +3 -0
  19. params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.1.mlp.gate_proj/_0.pt +3 -0
  20. params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.1.mlp.gate_proj/_s.pt +3 -0
  21. params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.1.mlp.up_proj/_0.pt +3 -0
  22. params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.1.mlp.up_proj/_s.pt +3 -0
  23. params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.1.self_attn.k_proj/_0.pt +3 -0
  24. params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.1.self_attn.k_proj/_s.pt +3 -0
  25. params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.1.self_attn.o_proj/_0.pt +3 -0
  26. params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.1.self_attn.o_proj/_s.pt +3 -0
  27. params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.1.self_attn.q_proj/_0.pt +3 -0
  28. params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.1.self_attn.q_proj/_s.pt +3 -0
  29. params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.1.self_attn.v_proj/_0.pt +3 -0
  30. params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.1.self_attn.v_proj/_s.pt +3 -0
  31. params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.10.mlp.down_proj/_0.pt +3 -0
  32. params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.10.mlp.down_proj/_s.pt +3 -0
  33. params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.10.mlp.gate_proj/_0.pt +3 -0
  34. params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.10.mlp.gate_proj/_s.pt +3 -0
  35. params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.10.mlp.up_proj/_0.pt +3 -0
  36. params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.10.mlp.up_proj/_s.pt +3 -0
  37. params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.10.self_attn.k_proj/_0.pt +3 -0
  38. params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.10.self_attn.k_proj/_s.pt +3 -0
  39. params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.10.self_attn.o_proj/_0.pt +3 -0
  40. params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.10.self_attn.o_proj/_s.pt +3 -0
  41. params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.10.self_attn.q_proj/_0.pt +3 -0
  42. params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.10.self_attn.q_proj/_s.pt +3 -0
  43. params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.10.self_attn.v_proj/_0.pt +3 -0
  44. params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.10.self_attn.v_proj/_s.pt +3 -0
  45. params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.11.mlp.down_proj/_0.pt +3 -0
  46. params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.11.mlp.down_proj/_s.pt +3 -0
  47. params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.11.mlp.gate_proj/_0.pt +3 -0
  48. params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.11.mlp.gate_proj/_s.pt +3 -0
  49. params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.11.mlp.up_proj/_0.pt +3 -0
  50. params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.11.mlp.up_proj/_s.pt +3 -0
params/tinyllama/8/nlr_t_no_sched/comb_12/init/lm_head/_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:905098e6144715c635fc1c23da6ddad1c9be2586b4cb724412fd8755bd5f5131
3
+ size 129091
params/tinyllama/8/nlr_t_no_sched/comb_12/init/lm_head/_s.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d67e3569700118f833c3257f4616498608da14118d1bc9a892a827961b63fa4a
3
+ size 129091
params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.0.mlp.down_proj/_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0e9756e98d67a99fa331d0125d20cc3067581ca4f77a9283d4b580490bd00ad0
3
+ size 9283
params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.0.mlp.down_proj/_s.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e37267f893b556da7315387d2f20372599d8283606872638af98d8c933b7fb2
3
+ size 9283
params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.0.mlp.gate_proj/_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1d91e050afe8386693413df3c404fcfb82b07d6a08769e41ace3a03c21d76584
3
+ size 23619
params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.0.mlp.gate_proj/_s.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c4c0e955b712b415de8c6e86af10bfbfea26a829884270ea251ab1d52c4d68d8
3
+ size 23619
params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.0.mlp.up_proj/_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:82d81217c0439b779ce2e052b4477641865e4ca3b99db37b003c7075fd1b65d1
3
+ size 23619
params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.0.mlp.up_proj/_s.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:70efe21f061953b91dd9c00d4a190dbd89e09d74115db1bf4c3d23e2f18a7c52
3
+ size 23619
params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.0.self_attn.k_proj/_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9a1ae55313d40f0aa85055b438744701188771bddb39fcd940fd65de131d72d8
3
+ size 2115
params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.0.self_attn.k_proj/_s.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:929a7b3ea8034d1349725ebb076d7ad5dc13cd89c4a7cf5b6569d0a2273a1e32
3
+ size 2115
params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.0.self_attn.o_proj/_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:79796dfed97469b60f019e686d8a9d9ce58318b4b9a0f5890e199b8c1400b950
3
+ size 9283
params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.0.self_attn.o_proj/_s.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:66b392ebbf2fa04c213beba00590954d9d250b6020daaab359a4053ddff4803a
3
+ size 9283
params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.0.self_attn.q_proj/_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e829f6ed76646d0b44a8e9aa9726ccf8516a2235b4e5f95e70243c6429576535
3
+ size 9283
params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.0.self_attn.q_proj/_s.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:72e5c48b9565aef083fc10e78023ba621e64997bbc10712f50688da4f51c986e
3
+ size 9283
params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.0.self_attn.v_proj/_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a9b1261d43d4e2aee70f4ca33444f4baf5bebfe48b7bf92b1ddf8377ad1d2ed
3
+ size 2115
params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.0.self_attn.v_proj/_s.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5477111e186888b6fc5788710db0b0fc32017badbae1f3515bcc826f5e7c28ef
3
+ size 2115
params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.1.mlp.down_proj/_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af428f2b19902da4b55719f257e36ceb98f10dc9aefa853ad78673e80f55b7a5
3
+ size 9283
params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.1.mlp.down_proj/_s.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bfb05cd2d69bedb8f80e28b877d15a4f65fb2569edfa50c3a4a9ed274103eb5b
3
+ size 9283
params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.1.mlp.gate_proj/_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:62c1623d9d5e3cf3454109006d1e84914f7893971cbd41ce2af43e47055cc0b7
3
+ size 23619
params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.1.mlp.gate_proj/_s.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1038bc4449521c0ab83876a16efa06312ff1f2dcf4c1b1e172872c79a78cc89f
3
+ size 23619
params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.1.mlp.up_proj/_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:295b576f899c53b49c988b4369099cd2bf94415e4f9bbb295db7587ec02aa3dd
3
+ size 23619
params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.1.mlp.up_proj/_s.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:26f65868f5e24e270e2e39b5fae24a0ddd9ce76fa7b66f190424f3ac8a8dee66
3
+ size 23619
params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.1.self_attn.k_proj/_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d22fc1d50d531e015c444d0a9adf9da0a3b5867f9e7d206ab4320fd085a18efd
3
+ size 2115
params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.1.self_attn.k_proj/_s.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f77745f8d2088b31fff7ac3d2f3d6a57f2de1165927dddbcd4d4848e2e09c2a9
3
+ size 2115
params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.1.self_attn.o_proj/_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9a987fc6ff36f47fa196597fbaed0b51e51e4c4fef739000d23d85cfe91b731c
3
+ size 9283
params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.1.self_attn.o_proj/_s.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8ab67ee3ed56769f69dc873703797a2900a4ec0546528b922561332754fe31c6
3
+ size 9283
params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.1.self_attn.q_proj/_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4c7769bb2d14e64cfd1b64fe2803545ddf34a574828225c8d9328b1ad39e866b
3
+ size 9283
params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.1.self_attn.q_proj/_s.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:65afa9fa8ef38b91dd439d94b93ebec2ad92a20ef1ac6d18e43e3cd9d7917d5c
3
+ size 9283
params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.1.self_attn.v_proj/_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e63ebcd43e9ea675df95a37d4fd6aee0b49ab7f5033a60ec539c2bc47ee4dd8a
3
+ size 2115
params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.1.self_attn.v_proj/_s.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0be3f8f814724416fde11088cd8bc82e55401dcbf501b5948a383cb4e42662cb
3
+ size 2115
params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.10.mlp.down_proj/_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:59b7418e043486968e0e0641b598b1c41a40679dbe87f80ed0ba6866402db1de
3
+ size 9283
params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.10.mlp.down_proj/_s.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:372eb56f31a25333d0e48209bacbf4910d636cd28bed2aeddd2394e3fd248394
3
+ size 9283
params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.10.mlp.gate_proj/_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5e1d855df580b720548ab374095d57adcb7bcbd4a911a3720182cb18a438ddd9
3
+ size 23619
params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.10.mlp.gate_proj/_s.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:523df9bc98a212d62973c0bd1bbbc420113f5eab513ac14cdd996d7b56a3feb0
3
+ size 23619
params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.10.mlp.up_proj/_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6cdabecc7d1b5b50b9df50445d3791b67f06a7509b244ae81c8aedb45da64fb7
3
+ size 23619
params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.10.mlp.up_proj/_s.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e307c92a32c213b49732eb3697ef980aca3a37e7826bffd519c39ba9d25bcfbe
3
+ size 23619
params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.10.self_attn.k_proj/_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2bf10ac68c8cdce1ef2cd7da8e035afedfdc31df9b1b68707651735605acfa62
3
+ size 2115
params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.10.self_attn.k_proj/_s.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fbd1f53aab130f5f43bc73465a3da7a198ee0cff676a27e7a15113194cdd7d09
3
+ size 2115
params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.10.self_attn.o_proj/_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e7e1c289e9027e190901c06cf6ff3dd0de554511e72ed30b642fdada5b879ae0
3
+ size 9283
params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.10.self_attn.o_proj/_s.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:466ca1dbe4e4dd09fad7daced949e0b8914cf9679825dcefb843f0fdb9926326
3
+ size 9283
params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.10.self_attn.q_proj/_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b48fc85bf9d0d7b7a8957455567dd076df7db8ad63474efb1bddca2584fddb21
3
+ size 9283
params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.10.self_attn.q_proj/_s.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e5304ddbed14f52a73c510d4bae9b223584bcc33b639e9a4b41c79da5844f18
3
+ size 9283
params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.10.self_attn.v_proj/_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d4cdb21390573d98c6040d8956224368099638231d16cdca09671d0ac6cb3bd
3
+ size 2115
params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.10.self_attn.v_proj/_s.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0647324ac2c352d909023aabb62ff7025ed03a8927cb07b2e908af6271056c89
3
+ size 2115
params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.11.mlp.down_proj/_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ac5a0cb7f458590cff9ca6f457fa3d05b992e38f7e3b7dc708da74ef910dfecc
3
+ size 9283
params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.11.mlp.down_proj/_s.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6c8c5d9263e3a793f5098bd86c368f8c92c74a50974060cc05149ca58efb6f6e
3
+ size 9283
params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.11.mlp.gate_proj/_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:22781fecc9f3e935110e464a4c0141343fd639357b796a0fbff58cde988c4810
3
+ size 23619
params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.11.mlp.gate_proj/_s.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5192e8fc99def9dded1550ce0b30e1a91fe96bbcdd4897b22f5141c25193d25c
3
+ size 23619
params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.11.mlp.up_proj/_0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dd914dd0bb3578732a04c97c0ac009418fcb9c79a3993b4309ba57f3cc763d8b
3
+ size 23619
params/tinyllama/8/nlr_t_no_sched/comb_12/init/model.layers.11.mlp.up_proj/_s.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e0bc58a33f4771aa37a15471eeafb149de55363b7e7ce2df672817417856c584
3
+ size 23619