sileod commited on
Commit
2995476
1 Parent(s): a16abee

Upload Adapter

Browse files
Files changed (2) hide show
  1. config.json +39 -0
  2. pytorch_model.bin +2 -2
config.json CHANGED
@@ -15,6 +15,7 @@
15
  2,
16
  2,
17
  2,
 
18
  3,
19
  3,
20
  3,
@@ -73,9 +74,12 @@
73
  2,
74
  3,
75
  2,
 
76
  3,
77
  3,
78
  3,
 
 
79
  1,
80
  2,
81
  2,
@@ -96,6 +100,7 @@
96
  2,
97
  2,
98
  2,
 
99
  3,
100
  4,
101
  3,
@@ -108,6 +113,7 @@
108
  2,
109
  2,
110
  2,
 
111
  4,
112
  3,
113
  2,
@@ -155,6 +161,10 @@
155
  "not_entailment",
156
  "entailment"
157
  ],
 
 
 
 
158
  [
159
  "entailment",
160
  "contradiction",
@@ -421,6 +431,17 @@
421
  "entailment",
422
  "not_entailment"
423
  ],
 
 
 
 
 
 
 
 
 
 
 
424
  [
425
  "entailment",
426
  "neutral",
@@ -431,6 +452,10 @@
431
  "neutral",
432
  "contradiction"
433
  ],
 
 
 
 
434
  [
435
  "entailment",
436
  "neutral",
@@ -519,6 +544,10 @@
519
  "entailment",
520
  "not_entailment"
521
  ],
 
 
 
 
522
  [
523
  "entailment",
524
  "not entailment"
@@ -578,6 +607,10 @@
578
  "entailment",
579
  "non_entailment"
580
  ],
 
 
 
 
581
  [
582
  "entailment",
583
  "not_entailment"
@@ -634,6 +667,7 @@
634
  "glue/qnli",
635
  "glue/rte",
636
  "glue/wnli",
 
637
  "super_glue/cb",
638
  "anli/a1",
639
  "anli/a2",
@@ -692,9 +726,12 @@
692
  "dadc-limit-nli",
693
  "folio",
694
  "tomi-nli",
 
695
  "temporal-nli",
696
  "counterfactually-augmented-snli",
697
  "cnli",
 
 
698
  "chaos-mnli-ambiguity",
699
  "logiqa-2.0-nli",
700
  "mindgames",
@@ -712,6 +749,7 @@
712
  "resnli",
713
  "semantic_fragments_nli",
714
  "dataset_train_nli",
 
715
  "ruletaker",
716
  "PARARULE-Plus",
717
  "logical-entailment",
@@ -725,6 +763,7 @@
725
  "FOL-nli",
726
  "doc-nli",
727
  "mctest-nli",
 
728
  "idioms-nli",
729
  "lifecycle-entailment",
730
  "MSciNLI",
 
15
  2,
16
  2,
17
  2,
18
+ 2,
19
  3,
20
  3,
21
  3,
 
74
  2,
75
  3,
76
  2,
77
+ 4,
78
  3,
79
  3,
80
  3,
81
+ 2,
82
+ 3,
83
  1,
84
  2,
85
  2,
 
100
  2,
101
  2,
102
  2,
103
+ 2,
104
  3,
105
  4,
106
  3,
 
113
  2,
114
  2,
115
  2,
116
+ 2,
117
  4,
118
  3,
119
  2,
 
161
  "not_entailment",
162
  "entailment"
163
  ],
164
+ [
165
+ "False",
166
+ "True"
167
+ ],
168
  [
169
  "entailment",
170
  "contradiction",
 
431
  "entailment",
432
  "not_entailment"
433
  ],
434
+ [
435
+ "entailment",
436
+ "non-entailment",
437
+ "not entailment - contradiction",
438
+ "not entailment - unknown"
439
+ ],
440
+ [
441
+ "entailment",
442
+ "neutral",
443
+ "contradiction"
444
+ ],
445
  [
446
  "entailment",
447
  "neutral",
 
452
  "neutral",
453
  "contradiction"
454
  ],
455
+ [
456
+ "False",
457
+ "True"
458
+ ],
459
  [
460
  "entailment",
461
  "neutral",
 
544
  "entailment",
545
  "not_entailment"
546
  ],
547
+ [
548
+ "False",
549
+ "True"
550
+ ],
551
  [
552
  "entailment",
553
  "not entailment"
 
607
  "entailment",
608
  "non_entailment"
609
  ],
610
+ [
611
+ "consistent",
612
+ "inconsistent"
613
+ ],
614
  [
615
  "entailment",
616
  "not_entailment"
 
667
  "glue/qnli",
668
  "glue/rte",
669
  "glue/wnli",
670
+ "super_glue/boolq",
671
  "super_glue/cb",
672
  "anli/a1",
673
  "anli/a2",
 
726
  "dadc-limit-nli",
727
  "folio",
728
  "tomi-nli",
729
+ "puzzte",
730
  "temporal-nli",
731
  "counterfactually-augmented-snli",
732
  "cnli",
733
+ "boolq-natural-perturbations",
734
+ "equate",
735
  "chaos-mnli-ambiguity",
736
  "logiqa-2.0-nli",
737
  "mindgames",
 
749
  "resnli",
750
  "semantic_fragments_nli",
751
  "dataset_train_nli",
752
+ "nlgraph",
753
  "ruletaker",
754
  "PARARULE-Plus",
755
  "logical-entailment",
 
763
  "FOL-nli",
764
  "doc-nli",
765
  "mctest-nli",
766
+ "natural-language-satisfiability",
767
  "idioms-nli",
768
  "lifecycle-entailment",
769
  "MSciNLI",
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9c6c8c0ff6648d9e04b0360aa9b83d054800531a44f6e5c749ec3f42f3feb8eb
3
- size 718274
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:085feb9865446d73ac3a32c6373f25f6e46f7eaa35cfaf6acd908ebab664ebf9
3
+ size 758050