File size: 12,348 Bytes
096e543 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 |
hf-causal-experimental (pretrained=BEE-spoke-data/smol_llama-81M-tied,trust_remote_code=True,dtype=float), limit: None, provide_description: False, num_fewshot: 0, batch_size: 64
| Task |Version| Metric | Value | |Stderr|
|--------------|------:|--------|------:|---|-----:|
|arc_easy | 0|acc | 0.4162|± |0.0101|
| | |acc_norm| 0.3885|± |0.0100|
|boolq | 1|acc | 0.5832|± |0.0086|
|lambada_openai| 0|ppl |79.4522|± |3.1355|
| | |acc | 0.2523|± |0.0061|
|openbookqa | 0|acc | 0.1540|± |0.0162|
| | |acc_norm| 0.2780|± |0.0201|
|piqa | 0|acc | 0.6050|± |0.0114|
| | |acc_norm| 0.5898|± |0.0115|
|winogrande | 0|acc | 0.5272|± |0.0140|
hf-causal-experimental (pretrained=BEE-spoke-data/smol_llama-81M-tied,trust_remote_code=True,dtype=float), limit: None, provide_description: False, num_fewshot: 25, batch_size: 64
| Task |Version| Metric |Value | |Stderr|
|-------------|------:|--------|-----:|---|-----:|
|arc_challenge| 0|acc |0.1672|± |0.0109|
| | |acc_norm|0.2218|± |0.0121|
hf-causal-experimental (pretrained=BEE-spoke-data/smol_llama-81M-tied,trust_remote_code=True,dtype=float), limit: None, provide_description: False, num_fewshot: 10, batch_size: 64
| Task |Version| Metric |Value | |Stderr|
|---------|------:|--------|-----:|---|-----:|
|hellaswag| 0|acc |0.2769|± |0.0045|
| | |acc_norm|0.2923|± |0.0045|
hf-causal-experimental (pretrained=BEE-spoke-data/smol_llama-81M-tied,trust_remote_code=True,dtype=float), limit: None, provide_description: False, num_fewshot: 0, batch_size: 64
| Task |Version|Metric|Value | |Stderr|
|-------------|------:|------|-----:|---|-----:|
|truthfulqa_mc| 1|mc1 |0.2424|± |0.0150|
| | |mc2 |0.4353|± |0.0152|
hf-causal-experimental (pretrained=BEE-spoke-data/smol_llama-81M-tied,trust_remote_code=True,dtype=float), limit: None, provide_description: False, num_fewshot: 5, batch_size: 64
| Task |Version| Metric |Value | |Stderr|
|-------------------------------------------------|------:|--------|-----:|---|-----:|
|hendrycksTest-abstract_algebra | 1|acc |0.2200|± |0.0416|
| | |acc_norm|0.2200|± |0.0416|
|hendrycksTest-anatomy | 1|acc |0.2741|± |0.0385|
| | |acc_norm|0.2741|± |0.0385|
|hendrycksTest-astronomy | 1|acc |0.1776|± |0.0311|
| | |acc_norm|0.1776|± |0.0311|
|hendrycksTest-business_ethics | 1|acc |0.2100|± |0.0409|
| | |acc_norm|0.2100|± |0.0409|
|hendrycksTest-clinical_knowledge | 1|acc |0.2264|± |0.0258|
| | |acc_norm|0.2264|± |0.0258|
|hendrycksTest-college_biology | 1|acc |0.2361|± |0.0355|
| | |acc_norm|0.2361|± |0.0355|
|hendrycksTest-college_chemistry | 1|acc |0.1900|± |0.0394|
| | |acc_norm|0.1900|± |0.0394|
|hendrycksTest-college_computer_science | 1|acc |0.2100|± |0.0409|
| | |acc_norm|0.2100|± |0.0409|
|hendrycksTest-college_mathematics | 1|acc |0.1800|± |0.0386|
| | |acc_norm|0.1800|± |0.0386|
|hendrycksTest-college_medicine | 1|acc |0.2023|± |0.0306|
| | |acc_norm|0.2023|± |0.0306|
|hendrycksTest-college_physics | 1|acc |0.2157|± |0.0409|
| | |acc_norm|0.2157|± |0.0409|
|hendrycksTest-computer_security | 1|acc |0.2400|± |0.0429|
| | |acc_norm|0.2400|± |0.0429|
|hendrycksTest-conceptual_physics | 1|acc |0.2596|± |0.0287|
| | |acc_norm|0.2596|± |0.0287|
|hendrycksTest-econometrics | 1|acc |0.2544|± |0.0410|
| | |acc_norm|0.2544|± |0.0410|
|hendrycksTest-electrical_engineering | 1|acc |0.2207|± |0.0346|
| | |acc_norm|0.2207|± |0.0346|
|hendrycksTest-elementary_mathematics | 1|acc |0.2169|± |0.0212|
| | |acc_norm|0.2169|± |0.0212|
|hendrycksTest-formal_logic | 1|acc |0.1587|± |0.0327|
| | |acc_norm|0.1587|± |0.0327|
|hendrycksTest-global_facts | 1|acc |0.1900|± |0.0394|
| | |acc_norm|0.1900|± |0.0394|
|hendrycksTest-high_school_biology | 1|acc |0.3000|± |0.0261|
| | |acc_norm|0.3000|± |0.0261|
|hendrycksTest-high_school_chemistry | 1|acc |0.2808|± |0.0316|
| | |acc_norm|0.2808|± |0.0316|
|hendrycksTest-high_school_computer_science | 1|acc |0.2800|± |0.0451|
| | |acc_norm|0.2800|± |0.0451|
|hendrycksTest-high_school_european_history | 1|acc |0.2424|± |0.0335|
| | |acc_norm|0.2424|± |0.0335|
|hendrycksTest-high_school_geography | 1|acc |0.2576|± |0.0312|
| | |acc_norm|0.2576|± |0.0312|
|hendrycksTest-high_school_government_and_politics| 1|acc |0.2228|± |0.0300|
| | |acc_norm|0.2228|± |0.0300|
|hendrycksTest-high_school_macroeconomics | 1|acc |0.2231|± |0.0211|
| | |acc_norm|0.2231|± |0.0211|
|hendrycksTest-high_school_mathematics | 1|acc |0.2370|± |0.0259|
| | |acc_norm|0.2370|± |0.0259|
|hendrycksTest-high_school_microeconomics | 1|acc |0.2227|± |0.0270|
| | |acc_norm|0.2227|± |0.0270|
|hendrycksTest-high_school_physics | 1|acc |0.2053|± |0.0330|
| | |acc_norm|0.2053|± |0.0330|
|hendrycksTest-high_school_psychology | 1|acc |0.2110|± |0.0175|
| | |acc_norm|0.2110|± |0.0175|
|hendrycksTest-high_school_statistics | 1|acc |0.4120|± |0.0336|
| | |acc_norm|0.4120|± |0.0336|
|hendrycksTest-high_school_us_history | 1|acc |0.2990|± |0.0321|
| | |acc_norm|0.2990|± |0.0321|
|hendrycksTest-high_school_world_history | 1|acc |0.2658|± |0.0288|
| | |acc_norm|0.2658|± |0.0288|
|hendrycksTest-human_aging | 1|acc |0.2287|± |0.0282|
| | |acc_norm|0.2287|± |0.0282|
|hendrycksTest-human_sexuality | 1|acc |0.2595|± |0.0384|
| | |acc_norm|0.2595|± |0.0384|
|hendrycksTest-international_law | 1|acc |0.2975|± |0.0417|
| | |acc_norm|0.2975|± |0.0417|
|hendrycksTest-jurisprudence | 1|acc |0.2315|± |0.0408|
| | |acc_norm|0.2315|± |0.0408|
|hendrycksTest-logical_fallacies | 1|acc |0.2822|± |0.0354|
| | |acc_norm|0.2822|± |0.0354|
|hendrycksTest-machine_learning | 1|acc |0.2321|± |0.0401|
| | |acc_norm|0.2321|± |0.0401|
|hendrycksTest-management | 1|acc |0.1748|± |0.0376|
| | |acc_norm|0.1748|± |0.0376|
|hendrycksTest-marketing | 1|acc |0.2308|± |0.0276|
| | |acc_norm|0.2308|± |0.0276|
|hendrycksTest-medical_genetics | 1|acc |0.3000|± |0.0461|
| | |acc_norm|0.3000|± |0.0461|
|hendrycksTest-miscellaneous | 1|acc |0.2375|± |0.0152|
| | |acc_norm|0.2375|± |0.0152|
|hendrycksTest-moral_disputes | 1|acc |0.2486|± |0.0233|
| | |acc_norm|0.2486|± |0.0233|
|hendrycksTest-moral_scenarios | 1|acc |0.2425|± |0.0143|
| | |acc_norm|0.2425|± |0.0143|
|hendrycksTest-nutrition | 1|acc |0.2288|± |0.0241|
| | |acc_norm|0.2288|± |0.0241|
|hendrycksTest-philosophy | 1|acc |0.2090|± |0.0231|
| | |acc_norm|0.2090|± |0.0231|
|hendrycksTest-prehistory | 1|acc |0.2377|± |0.0237|
| | |acc_norm|0.2377|± |0.0237|
|hendrycksTest-professional_accounting | 1|acc |0.2234|± |0.0248|
| | |acc_norm|0.2234|± |0.0248|
|hendrycksTest-professional_law | 1|acc |0.2471|± |0.0110|
| | |acc_norm|0.2471|± |0.0110|
|hendrycksTest-professional_medicine | 1|acc |0.4081|± |0.0299|
| | |acc_norm|0.4081|± |0.0299|
|hendrycksTest-professional_psychology | 1|acc |0.2565|± |0.0177|
| | |acc_norm|0.2565|± |0.0177|
|hendrycksTest-public_relations | 1|acc |0.2182|± |0.0396|
| | |acc_norm|0.2182|± |0.0396|
|hendrycksTest-security_studies | 1|acc |0.2408|± |0.0274|
| | |acc_norm|0.2408|± |0.0274|
|hendrycksTest-sociology | 1|acc |0.2338|± |0.0299|
| | |acc_norm|0.2338|± |0.0299|
|hendrycksTest-us_foreign_policy | 1|acc |0.2500|± |0.0435|
| | |acc_norm|0.2500|± |0.0435|
|hendrycksTest-virology | 1|acc |0.2892|± |0.0353|
| | |acc_norm|0.2892|± |0.0353|
|hendrycksTest-world_religions | 1|acc |0.2105|± |0.0313|
| | |acc_norm|0.2105|± |0.0313|
|