Lin-K76 commited on
Commit
46557f3
1 Parent(s): 2872cfa

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +68 -1
README.md CHANGED
@@ -218,4 +218,71 @@ This version of the lm-evaluation-harness includes versions of ARC-Challenge and
218
  <td><strong>99.97%</strong>
219
  </td>
220
  </tr>
221
- </table>
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
218
  <td><strong>99.97%</strong>
219
  </td>
220
  </tr>
221
+ </table>
222
+
223
+
224
+ ### Reproduction
225
+
226
+ The results were obtained using the following commands:
227
+
228
+ #### MMLU
229
+ ```
230
+ lm_eval \
231
+ --model vllm \
232
+ --model_args pretrained="neuralmagic/Meta-Llama-3.1-405B-Instruct-FP8-dynamic",dtype=auto,add_bos_token=True,max_model_len=4096,tensor_parallel_size=8 \
233
+ --tasks mmlu \
234
+ --num_fewshot 5 \
235
+ --batch_size auto
236
+ ```
237
+
238
+ #### ARC-Challenge
239
+ ```
240
+ lm_eval \
241
+ --model vllm \
242
+ --model_args pretrained="neuralmagic/Meta-Llama-3.1-405B-Instruct-FP8-dynamic",dtype=auto,add_bos_token=True,max_model_len=4096,tensor_parallel_size=8 \
243
+ --tasks arc_challenge_llama_3.1_instruct \
244
+ --apply_chat_template \
245
+ --num_fewshot 0 \
246
+ --batch_size auto
247
+ ```
248
+
249
+ #### GSM-8K
250
+ ```
251
+ lm_eval \
252
+ --model vllm \
253
+ --model_args pretrained="neuralmagic/Meta-Llama-3.1-405B-Instruct-FP8-dynamic",dtype=auto,add_bos_token=True,max_model_len=4096,tensor_parallel_size=8 \
254
+ --tasks gsm8k_cot_llama_3.1_instruct \
255
+ --apply_chat_template \
256
+ --num_fewshot 8 \
257
+ --batch_size auto
258
+ ```
259
+
260
+ #### Hellaswag
261
+ ```
262
+ lm_eval \
263
+ --model vllm \
264
+ --model_args pretrained="neuralmagic/Meta-Llama-3.1-405B-Instruct-FP8-dynamic",dtype=auto,add_bos_token=True,max_model_len=4096,tensor_parallel_size=8 \
265
+ --tasks hellaswag \
266
+ --num_fewshot 10 \
267
+ --batch_size auto
268
+ ```
269
+
270
+ #### Winogrande
271
+ ```
272
+ lm_eval \
273
+ --model vllm \
274
+ --model_args pretrained="neuralmagic/Meta-Llama-3.1-405B-Instruct-FP8-dynamic",dtype=auto,add_bos_token=True,max_model_len=4096,tensor_parallel_size=8 \
275
+ --tasks winogrande \
276
+ --num_fewshot 5 \
277
+ --batch_size auto
278
+ ```
279
+
280
+ #### Hellaswag
281
+ ```
282
+ lm_eval \
283
+ --model vllm \
284
+ --model_args pretrained="neuralmagic/Meta-Llama-3.1-405B-Instruct-FP8-dynamic",dtype=auto,add_bos_token=True,max_model_len=4096,tensor_parallel_size=8 \
285
+ --tasks truthfulqa_mc \
286
+ --num_fewshot 0 \
287
+ --batch_size auto
288
+ ```