prithivMLmods commited on
Commit
13d4bae
·
verified ·
1 Parent(s): d11d64b

Adding Evaluation Results (#3)

Browse files

- Adding Evaluation Results (2b7b57fa8f674c5c6d9b1afe7d917bd6dbbd86bc)

Files changed (1) hide show
  1. README.md +114 -1
README.md CHANGED
@@ -16,6 +16,105 @@ tags:
16
  - text-generation-inference
17
  - phi-4
18
  - qwq
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
19
  ---
20
  ![zsdfvdsfvasdfvsdrf.png](https://cdn-uploads.huggingface.co/production/uploads/65bb837dbfb878f46c77de4c/vchJIg-Y05Pb7AmCDgCkm.png)
21
 
@@ -112,4 +211,18 @@ While phi-4 o1 ft is a powerful and versatile model, it has certain limitations
112
  - The model should not be used for malicious purposes, such as generating harmful content, misinformation, or spam. Users are responsible for ensuring ethical use.
113
 
114
  7. **Domain-Specific Limitations**:
115
- - While the model performs well on general-purpose tasks, it may lack depth in highly specialized domains (e.g., medical, legal, or financial fields) without additional fine-tuning.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
16
  - text-generation-inference
17
  - phi-4
18
  - qwq
19
+ model-index:
20
+ - name: Phi-4-o1
21
+ results:
22
+ - task:
23
+ type: text-generation
24
+ name: Text Generation
25
+ dataset:
26
+ name: IFEval (0-Shot)
27
+ type: wis-k/instruction-following-eval
28
+ split: train
29
+ args:
30
+ num_few_shot: 0
31
+ metrics:
32
+ - type: inst_level_strict_acc and prompt_level_strict_acc
33
+ value: 2.9
34
+ name: averaged accuracy
35
+ source:
36
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=prithivMLmods%2FPhi-4-o1
37
+ name: Open LLM Leaderboard
38
+ - task:
39
+ type: text-generation
40
+ name: Text Generation
41
+ dataset:
42
+ name: BBH (3-Shot)
43
+ type: SaylorTwift/bbh
44
+ split: test
45
+ args:
46
+ num_few_shot: 3
47
+ metrics:
48
+ - type: acc_norm
49
+ value: 52.17
50
+ name: normalized accuracy
51
+ source:
52
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=prithivMLmods%2FPhi-4-o1
53
+ name: Open LLM Leaderboard
54
+ - task:
55
+ type: text-generation
56
+ name: Text Generation
57
+ dataset:
58
+ name: MATH Lvl 5 (4-Shot)
59
+ type: lighteval/MATH-Hard
60
+ split: test
61
+ args:
62
+ num_few_shot: 4
63
+ metrics:
64
+ - type: exact_match
65
+ value: 39.43
66
+ name: exact match
67
+ source:
68
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=prithivMLmods%2FPhi-4-o1
69
+ name: Open LLM Leaderboard
70
+ - task:
71
+ type: text-generation
72
+ name: Text Generation
73
+ dataset:
74
+ name: GPQA (0-shot)
75
+ type: Idavidrein/gpqa
76
+ split: train
77
+ args:
78
+ num_few_shot: 0
79
+ metrics:
80
+ - type: acc_norm
81
+ value: 17.67
82
+ name: acc_norm
83
+ source:
84
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=prithivMLmods%2FPhi-4-o1
85
+ name: Open LLM Leaderboard
86
+ - task:
87
+ type: text-generation
88
+ name: Text Generation
89
+ dataset:
90
+ name: MuSR (0-shot)
91
+ type: TAUR-Lab/MuSR
92
+ args:
93
+ num_few_shot: 0
94
+ metrics:
95
+ - type: acc_norm
96
+ value: 22.15
97
+ name: acc_norm
98
+ source:
99
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=prithivMLmods%2FPhi-4-o1
100
+ name: Open LLM Leaderboard
101
+ - task:
102
+ type: text-generation
103
+ name: Text Generation
104
+ dataset:
105
+ name: MMLU-PRO (5-shot)
106
+ type: TIGER-Lab/MMLU-Pro
107
+ config: main
108
+ split: test
109
+ args:
110
+ num_few_shot: 5
111
+ metrics:
112
+ - type: acc
113
+ value: 46.37
114
+ name: accuracy
115
+ source:
116
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=prithivMLmods%2FPhi-4-o1
117
+ name: Open LLM Leaderboard
118
  ---
119
  ![zsdfvdsfvasdfvsdrf.png](https://cdn-uploads.huggingface.co/production/uploads/65bb837dbfb878f46c77de4c/vchJIg-Y05Pb7AmCDgCkm.png)
120
 
 
211
  - The model should not be used for malicious purposes, such as generating harmful content, misinformation, or spam. Users are responsible for ensuring ethical use.
212
 
213
  7. **Domain-Specific Limitations**:
214
+ - While the model performs well on general-purpose tasks, it may lack depth in highly specialized domains (e.g., medical, legal, or financial fields) without additional fine-tuning.
215
+ # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard)
216
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/prithivMLmods__Phi-4-o1-details)!
217
+ Summarized results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/contents/viewer/default/train?q=prithivMLmods%2FPhi-4-o1&sort[column]=Average%20%E2%AC%86%EF%B8%8F&sort[direction]=desc)!
218
+
219
+ | Metric |Value (%)|
220
+ |-------------------|--------:|
221
+ |**Average** | 30.11|
222
+ |IFEval (0-Shot) | 2.90|
223
+ |BBH (3-Shot) | 52.17|
224
+ |MATH Lvl 5 (4-Shot)| 39.43|
225
+ |GPQA (0-shot) | 17.67|
226
+ |MuSR (0-shot) | 22.15|
227
+ |MMLU-PRO (5-shot) | 46.37|
228
+