Update src/display/about.py
Browse files- src/display/about.py +5 -6
src/display/about.py
CHANGED
@@ -38,7 +38,7 @@ We chose these benchmarks as they test a variety of reasoning and general knowle
|
|
38 |
|
39 |
---------------------------
|
40 |
|
41 |
-
##
|
42 |
To reproduce our results, here is the commands you can run, using [v0.4.2](https://github.com/EleutherAI/lm-evaluation-harness/tree/v0.4.2) of the Eleuther AI Harness:
|
43 |
```
|
44 |
python main.py --model=hf-causal-experimental
|
@@ -71,7 +71,7 @@ Side note on the baseline scores:
|
|
71 |
|
72 |
---------------------------
|
73 |
|
74 |
-
##
|
75 |
|
76 |
### Quantization
|
77 |
To get more information about quantization, see:
|
@@ -79,10 +79,9 @@ To get more information about quantization, see:
|
|
79 |
- 4 bits: [blog post](https://huggingface.co/blog/4bit-transformers-bitsandbytes), [paper](https://arxiv.org/abs/2305.14314)
|
80 |
|
81 |
### Other cool leaderboards:
|
|
|
82 |
- [LLM safety](https://huggingface.co/spaces/AI-Secure/llm-trustworthy-leaderboard)
|
83 |
- [LLM performance](https://huggingface.co/spaces/optimum/llm-perf-leaderboard)
|
84 |
-
- [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
|
85 |
-
|
86 |
|
87 |
"""
|
88 |
|
@@ -90,7 +89,7 @@ FAQ_TEXT = """
|
|
90 |
|
91 |
## SUBMISSIONS
|
92 |
My model requires `trust_remote_code=True`, can I submit it?
|
93 |
-
- *Yes,
|
94 |
|
95 |
How can I follow when my model is launched?
|
96 |
- *You can look for its request file [here](https://huggingface.co/datasets/Intel/ld_requests) and follow the status evolution, or directly in the queues above the submit form.*
|
@@ -182,7 +181,7 @@ The compute dtype will pass to `lm-eval` for the inference. Currently, we suppor
|
|
182 |
CITATION_BUTTON_LABEL = "Copy the following snippet to cite these results"
|
183 |
CITATION_BUTTON_TEXT = r"""
|
184 |
@software{auto-round,
|
185 |
-
title =
|
186 |
publisher = {Intel},
|
187 |
url = {https://github.com/intel/auto-round}
|
188 |
}
|
|
|
38 |
|
39 |
---------------------------
|
40 |
|
41 |
+
## RERODUCIBILITY
|
42 |
To reproduce our results, here is the commands you can run, using [v0.4.2](https://github.com/EleutherAI/lm-evaluation-harness/tree/v0.4.2) of the Eleuther AI Harness:
|
43 |
```
|
44 |
python main.py --model=hf-causal-experimental
|
|
|
71 |
|
72 |
---------------------------
|
73 |
|
74 |
+
## RESOURCES
|
75 |
|
76 |
### Quantization
|
77 |
To get more information about quantization, see:
|
|
|
79 |
- 4 bits: [blog post](https://huggingface.co/blog/4bit-transformers-bitsandbytes), [paper](https://arxiv.org/abs/2305.14314)
|
80 |
|
81 |
### Other cool leaderboards:
|
82 |
+
- [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
|
83 |
- [LLM safety](https://huggingface.co/spaces/AI-Secure/llm-trustworthy-leaderboard)
|
84 |
- [LLM performance](https://huggingface.co/spaces/optimum/llm-perf-leaderboard)
|
|
|
|
|
85 |
|
86 |
"""
|
87 |
|
|
|
89 |
|
90 |
## SUBMISSIONS
|
91 |
My model requires `trust_remote_code=True`, can I submit it?
|
92 |
+
- *Yes, the leaderboard supports the model with `trust_remote_code=True` well.*
|
93 |
|
94 |
How can I follow when my model is launched?
|
95 |
- *You can look for its request file [here](https://huggingface.co/datasets/Intel/ld_requests) and follow the status evolution, or directly in the queues above the submit form.*
|
|
|
181 |
CITATION_BUTTON_LABEL = "Copy the following snippet to cite these results"
|
182 |
CITATION_BUTTON_TEXT = r"""
|
183 |
@software{auto-round,
|
184 |
+
title = AutoRound,
|
185 |
publisher = {Intel},
|
186 |
url = {https://github.com/intel/auto-round}
|
187 |
}
|