|
The LLaMA based Pygmalion-7b model: |
|
|
|
https://huggingface.co/PygmalionAI/pygmalion-7b |
|
|
|
Merged alongside lmsys's Vicuna v1.1 deltas: |
|
|
|
https://huggingface.co/lmsys/vicuna-13b-delta-v1.1 |
|
|
|
This merge was done using an weighted average merge strategy, and the end result is a model composed of: |
|
|
|
Pygmalion-7b [60%] + LLaMA Vicuna v1.1 [40%] |
|
|
|
|
|
This was done under request, but the end result is intended to lean heavily towards Pygmalion's chatting + RP tendencies, and to inherit some of Vicuna's Assistant / Instruct / Helpful properties. |
|
|
|
Due to the influence of Pygmalion, this model will very likely generate content that is considered NSFW. |
|
|
|
The specific prompting is unknown, but try Pygmalion's prompt styles first, |
|
then a mix of the two to see what brings most interesting results. |
|
|
|
Treat this as a normal HF Transformers model. |
|
# [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard) |
|
Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_TehVenom__Pygmalion-Vicuna-1.1-7b) |
|
|
|
| Metric | Value | |
|
|-----------------------|---------------------------| |
|
| Avg. | 43.35 | |
|
| ARC (25-shot) | 52.82 | |
|
| HellaSwag (10-shot) | 78.66 | |
|
| MMLU (5-shot) | 43.61 | |
|
| TruthfulQA (0-shot) | 42.21 | |
|
| Winogrande (5-shot) | 71.98 | |
|
| GSM8K (5-shot) | 6.22 | |
|
| DROP (3-shot) | 7.97 | |
|
|