File size: 1,398 Bytes
ba5bd35 1c8b931 e28813b 3e5d016 e28813b b669fd0 9c59ad2 6ec6000 2dece0b b669fd0 2dece0b 3ecc7de 2dece0b b669fd0 cca530b 3ecc7de 2dece0b b669fd0 f2785e9 b669fd0 2dece0b 50164b6 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 |
---
license: llama2
base_model: beomi/llama-2-ko-7b
inference: false
datasets:
- Ash-Hun/Welfare-QA
library_name: peft
pipeline_tag: text-generation
tags:
- torch
- llama2
- domain-specific-lm
---
<div align='center'>
<img src="https://cdn-uploads.huggingface.co/production/uploads/6370a4e53d1bd47a4ebc2120/TQSWE0e3dAO_Ksbb8b5Xd.png" width='45%'/>
<h1>"WelSSiSKo : Welfare Domain Specific Model"</h1>
</div>
---
# Github โผ
> If you want to get how to use this model, please check my github repository :)
๐ [Github Repo](https://github.com/ash-hun/WelSSISKo)
[![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/ash-hun/WelSSISKo/blob/main/WelSSiSKo_Inference.ipynb)
# What is BaseModel โผ
> ๐ [beomi/llama-2-ko-7b](https://huggingface.co/beomi/llama-2-ko-7b)
# Training procedure โผ
The following `bitsandbytes` quantization config was used during training:
- **load_in_4bit**: True
- **bnb_4bit_quant_type**: nf4
- **bnb_4bit_use_double_quant**: False
- **bnb_4bit_compute_dtype**: float16
# Framework versions โผ
- PEFT 0.8.2.
# Evaluate Score
- ์ ์ ํ Domain Benchmark Set์ด ์๊ธฐ๋๋ฌธ์ ์ ์ฑํ๊ฐ๋ฅผ ์งํํ์๊ณ ๊ทธ์ ๋ฐ๋ฅธ **AVG Score๋ 0.74** ์
๋๋ค.
![image/png](https://cdn-uploads.huggingface.co/production/uploads/6370a4e53d1bd47a4ebc2120/HwIKWCJb3bT2pk_tP70e0.png)
|