File size: 961 Bytes
0c875c4
 
 
 
 
 
 
d912daa
0c875c4
 
 
454e3ac
 
e359677
 
 
 
 
 
0c3629b
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
---
license: apache-2.0
datasets:
- mlabonne/CodeLlama-2-20k
language:
- zh
---
# Chinese-CodeLlama-7B-SFT-V1

We implemented SFT based on our [Chinese-CodeLlama-7B-PT](https://huggingface.co/frankminors123/Chinese-CodeLlama-7B-PT). The dataset comes from [CodeLlama-2-20k](https://huggingface.co/datasets/mlabonne/CodeLlama-2-20k), we used Google Translate to translate it into Chinese. 

In addition, we designed appropriate Chinese prompt template for coding tasks, and during the fine-tuning stage, `memory efficient attention` was applied which save us a lot of GPU memory space.

The Chinese prompt template used is as follows:
```python
PROMPT_TEMPLATE = (
  "下面是描述一项任务的指令,并且与一则输入配对用来提供更多的上下文。请给出尽可能满足请求的回答.\n"
  "### 指令:\n{instruction}\n### 输入:\n{input}\n### 回答:\n"
)
```

If you are interested in our work, please follow our progress in the future.