metadata
license: other
license_name: tongyi-qianwen
license_link: https://huggingface.co/Qwen/Qwen1.5-7B-Chat/blob/main/LICENSE
datasets:
- REILX/extracted_tagengo_gpt4
- TigerResearch/sft_zh
- alexl83/AlpacaDataCleaned
- LooksJuicy/ruozhiba
- silk-road/alpaca-data-gpt4-chinese
- databricks/databricks-dolly-15k
- microsoft/orca-math-word-problems-200k
- Sao10K/Claude-3-Opus-Instruct-5K
language:
- zh
- en
数据集
使用以下8个数据集 对Qwen1.5-7B-Chat进行微调并测试,结果显示,微调后的模型在CEVAL和MMLU的评分上均有所提升。
基础模型:
训练工具
https://github.com/hiyouga/LLaMA-Factory
测评方式:
使用opencompass(https://github.com/open-compass/OpenCompass/ ), 测试工具基于CEval和MMLU对微调之后的模型和原始模型进行测试。
测试模型分别为:
- Qwen1.5-7B-Chat
- Qwen1.5-7B-Chat-750Mb-lora,使用8DataSets数据集对Qwen1.5-7B-Chat模型进行sft方式lora微调
8DataSets数据集:
大约750Mb的微调数据集
- https://huggingface.co/datasets/REILX/extracted_tagengo_gpt4
- https://huggingface.co/datasets/TigerResearch/sft_zh
- https://huggingface.co/datasets/silk-road/alpaca-data-gpt4-chinese
- https://huggingface.co/datasets/LooksJuicy/ruozhiba
- https://huggingface.co/datasets/microsoft/orca-math-word-problems-200k
- https://huggingface.co/datasets/alexl83/AlpacaDataCleaned
- https://huggingface.co/datasets/Sao10K/Claude-3-Opus-Instruct-5K
结果
模型名称 | CEVAL | MMLU |
---|---|---|
Qwen1.5-7B-Chat | 68.61 | 61.56 |
Qwen1.5-7B-Chat-750Mb-lora | 71.36 | 61.78 |
License
This project utilizes certain datasets and checkpoints that are subject to their respective original licenses. Users must comply with all terms and conditions of these original licenses. The content of this project itself is licensed under the Apache license 2.0.