--- datasets: - ML4SE2023-G1-WizardCoder/EvolInstruct-SCoT-1k language: - en tags: - code --- # ML4SE23_G1_WizardCoder-SCoT-1B-V1.0 IN4334 ML4SE Group1 WizardCoder This model is the result of the fine-tunign of the WizardCoder-1B-V1.0 model using Structured Chain-of-Though (S-CoT) enhanced instructions. S-CoT is used to enhance a sample of about 1200 entries from the Evol-Instruct 80k dataset. The resulting dataset is then used for the training task. The current WizardCoder model and the new S-CoT fine-tuned one are compared on both versions of HumanEval and MBPP (S-CoT enhanced and not) on the pass@1 metric. The S-CoT enhancement of the evaluation datasets allows to study its effect when used just as a prompting technique, independently of the S-CoT fine-tuning of the model. ## Fine-tuning Details | Hyperparameter | [WizardCoder-1B-V1.0](https://huggingface.co/WizardLM/WizardCoder-1B-V1.0) | |----------------|---------------------| | Batch size | 16 | | Learning rate | 2e-5 | | Epochs | 3 | | Max length | 2048 | | Warmup step | 30 | | LR scheduler | cosine | | Dataset | [ML4SE23_G1_EvolInstruct-SCoT-1k](https://huggingface.co/datasets/ML4SE2023-G1-WizardCoder/ML4SE23_G1_EvolInstruct-SCoT-1k) | The hardware consisted on a GPU instance rented from [DataCrunch](https://datacrunch.io/) with the following specifications: | NVidia RTX A6000 48GB 1A6000.10V | |----------------------------------| | 2 GPUs | | 48GB VRAM per GPU | | 60 GB RAM | | 10 CPUs | | 100GB SSD Storage | | Ubuntu 20.04 | | CUDA 11.6 | ## Results Results of pass@1(%) on HumanEval and MBPP compared to HumanEval-SCoT and MBPP-SCoT using WizardCoder-1B, WizardCoder-SCoT-1B and WizardCoder-15B. | **Dataset** | **WizardCoder-1B-V1.0** | **WizardCoder-SCoT-1B-V1.0** | **WizardCoder-15B-V1.0** | |----------------|-------------------------|------------------------------|--------------------------| | HumanEval | 23.78 | **17.68** | 57.3 | | HumanEval-SCoT | **44.51** | **27.44** | **57.3** | | MBPP | 23.4 | **19.4** | 51.8 | | MBPP-SCoT | **40** | **28** | **45.6** |