|
--- |
|
license: mit |
|
datasets: |
|
- yahma/alpaca-cleaned |
|
language: |
|
- en |
|
library_name: transformers |
|
pipeline_tag: text-generation |
|
--- |
|
|
|
# phi-1_5-alpaca-cleaned |
|
This model is an instruction-tuned version of the [microsoft/phi-1_5](https://huggingface.co/microsoft/phi-1_5) model fine-tuned on the [yahma/alpaca-cleaned](https://huggingface.co/datasets/yahma/alpaca-cleaned) dataset. |
|
|
|
In the training, full parameter fine-tuning of phi-2 was performed, and LoRA was not used. |
|
|
|
## Text Format |
|
``` |
|
Below is an instruction that describes a task. Write a response that appropriately completes the request. |
|
### Instruction: |
|
Based on the information provided, rewrite the sentence by changing its tense from past to future. |
|
|
|
### Input: |
|
She played the piano beautifully for hours and then stopped as it was midnight. |
|
|
|
### Response: |
|
She will play the piano beautifully for hours and then stop as it will be midnight. |
|
``` |
|
|
|
## Training |
|
- GPUs: 8 × A6000 48GB |
|
- per_device_train_batch_size: 8 |
|
- gradient_accumulation_steps: 8 |
|
- per_device_eval_batch_size: 8 |
|
- num_train_epochs: 3 |
|
- learning_rate: 2e-5 |
|
- warmup_ratio: 0.03 |
|
|
|
## Software |
|
- pytorch: 2.1.2 |
|
- transformers: 4.38.0.dev0 |
|
- accelerate: 0.26.1 |
|
- deepspeed: 0.13.1 |