ohashi56225 commited on
Commit
b9a8f7a
1 Parent(s): 3109a45

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +42 -0
README.md ADDED
@@ -0,0 +1,42 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ datasets:
4
+ - yahma/alpaca-cleaned
5
+ language:
6
+ - en
7
+ library_name: transformers
8
+ pipeline_tag: text-generation
9
+ ---
10
+
11
+ # phi-1_5-alpaca-cleaned
12
+ This model is an instruction-tuned version of the [microsoft/phi-1_5](https://huggingface.co/microsoft/phi-1_5) model fine-tuned on the [yahma/alpaca-cleaned](https://huggingface.co/datasets/yahma/alpaca-cleaned) dataset.
13
+
14
+ In the training, full parameter fine-tuning of phi-2 was performed, and LoRA was not used.
15
+
16
+ ## Text Format
17
+ ```
18
+ Below is an instruction that describes a task. Write a response that appropriately completes the request.
19
+ ### Instruction:
20
+ Based on the information provided, rewrite the sentence by changing its tense from past to future.
21
+
22
+ ### Input:
23
+ She played the piano beautifully for hours and then stopped as it was midnight.
24
+
25
+ ### Response:
26
+ She will play the piano beautifully for hours and then stop as it will be midnight.
27
+ ```
28
+
29
+ ## Training
30
+ - GPUs: 8 × A6000 48GB
31
+ - per_device_train_batch_size: 8
32
+ - gradient_accumulation_steps: 8
33
+ - per_device_eval_batch_size: 8
34
+ - num_train_epochs: 3
35
+ - learning_rate: 2e-5
36
+ - warmup_ratio: 0.03
37
+
38
+ ## Software
39
+ - pytorch: 2.1.2
40
+ - transformers: 4.38.0.dev0
41
+ - accelerate: 0.26.1
42
+ - deepspeed: 0.13.1