ohashi56225 commited on
Commit
afc4c53
1 Parent(s): 7be8b25

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +25 -6
README.md CHANGED
@@ -13,11 +13,30 @@ This model is an instruction-tuned version of the [microsoft/phi-2](https://hugg
13
 
14
  In the training, full parameter fine-tuning of phi-2 was performed, and LoRA was not used.
15
 
 
 
 
 
 
 
 
 
 
 
 
 
 
16
  ## Training
17
  - GPUs: 8 × A6000 48GB
18
- - per_device_train_batch_size 8
19
- - gradient_accumulation_steps 8
20
- - per_device_eval_batch_size 8
21
- - num_train_epochs 3
22
- - learning_rate 2e-5
23
- - warmup_ratio 0.03
 
 
 
 
 
 
 
13
 
14
  In the training, full parameter fine-tuning of phi-2 was performed, and LoRA was not used.
15
 
16
+ ## Text Format
17
+ ```
18
+ Below is an instruction that describes a task. Write a response that appropriately completes the request.
19
+ ### Instruction:
20
+ Based on the information provided, rewrite the sentence by changing its tense from past to future.
21
+
22
+ ### Input:
23
+ She played the piano beautifully for hours and then stopped as it was midnight.
24
+
25
+ ### Response:
26
+ She will play the piano beautifully for hours and then stop as it will be midnight.
27
+ ```
28
+
29
  ## Training
30
  - GPUs: 8 × A6000 48GB
31
+ - per_device_train_batch_size: 8
32
+ - gradient_accumulation_steps: 8
33
+ - per_device_eval_batch_size: 8
34
+ - num_train_epochs: 3
35
+ - learning_rate: 2e-5
36
+ - warmup_ratio: 0.03
37
+
38
+ ## Software
39
+ - pytorch: 2.1.2
40
+ - transformers: 4.38.0.dev0
41
+ - accelerate: 0.26.1
42
+ - deepspeed: 0.13.1