sridharps2 commited on
Commit
389cc8c
1 Parent(s): e727e1f

Model save

Browse files
Files changed (1) hide show
  1. README.md +8 -4
README.md CHANGED
@@ -1,5 +1,8 @@
1
  ---
 
2
  tags:
 
 
3
  - generated_from_trainer
4
  model-index:
5
  - name: llama-2-7b-chat-guanaco
@@ -31,14 +34,15 @@ More information needed
31
 
32
  The following hyperparameters were used during training:
33
  - learning_rate: 0.0002
34
- - train_batch_size: 8
35
  - eval_batch_size: 8
36
  - seed: 42
37
  - gradient_accumulation_steps: 4
38
- - total_train_batch_size: 32
39
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
40
  - lr_scheduler_type: linear
41
  - training_steps: 10
 
42
 
43
  ### Training results
44
 
@@ -46,7 +50,7 @@ The following hyperparameters were used during training:
46
 
47
  ### Framework versions
48
 
49
- - Transformers 4.30.2
50
  - Pytorch 2.1.0+cu121
51
  - Datasets 2.16.1
52
- - Tokenizers 0.13.3
 
1
  ---
2
+ base_model: meta-llama/Llama-2-7b-hf
3
  tags:
4
+ - trl
5
+ - sft
6
  - generated_from_trainer
7
  model-index:
8
  - name: llama-2-7b-chat-guanaco
 
34
 
35
  The following hyperparameters were used during training:
36
  - learning_rate: 0.0002
37
+ - train_batch_size: 2
38
  - eval_batch_size: 8
39
  - seed: 42
40
  - gradient_accumulation_steps: 4
41
+ - total_train_batch_size: 8
42
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
43
  - lr_scheduler_type: linear
44
  - training_steps: 10
45
+ - mixed_precision_training: Native AMP
46
 
47
  ### Training results
48
 
 
50
 
51
  ### Framework versions
52
 
53
+ - Transformers 4.35.2
54
  - Pytorch 2.1.0+cu121
55
  - Datasets 2.16.1
56
+ - Tokenizers 0.15.0