llama_2_alpaca_midset_helpful
This model is a fine-tuned version of meta-llama/Llama-2-7b-chat-hf on the None dataset. It achieves the following results on the evaluation set:
- Loss: 0.9658
Model description
More information needed
Intended uses & limitations
More information needed
Training and evaluation data
More information needed
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0002
- train_batch_size: 4
- eval_batch_size: 4
- seed: 42
- distributed_type: multi-GPU
- gradient_accumulation_steps: 2
- total_train_batch_size: 8
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine
- lr_scheduler_warmup_ratio: 0.1
- training_steps: 200
Training results
Training Loss | Epoch | Step | Validation Loss |
---|---|---|---|
1.7598 | 0.04 | 5 | 2.0204 |
1.8495 | 0.08 | 10 | 1.6179 |
1.3218 | 0.12 | 15 | 1.3287 |
1.1491 | 0.16 | 20 | 1.2307 |
1.0472 | 0.2 | 25 | 1.1143 |
0.8803 | 0.24 | 30 | 1.0921 |
0.9142 | 0.28 | 35 | 1.0660 |
0.9876 | 0.32 | 40 | 1.0418 |
0.9415 | 0.36 | 45 | 1.0230 |
0.8803 | 0.4 | 50 | 1.0050 |
0.8597 | 0.44 | 55 | 0.9965 |
0.912 | 0.48 | 60 | 0.9914 |
0.8499 | 0.52 | 65 | 0.9914 |
0.992 | 0.56 | 70 | 0.9915 |
0.9303 | 0.6 | 75 | 0.9914 |
0.8614 | 0.64 | 80 | 0.9887 |
0.8837 | 0.68 | 85 | 0.9819 |
0.8652 | 0.72 | 90 | 0.9804 |
0.9088 | 0.76 | 95 | 0.9782 |
0.875 | 0.8 | 100 | 0.9788 |
0.8589 | 0.84 | 105 | 0.9839 |
0.873 | 0.88 | 110 | 0.9817 |
0.9061 | 0.92 | 115 | 0.9767 |
0.8949 | 0.96 | 120 | 0.9719 |
0.8598 | 1.0 | 125 | 0.9682 |
0.8137 | 1.04 | 130 | 0.9683 |
0.7499 | 1.08 | 135 | 0.9730 |
0.7795 | 1.12 | 140 | 0.9758 |
0.7499 | 1.16 | 145 | 0.9743 |
0.7976 | 1.2 | 150 | 0.9712 |
0.7615 | 1.24 | 155 | 0.9691 |
0.7679 | 1.28 | 160 | 0.9683 |
0.7562 | 1.32 | 165 | 0.9682 |
0.7605 | 1.3600 | 170 | 0.9673 |
0.7964 | 1.4 | 175 | 0.9668 |
0.7692 | 1.44 | 180 | 0.9660 |
0.7825 | 1.48 | 185 | 0.9655 |
0.7905 | 1.52 | 190 | 0.9653 |
0.7625 | 1.56 | 195 | 0.9658 |
0.8549 | 1.6 | 200 | 0.9658 |
Framework versions
- PEFT 0.12.0
- Transformers 4.44.2
- Pytorch 2.4.1+cu121
- Datasets 3.0.0
- Tokenizers 0.19.1
- Downloads last month
- 4
Model tree for CharlesLi/llama_2_alpaca_midset_helpful
Base model
meta-llama/Llama-2-7b-chat-hf