AdamGrzesik commited on
Commit
0b28f24
1 Parent(s): 900a6af

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -28
README.md CHANGED
@@ -3,7 +3,7 @@ base_model: alpindale/Mistral-7B-v0.2-hf
3
  tags:
4
  - generated_from_trainer
5
  model-index:
6
- - name: workspace/dolphin-2.8-mistral-7b
7
  results: []
8
  ---
9
 
@@ -33,7 +33,7 @@ chat_template: chatml
33
 
34
  dataset_prepared_path: last_run_prepared
35
  val_set_size: 0.001
36
- output_dir: /workspace/dolphin-2.8-mistral-7b
37
 
38
  sequence_len: 16384
39
  sample_packing: true
@@ -94,25 +94,12 @@ tokens:
94
 
95
  </details><br>
96
 
97
- # workspace/dolphin-2.8-mistral-7b
98
 
99
  This model is a fine-tuned version of [alpindale/Mistral-7B-v0.2-hf](https://huggingface.co/alpindale/Mistral-7B-v0.2-hf) on the None dataset.
100
- It achieves the following results on the evaluation set:
101
- - Loss: 3.8281
102
 
103
- ## Model description
104
 
105
- More information needed
106
 
107
- ## Intended uses & limitations
108
-
109
- More information needed
110
-
111
- ## Training and evaluation data
112
-
113
- More information needed
114
-
115
- ## Training procedure
116
 
117
  ### Training hyperparameters
118
 
@@ -131,16 +118,4 @@ The following hyperparameters were used during training:
131
  - lr_scheduler_warmup_steps: 10
132
  - num_epochs: 4
133
 
134
- ### Training results
135
-
136
- | Training Loss | Epoch | Step | Validation Loss |
137
- |:-------------:|:-----:|:----:|:---------------:|
138
- | 3.6131 | 0.09 | 1 | 3.8281 |
139
-
140
-
141
- ### Framework versions
142
 
143
- - Transformers 4.40.0.dev0
144
- - Pytorch 2.2.0+cu121
145
- - Datasets 2.18.0
146
- - Tokenizers 0.15.0
 
3
  tags:
4
  - generated_from_trainer
5
  model-index:
6
+ - name: AdamGrzesik/Samantha-PL-AG-Mistral-7B-v0.2
7
  results: []
8
  ---
9
 
 
33
 
34
  dataset_prepared_path: last_run_prepared
35
  val_set_size: 0.001
36
+ output_dir: /workspace/Samantha
37
 
38
  sequence_len: 16384
39
  sample_packing: true
 
94
 
95
  </details><br>
96
 
97
+ # AdamGrzesik/Samantha-PL-AG-Mistral-7B-v0.2
98
 
99
  This model is a fine-tuned version of [alpindale/Mistral-7B-v0.2-hf](https://huggingface.co/alpindale/Mistral-7B-v0.2-hf) on the None dataset.
 
 
100
 
 
101
 
 
102
 
 
 
 
 
 
 
 
 
 
103
 
104
  ### Training hyperparameters
105
 
 
118
  - lr_scheduler_warmup_steps: 10
119
  - num_epochs: 4
120
 
 
 
 
 
 
 
 
 
121