geeknix commited on
Commit
54a6ecb
1 Parent(s): 8222b52

geeknix/geeknix_mistral_instruct_test_final

Browse files
README.md CHANGED
@@ -1,6 +1,6 @@
1
  ---
2
  license: apache-2.0
3
- library_name: transformers
4
  tags:
5
  - trl
6
  - sft
@@ -11,7 +11,6 @@ datasets:
11
  model-index:
12
  - name: mistral_instruct_generation
13
  results: []
14
- pipeline_tag: text-generation
15
  ---
16
 
17
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -21,7 +20,7 @@ should probably proofread and complete it, then remove this comment. -->
21
 
22
  This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.3](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.3) on the generator dataset.
23
  It achieves the following results on the evaluation set:
24
- - Loss: 0.7042
25
 
26
  ## Model description
27
 
@@ -41,7 +40,7 @@ More information needed
41
 
42
  The following hyperparameters were used during training:
43
  - learning_rate: 0.0002
44
- - train_batch_size: 4
45
  - eval_batch_size: 8
46
  - seed: 42
47
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
@@ -51,15 +50,15 @@ The following hyperparameters were used during training:
51
 
52
  ### Training results
53
 
54
- | Training Loss | Epoch | Step | Validation Loss |
55
- |:-------------:|:------:|:----:|:---------------:|
56
- | 0.9489 | 0.0524 | 11 | 0.7042 |
57
 
58
 
59
  ### Framework versions
60
 
61
  - PEFT 0.11.1
62
  - Transformers 4.41.1
63
- - Pytorch 2.1.2
64
  - Datasets 2.19.1
65
  - Tokenizers 0.19.1
 
1
  ---
2
  license: apache-2.0
3
+ library_name: peft
4
  tags:
5
  - trl
6
  - sft
 
11
  model-index:
12
  - name: mistral_instruct_generation
13
  results: []
 
14
  ---
15
 
16
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
20
 
21
  This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.3](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.3) on the generator dataset.
22
  It achieves the following results on the evaluation set:
23
+ - Loss: 0.4619
24
 
25
  ## Model description
26
 
 
40
 
41
  The following hyperparameters were used during training:
42
  - learning_rate: 0.0002
43
+ - train_batch_size: 2
44
  - eval_batch_size: 8
45
  - seed: 42
46
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
 
50
 
51
  ### Training results
52
 
53
+ | Training Loss | Epoch | Step | Validation Loss |
54
+ |:-------------:|:-----:|:----:|:---------------:|
55
+ | 0.5925 | 0.05 | 42 | 0.4619 |
56
 
57
 
58
  ### Framework versions
59
 
60
  - PEFT 0.11.1
61
  - Transformers 4.41.1
62
+ - Pytorch 2.3.0+cu121
63
  - Datasets 2.19.1
64
  - Tokenizers 0.19.1
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c29836be3414ac0f4bfec343923b802b8aa66da7eba7e7a7af6819bcb06e940e
3
  size 109069176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:44be91ee08e59ffe1ae92aee623b0e99f497350bec862f1e3a1f1ee1c04493ce
3
  size 109069176
runs/May31_23-54-15_25b160f765e5/events.out.tfevents.1717199669.25b160f765e5.405.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ba89f110f5da62d82063df37a5b317052d36b036b64df92f2c8b1cff2399ec6f
3
+ size 7608
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5985da9aecb9b8ef82c07213f4a3e6d122a6b0bced8c1f6c192c323783b0e2ed
3
- size 5176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7fc879a3c1f0fc1f23cf073eafcf74103a7156163da678a3fec0f9059860712a
3
+ size 5112