gultar commited on
Commit
38e6d26
1 Parent(s): 9931459

End of training

Browse files
Files changed (2) hide show
  1. README.md +187 -0
  2. adapter_model.bin +3 -0
README.md CHANGED
@@ -1,3 +1,190 @@
1
  ---
2
  license: apache-2.0
 
 
 
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: apache-2.0
3
+ library_name: peft
4
+ tags:
5
+ - axolotl
6
+ - generated_from_trainer
7
+ base_model: GeneZC/MiniChat-2-3B
8
+ model-index:
9
+ - name: Automata-Minichat-3b
10
+ results: []
11
  ---
12
+
13
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
14
+ should probably proofread and complete it, then remove this comment. -->
15
+
16
+ [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
17
+ <details><summary>See axolotl config</summary>
18
+
19
+ axolotl version: `0.4.0`
20
+ ```yaml
21
+ base_model: GeneZC/MiniChat-2-3B
22
+ base_model_config: GeneZC/MiniChat-2-3B
23
+ model_type: LlamaForCausalLM
24
+ tokenizer_type: LlamaTokenizer
25
+ is_llama_derived_model: true
26
+
27
+ load_in_8bit: false
28
+ load_in_4bit: true
29
+ strict: false
30
+
31
+ datasets:
32
+ - path: THUDM/AgentInstruct
33
+ type: sharegpt
34
+ conversation: llama-2
35
+ split: os
36
+ - path: THUDM/AgentInstruct
37
+ type: sharegpt
38
+ conversation: llama-2
39
+ split: db
40
+ - path: THUDM/AgentInstruct
41
+ type: sharegpt
42
+ conversation: llama-2
43
+ split: alfworld
44
+ - path: THUDM/AgentInstruct
45
+ type: sharegpt
46
+ conversation: llama-2
47
+ split: webshop
48
+ - path: THUDM/AgentInstruct
49
+ type: sharegpt
50
+ conversation: llama-2
51
+ split: kg
52
+ - path: THUDM/AgentInstruct
53
+ type: sharegpt
54
+ conversation: llama-2
55
+ split: mind2web
56
+
57
+ dataset_prepared_path: last_run_prepared
58
+ val_set_size: 0.1
59
+ output_dir: ./qlora-out
60
+
61
+ hub_model_id: gultar/Automata-Minichat-3b
62
+
63
+ wandb_project: "Mistral-Agent"
64
+ wandb_log_model: "checkpoint"
65
+
66
+ chat_template: inst
67
+
68
+ adapter: qlora
69
+ lora_model_dir:
70
+
71
+ sequence_len: 4096
72
+ sample_packing: true
73
+ pad_to_sequence_len: true
74
+
75
+ lora_r: 8
76
+ lora_alpha: 16
77
+ lora_dropout: 0.05
78
+ lora_target_linear: true
79
+ lora_fan_in_fan_out:
80
+ lora_target_modules:
81
+ - gate_proj
82
+ - down_proj
83
+ - up_proj
84
+ - q_proj
85
+ - v_proj
86
+ - k_proj
87
+ - o_proj
88
+
89
+ wandb_project:
90
+ wandb_entity:
91
+ wandb_watch:
92
+ wandb_name:
93
+ wandb_log_model:
94
+
95
+ gradient_accumulation_steps: 4
96
+ micro_batch_size: 2
97
+ num_epochs: 1
98
+ optimizer: adamw_bnb_8bit
99
+ lr_scheduler: cosine
100
+ learning_rate: 0.0002
101
+
102
+ train_on_inputs: false
103
+ group_by_length: false
104
+ bf16: true
105
+ fp16: false
106
+ tf32: false
107
+
108
+ gradient_checkpointing: true
109
+ early_stopping_patience:
110
+ resume_from_checkpoint:
111
+ local_rank:
112
+ logging_steps: 1
113
+ xformers_attention:
114
+ flash_attention: true
115
+
116
+ loss_watchdog_threshold: 5.0
117
+ loss_watchdog_patience: 3
118
+
119
+ warmup_steps: 10
120
+ evals_per_epoch: 4
121
+ eval_table_size:
122
+ eval_table_max_new_tokens: 128
123
+ saves_per_epoch: 1
124
+ debug:
125
+ deepspeed:
126
+ weight_decay: 0.0
127
+ fsdp:
128
+ fsdp_config:
129
+ special_tokens:
130
+ bos_token: "<s>"
131
+ eos_token: "</s>"
132
+ unk_token: "<unk>"
133
+
134
+
135
+
136
+ ```
137
+
138
+ </details><br>
139
+
140
+ # Automata-Minichat-3b
141
+
142
+ This model is a fine-tuned version of [GeneZC/MiniChat-2-3B](https://huggingface.co/GeneZC/MiniChat-2-3B) on the None dataset.
143
+ It achieves the following results on the evaluation set:
144
+ - Loss: 0.3139
145
+
146
+ ## Model description
147
+
148
+ More information needed
149
+
150
+ ## Intended uses & limitations
151
+
152
+ More information needed
153
+
154
+ ## Training and evaluation data
155
+
156
+ More information needed
157
+
158
+ ## Training procedure
159
+
160
+ ### Training hyperparameters
161
+
162
+ The following hyperparameters were used during training:
163
+ - learning_rate: 0.0002
164
+ - train_batch_size: 2
165
+ - eval_batch_size: 2
166
+ - seed: 42
167
+ - gradient_accumulation_steps: 4
168
+ - total_train_batch_size: 8
169
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
170
+ - lr_scheduler_type: cosine
171
+ - lr_scheduler_warmup_steps: 10
172
+ - num_epochs: 1
173
+
174
+ ### Training results
175
+
176
+ | Training Loss | Epoch | Step | Validation Loss |
177
+ |:-------------:|:-----:|:----:|:---------------:|
178
+ | 0.9648 | 0.01 | 1 | 0.9762 |
179
+ | 0.5564 | 0.26 | 19 | 0.5018 |
180
+ | 0.2629 | 0.52 | 38 | 0.3400 |
181
+ | 0.2789 | 0.78 | 57 | 0.3139 |
182
+
183
+
184
+ ### Framework versions
185
+
186
+ - PEFT 0.8.2.dev0
187
+ - Transformers 4.37.0
188
+ - Pytorch 2.0.1+cu118
189
+ - Datasets 2.16.1
190
+ - Tokenizers 0.15.0
adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9db77f3519b2476b40ea9fcce6a9fe2371077b53fcc087a8fba34db5c34096eb
3
+ size 44947085