tonyshark commited on
Commit
a81beac
·
verified ·
1 Parent(s): a5c01c3

End of training

Browse files
README.md CHANGED
@@ -1,9 +1,58 @@
1
  ---
2
- license: mit
3
- base_model:
4
- - deepseek-ai/DeepSeek-V3
5
- pipeline_tag: text-generation
6
  library_name: transformers
 
 
 
 
 
 
7
  ---
8
- # DeepSeek V3 1B Test
9
- This model is randomly initialized for testing implementations, it's **not** a trained model and it will only generate random tokens.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ base_model: tonyshark/deepdeek-v3-1b
 
 
 
3
  library_name: transformers
4
+ model_name: deepdeek-v3-1b
5
+ tags:
6
+ - generated_from_trainer
7
+ - trl
8
+ - sft
9
+ licence: license
10
  ---
11
+
12
+ # Model Card for deepdeek-v3-1b
13
+
14
+ This model is a fine-tuned version of [tonyshark/deepdeek-v3-1b](https://huggingface.co/tonyshark/deepdeek-v3-1b).
15
+ It has been trained using [TRL](https://github.com/huggingface/trl).
16
+
17
+ ## Quick start
18
+
19
+ ```python
20
+ from transformers import pipeline
21
+
22
+ question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
23
+ generator = pipeline("text-generation", model="tonyshark/deepdeek-v3-1b", device="cuda")
24
+ output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
25
+ print(output["generated_text"])
26
+ ```
27
+
28
+ ## Training procedure
29
+
30
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/tonyshark365-aixblock/huggingface/runs/fd6g1fre)
31
+
32
+
33
+ This model was trained with SFT.
34
+
35
+ ### Framework versions
36
+
37
+ - TRL: 0.13.0
38
+ - Transformers: 4.47.1
39
+ - Pytorch: 2.5.1+cu121
40
+ - Datasets: 3.2.0
41
+ - Tokenizers: 0.21.0
42
+
43
+ ## Citations
44
+
45
+
46
+
47
+ Cite TRL as:
48
+
49
+ ```bibtex
50
+ @misc{vonwerra2022trl,
51
+ title = {{TRL: Transformer Reinforcement Learning}},
52
+ author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallouédec},
53
+ year = 2020,
54
+ journal = {GitHub repository},
55
+ publisher = {GitHub},
56
+ howpublished = {\url{https://github.com/huggingface/trl}}
57
+ }
58
+ ```
config.json CHANGED
@@ -1,14 +1,14 @@
1
  {
2
- "_name_or_path": "deepseek-v3-1b",
3
  "architectures": [
4
  "DeepseekV3ForCausalLM"
5
  ],
6
  "attention_bias": false,
7
  "attention_dropout": 0.0,
8
  "auto_map": {
9
- "AutoConfig": "configuration_deepseek.DeepseekV3Config",
10
- "AutoModel": "modeling_deepseek.DeepseekV3Model",
11
- "AutoModelForCausalLM": "modeling_deepseek.DeepseekV3ForCausalLM"
12
  },
13
  "aux_loss_alpha": 0.001,
14
  "bos_token_id": 0,
@@ -54,7 +54,7 @@
54
  "tie_word_embeddings": false,
55
  "topk_group": 4,
56
  "topk_method": "noaux_tc",
57
- "torch_dtype": "bfloat16",
58
  "transformers_version": "4.47.1",
59
  "use_cache": true,
60
  "v_head_dim": 128,
 
1
  {
2
+ "_name_or_path": "tonyshark/deepdeek-v3-1b",
3
  "architectures": [
4
  "DeepseekV3ForCausalLM"
5
  ],
6
  "attention_bias": false,
7
  "attention_dropout": 0.0,
8
  "auto_map": {
9
+ "AutoConfig": "tonyshark/deepdeek-v3-1b--configuration_deepseek.DeepseekV3Config",
10
+ "AutoModel": "tonyshark/deepdeek-v3-1b--modeling_deepseek.DeepseekV3Model",
11
+ "AutoModelForCausalLM": "tonyshark/deepdeek-v3-1b--modeling_deepseek.DeepseekV3ForCausalLM"
12
  },
13
  "aux_loss_alpha": 0.001,
14
  "bos_token_id": 0,
 
54
  "tie_word_embeddings": false,
55
  "topk_group": 4,
56
  "topk_method": "noaux_tc",
57
+ "torch_dtype": "float32",
58
  "transformers_version": "4.47.1",
59
  "use_cache": true,
60
  "v_head_dim": 128,
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a9db5309afa023828200507bbe04d1aaff8667510b47b24ac999f340876da1ee
3
- size 2099235336
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:98055ce78ca046c4d239892034abb11415818575dff3fccd3effe00d4bd8067e
3
+ size 4198331024
runs/Jan06_14-42-19_568fc3dd2737/events.out.tfevents.1736174543.568fc3dd2737.2792.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6a7bfc97850606cfd2502d6e5c08af8c2744ea1d1f27c56e4c42aa5c8578fb6b
3
+ size 6390
runs/Jan06_15-00-00_568fc3dd2737/events.out.tfevents.1736175601.568fc3dd2737.2792.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:965874718edfee6555923e5222f0a60f41d85d93ae0e55e74ecfc0f33672aa01
3
+ size 6390
runs/Jan06_15-00-57_568fc3dd2737/events.out.tfevents.1736175658.568fc3dd2737.2792.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c24676654cbafecb33796eaa7cbe8fccda6d1dcdf42b5e25a402bb1c961d03ef
3
+ size 6390
runs/Jan06_15-01-12_568fc3dd2737/events.out.tfevents.1736175673.568fc3dd2737.2792.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:269299b06f43c7be46a8b3211e95c4859e0d2a6cf34784ed6381c453b3c77f70
3
+ size 6390
runs/Jan06_15-05-48_568fc3dd2737/events.out.tfevents.1736175949.568fc3dd2737.2792.4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3f2e5c25c9d5f2e6ba1ab78cb7aeb9394e342df03d94eca47813d62466ffd202
3
+ size 12519
runs/Jan06_15-28-47_568fc3dd2737/events.out.tfevents.1736177328.568fc3dd2737.2792.5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c362b0ff8146487b3771ced01c8db20e0ccdb8dfa4b60b590cd964d55de79e74
3
+ size 12519
tokenizer.json CHANGED
@@ -7419,6 +7419,12 @@
7419
  "id": "A",
7420
  "type_id": 0
7421
  }
 
 
 
 
 
 
7422
  }
7423
  ],
7424
  "pair": [
@@ -7434,6 +7440,12 @@
7434
  "type_id": 0
7435
  }
7436
  },
 
 
 
 
 
 
7437
  {
7438
  "SpecialToken": {
7439
  "id": "<|begin▁of▁sentence|>",
@@ -7445,6 +7457,12 @@
7445
  "id": "B",
7446
  "type_id": 1
7447
  }
 
 
 
 
 
 
7448
  }
7449
  ],
7450
  "special_tokens": {
@@ -7456,6 +7474,15 @@
7456
  "tokens": [
7457
  "<|begin▁of▁sentence|>"
7458
  ]
 
 
 
 
 
 
 
 
 
7459
  }
7460
  }
7461
  },
 
7419
  "id": "A",
7420
  "type_id": 0
7421
  }
7422
+ },
7423
+ {
7424
+ "SpecialToken": {
7425
+ "id": "<|end▁of▁sentence|>",
7426
+ "type_id": 0
7427
+ }
7428
  }
7429
  ],
7430
  "pair": [
 
7440
  "type_id": 0
7441
  }
7442
  },
7443
+ {
7444
+ "SpecialToken": {
7445
+ "id": "<|end▁of▁sentence|>",
7446
+ "type_id": 0
7447
+ }
7448
+ },
7449
  {
7450
  "SpecialToken": {
7451
  "id": "<|begin▁of▁sentence|>",
 
7457
  "id": "B",
7458
  "type_id": 1
7459
  }
7460
+ },
7461
+ {
7462
+ "SpecialToken": {
7463
+ "id": "<|end▁of▁sentence|>",
7464
+ "type_id": 1
7465
+ }
7466
  }
7467
  ],
7468
  "special_tokens": {
 
7474
  "tokens": [
7475
  "<|begin▁of▁sentence|>"
7476
  ]
7477
+ },
7478
+ "<|end▁of▁sentence|>": {
7479
+ "id": "<|end▁of▁sentence|>",
7480
+ "ids": [
7481
+ 1
7482
+ ],
7483
+ "tokens": [
7484
+ "<|end▁of▁sentence|>"
7485
+ ]
7486
  }
7487
  }
7488
  },
tokenizer_config.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "add_bos_token": true,
3
- "add_eos_token": false,
4
  "add_prefix_space": null,
5
  "added_tokens_decoder": {
6
  "0": {
 
1
  {
2
  "add_bos_token": true,
3
+ "add_eos_token": true,
4
  "add_prefix_space": null,
5
  "added_tokens_decoder": {
6
  "0": {
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0480284713e598a4a5c2a8cea0dfb8e429e3e644cefc1bb93573453da3a97573
3
+ size 5560