TheRamsay commited on
Commit
511ec8e
·
verified ·
1 Parent(s): 4257ac8

Training in progress, epoch 0

Browse files
config.json CHANGED
@@ -3,8 +3,8 @@
3
  "SpeechEncoderDecoderModel"
4
  ],
5
  "decoder": {
6
- "_name_or_path": "spital/gpt2-small-czech-cs",
7
- "activation_function": "gelu_new",
8
  "add_cross_attention": true,
9
  "architectures": [
10
  "GPT2LMHeadModel"
@@ -12,7 +12,7 @@
12
  "attn_pdrop": 0.1,
13
  "bad_words_ids": null,
14
  "begin_suppress_tokens": null,
15
- "bos_token_id": 50256,
16
  "chunk_size_feed_forward": 0,
17
  "cross_attention_hidden_size": null,
18
  "decoder_start_token_id": null,
@@ -21,11 +21,12 @@
21
  "early_stopping": false,
22
  "embd_pdrop": 0.1,
23
  "encoder_no_repeat_ngram_size": 0,
24
- "eos_token_id": 50256,
25
  "exponential_decay_length_penalty": null,
26
  "finetuning_task": null,
27
  "forced_bos_token_id": null,
28
  "forced_eos_token_id": null,
 
29
  "id2label": {
30
  "0": "LABEL_0",
31
  "1": "LABEL_1"
@@ -43,10 +44,10 @@
43
  "min_length": 0,
44
  "model_type": "gpt2",
45
  "n_ctx": 1024,
46
- "n_embd": 768,
47
- "n_head": 12,
48
  "n_inner": null,
49
- "n_layer": 12,
50
  "n_positions": 1024,
51
  "no_repeat_ngram_size": 0,
52
  "num_beam_groups": 1,
@@ -77,7 +78,7 @@
77
  "task_specific_params": {
78
  "text-generation": {
79
  "do_sample": true,
80
- "max_length": 50
81
  }
82
  },
83
  "temperature": 1.0,
@@ -92,7 +93,7 @@
92
  "typical_p": 1.0,
93
  "use_bfloat16": false,
94
  "use_cache": true,
95
- "vocab_size": 50257
96
  },
97
  "decoder_start_token_id": 0,
98
  "encoder": {
@@ -255,7 +256,7 @@
255
  "vocab_size": 32,
256
  "xvector_output_dim": 512
257
  },
258
- "eos_token_id": 50256,
259
  "is_encoder_decoder": true,
260
  "model_type": "speech-encoder-decoder",
261
  "pad_token_id": 0,
 
3
  "SpeechEncoderDecoderModel"
4
  ],
5
  "decoder": {
6
+ "_name_or_path": "fav-kky/gpt2-small-cs",
7
+ "activation_function": "gelu",
8
  "add_cross_attention": true,
9
  "architectures": [
10
  "GPT2LMHeadModel"
 
12
  "attn_pdrop": 0.1,
13
  "bad_words_ids": null,
14
  "begin_suppress_tokens": null,
15
+ "bos_token_id": 0,
16
  "chunk_size_feed_forward": 0,
17
  "cross_attention_hidden_size": null,
18
  "decoder_start_token_id": null,
 
21
  "early_stopping": false,
22
  "embd_pdrop": 0.1,
23
  "encoder_no_repeat_ngram_size": 0,
24
+ "eos_token_id": 0,
25
  "exponential_decay_length_penalty": null,
26
  "finetuning_task": null,
27
  "forced_bos_token_id": null,
28
  "forced_eos_token_id": null,
29
+ "gradient_checkpointing": false,
30
  "id2label": {
31
  "0": "LABEL_0",
32
  "1": "LABEL_1"
 
44
  "min_length": 0,
45
  "model_type": "gpt2",
46
  "n_ctx": 1024,
47
+ "n_embd": 512,
48
+ "n_head": 8,
49
  "n_inner": null,
50
+ "n_layer": 4,
51
  "n_positions": 1024,
52
  "no_repeat_ngram_size": 0,
53
  "num_beam_groups": 1,
 
78
  "task_specific_params": {
79
  "text-generation": {
80
  "do_sample": true,
81
+ "max_length": 1024
82
  }
83
  },
84
  "temperature": 1.0,
 
93
  "typical_p": 1.0,
94
  "use_bfloat16": false,
95
  "use_cache": true,
96
+ "vocab_size": 50000
97
  },
98
  "decoder_start_token_id": 0,
99
  "encoder": {
 
256
  "vocab_size": 32,
257
  "xvector_output_dim": 512
258
  },
259
+ "eos_token_id": 0,
260
  "is_encoder_decoder": true,
261
  "model_type": "speech-encoder-decoder",
262
  "pad_token_id": 0,
merges.txt CHANGED
The diff for this file is too large to render. See raw diff
 
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3cd35ca8e9c67c36e263512e754494048d9dd66cd72ca8def638ca69c920acdc
3
- size 988766816
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:891b901803dfe40cc3a5d9d5b5feb4444a76ce2ed1155fd664a05d3311059532
3
+ size 550862672
special_tokens_map.json CHANGED
@@ -2,14 +2,14 @@
2
  "bos_token": {
3
  "content": "<|endoftext|>",
4
  "lstrip": false,
5
- "normalized": false,
6
  "rstrip": false,
7
  "single_word": false
8
  },
9
  "eos_token": {
10
  "content": "<|endoftext|>",
11
  "lstrip": false,
12
- "normalized": false,
13
  "rstrip": false,
14
  "single_word": false
15
  },
@@ -17,7 +17,7 @@
17
  "unk_token": {
18
  "content": "<|endoftext|>",
19
  "lstrip": false,
20
- "normalized": false,
21
  "rstrip": false,
22
  "single_word": false
23
  }
 
2
  "bos_token": {
3
  "content": "<|endoftext|>",
4
  "lstrip": false,
5
+ "normalized": true,
6
  "rstrip": false,
7
  "single_word": false
8
  },
9
  "eos_token": {
10
  "content": "<|endoftext|>",
11
  "lstrip": false,
12
+ "normalized": true,
13
  "rstrip": false,
14
  "single_word": false
15
  },
 
17
  "unk_token": {
18
  "content": "<|endoftext|>",
19
  "lstrip": false,
20
+ "normalized": true,
21
  "rstrip": false,
22
  "single_word": false
23
  }
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -1,10 +1,11 @@
1
  {
 
2
  "add_prefix_space": false,
3
  "added_tokens_decoder": {
4
  "0": {
5
  "content": "<|endoftext|>",
6
  "lstrip": false,
7
- "normalized": false,
8
  "rstrip": false,
9
  "single_word": false,
10
  "special": true
 
1
  {
2
+ "add_bos_token": false,
3
  "add_prefix_space": false,
4
  "added_tokens_decoder": {
5
  "0": {
6
  "content": "<|endoftext|>",
7
  "lstrip": false,
8
+ "normalized": true,
9
  "rstrip": false,
10
  "single_word": false,
11
  "special": true
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:28bdce1201546204b4c9e32c420b3ab1754ee964788c8eacb549a55d30826eea
3
  size 5176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:25bcb2f8affe3e0b5620c12a51a066c661187b7b619b589d43f303545c820caa
3
  size 5176
vocab.json CHANGED
The diff for this file is too large to render. See raw diff