javadr commited on
Commit
ebfcfae
1 Parent(s): 3d2b750

Training in progress, step 500

Browse files
config.json CHANGED
@@ -25,7 +25,20 @@
25
  "encoder_layerdrop": 0.0,
26
  "encoder_layers": 4,
27
  "eos_token_id": 50257,
28
- "forced_decoder_ids": null,
 
 
 
 
 
 
 
 
 
 
 
 
 
29
  "init_std": 0.02,
30
  "is_encoder_decoder": true,
31
  "mask_feature_length": 10,
@@ -43,7 +56,96 @@
43
  "num_mel_bins": 80,
44
  "pad_token_id": 50257,
45
  "scale_embedding": false,
46
- "suppress_tokens": [],
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
47
  "torch_dtype": "float32",
48
  "transformers_version": "4.37.0.dev0",
49
  "use_cache": true,
 
25
  "encoder_layerdrop": 0.0,
26
  "encoder_layers": 4,
27
  "eos_token_id": 50257,
28
+ "forced_decoder_ids": [
29
+ [
30
+ 1,
31
+ 50259
32
+ ],
33
+ [
34
+ 2,
35
+ 50359
36
+ ],
37
+ [
38
+ 3,
39
+ 50363
40
+ ]
41
+ ],
42
  "init_std": 0.02,
43
  "is_encoder_decoder": true,
44
  "mask_feature_length": 10,
 
56
  "num_mel_bins": 80,
57
  "pad_token_id": 50257,
58
  "scale_embedding": false,
59
+ "suppress_tokens": [
60
+ 1,
61
+ 2,
62
+ 7,
63
+ 8,
64
+ 9,
65
+ 10,
66
+ 14,
67
+ 25,
68
+ 26,
69
+ 27,
70
+ 28,
71
+ 29,
72
+ 31,
73
+ 58,
74
+ 59,
75
+ 60,
76
+ 61,
77
+ 62,
78
+ 63,
79
+ 90,
80
+ 91,
81
+ 92,
82
+ 93,
83
+ 359,
84
+ 503,
85
+ 522,
86
+ 542,
87
+ 873,
88
+ 893,
89
+ 902,
90
+ 918,
91
+ 922,
92
+ 931,
93
+ 1350,
94
+ 1853,
95
+ 1982,
96
+ 2460,
97
+ 2627,
98
+ 3246,
99
+ 3253,
100
+ 3268,
101
+ 3536,
102
+ 3846,
103
+ 3961,
104
+ 4183,
105
+ 4667,
106
+ 6585,
107
+ 6647,
108
+ 7273,
109
+ 9061,
110
+ 9383,
111
+ 10428,
112
+ 10929,
113
+ 11938,
114
+ 12033,
115
+ 12331,
116
+ 12562,
117
+ 13793,
118
+ 14157,
119
+ 14635,
120
+ 15265,
121
+ 15618,
122
+ 16553,
123
+ 16604,
124
+ 18362,
125
+ 18956,
126
+ 20075,
127
+ 21675,
128
+ 22520,
129
+ 26130,
130
+ 26161,
131
+ 26435,
132
+ 28279,
133
+ 29464,
134
+ 31650,
135
+ 32302,
136
+ 32470,
137
+ 36865,
138
+ 42863,
139
+ 47425,
140
+ 49870,
141
+ 50254,
142
+ 50258,
143
+ 50358,
144
+ 50359,
145
+ 50360,
146
+ 50361,
147
+ 50362
148
+ ],
149
  "torch_dtype": "float32",
150
  "transformers_version": "4.37.0.dev0",
151
  "use_cache": true,
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:06ec4d19756bbd65a43a643285d448c82a1149f8f623d7c3810439d30c91109f
3
  size 151061672
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bed1ff0a055c53c3923b5c87041260dcfe9b69f92e104c8861b4c4935a8494cb
3
  size 151061672
runs/Dec30_08-15-58_d00b0d360d05/events.out.tfevents.1703924170.d00b0d360d05.918.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0a65cd50381a315bbab2d1b6e69cb64c72e3f54c26de9c11adbf6c2f3b68082b
3
+ size 6297
runs/Dec30_08-18-32_d00b0d360d05/events.out.tfevents.1703924325.d00b0d360d05.918.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fdf385640b3f7a3ae7fc54742ca5385a5ab72360e3e408b4e8b2284903c2009f
3
+ size 6143
runs/Dec30_08-22-23_d00b0d360d05/events.out.tfevents.1703924553.d00b0d360d05.918.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f5bd0c96ed16945a10bbf47b1abfb8ebf5fad949bfc4f6f4e60f8600574bf589
3
+ size 10852
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:810c5827ccfb866bc35c44b0a7987ca4b3ff94a85c459102576411383f93e437
3
  size 4856
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7a244d32a69778dda52dbe57e9cc84be789553c2ae73bba64ce2c587ccb78d35
3
  size 4856