TeamNL commited on
Commit
d1dbd09
·
verified ·
1 Parent(s): 56fc7b9

Training in progress, epoch 19, checkpoint

Browse files
last-checkpoint/added_tokens.json ADDED
@@ -0,0 +1,927 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "##ख𑀢": 106740,
3
+ "##ख𑁦": 106738,
4
+ "##चढ": 106076,
5
+ "##चढच": 106115,
6
+ "##चढ𑀢": 106219,
7
+ "##चलच": 106352,
8
+ "##चललच": 106051,
9
+ "##चल𑀢": 106353,
10
+ "##च𑀟": 106661,
11
+ "##च𑀟च": 106483,
12
+ "##च𑀟च𑀟": 106794,
13
+ "##च𑀟च𑀤": 106755,
14
+ "##च𑀟च𑀤च𑀪पच": 106510,
15
+ "##च𑀟च𑀪": 106769,
16
+ "##च𑀟त": 106363,
17
+ "##च𑀟त𑀢": 105978,
18
+ "##च𑀟त𑀢𑀟": 106373,
19
+ "##च𑀟त𑁦": 105988,
20
+ "##च𑀟पच": 106479,
21
+ "##च𑀟पच𑀪": 106599,
22
+ "##च𑀟𑀟": 106701,
23
+ "##च𑀟𑀟च𑀟": 106640,
24
+ "##च𑀟𑀟𑀢𑀟": 106298,
25
+ "##च𑀟𑀢": 106127,
26
+ "##च𑀟𑀢𑀟": 106227,
27
+ "##च𑀟𑀣": 106146,
28
+ "##च𑀟𑀣च": 106121,
29
+ "##च𑀟𑀳च": 106629,
30
+ "##च𑀟𑁦": 106272,
31
+ "##च𑀠": 106315,
32
+ "##च𑀠च": 106536,
33
+ "##च𑀠च𑀪": 106134,
34
+ "##च𑀠ढ": 106275,
35
+ "##च𑀠ढच": 106164,
36
+ "##च𑀠𑀠च": 106085,
37
+ "##च𑀠𑀠च𑀟": 106113,
38
+ "##च𑀢": 106480,
39
+ "##च𑀢𑀞च": 106767,
40
+ "##च𑀪": 106620,
41
+ "##च𑀪च": 106074,
42
+ "##च𑀪पच": 106496,
43
+ "##च𑀪𑀢": 106403,
44
+ "##च𑀪𑀢𑀟": 106062,
45
+ "##च𑀪𑀪": 106197,
46
+ "##च𑀪𑁦": 106415,
47
+ "##च𑀲": 106133,
48
+ "##च𑀲च": 106662,
49
+ "##च𑀲𑀢𑀟": 106791,
50
+ "##च𑀳": 105960,
51
+ "##च𑀳च": 106368,
52
+ "##च𑀳च𑀪": 106055,
53
+ "##च𑀳𑀫𑀢": 106128,
54
+ "##च𑀳𑀫𑀢𑀟": 106560,
55
+ "##च𑀳𑀫𑁦": 106412,
56
+ "##च𑀳𑀫𑁦𑀟": 106064,
57
+ "##च𑀳𑀳": 105928,
58
+ "##झच": 106453,
59
+ "##झ𑀢𑀟": 105932,
60
+ "##ञच": 106159,
61
+ "##ठ𑀦": 106730,
62
+ "##ठ𑀧": 106512,
63
+ "##ठ𑀧𑀧": 106540,
64
+ "##ठ𑀧𑀭": 106505,
65
+ "##ठ𑀯": 106011,
66
+ "##ढच": 106426,
67
+ "##ढच𑀟": 106195,
68
+ "##ढच𑀪": 106524,
69
+ "##ढन": 106163,
70
+ "##ढन𑀱च𑀟": 106492,
71
+ "##ढ𑀢": 106117,
72
+ "##ढ𑀢𑀟": 106015,
73
+ "##ढ𑁣": 106458,
74
+ "##ढ𑁦": 106398,
75
+ "##ढ𑁦𑀪": 105917,
76
+ "##णच": 106335,
77
+ "##णचणण": 106246,
78
+ "##णचणणन𑀟": 106673,
79
+ "##णच𑀟": 106507,
80
+ "##णच𑀪": 106443,
81
+ "##णण": 106144,
82
+ "##णणच": 106369,
83
+ "##णणच𑀟च": 106711,
84
+ "##ण𑀢": 106563,
85
+ "##ण𑀢𑀟": 106199,
86
+ "##ण𑀦": 105890,
87
+ "##ण𑁣": 106283,
88
+ "##ण𑁣ण𑀢𑀟": 106168,
89
+ "##ण𑁦": 106785,
90
+ "##ण𑁦𑀟": 105926,
91
+ "##तच": 106771,
92
+ "##तत𑁦": 105883,
93
+ "##त𑀞": 106364,
94
+ "##त𑀢": 106802,
95
+ "##त𑀢𑀞": 106548,
96
+ "##त𑀢𑀟": 106200,
97
+ "##त𑀫": 106301,
98
+ "##त𑁣": 106580,
99
+ "##त𑁦": 106251,
100
+ "##धच": 106522,
101
+ "##ध𑀫": 106225,
102
+ "##ध𑁦": 106295,
103
+ "##नढन": 106092,
104
+ "##नपच": 106569,
105
+ "##नपच𑀪": 106547,
106
+ "##नपन": 106280,
107
+ "##नबच": 106672,
108
+ "##नल": 106761,
109
+ "##न𑀞": 106464,
110
+ "##न𑀞च": 106448,
111
+ "##न𑀞च𑀟": 106684,
112
+ "##न𑀞च𑀪": 105910,
113
+ "##न𑀟": 106789,
114
+ "##न𑀟च": 106633,
115
+ "##न𑀟च𑀟": 106132,
116
+ "##न𑀟त𑀢": 105925,
117
+ "##न𑀟ब": 106567,
118
+ "##न𑀟ब𑀢णच𑀪": 106239,
119
+ "##न𑀟𑀢": 106215,
120
+ "##न𑀟𑀢𑀟": 106046,
121
+ "##न𑀟𑀣": 106058,
122
+ "##न𑀠": 105952,
123
+ "##न𑀠च𑀪": 106276,
124
+ "##न𑀣": 106255,
125
+ "##न𑀦": 106681,
126
+ "##न𑀪": 105963,
127
+ "##न𑀪च": 106347,
128
+ "##न𑀪च𑀟": 105931,
129
+ "##न𑀪𑀞च": 106004,
130
+ "##न𑀪𑀢": 106048,
131
+ "##न𑀪𑁦": 106758,
132
+ "##न𑀯": 106089,
133
+ "##न𑀱": 106148,
134
+ "##न𑀱च": 106206,
135
+ "##न𑀱च𑀪": 106606,
136
+ "##न𑀳": 105965,
137
+ "##पच": 106261,
138
+ "##पच𑀟": 105987,
139
+ "##पच𑀦": 106351,
140
+ "##पच𑀪": 105886,
141
+ "##पच𑀯": 106124,
142
+ "##पन𑀟": 105954,
143
+ "##पन𑀠": 105990,
144
+ "##पप": 106096,
145
+ "##पपच": 106487,
146
+ "##प𑀢": 106207,
147
+ "##प𑀢त": 105961,
148
+ "##प𑀢𑀟": 106278,
149
+ "##प𑀢𑁣𑀟": 106122,
150
+ "##प𑀢𑁣𑀟चल": 106574,
151
+ "##प𑀫": 105973,
152
+ "##प𑀳": 106748,
153
+ "##प𑀳च": 106717,
154
+ "##प𑁣": 106607,
155
+ "##प𑁣𑀟": 105994,
156
+ "##प𑁣𑀪": 106766,
157
+ "##प𑁦": 106434,
158
+ "##प𑁦𑀪": 106742,
159
+ "##बच": 106663,
160
+ "##बच𑀪": 106726,
161
+ "##बच𑀳": 106423,
162
+ "##ब𑀢": 105958,
163
+ "##ब𑀢𑀟": 106595,
164
+ "##ब𑁣": 106359,
165
+ "##ब𑁣𑀪": 106289,
166
+ "##ब𑁦": 106035,
167
+ "##ब𑁦𑀟": 106226,
168
+ "##र𑀦": 106478,
169
+ "##र𑀯": 106439,
170
+ "##लच": 106695,
171
+ "##लच𑀟𑀣": 106390,
172
+ "##लच𑀠": 105945,
173
+ "##लच𑀪": 106666,
174
+ "##लण": 106189,
175
+ "##ललच": 106700,
176
+ "##लल𑁣𑀟": 106615,
177
+ "##ल𑀢": 106086,
178
+ "##ल𑀢𑀟": 105983,
179
+ "##ल𑁣": 106557,
180
+ "##ल𑁣ढ": 106024,
181
+ "##ल𑁦": 105985,
182
+ "##व𑀯": 106502,
183
+ "##ष𑀣च": 106637,
184
+ "##𑀓": 106549,
185
+ "##𑀕": 106243,
186
+ "##𑀖": 106554,
187
+ "##𑀖𑀦": 106231,
188
+ "##𑀖𑀯": 106778,
189
+ "##𑀗": 106210,
190
+ "##𑀗𑀦": 106097,
191
+ "##𑀗𑀯": 106564,
192
+ "##𑀘": 106427,
193
+ "##𑀘च": 106068,
194
+ "##𑀘चल": 105984,
195
+ "##𑀘𑀢": 106027,
196
+ "##𑀘𑀢𑀟": 106521,
197
+ "##𑀘𑁣": 106084,
198
+ "##𑀘𑁦": 106800,
199
+ "##𑀘𑁦𑀟": 106490,
200
+ "##𑀘𑁦𑀪": 106350,
201
+ "##𑀘𑁦𑀪𑀢णच": 106589,
202
+ "##𑀙": 106421,
203
+ "##𑀙च": 106546,
204
+ "##𑀙च𑀪": 106699,
205
+ "##𑀙𑀢": 106249,
206
+ "##𑀙𑀢𑀟": 106400,
207
+ "##𑀚": 106138,
208
+ "##𑀛": 106208,
209
+ "##𑀜": 106330,
210
+ "##𑀜न": 106798,
211
+ "##𑀝": 106286,
212
+ "##𑀞": 106190,
213
+ "##𑀞च": 106597,
214
+ "##𑀞च𑀟": 106425,
215
+ "##𑀞च𑀟𑀢𑀟": 105995,
216
+ "##𑀞च𑀠": 106038,
217
+ "##𑀞च𑀠च": 106516,
218
+ "##𑀞च𑀢": 106555,
219
+ "##𑀞च𑀪": 106677,
220
+ "##𑀞च𑀪च": 106678,
221
+ "##𑀞न": 106136,
222
+ "##𑀞न𑀟": 106238,
223
+ "##𑀞𑀢": 106307,
224
+ "##𑀞𑀢ल": 106332,
225
+ "##𑀞𑀢𑀟": 106265,
226
+ "##𑀞𑁣": 106268,
227
+ "##𑀞𑁣𑀞𑀢𑀟": 106665,
228
+ "##𑀞𑁣𑀟": 106336,
229
+ "##𑀞𑁦": 106651,
230
+ "##𑀞𑁦𑀟": 106756,
231
+ "##𑀟": 106102,
232
+ "##𑀟च": 106183,
233
+ "##𑀟च𑀟": 106765,
234
+ "##𑀟𑀢": 106744,
235
+ "##𑀟𑀢𑀟": 106106,
236
+ "##𑀟𑁦": 106576,
237
+ "##𑀠": 105996,
238
+ "##𑀠च": 106033,
239
+ "##𑀠चलच": 106551,
240
+ "##𑀠च𑀟": 106803,
241
+ "##𑀠च𑀪": 106381,
242
+ "##𑀠𑀢": 105934,
243
+ "##𑀠𑀢𑀟": 106690,
244
+ "##𑀡": 106341,
245
+ "##𑀢": 106712,
246
+ "##𑀢च": 105897,
247
+ "##𑀢ढ": 106562,
248
+ "##𑀢ण": 106008,
249
+ "##𑀢णच": 106257,
250
+ "##𑀢णच𑀟": 106494,
251
+ "##𑀢णच𑀦": 106659,
252
+ "##𑀢णच𑀪": 106721,
253
+ "##𑀢णच𑀯": 106457,
254
+ "##𑀢णणच": 106550,
255
+ "##𑀢णणच𑀪": 106553,
256
+ "##𑀢ण𑁣": 106292,
257
+ "##𑀢ण𑁣ण𑀢𑀟": 106039,
258
+ "##𑀢ण𑁦": 106638,
259
+ "##𑀢त": 106611,
260
+ "##𑀢प": 106214,
261
+ "##𑀢पच": 106642,
262
+ "##𑀢पपच": 105936,
263
+ "##𑀢प𑀢𑀟": 106414,
264
+ "##𑀢प𑁣": 105916,
265
+ "##𑀢ब": 106463,
266
+ "##𑀢ल": 106229,
267
+ "##𑀢𑀞": 105980,
268
+ "##𑀢𑀞𑀢": 106630,
269
+ "##𑀢𑀞𑀢𑀟": 106327,
270
+ "##𑀢𑀞𑀱च": 105979,
271
+ "##𑀢𑀟": 106526,
272
+ "##𑀢𑀟च": 106023,
273
+ "##𑀢𑀟ब": 106378,
274
+ "##𑀢𑀟𑁦": 106093,
275
+ "##𑀢𑀠": 106586,
276
+ "##𑀢𑀠च": 106739,
277
+ "##𑀢𑀣": 106130,
278
+ "##𑀢𑀣च": 106078,
279
+ "##𑀢𑀣च𑀟": 106201,
280
+ "##𑀢𑀤": 106713,
281
+ "##𑀢𑀦": 106667,
282
+ "##𑀢𑀪": 106329,
283
+ "##𑀢𑀪च": 106617,
284
+ "##𑀢𑀪𑀞च": 105950,
285
+ "##𑀢𑀪𑀢": 106411,
286
+ "##𑀢𑀪𑀢𑀟": 106770,
287
+ "##𑀢𑀪𑁦": 106188,
288
+ "##𑀢𑀯": 106745,
289
+ "##𑀢𑀱च": 106445,
290
+ "##𑀢𑀳": 106533,
291
+ "##𑀢𑀳च": 106741,
292
+ "##𑀢𑀳पच𑀟": 106252,
293
+ "##𑀢𑀳𑀫": 105941,
294
+ "##𑀣": 106462,
295
+ "##𑀣च": 106041,
296
+ "##𑀣च𑀟": 106191,
297
+ "##𑀣च𑀪": 106180,
298
+ "##𑀣न": 106173,
299
+ "##𑀣𑀢": 105966,
300
+ "##𑀣𑀢𑀟": 106218,
301
+ "##𑀣𑀣": 106310,
302
+ "##𑀣𑀣च": 106783,
303
+ "##𑀣𑁣": 106362,
304
+ "##𑀣𑁣𑀟": 106634,
305
+ "##𑀣𑁦": 106689,
306
+ "##𑀤": 106372,
307
+ "##𑀤च": 106087,
308
+ "##𑀤न": 106486,
309
+ "##𑀥": 106724,
310
+ "##𑀦": 106259,
311
+ "##𑀧": 106107,
312
+ "##𑀧𑀦": 106237,
313
+ "##𑀧𑀧": 106514,
314
+ "##𑀧𑀯": 106339,
315
+ "##𑀨": 106100,
316
+ "##𑀩": 106587,
317
+ "##𑀪": 105911,
318
+ "##𑀪च": 106169,
319
+ "##𑀪च𑀟": 106287,
320
+ "##𑀪𑀢": 106537,
321
+ "##𑀪𑀢ल": 106402,
322
+ "##𑀪𑀢𑀟": 106430,
323
+ "##𑀪𑁣": 106175,
324
+ "##𑀪𑁦": 106566,
325
+ "##𑀫": 106236,
326
+ "##𑀫च": 106493,
327
+ "##𑀫चल": 106774,
328
+ "##𑀫च𑀟": 106784,
329
+ "##𑀫च𑀟ण𑁣ण𑀢𑀟": 106165,
330
+ "##𑀫च𑀢": 106703,
331
+ "##𑀫च𑀪": 106768,
332
+ "##𑀫न": 106037,
333
+ "##𑀫𑀢": 106509,
334
+ "##𑀫𑀢𑀟": 106763,
335
+ "##𑀫𑀢𑀠": 106405,
336
+ "##𑀫𑁣": 106308,
337
+ "##𑀫𑁦": 105885,
338
+ "##𑀫𑁦𑀟": 106473,
339
+ "##𑀬": 106036,
340
+ "##𑀬𑀦": 106149,
341
+ "##𑀬𑀯": 106718,
342
+ "##𑀭": 106288,
343
+ "##𑀭थ": 106157,
344
+ "##𑀭𑀦": 106222,
345
+ "##𑀭𑀯": 106009,
346
+ "##𑀮": 105969,
347
+ "##𑀮𑀦": 106342,
348
+ "##𑀮𑀯": 105923,
349
+ "##𑀯": 106357,
350
+ "##𑀰": 105905,
351
+ "##𑀰𑀦": 106230,
352
+ "##𑀰𑀯": 106590,
353
+ "##𑀱": 106178,
354
+ "##𑀱च": 106099,
355
+ "##𑀱च𑀟": 106184,
356
+ "##𑀱च𑀟𑁦": 106754,
357
+ "##𑀱च𑀠": 106538,
358
+ "##𑀱च𑀢": 105913,
359
+ "##𑀱च𑀦": 106095,
360
+ "##𑀱च𑀪": 106209,
361
+ "##𑀱च𑀯": 106250,
362
+ "##𑀱𑁣": 106012,
363
+ "##𑀱𑁣𑀟": 106657,
364
+ "##𑀱𑁦": 106176,
365
+ "##𑀲": 106202,
366
+ "##𑀲च": 106312,
367
+ "##𑀲च𑀪": 105956,
368
+ "##𑀲च𑀲": 106131,
369
+ "##𑀲𑀢": 106797,
370
+ "##𑀲𑀢𑀟": 106320,
371
+ "##𑀲𑁣": 106355,
372
+ "##𑀲𑁦": 106625,
373
+ "##𑀳": 106198,
374
+ "##𑀳च": 106258,
375
+ "##𑀳च𑀟": 105882,
376
+ "##𑀳च𑀦": 106499,
377
+ "##𑀳न": 106234,
378
+ "##𑀳प": 106670,
379
+ "##𑀳𑀦": 105971,
380
+ "##𑀳𑀫": 106680,
381
+ "##𑀳𑀫𑀢": 105957,
382
+ "##𑀳𑀫𑁦": 106147,
383
+ "##𑀳𑀫𑁦𑀟": 106459,
384
+ "##𑀳𑀯": 106476,
385
+ "##𑁢": 106140,
386
+ "##𑁢𑀦": 106019,
387
+ "##𑁢𑀯": 106675,
388
+ "##𑁣": 106192,
389
+ "##𑁣ल": 106440,
390
+ "##𑁣𑀟": 106696,
391
+ "##𑁣𑀟च": 106652,
392
+ "##𑁣𑀠": 106040,
393
+ "##𑁣𑀠𑀢𑀟": 106172,
394
+ "##𑁣𑀣च": 106182,
395
+ "##𑁣𑀦": 106786,
396
+ "##𑁣𑀪": 106467,
397
+ "##𑁣𑀪𑀢𑀟": 106734,
398
+ "##𑁣𑀫": 106519,
399
+ "##𑁣𑀳": 106034,
400
+ "##𑁤": 106267,
401
+ "##𑁥": 106419,
402
+ "##𑁥च": 106396,
403
+ "##𑁦": 106508,
404
+ "##𑁦प": 106603,
405
+ "##𑁦ल": 106609,
406
+ "##𑁦𑀟": 106057,
407
+ "##𑁦𑀟प": 106397,
408
+ "##𑁦𑀠": 106020,
409
+ "##𑁦𑀣": 106627,
410
+ "##𑁦𑀦": 106429,
411
+ "##𑁦𑀪": 106349,
412
+ "##𑁦𑀪च": 106747,
413
+ "##𑁦𑀪𑀳": 106762,
414
+ "##𑁦𑀯": 106433,
415
+ "##𑁦𑀳": 106108,
416
+ "##𑁦𑀳𑀳": 106242,
417
+ "चढ": 105955,
418
+ "चढन": 105898,
419
+ "चढनढन𑀱च𑀟": 106103,
420
+ "चढ𑀢𑀟": 106622,
421
+ "चढ𑁣": 106141,
422
+ "चणण": 105891,
423
+ "चणणन𑀞च𑀟": 106375,
424
+ "चप": 106787,
425
+ "चब": 106114,
426
+ "चलष": 105903,
427
+ "चल𑀙": 106160,
428
+ "चल𑀙च": 106655,
429
+ "चल𑀙न𑀠": 106671,
430
+ "चल𑀢": 106731,
431
+ "च𑀞": 106583,
432
+ "च𑀞च": 106704,
433
+ "च𑀞च𑀟": 106614,
434
+ "च𑀞𑀱च𑀢": 106779,
435
+ "च𑀞𑁦": 106447,
436
+ "च𑀟": 106643,
437
+ "च𑀟च": 106066,
438
+ "च𑀟𑀣": 106126,
439
+ "च𑀠": 106032,
440
+ "च𑀠न𑀪𑀞च": 106649,
441
+ "च𑀠𑀠च": 106750,
442
+ "च𑀠𑀢𑀟": 106477,
443
+ "च𑀠𑀲": 106271,
444
+ "च𑀠𑀲च𑀟𑀢": 106002,
445
+ "च𑀢𑀞𑀢": 106167,
446
+ "च𑀢𑀞𑀢𑀟": 105976,
447
+ "च𑀣": 106683,
448
+ "च𑀣च": 106010,
449
+ "च𑀣च𑀠": 106725,
450
+ "च𑀣𑀣𑀢𑀟": 106442,
451
+ "च𑀤": 106460,
452
+ "च𑀪": 106503,
453
+ "च𑀪𑀤": 106304,
454
+ "च𑀪𑁦": 106105,
455
+ "च𑀪𑁦𑀱च": 106285,
456
+ "च𑀲": 106656,
457
+ "च𑀲𑀢𑀪𑀞च": 106161,
458
+ "च𑀳": 106799,
459
+ "च𑀳च": 106531,
460
+ "झच": 106152,
461
+ "झच𑀪": 106455,
462
+ "झच𑀲च": 106579,
463
+ "झच𑀳च": 105981,
464
+ "झच𑀳च𑀪": 105943,
465
+ "झन𑀟ब": 106228,
466
+ "झन𑀟ब𑀢णच𑀪": 106079,
467
+ "झ𑀱च": 106162,
468
+ "झ𑀱चलल𑁣𑀟": 106585,
469
+ "झ𑀱च𑀪": 106751,
470
+ "ञच": 106281,
471
+ "ञचणच": 105912,
472
+ "ञच𑀟": 106565,
473
+ "ञ𑀢𑀟": 106498,
474
+ "ठ𑀧": 106444,
475
+ "ठ𑀧ठ": 106223,
476
+ "ठ𑀧ठ𑀧": 106685,
477
+ "ठ𑀧𑀧": 106705,
478
+ "ठ𑀧𑀭": 105893,
479
+ "ठ𑀧𑀭थ": 106306,
480
+ "ढच": 106098,
481
+ "ढचढ": 105999,
482
+ "ढचढढच𑀟": 106170,
483
+ "ढचण": 106776,
484
+ "ढचणच": 106715,
485
+ "ढचणच𑀟": 106641,
486
+ "ढचणच𑀟च𑀟": 106775,
487
+ "ढचणणच𑀟च": 106753,
488
+ "ढच𑀟": 106077,
489
+ "ढच𑀦": 105970,
490
+ "ढच𑀪": 106408,
491
+ "ढच𑀯": 106706,
492
+ "ढन": 106601,
493
+ "ढनबच": 106083,
494
+ "ढन𑀪": 106646,
495
+ "ढ𑀢": 106485,
496
+ "ढ𑀢ण": 106385,
497
+ "ढ𑀢णन": 106213,
498
+ "ढ𑀢𑀟": 106693,
499
+ "ढ𑀢𑀟त𑀢𑀞": 106376,
500
+ "ढ𑀢𑀪": 106014,
501
+ "ढ𑀢𑀪𑀟𑀢𑀟": 106796,
502
+ "ढ𑁣": 105906,
503
+ "ढ𑁦": 106067,
504
+ "णच": 106417,
505
+ "णचण𑀢𑀟": 106392,
506
+ "णच𑀞𑁦": 105915,
507
+ "णच𑀟": 105880,
508
+ "णच𑀟च": 106573,
509
+ "णच𑀟चण𑀢": 106293,
510
+ "णच𑀟चण𑀢𑀟": 106367,
511
+ "णच𑀟𑀞𑀢𑀟": 106145,
512
+ "णच𑀟𑀤न": 106094,
513
+ "णच𑀠𑀠च": 106520,
514
+ "णच𑀣𑀣च": 106735,
515
+ "णच𑀪": 106129,
516
+ "णच𑀱": 106091,
517
+ "णच𑀱च": 106465,
518
+ "णच𑀱च𑀟": 105904,
519
+ "णच𑀱च𑀟त𑀢𑀟": 105953,
520
+ "णन": 106450,
521
+ "ण𑀢": 106370,
522
+ "ण𑀢𑀟": 106790,
523
+ "तच𑀟": 105942,
524
+ "त𑀢": 106575,
525
+ "त𑀢ढ": 106404,
526
+ "त𑀢ढ𑀢णच𑀪": 105930,
527
+ "त𑀢𑀞𑀢": 106333,
528
+ "त𑀢𑀞𑀢𑀟": 106436,
529
+ "त𑀢𑀟": 106437,
530
+ "त𑀫": 106600,
531
+ "त𑁣": 106080,
532
+ "त𑁣𑀟": 106489,
533
+ "त𑁣𑀠": 106777,
534
+ "त𑁦": 106240,
535
+ "त𑁦𑀱च": 106334,
536
+ "धच": 106470,
537
+ "धच𑀪": 106030,
538
+ "न𑀞न": 106050,
539
+ "न𑀟": 106781,
540
+ "पच": 106217,
541
+ "पचढ": 106171,
542
+ "पचपपच": 106366,
543
+ "पचपपचल𑀢𑀟": 106644,
544
+ "पच𑀞च": 106468,
545
+ "पच𑀟च": 106344,
546
+ "पच𑀢𑀠च": 106361,
547
+ "पच𑀪": 106645,
548
+ "पच𑀪च": 105922,
549
+ "पच𑀪𑀢": 106692,
550
+ "पच𑀪𑀢𑀫𑀢": 106399,
551
+ "पच𑀪𑀢𑀫𑀢𑀟": 105896,
552
+ "पच𑀪𑁦": 106260,
553
+ "पच𑀱च": 105884,
554
+ "पच𑀲": 106475,
555
+ "पच𑀳": 106732,
556
+ "पन𑀟": 105881,
557
+ "पन𑀪": 105982,
558
+ "पन𑀪च𑀢": 106374,
559
+ "प𑀫𑁦": 106264,
560
+ "प𑀳": 106451,
561
+ "प𑀳च": 106070,
562
+ "प𑀳च𑀞": 106782,
563
+ "प𑀳च𑀞च𑀟𑀢𑀟": 106003,
564
+ "प𑀳च𑀪": 106116,
565
+ "प𑀳च𑀪𑀢𑀟": 106282,
566
+ "प𑀳𑁣𑀫": 105959,
567
+ "प𑁣": 106605,
568
+ "प𑁦": 106511,
569
+ "बच": 106506,
570
+ "बचढ": 105940,
571
+ "बचढच": 106393,
572
+ "बचढचपच𑀪": 106608,
573
+ "बच𑀟": 105935,
574
+ "बच𑀠": 106716,
575
+ "बच𑀠𑁦": 105962,
576
+ "बच𑀪": 106060,
577
+ "बच𑀪𑀢𑀟": 106558,
578
+ "बच𑀳च𑀪": 106090,
579
+ "बन𑀣": 106582,
580
+ "बन𑀣च": 105895,
581
+ "बन𑀣च𑀟च𑀪": 106621,
582
+ "बन𑀪": 106504,
583
+ "ब𑀢𑀣च": 105992,
584
+ "ब𑀢𑀣च𑀟": 106395,
585
+ "ब𑀢𑀪": 106059,
586
+ "ब𑀢𑀪𑀠च": 105944,
587
+ "ब𑀫": 106232,
588
+ "ब𑀫च𑀟च": 106000,
589
+ "ब𑀱च𑀠": 106616,
590
+ "ब𑀱च𑀠𑀟च": 106111,
591
+ "ब𑀱च𑀠𑀟चप𑀢": 105947,
592
+ "ब𑀱च𑀠𑀟चप𑀢𑀟": 105989,
593
+ "ब𑁣": 106382,
594
+ "ब𑁦": 106323,
595
+ "लच": 106736,
596
+ "लचढ": 106112,
597
+ "लच𑀠": 106407,
598
+ "लच𑀠ढच𑀪": 106610,
599
+ "लच𑀲": 106541,
600
+ "लच𑀳𑀫𑁦": 105929,
601
+ "ल𑀢पपच": 106204,
602
+ "ल𑁣": 106174,
603
+ "ल𑁣𑀞च": 106639,
604
+ "ल𑁣𑀞चत𑀢": 106788,
605
+ "ल𑁣𑀞चत𑀢𑀟": 106383,
606
+ "ल𑁦": 106571,
607
+ "𑀓": 105920,
608
+ "𑀕": 106391,
609
+ "𑀖": 105949,
610
+ "𑀗": 106527,
611
+ "𑀘": 105900,
612
+ "𑀘च": 106317,
613
+ "𑀘चब𑁣𑀪": 106325,
614
+ "𑀘च𑀟च": 106801,
615
+ "𑀘च𑀟च𑀢𑀪": 106256,
616
+ "𑀘च𑀠": 106723,
617
+ "𑀘च𑀠च": 106047,
618
+ "𑀘च𑀠च𑀙च": 106584,
619
+ "𑀘च𑀠𑀢": 106221,
620
+ "𑀘च𑀠𑀢𑀙च𑀪": 106073,
621
+ "𑀘च𑀪𑀢": 106110,
622
+ "𑀘न": 106441,
623
+ "𑀘𑀢": 106623,
624
+ "𑀘𑀢𑀟": 106296,
625
+ "𑀘𑀢𑀪": 106263,
626
+ "𑀘𑀢𑀪ब𑀢𑀟": 106420,
627
+ "𑀘𑀢𑀫च𑀪": 106456,
628
+ "𑀘𑁦𑀪": 106018,
629
+ "𑀘𑁦𑀪𑀢𑀟": 106045,
630
+ "𑀙": 106052,
631
+ "𑀙णच𑀟": 106577,
632
+ "𑀚": 105993,
633
+ "𑀛": 106053,
634
+ "𑀜": 106530,
635
+ "𑀝": 106674,
636
+ "𑀝च𑀟": 106299,
637
+ "𑀝𑀭थ": 106495,
638
+ "𑀞": 106737,
639
+ "𑀞च": 106697,
640
+ "𑀞चणच𑀟": 106338,
641
+ "𑀞चल": 106491,
642
+ "𑀞च𑀟": 106269,
643
+ "𑀞च𑀠": 106559,
644
+ "𑀞च𑀠च𑀪": 106691,
645
+ "𑀞च𑀠𑀠चलच": 106253,
646
+ "𑀞च𑀠𑀲": 106752,
647
+ "𑀞च𑀠𑀲च𑀟𑀢𑀟": 106515,
648
+ "𑀞च𑀢": 106410,
649
+ "𑀞च𑀪": 106709,
650
+ "𑀞च𑀪च": 105918,
651
+ "𑀞च𑀪चपन𑀟": 106042,
652
+ "𑀞च𑀪𑁦": 106013,
653
+ "𑀞च𑀲च": 106626,
654
+ "𑀞च𑀲𑀢𑀟": 105914,
655
+ "𑀞च𑀳": 106120,
656
+ "𑀞च𑀳च": 106719,
657
+ "𑀞च𑀳च𑀟त𑁦": 106151,
658
+ "𑀞च𑀳च𑀪": 105968,
659
+ "𑀞च𑀳न𑀱": 106474,
660
+ "𑀞च𑀳𑀫𑀢": 106025,
661
+ "𑀞णच": 106224,
662
+ "𑀞न": 105964,
663
+ "𑀞नलच": 105889,
664
+ "𑀞न𑀟": 106322,
665
+ "𑀞न𑀟ब𑀢णच𑀪": 106428,
666
+ "𑀞न𑀠च": 106154,
667
+ "𑀞न𑀣न": 106572,
668
+ "𑀞न𑀳च": 106535,
669
+ "𑀞न𑀳च𑀟": 106513,
670
+ "𑀞𑀢": 105939,
671
+ "𑀞𑀢𑀠": 106005,
672
+ "𑀞𑀢𑀪": 106694,
673
+ "𑀞𑀢𑀪च": 106598,
674
+ "𑀞𑀱": 106156,
675
+ "𑀞𑀱च": 106461,
676
+ "𑀞𑀱चलल𑁣𑀟": 105991,
677
+ "𑀞𑀱च𑀟": 106290,
678
+ "𑀞𑀱च𑀠": 106380,
679
+ "𑀞𑀱च𑀢𑀞𑀱च": 106592,
680
+ "𑀞𑀱च𑀢𑀞𑀱चण𑁣": 106728,
681
+ "𑀞𑁣": 106418,
682
+ "𑀞𑁣ब𑀢𑀟": 106792,
683
+ "𑀞𑁣𑀠च": 106321,
684
+ "𑀞𑁣𑀲𑀢𑀟": 106631,
685
+ "𑀞𑁦": 106698,
686
+ "𑀟": 106591,
687
+ "𑀟च": 106273,
688
+ "𑀟चन": 106360,
689
+ "𑀟च𑀘𑁦𑀪𑀢णच": 106233,
690
+ "𑀟च𑀘𑁦𑀪𑀢णच𑀯": 106212,
691
+ "𑀟च𑀟": 106387,
692
+ "𑀟च𑀳": 106297,
693
+ "𑀟च𑀳च𑀪च": 106277,
694
+ "𑀟न": 106632,
695
+ "𑀟न𑀟च": 106679,
696
+ "𑀟𑀢": 106764,
697
+ "𑀟𑁣": 106532,
698
+ "𑀟𑁦": 106193,
699
+ "𑀟𑁦𑀱": 105933,
700
+ "𑀠": 106523,
701
+ "𑀠च": 106529,
702
+ "𑀠चण": 106594,
703
+ "𑀠चप": 106356,
704
+ "𑀠चपच": 106542,
705
+ "𑀠चप𑀳च": 106143,
706
+ "𑀠चप𑀳चण𑀢𑀟": 106772,
707
+ "𑀠चब": 106109,
708
+ "𑀠चललच": 106101,
709
+ "𑀠च𑀘चल": 106185,
710
+ "𑀠च𑀘चल𑀢𑀳": 106528,
711
+ "𑀠च𑀘चल𑀢𑀳च𑀪": 106720,
712
+ "𑀠च𑀙": 106071,
713
+ "𑀠च𑀙च𑀢𑀞च": 105986,
714
+ "𑀠च𑀞च𑀪": 106266,
715
+ "𑀠च𑀞च𑀪च𑀟पच𑀪": 106648,
716
+ "𑀠च𑀟": 106570,
717
+ "𑀠च𑀟च𑀤च𑀪पच": 106248,
718
+ "𑀠च𑀟णच𑀟": 105921,
719
+ "𑀠च𑀢": 106484,
720
+ "𑀠च𑀤च": 106544,
721
+ "𑀠च𑀪": 105879,
722
+ "𑀠च𑀫च𑀢": 105948,
723
+ "𑀠च𑀲𑀢": 106707,
724
+ "𑀠च𑀳": 105997,
725
+ "𑀠च𑀳च𑀟च": 106710,
726
+ "𑀠च𑀳न": 106795,
727
+ "𑀠न": 106291,
728
+ "𑀠नपच𑀟": 106049,
729
+ "𑀠नपच𑀟𑁦": 105937,
730
+ "𑀠नपच𑀟𑁦𑀟": 106354,
731
+ "𑀠नपन": 106647,
732
+ "𑀠नपन𑀠": 106158,
733
+ "𑀠नल": 106021,
734
+ "𑀠नल𑀞𑀢𑀟": 105924,
735
+ "𑀠न𑀫चल": 106669,
736
+ "𑀠न𑀫चलल𑀢": 106054,
737
+ "𑀠न𑀳": 106682,
738
+ "𑀠न𑀳च𑀠𑀠च𑀟": 106300,
739
+ "𑀠न𑀳न": 106552,
740
+ "𑀠न𑀳नल": 105927,
741
+ "𑀠𑀢ल": 106619,
742
+ "𑀠𑀢���": 106722,
743
+ "𑀠𑁣": 106166,
744
+ "𑀠𑁦": 105972,
745
+ "𑀠𑁦𑀠": 106438,
746
+ "𑀡": 105974,
747
+ "𑀢": 106235,
748
+ "𑀢ढ": 106294,
749
+ "𑀢णच": 105899,
750
+ "𑀢पच": 106135,
751
+ "𑀢ल𑀢": 106568,
752
+ "𑀢ल𑀢𑀠𑀢": 106274,
753
+ "𑀢ल𑀢𑀠𑀢𑀟": 106319,
754
+ "𑀢𑀟": 106593,
755
+ "𑀢𑀟ब": 106153,
756
+ "𑀢𑀟𑀣च": 106082,
757
+ "𑀢𑀪𑀢𑀟": 106187,
758
+ "𑀢𑀳": 106452,
759
+ "𑀣": 106660,
760
+ "𑀣च": 106687,
761
+ "𑀣चढच𑀟": 106181,
762
+ "𑀣चणच": 106688,
763
+ "𑀣चबच": 106371,
764
+ "𑀣च𑀘𑀢": 106061,
765
+ "𑀣च𑀟": 106006,
766
+ "𑀣च𑀠च𑀪": 106578,
767
+ "𑀣च𑀢𑀣च": 106318,
768
+ "𑀣न": 106581,
769
+ "𑀣नढच": 106488,
770
+ "𑀣न𑀞": 106409,
771
+ "𑀣न𑀟": 106733,
772
+ "𑀣न𑀟𑀢णच": 106194,
773
+ "𑀣𑀢": 106618,
774
+ "𑀣𑀢ब": 106757,
775
+ "𑀣𑀢𑀟": 106525,
776
+ "𑀣𑀢𑀳": 106205,
777
+ "𑀣𑁣": 106254,
778
+ "𑀣𑁣𑀟": 106472,
779
+ "𑀣𑁦": 106539,
780
+ "𑀤": 106137,
781
+ "𑀤च": 106284,
782
+ "𑀤चढ": 106337,
783
+ "𑀤च𑀠": 106517,
784
+ "𑀤च𑀠च": 106324,
785
+ "𑀤च𑀠च𑀟": 106604,
786
+ "𑀤च𑀢": 106469,
787
+ "𑀤च𑀪": 106031,
788
+ "𑀤च𑁥": 106028,
789
+ "𑀤न𑀱च": 106454,
790
+ "𑀥": 106613,
791
+ "𑀦": 106104,
792
+ "𑀧": 106658,
793
+ "𑀨": 106016,
794
+ "𑀩": 106624,
795
+ "𑀪": 106628,
796
+ "𑀪च": 106432,
797
+ "𑀪चण": 106313,
798
+ "𑀪चणन𑀱च": 106186,
799
+ "𑀪चणन𑀱च𑀪": 106389,
800
+ "𑀪च𑀟च𑀪": 106340,
801
+ "𑀪च𑀳𑀫𑀢𑀟": 106416,
802
+ "𑀪न": 106142,
803
+ "𑀪नढन": 106377,
804
+ "𑀪न𑀱च": 106311,
805
+ "𑀪𑀢": 106388,
806
+ "𑀪𑁣": 106081,
807
+ "𑀪𑁦": 106365,
808
+ "𑀫": 105892,
809
+ "𑀫च": 106729,
810
+ "𑀫चझ": 106588,
811
+ "𑀫चञ": 105998,
812
+ "𑀫चञच": 105908,
813
+ "𑀫चल": 106179,
814
+ "𑀫च𑀞च": 106279,
815
+ "𑀫च𑀟": 106270,
816
+ "𑀫च𑀟णच𑀪": 106326,
817
+ "𑀫च𑀟𑀞च": 106069,
818
+ "𑀫च𑀟𑀞चल𑀢": 105888,
819
+ "𑀫च𑀟𑀟": 106044,
820
+ "𑀫च𑀟𑀟न": 106449,
821
+ "𑀫च𑀢": 106386,
822
+ "𑀫च𑀢𑀲च𑀪": 106348,
823
+ "𑀫च𑀢𑀲𑁦": 106556,
824
+ "𑀫च𑀣च": 106497,
825
+ "𑀫च𑀪": 106612,
826
+ "𑀫न": 106727,
827
+ "𑀫न𑀞": 106759,
828
+ "𑀫न𑀞न𑀠च𑀪": 106343,
829
+ "𑀫𑁣": 106247,
830
+ "𑀬": 106635,
831
+ "𑀭": 106203,
832
+ "𑀭थ": 106435,
833
+ "𑀭थथ": 106119,
834
+ "𑀭थ𑀖": 106216,
835
+ "𑀭थ𑀗": 106029,
836
+ "𑀭थ𑀬": 106668,
837
+ "𑀭थ𑁢": 106481,
838
+ "𑀭𑀗": 105907,
839
+ "𑀮": 106196,
840
+ "𑀯": 106177,
841
+ "𑀰": 105901,
842
+ "𑀱": 106482,
843
+ "𑀱च": 106065,
844
+ "𑀱चञ": 106328,
845
+ "𑀱चञच𑀟𑀟च𑀟": 106773,
846
+ "𑀱चञच𑀟𑀣च": 106466,
847
+ "𑀱चपच": 106500,
848
+ "𑀱चपच𑀟": 106345,
849
+ "𑀱च𑀘𑁦": 106793,
850
+ "𑀱च𑀘𑁦𑀟": 106471,
851
+ "𑀱च𑀞𑀢ल": 106007,
852
+ "𑀱च𑀟": 106406,
853
+ "𑀱च𑀟𑀟च𑀟": 106220,
854
+ "𑀱च𑀟𑀢": 105975,
855
+ "𑀱च𑀟𑀣च": 106150,
856
+ "𑀱च𑀳": 106561,
857
+ "𑀱च𑀳च": 106118,
858
+ "𑀱च𑀳च𑀟": 106241,
859
+ "𑀱च𑀳न": 105967,
860
+ "𑀱न𑀪": 105977,
861
+ "𑀱न𑀪𑀢𑀟": 106022,
862
+ "𑀲": 106596,
863
+ "𑀲च": 106545,
864
+ "𑀲च𑀪": 106413,
865
+ "𑀲च𑀪च": 106358,
866
+ "𑀲च𑀪𑀞𑁣": 106072,
867
+ "𑀲च𑀪𑀞𑁣𑀟": 106518,
868
+ "𑀲च𑀳च": 106001,
869
+ "𑀲𑀢": 106075,
870
+ "𑀲𑀢प𑁣": 106746,
871
+ "𑀲𑀢ल": 106686,
872
+ "𑀲𑀢𑀟च": 106262,
873
+ "𑀲𑀢𑀠": 106139,
874
+ "𑀲𑁣𑀪": 106653,
875
+ "𑀳": 106331,
876
+ "𑀳च": 106211,
877
+ "𑀳चढ": 106303,
878
+ "𑀳चढ𑁣𑀣च": 106676,
879
+ "𑀳चन": 106026,
880
+ "𑀳चन𑀪च𑀟": 106760,
881
+ "𑀳चल": 106534,
882
+ "𑀳च𑀞𑁦": 106384,
883
+ "𑀳च𑀟": 105909,
884
+ "𑀳च𑀟च": 106043,
885
+ "𑀳च𑀟णच": 105951,
886
+ "𑀳च𑀟𑀟च𑀟": 106654,
887
+ "𑀳च𑀟𑀢": 106401,
888
+ "𑀳च𑀠": 105894,
889
+ "𑀳च𑀠च": 106749,
890
+ "𑀳च𑀠च𑀪": 106302,
891
+ "𑀳च𑀠न": 105902,
892
+ "𑀳च𑀠न𑀟": 106664,
893
+ "𑀳च𑀠𑀢": 105887,
894
+ "𑀳च𑀢": 106543,
895
+ "𑀳च𑀪": 106743,
896
+ "𑀳न": 106714,
897
+ "𑀳न𑀞च": 106346,
898
+ "𑀳न𑀟": 106056,
899
+ "𑀳न𑀟च": 106708,
900
+ "𑀳न𑀟च𑀟": 106379,
901
+ "𑀳𑀢ण": 106780,
902
+ "𑀳𑀢णच𑀳च": 106602,
903
+ "𑀳𑀫": 106422,
904
+ "𑀳𑀫च": 106305,
905
+ "𑀳𑀫च𑀪𑀢": 105938,
906
+ "𑀳𑀫च𑀲": 106309,
907
+ "𑀳𑀫नबच": 106063,
908
+ "𑀳𑀫नबचढच𑀟": 106446,
909
+ "𑀳𑀫𑀢": 106123,
910
+ "𑀳𑀫𑀢बच": 106702,
911
+ "𑀳𑀫𑀢𑀟𑁦": 106394,
912
+ "𑀳𑀫𑀢𑀪": 106650,
913
+ "𑀳𑀫𑁦": 106088,
914
+ "𑀳𑀫𑁦𑀞च𑀪": 106155,
915
+ "𑀳𑀫𑁦𑀞च𑀪च": 106316,
916
+ "𑀳𑀫𑁦𑀞च𑀪च𑀪": 106314,
917
+ "𑀳𑀫𑁦𑀞च𑀪न": 106636,
918
+ "𑀳𑁣": 106244,
919
+ "𑀳𑁦": 105946,
920
+ "𑁢": 105919,
921
+ "𑁣": 106125,
922
+ "𑁣𑀞": 106017,
923
+ "𑁣𑀲": 106245,
924
+ "𑁤": 106431,
925
+ "𑁥": 106424,
926
+ "𑁦": 106501
927
+ }
last-checkpoint/config.json ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "google-bert/bert-base-multilingual-uncased",
3
+ "architectures": [
4
+ "BertForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "directionality": "bidi",
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "id2label": {
13
+ "0": "LABEL_0",
14
+ "1": "LABEL_1",
15
+ "2": "LABEL_2",
16
+ "3": "LABEL_3",
17
+ "4": "LABEL_4"
18
+ },
19
+ "initializer_range": 0.02,
20
+ "intermediate_size": 3072,
21
+ "label2id": {
22
+ "LABEL_0": 0,
23
+ "LABEL_1": 1,
24
+ "LABEL_2": 2,
25
+ "LABEL_3": 3,
26
+ "LABEL_4": 4
27
+ },
28
+ "layer_norm_eps": 1e-12,
29
+ "max_position_embeddings": 512,
30
+ "model_type": "bert",
31
+ "num_attention_heads": 12,
32
+ "num_hidden_layers": 12,
33
+ "pad_token_id": 0,
34
+ "pooler_fc_size": 768,
35
+ "pooler_num_attention_heads": 12,
36
+ "pooler_num_fc_layers": 3,
37
+ "pooler_size_per_head": 128,
38
+ "pooler_type": "first_token_transform",
39
+ "position_embedding_type": "absolute",
40
+ "problem_type": "single_label_classification",
41
+ "torch_dtype": "float32",
42
+ "transformers_version": "4.41.2",
43
+ "type_vocab_size": 2,
44
+ "use_cache": true,
45
+ "vocab_size": 106804
46
+ }
last-checkpoint/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2d7b60a28934db5de15edd4b9af17a8e9a1a613be4aceb5b87391523e3de406d
3
+ size 672306188
last-checkpoint/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5fed01eb4bb307d16e1a3272e96fbd213d82ebddc8f4e6afa3d4a9bbf4c09b1d
3
+ size 1344733434
last-checkpoint/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9bdeb682a8898e3848e8e66378d3969d13e44b8b0f8f6b725f583a34797210a6
3
+ size 14244
last-checkpoint/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b44c59300d2d7f1a454fae0e4066f6b6797dd30a81acb9effd90e8eed3ea9e7e
3
+ size 1064
last-checkpoint/special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
last-checkpoint/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
last-checkpoint/tokenizer_config.json ADDED
The diff for this file is too large to render. See raw diff
 
last-checkpoint/trainer_state.json ADDED
@@ -0,0 +1,337 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.827338266658216,
3
+ "best_model_checkpoint": "./results/checkpoint-456",
4
+ "epoch": 19.0,
5
+ "eval_steps": 500,
6
+ "global_step": 456,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 1.0,
13
+ "grad_norm": 4.450119972229004,
14
+ "learning_rate": 9.5e-06,
15
+ "loss": 1.5028,
16
+ "step": 24
17
+ },
18
+ {
19
+ "epoch": 1.0,
20
+ "eval_f1": 0.38772036474164134,
21
+ "eval_loss": 1.346899151802063,
22
+ "eval_runtime": 0.1331,
23
+ "eval_samples_per_second": 1637.312,
24
+ "eval_steps_per_second": 30.042,
25
+ "step": 24
26
+ },
27
+ {
28
+ "epoch": 2.0,
29
+ "grad_norm": 5.728583335876465,
30
+ "learning_rate": 9e-06,
31
+ "loss": 1.2709,
32
+ "step": 48
33
+ },
34
+ {
35
+ "epoch": 2.0,
36
+ "eval_f1": 0.421720746842489,
37
+ "eval_loss": 1.1574269533157349,
38
+ "eval_runtime": 0.0998,
39
+ "eval_samples_per_second": 2183.766,
40
+ "eval_steps_per_second": 40.069,
41
+ "step": 48
42
+ },
43
+ {
44
+ "epoch": 3.0,
45
+ "grad_norm": 8.520730018615723,
46
+ "learning_rate": 8.5e-06,
47
+ "loss": 1.0793,
48
+ "step": 72
49
+ },
50
+ {
51
+ "epoch": 3.0,
52
+ "eval_f1": 0.5100182149362478,
53
+ "eval_loss": 0.9754979610443115,
54
+ "eval_runtime": 0.0959,
55
+ "eval_samples_per_second": 2274.042,
56
+ "eval_steps_per_second": 41.726,
57
+ "step": 72
58
+ },
59
+ {
60
+ "epoch": 4.0,
61
+ "grad_norm": 7.030355930328369,
62
+ "learning_rate": 8.000000000000001e-06,
63
+ "loss": 0.9111,
64
+ "step": 96
65
+ },
66
+ {
67
+ "epoch": 4.0,
68
+ "eval_f1": 0.6386053201801735,
69
+ "eval_loss": 0.8359459638595581,
70
+ "eval_runtime": 0.096,
71
+ "eval_samples_per_second": 2270.214,
72
+ "eval_steps_per_second": 41.655,
73
+ "step": 96
74
+ },
75
+ {
76
+ "epoch": 5.0,
77
+ "grad_norm": 4.342397689819336,
78
+ "learning_rate": 7.500000000000001e-06,
79
+ "loss": 0.7795,
80
+ "step": 120
81
+ },
82
+ {
83
+ "epoch": 5.0,
84
+ "eval_f1": 0.6719327398318995,
85
+ "eval_loss": 0.7689192295074463,
86
+ "eval_runtime": 0.0959,
87
+ "eval_samples_per_second": 2272.996,
88
+ "eval_steps_per_second": 41.706,
89
+ "step": 120
90
+ },
91
+ {
92
+ "epoch": 6.0,
93
+ "grad_norm": 7.117249488830566,
94
+ "learning_rate": 7e-06,
95
+ "loss": 0.6623,
96
+ "step": 144
97
+ },
98
+ {
99
+ "epoch": 6.0,
100
+ "eval_f1": 0.7363490153785023,
101
+ "eval_loss": 0.6683973670005798,
102
+ "eval_runtime": 0.1008,
103
+ "eval_samples_per_second": 2163.547,
104
+ "eval_steps_per_second": 39.698,
105
+ "step": 144
106
+ },
107
+ {
108
+ "epoch": 7.0,
109
+ "grad_norm": 5.1266303062438965,
110
+ "learning_rate": 6.5000000000000004e-06,
111
+ "loss": 0.5661,
112
+ "step": 168
113
+ },
114
+ {
115
+ "epoch": 7.0,
116
+ "eval_f1": 0.7721861083416417,
117
+ "eval_loss": 0.6161485314369202,
118
+ "eval_runtime": 0.0971,
119
+ "eval_samples_per_second": 2243.984,
120
+ "eval_steps_per_second": 41.174,
121
+ "step": 168
122
+ },
123
+ {
124
+ "epoch": 8.0,
125
+ "grad_norm": 5.100093364715576,
126
+ "learning_rate": 6e-06,
127
+ "loss": 0.4879,
128
+ "step": 192
129
+ },
130
+ {
131
+ "epoch": 8.0,
132
+ "eval_f1": 0.7929782861547567,
133
+ "eval_loss": 0.5681259632110596,
134
+ "eval_runtime": 0.0977,
135
+ "eval_samples_per_second": 2230.333,
136
+ "eval_steps_per_second": 40.924,
137
+ "step": 192
138
+ },
139
+ {
140
+ "epoch": 9.0,
141
+ "grad_norm": 4.375860691070557,
142
+ "learning_rate": 5.500000000000001e-06,
143
+ "loss": 0.4076,
144
+ "step": 216
145
+ },
146
+ {
147
+ "epoch": 9.0,
148
+ "eval_f1": 0.7979355926308811,
149
+ "eval_loss": 0.5458353161811829,
150
+ "eval_runtime": 0.0962,
151
+ "eval_samples_per_second": 2265.798,
152
+ "eval_steps_per_second": 41.574,
153
+ "step": 216
154
+ },
155
+ {
156
+ "epoch": 10.0,
157
+ "grad_norm": 5.850133419036865,
158
+ "learning_rate": 5e-06,
159
+ "loss": 0.359,
160
+ "step": 240
161
+ },
162
+ {
163
+ "epoch": 10.0,
164
+ "eval_f1": 0.8038855056192519,
165
+ "eval_loss": 0.5337401032447815,
166
+ "eval_runtime": 0.0993,
167
+ "eval_samples_per_second": 2194.426,
168
+ "eval_steps_per_second": 40.265,
169
+ "step": 240
170
+ },
171
+ {
172
+ "epoch": 11.0,
173
+ "grad_norm": 5.326167106628418,
174
+ "learning_rate": 4.5e-06,
175
+ "loss": 0.3049,
176
+ "step": 264
177
+ },
178
+ {
179
+ "epoch": 11.0,
180
+ "eval_f1": 0.808985274271652,
181
+ "eval_loss": 0.5208148956298828,
182
+ "eval_runtime": 0.0986,
183
+ "eval_samples_per_second": 2210.709,
184
+ "eval_steps_per_second": 40.563,
185
+ "step": 264
186
+ },
187
+ {
188
+ "epoch": 12.0,
189
+ "grad_norm": 4.419217109680176,
190
+ "learning_rate": 4.000000000000001e-06,
191
+ "loss": 0.2707,
192
+ "step": 288
193
+ },
194
+ {
195
+ "epoch": 12.0,
196
+ "eval_f1": 0.7915964367931138,
197
+ "eval_loss": 0.5140932202339172,
198
+ "eval_runtime": 0.1131,
199
+ "eval_samples_per_second": 1928.281,
200
+ "eval_steps_per_second": 35.381,
201
+ "step": 288
202
+ },
203
+ {
204
+ "epoch": 13.0,
205
+ "grad_norm": 3.2652978897094727,
206
+ "learning_rate": 3.5e-06,
207
+ "loss": 0.2426,
208
+ "step": 312
209
+ },
210
+ {
211
+ "epoch": 13.0,
212
+ "eval_f1": 0.7960978274579872,
213
+ "eval_loss": 0.5064918398857117,
214
+ "eval_runtime": 0.0963,
215
+ "eval_samples_per_second": 2262.854,
216
+ "eval_steps_per_second": 41.52,
217
+ "step": 312
218
+ },
219
+ {
220
+ "epoch": 14.0,
221
+ "grad_norm": 8.18807315826416,
222
+ "learning_rate": 3e-06,
223
+ "loss": 0.2111,
224
+ "step": 336
225
+ },
226
+ {
227
+ "epoch": 14.0,
228
+ "eval_f1": 0.8024928631739778,
229
+ "eval_loss": 0.5025272965431213,
230
+ "eval_runtime": 0.1145,
231
+ "eval_samples_per_second": 1904.663,
232
+ "eval_steps_per_second": 34.948,
233
+ "step": 336
234
+ },
235
+ {
236
+ "epoch": 15.0,
237
+ "grad_norm": 3.970165491104126,
238
+ "learning_rate": 2.5e-06,
239
+ "loss": 0.1919,
240
+ "step": 360
241
+ },
242
+ {
243
+ "epoch": 15.0,
244
+ "eval_f1": 0.8067770829812405,
245
+ "eval_loss": 0.5004199147224426,
246
+ "eval_runtime": 0.0966,
247
+ "eval_samples_per_second": 2257.346,
248
+ "eval_steps_per_second": 41.419,
249
+ "step": 360
250
+ },
251
+ {
252
+ "epoch": 16.0,
253
+ "grad_norm": 6.731950283050537,
254
+ "learning_rate": 2.0000000000000003e-06,
255
+ "loss": 0.1764,
256
+ "step": 384
257
+ },
258
+ {
259
+ "epoch": 16.0,
260
+ "eval_f1": 0.8175947589100725,
261
+ "eval_loss": 0.4881379008293152,
262
+ "eval_runtime": 0.0953,
263
+ "eval_samples_per_second": 2287.302,
264
+ "eval_steps_per_second": 41.969,
265
+ "step": 384
266
+ },
267
+ {
268
+ "epoch": 17.0,
269
+ "grad_norm": 3.039834976196289,
270
+ "learning_rate": 1.5e-06,
271
+ "loss": 0.1694,
272
+ "step": 408
273
+ },
274
+ {
275
+ "epoch": 17.0,
276
+ "eval_f1": 0.8178525944285818,
277
+ "eval_loss": 0.49215319752693176,
278
+ "eval_runtime": 0.1114,
279
+ "eval_samples_per_second": 1956.081,
280
+ "eval_steps_per_second": 35.891,
281
+ "step": 408
282
+ },
283
+ {
284
+ "epoch": 18.0,
285
+ "grad_norm": 5.969228744506836,
286
+ "learning_rate": 1.0000000000000002e-06,
287
+ "loss": 0.155,
288
+ "step": 432
289
+ },
290
+ {
291
+ "epoch": 18.0,
292
+ "eval_f1": 0.8217110120593946,
293
+ "eval_loss": 0.4937782883644104,
294
+ "eval_runtime": 0.0976,
295
+ "eval_samples_per_second": 2232.832,
296
+ "eval_steps_per_second": 40.969,
297
+ "step": 432
298
+ },
299
+ {
300
+ "epoch": 19.0,
301
+ "grad_norm": 3.987272262573242,
302
+ "learning_rate": 5.000000000000001e-07,
303
+ "loss": 0.147,
304
+ "step": 456
305
+ },
306
+ {
307
+ "epoch": 19.0,
308
+ "eval_f1": 0.827338266658216,
309
+ "eval_loss": 0.4907718598842621,
310
+ "eval_runtime": 0.096,
311
+ "eval_samples_per_second": 2269.814,
312
+ "eval_steps_per_second": 41.648,
313
+ "step": 456
314
+ }
315
+ ],
316
+ "logging_steps": 500,
317
+ "max_steps": 480,
318
+ "num_input_tokens_seen": 0,
319
+ "num_train_epochs": 20,
320
+ "save_steps": 500,
321
+ "stateful_callbacks": {
322
+ "TrainerControl": {
323
+ "args": {
324
+ "should_epoch_stop": false,
325
+ "should_evaluate": false,
326
+ "should_log": false,
327
+ "should_save": true,
328
+ "should_training_stop": false
329
+ },
330
+ "attributes": {}
331
+ }
332
+ },
333
+ "total_flos": 711531069926808.0,
334
+ "train_batch_size": 64,
335
+ "trial_name": null,
336
+ "trial_params": null
337
+ }
last-checkpoint/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d8de40f27c74c1baf574f62e81d0dc7e9ff066d1a50a2ab5bf7d54ee24164f8a
3
+ size 5112
last-checkpoint/vocab.txt ADDED
The diff for this file is too large to render. See raw diff