mujadid-syahbana commited on
Commit
ea6f1e0
1 Parent(s): d4fcc17

Training in progress, epoch 1

Browse files
config.json ADDED
@@ -0,0 +1,115 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "mujadid-syahbana/whisper-small-qur-base",
3
+ "activation_dropout": 0.0,
4
+ "activation_function": "gelu",
5
+ "apply_spec_augment": false,
6
+ "architectures": [
7
+ "WhisperForAudioClassification"
8
+ ],
9
+ "attention_dropout": 0.0,
10
+ "begin_suppress_tokens": [
11
+ 220,
12
+ 50257
13
+ ],
14
+ "bos_token_id": 50257,
15
+ "classifier_proj_size": 256,
16
+ "d_model": 512,
17
+ "decoder_attention_heads": 8,
18
+ "decoder_ffn_dim": 2048,
19
+ "decoder_layerdrop": 0.0,
20
+ "decoder_layers": 6,
21
+ "decoder_start_token_id": 50258,
22
+ "dropout": 0.0,
23
+ "encoder_attention_heads": 8,
24
+ "encoder_ffn_dim": 2048,
25
+ "encoder_layerdrop": 0.0,
26
+ "encoder_layers": 6,
27
+ "eos_token_id": 50257,
28
+ "finetuning_task": "audio-classification",
29
+ "forced_decoder_ids": null,
30
+ "id2label": {
31
+ "0": "alif",
32
+ "1": "seen",
33
+ "10": "wao",
34
+ "11": "kaif",
35
+ "12": "duad",
36
+ "13": "faa",
37
+ "14": "tua",
38
+ "15": "suad",
39
+ "16": "ba",
40
+ "17": "noon",
41
+ "18": "yaa",
42
+ "19": "hha",
43
+ "2": "zua",
44
+ "20": "hamzah",
45
+ "21": "sheen",
46
+ "22": "meem",
47
+ "23": "quaf",
48
+ "24": "zhal",
49
+ "25": "sa",
50
+ "26": "jeem",
51
+ "27": "aain",
52
+ "28": "zaa",
53
+ "3": "ta",
54
+ "4": "kha",
55
+ "5": "dal",
56
+ "6": "haa",
57
+ "7": "laam",
58
+ "8": "raa",
59
+ "9": "ghain"
60
+ },
61
+ "init_std": 0.02,
62
+ "is_encoder_decoder": true,
63
+ "label2id": {
64
+ "aain": "27",
65
+ "alif": "0",
66
+ "ba": "16",
67
+ "dal": "5",
68
+ "duad": "12",
69
+ "faa": "13",
70
+ "ghain": "9",
71
+ "haa": "6",
72
+ "hamzah": "20",
73
+ "hha": "19",
74
+ "jeem": "26",
75
+ "kaif": "11",
76
+ "kha": "4",
77
+ "laam": "7",
78
+ "meem": "22",
79
+ "noon": "17",
80
+ "quaf": "23",
81
+ "raa": "8",
82
+ "sa": "25",
83
+ "seen": "1",
84
+ "sheen": "21",
85
+ "suad": "15",
86
+ "ta": "3",
87
+ "tua": "14",
88
+ "wao": "10",
89
+ "yaa": "18",
90
+ "zaa": "28",
91
+ "zhal": "24",
92
+ "zua": "2"
93
+ },
94
+ "mask_feature_length": 10,
95
+ "mask_feature_min_masks": 0,
96
+ "mask_feature_prob": 0.0,
97
+ "mask_time_length": 10,
98
+ "mask_time_min_masks": 2,
99
+ "mask_time_prob": 0.05,
100
+ "max_length": 448,
101
+ "max_source_positions": 1500,
102
+ "max_target_positions": 448,
103
+ "median_filter_width": 7,
104
+ "model_type": "whisper",
105
+ "num_hidden_layers": 6,
106
+ "num_mel_bins": 80,
107
+ "pad_token_id": 50257,
108
+ "scale_embedding": false,
109
+ "suppress_tokens": [],
110
+ "torch_dtype": "float32",
111
+ "transformers_version": "4.36.0.dev0",
112
+ "use_cache": true,
113
+ "use_weighted_layer_sum": false,
114
+ "vocab_size": 51865
115
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:24cc56fd64b5518197381a0e7b48c8aca92f7b2ba6ce18d1e1b7db8761d4873a
3
+ size 82928036
preprocessor_config.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "chunk_length": 30,
3
+ "feature_extractor_type": "WhisperFeatureExtractor",
4
+ "feature_size": 80,
5
+ "hop_length": 160,
6
+ "n_fft": 400,
7
+ "n_samples": 480000,
8
+ "nb_max_frames": 3000,
9
+ "padding_side": "right",
10
+ "padding_value": 0.0,
11
+ "processor_class": "WhisperProcessor",
12
+ "return_attention_mask": false,
13
+ "sampling_rate": 16000
14
+ }
runs/Nov06_14-09-09_202e33c68888/events.out.tfevents.1699279757.202e33c68888.642.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d047c25232897e45dd45b5339840242740600f16c588b0a4f69a4301b08f6a8a
3
+ size 6502
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ba04a5b5cf58e4af1aa8726a6fbfdda9f320cbe569df983ff4a1b9f5c478cade
3
+ size 4600