File size: 1,493 Bytes
092d812 44600eb 092d812 4222f27 092d812 4222f27 092d812 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 |
# ################################
# Model: Whisper (Encoder-Decoder) + NLL
# Augmentation: TimeDomainSpecAugment
# Authors: Pooneh Mousavi 2022
# ################################
# URL for the biggest Fairseq english whisper model.
whisper_hub: openai/whisper-large-v2
# Normalize inputs with
# the same normalization done in the paper. Refer to Appendix C for further information.
normalized_transcripts: True
language: mongolian
auto_mix_prec: False
sample_rate: 16000
# Decoding parameters
min_decode_ratio: 0.0
max_decode_ratio: 0.1
test_beam_size: 8
# Model parameters
freeze_whisper: True
freeze_encoder: True
whisper: !new:speechbrain.lobes.models.huggingface_transformers.whisper.Whisper
source: !ref <whisper_hub>
freeze: !ref <freeze_whisper>
freeze_encoder: !ref <freeze_encoder>
save_path: whisper_checkpoints
encoder_only: False
decoder: !new:speechbrain.decoders.seq2seq.S2SWhisperGreedySearcher
model: !ref <whisper>
min_decode_ratio: !ref <min_decode_ratio>
max_decode_ratio: !ref <max_decode_ratio>
# test_beam_searcher: !new:speechbrain.decoders.seq2seq.S2SWhisperBeamSearcher
# module: [!ref <whisper>]
# min_decode_ratio: !ref <min_decode_ratio>
# max_decode_ratio: !ref <max_decode_ratio>
# beam_size: !ref <test_beam_size>
modules:
whisper: !ref <whisper>
decoder: !ref <decoder>
pretrainer: !new:speechbrain.utils.parameter_transfer.Pretrainer
loadables:
whisper: !ref <whisper>
|