SarwarShafee's picture
Update README.md
2a87aff
|
raw
history blame
3.87 kB
metadata
language:
  - bn
license: cc-by-nc-4.0
task_categories:
  - automatic-speech-recognition
dataset_info:
  features:
    - name: audio
      dtype: audio
    - name: text
      dtype: string
    - name: duration
      dtype: float64
    - name: category
      dtype: string
    - name: source
      dtype: string
  splits:
    - name: train
      num_bytes: 219091915.875
      num_examples: 1753
  download_size: 214321460
  dataset_size: 219091915.875
configs:
  - config_name: default
    data_files:
      - split: train
        path: data/train-*

MegaBNSpeech

To evaluate the performance of the models, we used four test sets. Two of these were developed as part of the MegaBNSpeech corpus, while the remaining two (Fleurs and Common Voice) are commonly used test sets that are widely recognized by the speech community.

How to use:

The datasets library provides the capability to load and process your dataset efficiently using just Python. You can easily download and set up the dataset on your local drive with a single call using the load_dataset function.

from datasets import load_dataset
dataset = load_dataset("hishab/MegaBNSpeech", split="train")

With the datasets library, you have the option to stream the dataset in real-time by appending the streaming=True parameter to the load_dataset function. In streaming mode, the dataset loads one sample at a time instead of storing the whole dataset on the disk.

from datasets import load_dataset
dataset = load_dataset("hishab/MegaBNSpeech", split="train", streaming=True)
print(next(iter(dataset)))

Speech Recognition (ASR)

from datasets import load_dataset

mega_bn_asr = load_dataset("hishab/MegaBNSpeech")

# see structure
print(mega_bn_asr)

# load audio sample on the fly
audio_input = mega_bn_asr["train"][0]["audio"]  # first decoded audio sample
transcription = mega_bn_asr["train"][0]["transcription"]  # first transcription
# use `audio_input` and `transcription` to fine-tune your model for ASR

Data Structure

  • The dataset was developed using a pseudo-labeling approach.
  • The largest collection of Bangla audio-video data was curated and cleaned from various Bangla TV channels on YouTube. This data covers varying domains, speaking styles, dialects, and communication channels.
  • Alignments from two ASR systems were leveraged to segment and automatically annotate the audio segments.
  • The created dataset was used to design an end-to-end state-of-the-art Bangla ASR system.

Data Instances

  • Size of downloaded dataset files: ___ GB
  • Size of the generated dataset: ___ MB
  • Total amount of disk used: ___ GB

An example of a data instance looks as follows:

 {
  "id": 0,
  "audio_path": "data/train/wav/UCPREnbhKQP-hsVfsfKP-mCw_id_2kux6rFXMeM_85.wav",
  "transcription": "পরীক্ষার মূল্য তালিকা উন্মুক্ত স্থানে প্রদর্শনের আদেশ দেন এই আদেশ পাওয়ার",
  "duration": 5.055
 }

Data Fields

The data fields are written below.

  • id (int): ID of audio sample
  • audio_path (str): Path to the audio file
  • transcription (str): Transcription of the audio file
  • duration : 5.055

Dataset Creation

The dataset was developed using a pseudo-labeling approach. An extensive, large-scale, and high-quality speech dataset of approximately 20,000 hours was developed for domain-agnostic Bangla ASR.

Social Impact of Dataset

Limitations

Citation Information

You can access the MegaBNSpeech paper at _________________ Please cite the paper when referencing the MegaBNSpeech corpus as:

@article{_______________,
  title = {_______________________________},
  author = {___,___,___,___,___,___,___,___},
  journal={_______________________________},
  url = {_________________________________},
  year = {2023},