anim400k / README.md
davidchan's picture
Update README.md
272b1a4 verified
metadata
task_categories:
  - text-to-speech
  - automatic-speech-recognition
  - audio-to-audio
  - audio-classification
  - text-classification
  - text2text-generation
  - video-classification
  - summarization
language:
  - en
  - ja
pretty_name: Anim-400K
size_categories:
  - 100K<n<1M
extra_gated_prompt: >-
  Due to copyright limitations, all prospect users of the Anim400K Datasets must
  sign both a Terms of Use Agreement (TOU) and a Non Disclosure Agreement (NDA).
  The form MUST be filled out for all users of the dataset. The answers to this
  form will auto-complete and sign the template TOU
  (https://docs.google.com/document/d/1MNAU12i4XIXj8O8ThUuep8jJw5WOGArCvZdcR2UmJGM/edit?usp=sharing)
  and NDA
  (https://docs.google.com/document/d/1cLtFX2GarMEzZn5RwL-gEuAthS1qaKOfb3r5m1icOX0/edit?usp=sharing)
extra_gated_fields:
  Full Name: text
  Email: text
  Researcher Google Scholar Page: text
  Affiliation: text
  Name of Principal Investigator or Supervisor: text
  Principle Investigator/Supervisior Google Scholar Page: text
  Purpose of Intended Use: text
  I understand that the Anim400K team and the Regents of the University of California make no warranties, express or implied,  regarding  the  Dataset,  including  but  not  limited  to  being  up-to-­date,  correct  or  complete․  ​Neither the Anim400K team nor the Regents of the University of California can be  held  liable  for  providing  access  to  the  Dataset  ​or  usage  of  the  ​Dataset: checkbox
  I understand and agree that the use of this Dataset is for scientific or research purposes only․ Any other use is explicitly prohibited: checkbox
  I understand and agree that this Dataset and the videos are protected by copyrights: checkbox
  I understand and agree not to share the dataset with any third party: checkbox
  I understand and agree that I (as the researcher)  takes  full  responsibility  for  usage  of    the  Dataset  and  processing  the  Dataset: checkbox
  I have read, and I agree and sign the Non Disclosure Agreement: checkbox
  I have read, and I agree and sign the Terms of Use Agreement: checkbox

Anim-400K: A dataset designed from the ground up for automated dubbing of video

image

What is Anim-400K?

Anim-400K is a large-scale dataset of aligned audio-video clips in both the English and Japanese languages. It is comprised of over 425K aligned clips (763 hours) consisting of both video and audio drawn from over 190 properties covering hundreds of themes and genres. Anim400K is further augmented with metadata including genres, themes, show-ratings, character profiles, and animation styles at a property level, episode synopses, ratings, and subtitles at an episode level, and pre-computed ASR at an aligned clip level to enable in-depth research into several audio-visual tasks.

Read the [ArXiv Preprint]

Check us out on [Github]

News

[January 2024] Anim-400K (v1) available on Huggingface Datasets.
[January 2024] Anim-400K (v1) release.
[January 2024] Anim-400K (v1) accepted at ICASSP2024.

Using the Data

Use the Huggingface CLI/tools to download the data. The tar.gz files are provided as .tar.gz-part files, which need to be assembled before unzipping. This can be done following the instructions here.

Citation

If any part of our paper is helpful to your work, please cite with:

@inproceedings{cai2024anim400k,
  title={ANIM-400K: A Large-Scale Dataset for Automated End to End Dubbing of Video},
  author={Cai, Kevin and Liu, Chonghua and Chan, David M.},
  booktitle={ICASSP 2024-2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)},
  pages={1--5},
  year={2024},
  organization={IEEE}
}

Acknowledgements

This repository, and data release model is modeled on that used by the MAD dataset.