|
--- |
|
license: afl-3.0 |
|
--- |
|
|
|
# Dataset Card for ImageCoDe |
|
|
|
To get started quickly, load descriptions via: |
|
``` |
|
from datasets import load_dataset |
|
examples = load_dataset('BennoKrojer/ImageCoDe') |
|
``` |
|
And download `image_sets.zip` for all images sets (each directory consisting of 10 images). |
|
|
|
## Dataset Description |
|
|
|
- **Homepage & Leaderboard:** https://mcgill-nlp.github.io/imagecode/ |
|
- **Repository:** https://github.com/McGill-NLP/imagecode |
|
- **Paper:** https://arxiv.org/abs/2203.15867 |
|
- **Point of Contact:** benno DOT krojer ÄT gmail DOT com |
|
|
|
### Dataset Summary |
|
|
|
We introduce ImageCoDe, a vision-and-language benchmark that requires contextual language understanding in the form of pragmatics, temporality, long descriptions and visual nuances. The task: Given a detailed description, retrieve the target image among 10 minimally contrastive images. ImageCoDe contains 21K descriptions and 94K images. THe images are primarily frames based on video datasets. |
|
|
|
## Dataset Structure |
|
|
|
### Data Instances |
|
|
|
An instance contains a description, the corresponding image set name, and the target index: |
|
``` |
|
{"image_set": "video-storytelling-videowedding_de8dLXvgV-I-shot6_0", |
|
"image_index": "8", |
|
"description": "The flowers the woman in the teal strapless dress is carrying are completely obscured by the man in the black shirt's head. "} |
|
``` |
|
### Data Splits |
|
|
|
| Dataset Split | Number of Descriptions in Split | |
|
| ------------- |----------------------------- | |
|
| Train | 16,594 | |
|
| Validation | 2,302 | |
|
| Test | 2,306 | |
|
|
|
## Dataset Creation |
|
|
|
### Curation Rationale |
|
|
|
The main goal of ImageCoDe is to highlight weaknesses of recent Vision-and-Language models regarding complex language and fine-grained visual representations. In addition, we found that the dataset offers plenty of pragmatic examples and is therefore suitable for studying pragmatics. |