Interleaved Scene Graph for Interleaved Text-and-Image Generation Assessment
Abstract
Many real-world user queries (e.g. "How do to make egg fried rice?") could benefit from systems capable of generating responses with both textual steps with accompanying images, similar to a cookbook. Models designed to generate interleaved text and images face challenges in ensuring consistency within and across these modalities. To address these challenges, we present ISG, a comprehensive evaluation framework for interleaved text-and-image generation. ISG leverages a scene graph structure to capture relationships between text and image blocks, evaluating responses on four levels of granularity: holistic, structural, block-level, and image-specific. This multi-tiered evaluation allows for a nuanced assessment of consistency, coherence, and accuracy, and provides interpretable question-answer feedback. In conjunction with ISG, we introduce a benchmark, ISG-Bench, encompassing 1,150 samples across 8 categories and 21 subcategories. This benchmark dataset includes complex language-vision dependencies and golden answers to evaluate models effectively on vision-centric tasks such as style transfer, a challenging area for current models. Using ISG-Bench, we demonstrate that recent unified vision-language models perform poorly on generating interleaved content. While compositional approaches that combine separate language and image models show a 111% improvement over unified models at the holistic level, their performance remains suboptimal at both block and image levels. To facilitate future work, we develop ISG-Agent, a baseline agent employing a "plan-execute-refine" pipeline to invoke tools, achieving a 122% performance improvement.
Community
TLDR: This paper introduces a fine-grained automatic evaluation framework and a new benchmark for interleaved text-and-image generation, offering valuable insights for future research in unified models and compositional frameworks that can generate interleaved content.
Website: https://interleave-eval.github.io/
GitHub: https://github.com/Dongping-Chen/ISG
ArXIv: https://arxiv.org/abs/2411.17188
Dataset: https://huggingface.co./datasets/shuaishuaicdp/ISG-Bench
Thanks for your interest in our paper!
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- What Makes a Scene ? Scene Graph-based Evaluation and Feedback for Controllable Generation (2024)
- GATE OpenING: A Comprehensive Benchmark for Judging Open-ended Interleaved Image-Text Generation (2024)
- Conditional Text-to-Image Generation with Reference Guidance (2024)
- Precision or Recall? An Analysis of Image Captions for Training Text-to-Image Generation Model (2024)
- From Words to Structured Visuals: A Benchmark and Framework for Text-to-Diagram Generation and Editing (2024)
- MMIE: Massive Multimodal Interleaved Comprehension Benchmark for Large Vision-Language Models (2024)
- Text-to-Image Synthesis: A Decade Survey (2024)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 1
Spaces citing this paper 0
No Space linking this paper