volcano-7b / README.md
Seongyun's picture
Update README.md
2095c91
|
raw
history blame
1.95 kB
---
tags:
- image-to-text
- visual-question-answering
- image-captioning
datasets:
- kaist-ai/volcano-train
language:
- en
pipeline_tag: image-to-text
library_name: transformers
---
## Links for Reference
- **Repository: https://github.com/kaistAI/Volcano**
- **Paper: https://arxiv.org/abs/2311.07362**
# Overview
![image/png](https://cdn-uploads.huggingface.co/production/uploads/6550c4f27bbfce1878f5f280/AnqbCNf6pRiQ_5uNX0r4d.png)
Volcano employs a single LMM to generate initial responses, feedback, and revisions, as well as decisions to accept revisions. It follows a sequential procedure of an iterative critique-revision-decide loop.
# Model details
**Model type:**
Volcano-7b is a multimodal self-feedback guided revision model that was fine-tuned by mixing the visual instruction tuning dataset used in [LLaVA-v1.5](https://llava-vl.github.io/) with multimodal feedback and revision data collected through [gpt-3.5-turbo](https://platform.openai.com/docs/models/gpt-3-5), applied to the [vicuna-7b-v1.5](https://huggingface.co./lmsys/vicuna-7b-v1.5) model.
**Model date:**
Volcano-7b was trained in October 2023.
# Training dataset
- **274K multimodal feedback and revision data**
- 558K filtered image-text pairs from LAION/CC/SBU, captioned by BLIP.
- 158K GPT-generated multimodal instruction-following data.
- 450K academic-task-oriented VQA data mixture.
- 40K ShareGPT data
You can find [here](https://huggingface.co./datasets/kaist-ai/volcano-train) the dataset used to train Volcano, which includes all the aforementioned datasets.
# Evaluation dataset
A collection of three multimodal hallucination benchmarks ([MMHal-Bench](https://huggingface.co./datasets/Shengcao1006/MMHal-Bench), [Pope](https://github.com/RUCAIBox/POPE), [GAVIE](https://github.com/FuxiaoLiu/LRV-Instruction)) and two multimodal understanding benchmarks ([MM-Vet](https://github.com/yuweihao/MM-Vet), [MMBench](https://github.com/open-compass/MMBench)).