3D-GRAND: A Million-Scale Dataset for 3D-LLMs with Better Grounding and Less Hallucination
Abstract
The integration of language and 3D perception is crucial for developing embodied agents and robots that comprehend and interact with the physical world. While large language models (LLMs) have demonstrated impressive language understanding and generation capabilities, their adaptation to 3D environments (3D-LLMs) remains in its early stages. A primary challenge is the absence of large-scale datasets that provide dense grounding between language and 3D scenes. In this paper, we introduce 3D-GRAND, a pioneering large-scale dataset comprising 40,087 household scenes paired with 6.2 million densely-grounded scene-language instructions. Our results show that instruction tuning with 3D-GRAND significantly enhances grounding capabilities and reduces hallucinations in 3D-LLMs. As part of our contributions, we propose a comprehensive benchmark 3D-POPE to systematically evaluate hallucination in 3D-LLMs, enabling fair comparisons among future models. Our experiments highlight a scaling effect between dataset size and 3D-LLM performance, emphasizing the critical role of large-scale 3D-text datasets in advancing embodied AI research. Notably, our results demonstrate early signals for effective sim-to-real transfer, indicating that models trained on large synthetic data can perform well on real-world 3D scans. Through 3D-GRAND and 3D-POPE, we aim to equip the embodied AI community with essential resources and insights, setting the stage for more reliable and better-grounded 3D-LLMs. Project website: https://3d-grand.github.io
Community
๐ฅ 3D-LLMs go brrrr! ๐ Excited to announce our latest research on scaling 3D-LLM training data to million-scale with dense grounding.
๐ Introducing 3D-GRAND: a pioneering dataset featuring 40,087 household scenes paired with 6.2 million densely-grounded 3D-text pairs. ๐ ๐ฌ https://3d-grand.github.io
๐ We envision 3D-GRAND to be the bedrock for future 3D-LLMs! ๐ ๐ฌ
- 6.2 million instructions + 40k 3D household scenes ๐ฅ
- Significantly enhances grounding & reduces hallucinations for 3D-LLMs ๐
- 3D-POPE: the first benchmark for systematic evaluation of hallucinations in 3D-LLMs ๐ฏ
- Data scaling law and sim-to-real transfer provide strong early signals for a low-cost, scalable future for 3D-LLMs ๐
๐ What's special about this data? ๐ค
- Dense Grounding: Unlike traditional 3D-text datasets, ours connects every noun to an object in the 3D world. ๐ ๐
- Large-scale: We provide million-scale data, bridging the gap between 3D and 2D datasets. ๐
- Diverse Tasks: Curated 8 diverse tasks to cover future 3D-LLM challenges. ๐
- Hallucination: Special attention was given to curate a balanced dataset to help reduce hallucinations & Introduced a benchmark for evaluating hallucinations in 3D-LLMs. ๐ง ๐
๐ Results of 3D-LLMs trained on 3D-GRAND:
- Stronger Grounding
- Less Hallucination (huge improvement over prev. 3D-LLMs)
- Data Scaling Law: More data -> better performance. ๐
- Sim-to-real Transfer: Trained on synthetic 3D scenes -> effective transfer to real 3D scans in ScanNet. ๐
๐ Let's build better 3D-LLMs together! ๐
๐ Paper: http://arxiv.org/abs/2406.05132
๐ Website & Data: http://3d-grand.github.io
๐ป Demo: http://huggingface.co/spaces/jedyang97/3D-GRAND
๐ 3D-POPE Leaderboard: http://huggingface.co/spaces/sled-umich/3D-POPE-leaderboard
๐ง Code: http://github.com/sled-group/3D-GRAND
Thanks for sharing all details!! This helps : )
Models citing this paper 0
No model linking this paper