Backpack
|
Keyboard
|
Pillow
|
Ladder
|
paper
|
Keyboard
|
shoes
|
Trash Can
|
TL;DR: We propose a lightweight and scalable regional point-language contrastive learning framework for open-world 3D scene understanding.
We propose a lightweight and scalable Regional Point-Language Contrastive learning framework, namely RegionPLC, for open-world 3D scene understanding, aiming to identify and recognize open-set objects and categories. Specifically, based on our empirical studies, we introduce a 3D-aware SFusion strategy that fuses 3D vision-language pairs derived from multiple 2D foundation models, yielding high-quality, dense region-level language descriptions without human 3D annotations. Subsequently, we devise a region-aware point-discriminative contrastive learning objective to enable robust and effective 3D learning from dense regional language supervision. We carry out extensive experiments on ScanNet, ScanNet200, and nuScenes datasets, and our model outperforms prior 3D open-world scene understanding approaches by an average of 17.2% and 9.1% for semantic and instance segmentation, respectively, while maintaining greater scalability and lower resource demands. Furthermore, our method has the flexibility to be effortlessly integrated with language models to enable open-ended grounded 3D reasoning without extra task-specific training.
@article{yang2023regionplc,
title={RegionPLC: Regional Point-Language Contrastive Learning for Open-World 3D Scene Understanding},
author={Yang, Jihan and Ding, Runyu and Wang, Zhe and Qi, Xiaojuan},
journal={arXiv preprint arXiv:2304.00962},
year={2023}
}
}