[Sample data]: This contains 167MB sample data
[InSpaceType Eval set]: This contains 1260 RGBD pairs for evaluation use about 11.5G. For evaluation please go to our codebase
[InSpaceType all data]:This contains 40K RGBD pairs, about 500G the whole InSpaceType dataset. The whole data is split into 8 chunks. Please download all chunks in the folder and extract them.
This work introduces a dataset and benchmark that reconsiders an important but usually overlooked factor- space type. We detailedly analyze 12 SOTA models and four popular training dataset to unveil their potential biases. Further, we study cross-type generalization and domain generalization techniques.
Indoor monocular depth estimation attracted higher research interest in indoor robots to aid navigation and perception. Most previous methods primarily experiment with NYUv2 Dataset and concentrate on the overall evaluation performance. However, little is known regarding robustness and generalization in the real world where highly varying and diverse functional \textit{space types}, such as library or kitchen, are present as tailed types. This work studies the common but easily overlooked factor- space type and realizes a model's performance variance. We present InSpaceType Dataset, a high-quality and high-resolution RGBD dataset for general indoor environments. We study 12 recent methods on InSpaceType and find most of them severely suffer from performance imbalance between head and tailed types and reveal their underlying bias. We extend analysis to total 4 datasets and organize their characteristics to enlighten further research directions on proper usage of them. Further, we study interplays between types and generalization to unseen spaces. Our work marks the first in-depth investigation of performance variance across space types and, more importantly, releases useful tools, including datasets and codes, to closely examine a given pretrained model.