Home Abstract Demo Paper Code

Abstract

Despite that significant progress has been made in room layout estimation, most methods aim to reduce the loss in the 2D pixel coordinate rather than exploiting the room structure in the 3D space. Towards reconstructing the room layout in 3D, we formulate the task of 360˚ layout estimation as a problem of predicting depth on the horizon line of a panorama. Specifically, we propose the Differentiable Depth Rendering procedure to make the conversion from layout to depth prediction differentiable, thus making our proposed model end-to-end trainable while leveraging the 3D geometric information, without the need of providing the ground truth depth. We show that our method achieves state-of-the-art performance on numerous 360˚ layout benchmark datasets. Moreover, our formulation enables a pre-training step on the depth dataset, which further improves the generalizability of our layout estimation model.

Upload your own panorama!

You can capture your own panorama, and upload it to see the 3D result.

* When capturing, try your best to keep the camera parallel to the ground like the demo figures below. This will improve the accuracy!

* It takes about 10 seconds. If you wait longer than 20 sec, our server may be dead! RIP my server...

Image Format: jpg or png.

Size limit: 2MB

Online Demo

Click the link for 3D visualization!

CVPR 2021 Oral

LED2-Net: Monocular 360˚ Layout Estimation via Differentiable Depth

Fu-En Wang*, Yu-Hsuan Yeh*, Min Sun, Wei-Chen Chiu, Yi-Hsuan Tsai

Links

[Paper (arXiv)] [Supplementary Marterial] [Code]

Citation

@misc{wang2021led2net,
      title={LED2-Net: Monocular 360 Layout Estimation via Differentiable Depth Rendering}, 
      author={Fu-En Wang and Yu-Hsuan Yeh and Min Sun and Wei-Chen Chiu and Yi-Hsuan Tsai},
      year={2021},
      eprint={2104.00568},
      archivePrefix={arXiv},
      primaryClass={cs.CV}
}