Please use this identifier to cite or link to this item: https://hdl.handle.net/2440/137141
Citations
Scopus Web of Science® Altmetric
?
?
Full metadata record
DC FieldValueLanguage
dc.contributor.authorBian, J.W.-
dc.contributor.authorZhan, H.-
dc.contributor.authorWang, N.-
dc.contributor.authorChin, T.J.-
dc.contributor.authorShen, C.-
dc.contributor.authorReid, I.-
dc.date.issued2021-
dc.identifier.citationIEEE Transactions on Pattern Analysis and Machine Intelligence, 2021; 44(12)-
dc.identifier.issn0162-8828-
dc.identifier.issn1939-3539-
dc.identifier.urihttps://hdl.handle.net/2440/137141-
dc.description.abstractSingle-View depth estimation using the CNNs trained from unlabelled videos has shown significant promise. However, excellent results have mostly been obtained in street-scene driving scenarios, and such methods often fail in other settings, particularly indoor videos taken by handheld devices. In this work,we establish that the complex ego-motions exhibited in handheld settings are a critical obstacle for learning depth. Our fundamental analysis suggests that the rotation behaves as noise during training, as opposed to the translation (baseline) which provides supervision signals. To address the challenge, we propose a data pre-processing method that rectifies training images by removing their relative rotations for effective learning. The significantly improved performance validates our motivation. Towards end-to-end learning without requiring pre-processing, we propose an Auto-Rectify Network with novel loss functions, which can automatically learn to rectify images during training. Consequently, our results out perform the previous unsupervised SOTA method by a large margin on the challenging NYUv2 dataset.We also demonstrate the generalization of our trained model in ScanNet and Make3D, and the universality of our proposed learning method on 7-Scenes and KITTI datasets.-
dc.description.statementofresponsibilityJia-Wang Bian, Huangying Zhan, Naiyan Wang, Tat-Jun Chin, Chunhua Shen, and Ian Reid-
dc.language.isoen-
dc.publisherInstitute of Electrical and Electronics Engineers (IEEE)-
dc.rights© 2021 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission.-
dc.source.urihttp://dx.doi.org/10.1109/tpami.2021.3136220-
dc.subjectSingle-view depth estimation; unsupervised learning; image rectification-
dc.subject.meshAlgorithms-
dc.titleAuto-Rectify Network for Unsupervised Indoor Depth Estimation-
dc.typeJournal article-
dc.identifier.doi10.1109/TPAMI.2021.3136220-
dc.relation.granthttp://purl.org/au-research/grants/arc/FL130100102-
pubs.publication-statusPublished-
dc.identifier.orcidBian, J.W. [0000-0003-2046-3363]-
dc.identifier.orcidShen, C. [0000-0002-8648-8718]-
dc.identifier.orcidReid, I. [0000-0001-7790-6423]-
Appears in Collections:Computer Science publications

Files in This Item:
There are no files associated with this item.


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.