Item request has been placed!
×
Item request cannot be made.
×
Processing Request
A self‐supervised monocular depth estimation model with scale recovery and transfer learning for construction scene analysis.
Item request has been placed!
×
Item request cannot be made.
×
Processing Request
- Additional Information
- Subject Terms:
- Abstract:
Estimating the depth of a construction scene from a single red‐green‐blue image is a crucial prerequisite for various applications, including work zone safety, localization, productivity analysis, activity recognition, and scene understanding. Recently, self‐supervised representation learning methods have made significant progress and demonstrated state‐of‐the‐art performance on monocular depth estimation. However, the two leading open challenges are the ambiguity of estimated depth up to an unknown scale and representation transferability for a downstream task, which severely hinders the practical deployment of self‐supervised methods. We propose a prior information‐based method, not depending on additional sensors, to recover the unknown scale in monocular vision and predict per‐pixel absolute depth. Moreover, a new learning paradigm for a self‐supervised monocular depth estimation model is constructed to transfer the pre‐trained self‐supervised model to other downstream construction scene analysis tasks. Meanwhile, we also propose a novel depth loss to enforce depth consistency when transferring to a new downstream task and two new metrics to measure transfer performance. Finally, we verify the effectiveness of scale recovery and representation transferability in isolation. The new learning paradigm with our new metrics and depth loss is expected to estimate the monocular depth of a construction scene without depth ground truth like light detection and ranging. Our models will serve as a good foundation for further construction scene analysis tasks. [ABSTRACT FROM AUTHOR]
- Abstract:
Copyright of Computer-Aided Civil & Infrastructure Engineering is the property of Wiley-Blackwell and its content may not be copied or emailed to multiple sites or posted to a listserv without the copyright holder's express written permission. However, users may print, download, or email articles for individual use. This abstract may be abridged. No warranty is given about the accuracy of the copy. Users should refer to the original published version of the material for the full abstract. (Copyright applies to all Abstracts.)
No Comments.