7 March 2024 Lightweight monocular absolute depth estimation based on attention mechanism
Jiayu Jin, Bo Tao, Xinbo Qian, Jiaxin Hu, Gongfa Li
Author Affiliations +
Abstract

To solve the problem of obtaining a higher accuracy at the expense of redundant models, we propose a network architecture. We utilize a lightweight network that retains the high-precision advantage of the transformer and effectively combines it with convolutional neural network. By greatly reducing the training parameters, this approach achieves high precision, making it well suited for deployment on edge devices. A detail highlight module (DHM) is added to effectively fuse information from multiple scales, making the depth of prediction more accurate and clearer. A dense geometric constraints module is introduced to recover accurate scale factors in autonomous driving without additional sensors. Experimental results demonstrate that our model improves the accuracy from 98.1% to 98.3% compared with Monodepth2, and the model parameters are reduced by about 80%.

© 2024 SPIE and IS&T
Jiayu Jin, Bo Tao, Xinbo Qian, Jiaxin Hu, and Gongfa Li "Lightweight monocular absolute depth estimation based on attention mechanism," Journal of Electronic Imaging 33(2), 023010 (7 March 2024). https://doi.org/10.1117/1.JEI.33.2.023010
Received: 2 November 2023; Accepted: 8 February 2024; Published: 7 March 2024
Advertisement
Advertisement
RIGHTS & PERMISSIONS
Get copyright permission  Get copyright permission on Copyright Marketplace
KEYWORDS
Convolution

Cameras

Education and training

Image processing

Data modeling

Depth maps

Transformers

Back to Top