Open Access iconOpen Access

ARTICLE

Unsupervised Monocular Depth Estimation with Edge Enhancement for Dynamic Scenes

Peicheng Shi1,*, Yueyue Tang1, Yi Li1, Xinlong Dong1, Yu Sun2, Aixi Yang3

1 School of Mechanical and Automotive Engineering, Anhui Polytechnic University, Wuhu, 241000, China
2 Chery New Energy Automobile Co., Ltd., Wuhu, 241000, China
3 Polytechnic Institute, Zhejiang University, Hangzhou, 310015, China

* Corresponding Author: Peicheng Shi. Email: email

(This article belongs to the Special Issue: Advances in Deep Learning and Neural Networks: Architectures, Applications, and Challenges)

Computers, Materials & Continua 2025, 84(2), 3321-3343. https://doi.org/10.32604/cmc.2025.065297

Abstract

In the dynamic scene of autonomous vehicles, the depth estimation of monocular cameras often faces the problem of inaccurate edge depth estimation. To solve this problem, we propose an unsupervised monocular depth estimation model based on edge enhancement, which is specifically aimed at the depth perception challenge in dynamic scenes. The model consists of two core networks: a deep prediction network and a motion estimation network, both of which adopt an encoder-decoder architecture. The depth prediction network is based on the U-Net structure of ResNet18, which is responsible for generating the depth map of the scene. The motion estimation network is based on the U-Net structure of Flow-Net, focusing on the motion estimation of dynamic targets. In the decoding stage of the motion estimation network, we innovatively introduce an edge-enhanced decoder, which integrates a convolutional block attention module (CBAM) in the decoding process to enhance the recognition ability of the edge features of moving objects. In addition, we also designed a strip convolution module to improve the model’s capture efficiency of discrete moving targets. To further improve the performance of the model, we propose a novel edge regularization method based on the Laplace operator, which effectively accelerates the convergence process of the model. Experimental results on the KITTI and Cityscapes datasets show that compared with the current advanced dynamic unsupervised monocular model, the proposed model has a significant improvement in depth estimation accuracy and convergence speed. Specifically, the root mean square error (RMSE) is reduced by 4.8% compared with the DepthMotion algorithm, while the training convergence speed is increased by 36%, which shows the superior performance of the model in the depth estimation task in dynamic scenes.

Keywords

Dynamic scenes; unsupervised learning; monocular depth; edge enhancement

Cite This Article

APA Style
Shi, P., Tang, Y., Li, Y., Dong, X., Sun, Y. et al. (2025). Unsupervised Monocular Depth Estimation with Edge Enhancement for Dynamic Scenes. Computers, Materials & Continua, 84(2), 3321–3343. https://doi.org/10.32604/cmc.2025.065297
Vancouver Style
Shi P, Tang Y, Li Y, Dong X, Sun Y, Yang A. Unsupervised Monocular Depth Estimation with Edge Enhancement for Dynamic Scenes. Comput Mater Contin. 2025;84(2):3321–3343. https://doi.org/10.32604/cmc.2025.065297
IEEE Style
P. Shi, Y. Tang, Y. Li, X. Dong, Y. Sun, and A. Yang, “Unsupervised Monocular Depth Estimation with Edge Enhancement for Dynamic Scenes,” Comput. Mater. Contin., vol. 84, no. 2, pp. 3321–3343, 2025. https://doi.org/10.32604/cmc.2025.065297



cc Copyright © 2025 The Author(s). Published by Tech Science Press.
This work is licensed under a Creative Commons Attribution 4.0 International License , which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
  • 332

    View

  • 149

    Download

  • 0

    Like

Share Link