计算机科学
点云
残余物
人工智能
正规化(语言学)
深度学习
特征提取
像素
模式识别(心理学)
数据挖掘
计算机视觉
算法
作者
Guangchen Li,Kefeng Li,Shouxin Zhang,Zhenfang Zhu,Peng Wang,Zhenfei Wang,Fu Chen
标识
DOI:10.1038/s41598-024-64805-y
摘要
Abstract Although 3D reconstruction has been widely used in many fields as a key component of environment perception, existing technologies still have the potential for further improvement in 3D scene reconstruction. We propose an improved reconstruction algorithm based on the MVSNet network architecture. To glean richer pixel details from images, we suggest deploying a DE module integrated with a residual framework, which supplants the prevailing feature extraction mechanism. The DE module uses ECA-Net and dilated convolution to expand the receptive field range, performing feature splicing and fusion through the residual structure to retain the global information of the original image. Moreover, harnessing attention mechanisms refines the 3D cost volume's regularization process, bolstering the integration of information across multi-scale feature volumes, consequently enhancing depth estimation precision. When assessed our model using the DTU dataset, findings highlight the network's 3D reconstruction scoring a completeness (comp) of 0.411 mm and an overall quality of 0.418 mm. This performance is higher than that of traditional methods and other deep learning-based methods. Additionally, the visual representation of the point cloud model exhibits marked advancements. Trials on the Blended MVS dataset signify that our network exhibits commendable generalization prowess.
科研通智能强力驱动
Strongly Powered by AbleSci AI