计算机科学
人工智能
分割
解析
棱锥(几何)
特征(语言学)
计算机视觉
GSM演进的增强数据速率
背景(考古学)
试验装置
集合(抽象数据类型)
对象(语法)
模式识别(心理学)
地理
语言学
哲学
物理
光学
程序设计语言
考古
作者
Xianwei Zheng,Linxi Huan,Gui-Song Xia,Jianya Gong
出处
期刊:Isprs Journal of Photogrammetry and Remote Sensing
日期:2020-10-14
卷期号:170: 15-28
被引量:104
标识
DOI:10.1016/j.isprsjprs.2020.09.019
摘要
Parsing very high resolution (VHR) urban scene images into regions with semantic meaning, e.g. buildings and cars, is a fundamental task in urban scene understanding. However, due to the huge quantity of details contained in an image and the large variations of objects in scale and appearance, the existing semantic segmentation methods often break one object into pieces, or confuse adjacent objects and thus fail to depict these objects consistently. To address these issues uniformly, we propose a standalone end-to-end edge-aware neural network (EaNet) for urban scene semantic segmentation. For semantic consistency preservation inside objects, the EaNet model incorporates a large kernel pyramid pooling (LKPP) module to capture rich multi-scale context with strong continuous feature relations. To effectively separate confusing objects with sharp contours, a Dice-based edge-aware loss function (EA loss) is devised to guide the EaNet to refine both the pixel- and image-level edge information directly from semantic segmentation prediction. In the proposed EaNet model, the LKPP and the EA loss couple to enable comprehensive feature learning across an entire semantic object. Extensive experiments on three challenging datasets demonstrate that our method can be readily generalized to multi-scale ground/aerial urban scene images, achieving 81.7% in mIoU on Cityscapes Test set and 90.8% in the mean F1-score on the ISPRS Vaihingen 2D Test set. Code is available at: https://github.com/geovsion/EaNet.
科研通智能强力驱动
Strongly Powered by AbleSci AI