计算机科学
目标检测
比例(比率)
航空影像
人工智能
搜救
遥感
计算机视觉
基于对象
卫星图像
人工神经网络
对象(语法)
实时计算
模式识别(心理学)
地质学
地图学
地理
机器人
作者
Jianhao Xu,Xiangtao Fan,Hongdeng Jian,Chen Xu,Weijia Bei,Qifeng Ge,Teng Zhao
出处
期刊:IEEE Transactions on Geoscience and Remote Sensing
[Institute of Electrical and Electronics Engineers]
日期:2024-01-01
卷期号:: 1-1
被引量:3
标识
DOI:10.1109/tgrs.2024.3395483
摘要
Personnel and boat detection in Unmanned Aerial Vehicles (UAVs) imagery plays a crucial role in Open Water Search and Rescue Missions. The diverse perspectives and altitudes of UAV images often result in significant variations in the imagery's appearance and dimensions of personnel and boats, and the false detections arising from water surface flares are acknowledged as a great challenge as well. Existing deep learning-based detection methods employ convolutional blocks with fixed kernel sizes to extract features from the imagery at a fixed spatial scale, which will lead to missed and false detections, and severely affect detection accuracy when there are substantial differences in the appearance and size of the target objects. In this paper, a spatial scale adaptive real-time object detection neural network, namely YoloOW, was proposed to tackle the challenge of personnel and boat detection amidst the diverse UAV imagery, which comprises a feature extractor, a feature enhancer, and a postprocessor. The OaohRep convolutional block was proposed as a pivotal component in constructing the YoloOW and applied to the feature extractor and the feature enhancer. Compared with general convolution blocks, the OaohRep convolution block can extract image features across a wide range of spatial scales, show better scale adaptability, and achieve faster detection speed due to its unique merged convolution layer design. OaohRepBi-PAN was proposed in the feature enhancer, which imitated the architecture of the classic algorithm SIFT and was successfully applied to deep learning models, showing better scale adaptability. A novel UAV detection box filter (UDBF) module was proposed in the postprocessor, which can effectively remove false detections caused by water surface flares. Experimental results demonstrate that our YoloOW model achieves 37.18% mAP on the SeaDronesSee dataset, surpassing the baseline by 8.43%. This notable improvement positions our model at the first of the leaderboard. The code will be available at https://github.com/Xjh-UCAS/YoloOW.
科研通智能强力驱动
Strongly Powered by AbleSci AI