An improved YOLOv5 method for large objects detection with multi-scale feature cross-layer fusion network

计算机科学 模式识别(心理学) 人工智能 融合 特征(语言学) 比例(比率) 图层(电子) 计算机视觉 材料科学 物理 语言学 哲学 量子力学 复合材料
作者
Zhong Qu,Le-yuan Gao,Shengye Wang,Haonan Yin,Tuming Yi
出处
期刊:Image and Vision Computing [Elsevier]
卷期号:125: 104518-104518 被引量:9
标识
DOI:10.1016/j.imavis.2022.104518
摘要

SSD and YOLOv5 are the one-stage object detector representative algorithms. An improved one-stage object detector based on the YOLOv5 method is proposed in this paper, named Multi-scale Feature Cross-layer Fusion Network (M-FCFN). Firstly, we extract shallow features and deep features from the PANet structure for cross-layer fusion and obtain a feature scale different from 80 × 80, 40 × 40, and 20 × 20 as output. Then, according to the single shot multi-box detector, we propose the different scale features which are obtained by cross-layer fusion for dimension reduction and use it as another output for prediction. Therefore, two completely different feature scales are added as the output. Features of different scales are necessary for detecting objects of different sizes, which can increase the probability of object detection and significantly improve detection accuracy. Finally, aiming at the Autoanchor mechanism proposed by YOLOv5, we propose an EIOU k-means calculation. We have compared the four model structures of S , M , L , and X of YOLOv5 respectively. The problem of missed and false detections for large objects is improved which has better detection results. The experimental results show that our methods achieve 89.1% and 67.8% mAP @0.5 on the PASCAL VOC and MS COCO datasets. Compared with the YOLOv5_S, our methods improve by 4.4% and 1.4% mAP @ [0.5:0.95] on the PASCAL VOC and MS COCO datasets. Compared with the four models of YOLOv5, our methods have better detection accuracy for large objects. It should be more attention that our method on the large-scale mAP @ [0.5:0.95] is 5.4% higher than YOLOv5_S on the MS COCO datasets. • We proposed Multi-scale Feature Cross-layer Fusion Network (M-FCFN). • Two completely different feature scales are added as the output. • We propose an EIOU k-means Autoanchor calculation. • The problem of missed and false detections for large objects is improved. • Our method on the large-scale mAP @[0.5:0.95] is 5.4% higher than YOLOv5_S.

科研通智能强力驱动
Strongly Powered by AbleSci AI
科研通是完全免费的文献互助平台,具备全网最快的应助速度,最高的求助完成率。 对每一个文献求助,科研通都将尽心尽力,给求助人一个满意的交代。
实时播报
刚刚
科研通AI6.1应助晾猫人采纳,获得10
1秒前
FashionBoy应助受伤灵薇采纳,获得10
1秒前
1秒前
2秒前
TheGreatBug发布了新的文献求助10
2秒前
2秒前
2秒前
绛春寒发布了新的文献求助10
3秒前
可爱的函函应助zhou采纳,获得10
3秒前
斯文复天发布了新的文献求助10
4秒前
NexusExplorer应助huangchenxi采纳,获得10
4秒前
4秒前
5秒前
谈笑间应助ll采纳,获得20
6秒前
6秒前
种喜欢的花完成签到 ,获得积分10
6秒前
星期八发布了新的文献求助10
6秒前
蓝天应助SJ_Wang采纳,获得10
7秒前
7秒前
科研通AI6.2应助晾猫人采纳,获得10
8秒前
ppc524完成签到,获得积分10
9秒前
cnyyp发布了新的文献求助10
9秒前
Rafayel应助yolo采纳,获得10
10秒前
liu11发布了新的文献求助10
10秒前
性静H情逸发布了新的文献求助10
10秒前
10秒前
健忘碧灵发布了新的文献求助10
10秒前
11秒前
小张完成签到,获得积分20
12秒前
ky废品应助老实灵安采纳,获得10
12秒前
疯狂的梦琪完成签到,获得积分10
12秒前
田様应助狗猪仔采纳,获得10
12秒前
12秒前
脑洞疼应助fan采纳,获得30
12秒前
慕青应助彭哒哒采纳,获得10
12秒前
受伤灵薇发布了新的文献求助10
13秒前
华仔应助雷小哈采纳,获得10
13秒前
鱼肠发布了新的文献求助10
14秒前
14秒前
高分求助中
(应助此贴封号)【重要!!请各用户(尤其是新用户)详细阅读】【科研通的精品贴汇总】 10000
Modern Epidemiology, Fourth Edition 5000
Digital Twins of Advanced Materials Processing 2000
Weaponeering, Fourth Edition – Two Volume SET 2000
Polymorphism and polytypism in crystals 1000
Social Cognition: Understanding People and Events 800
Signals, Systems, and Signal Processing 610
热门求助领域 (近24小时)
化学 材料科学 医学 生物 工程类 纳米技术 有机化学 物理 生物化学 化学工程 计算机科学 复合材料 内科学 催化作用 光电子学 物理化学 电极 冶金 遗传学 细胞生物学
热门帖子
关注 科研通微信公众号,转发送积分 6026189
求助须知:如何正确求助?哪些是违规求助? 7667883
关于积分的说明 16181862
捐赠科研通 5174187
什么是DOI,文献DOI怎么找? 2768632
邀请新用户注册赠送积分活动 1751924
关于科研通互助平台的介绍 1637936