Identifying and mapping individual plants in a highly diverse high-elevation ecosystem using UAV imagery and deep learning

稳健性(进化) 人工智能 分割 残余物 计算机科学 深度学习 比例(比率) 仰角(弹道) 样品(材料) 水准点(测量) 遥感 地图学 地理 数学 生物化学 化学 几何学 算法 色谱法 基因
作者
Ce Zhang,Peter M. Atkinson,Charles George,Zhaofei Wen,Mauricio Diazgranados,France Gerard
出处
期刊:Isprs Journal of Photogrammetry and Remote Sensing 卷期号:169: 280-291 被引量:64
标识
DOI:10.1016/j.isprsjprs.2020.09.025
摘要

The identification and counting of plant individuals is essential for environmental monitoring. UAV based imagery offer ultra-fine spatial resolution and flexibility in data acquisition, and so provide a great opportunity to enhance current plant and in-situ field surveying. However, accurate mapping of individual plants from UAV imagery remains challenging, given the great variation in the sizes and geometries of individual plants and in their distribution. This is true even for deep learning based semantic segmentation and classification methods. In this research, a novel Scale Sequence Residual U-Net (SS Res U-Net) deep learning method was proposed, which integrates a set of Residual U-Nets with a sequence of input scales that can be derived automatically. The SS Res U-Net classifies individual plants by continuously increasing the patch scale, with features learned at small scales passing gradually to larger scales, thus, achieving multi-scale information fusion while retaining fine spatial details of interest. The SS Res U-Net was tested to identify and map frailejones (all plant species of the subtribe Espeletiinae), the dominant plants in one of the world’s most biodiverse high-elevation ecosystems (i.e. the páramos) from UAV imagery. Results demonstrate that the SS Res U-Net has the ability to self-adapt to variation in objects, and consistently achieved the highest classification accuracy (91.67% on average) compared with four state-of-the-art benchmark approaches. In addition, SS Res U-Net produced the best performances in terms of both robustness to training sample size reduction and computational efficiency compared with the benchmarks. Thus, SS Res U-Net shows great promise for solving remotely sensed semantic segmentation and classification tasks, and more general machine intelligence. The prospective implementation of this method to identify and map frailejones in the páramos will benefit immensely the monitoring of their populations for conservation assessments and management, among many other applications.

科研通智能强力驱动
Strongly Powered by AbleSci AI
科研通是完全免费的文献互助平台,具备全网最快的应助速度,最高的求助完成率。 对每一个文献求助,科研通都将尽心尽力,给求助人一个满意的交代。
实时播报
傻傻的小虾米完成签到 ,获得积分10
1秒前
sad发布了新的文献求助10
2秒前
执着从灵发布了新的文献求助20
2秒前
2秒前
扶苏完成签到,获得积分10
4秒前
谨慎的果汁完成签到 ,获得积分10
4秒前
4秒前
orixero应助郦乞采纳,获得10
5秒前
屈绮兰发布了新的文献求助50
5秒前
含着朵白云完成签到,获得积分10
6秒前
科研通AI6.2应助天才玩家H采纳,获得10
6秒前
6秒前
7秒前
taozi完成签到,获得积分10
8秒前
香蕉觅云应助滴滴滴滴采纳,获得30
8秒前
付辛博boo完成签到,获得积分10
8秒前
所所应助今我来思采纳,获得10
9秒前
hhhuan发布了新的文献求助10
9秒前
今后应助负责以山采纳,获得10
10秒前
咕噜噜完成签到 ,获得积分10
11秒前
11秒前
CodeCraft应助杨廷友采纳,获得10
12秒前
张凯发布了新的文献求助10
13秒前
Eve完成签到,获得积分10
13秒前
ding完成签到,获得积分10
14秒前
Dang发布了新的文献求助30
14秒前
JamesPei应助Wang采纳,获得10
15秒前
suiwuya给suiwuya的求助进行了留言
15秒前
科研通AI6.3应助zkkk采纳,获得10
16秒前
top发布了新的文献求助10
16秒前
科研通AI2S应助懵懂的梦容采纳,获得10
17秒前
L_完成签到,获得积分10
18秒前
江11111完成签到,获得积分10
18秒前
Hello应助明亮的lunacake采纳,获得10
18秒前
19秒前
量子星尘发布了新的文献求助10
19秒前
21秒前
科研通AI6.1应助天道酬勤采纳,获得10
21秒前
black的hole完成签到,获得积分10
23秒前
吉祥高趙发布了新的文献求助20
23秒前
高分求助中
(应助此贴封号)【重要!!请各用户(尤其是新用户)详细阅读】【科研通的精品贴汇总】 10000
Relation between chemical structure and local anesthetic action: tertiary alkylamine derivatives of diphenylhydantoin 1000
Signals, Systems, and Signal Processing 610
Discrete-Time Signals and Systems 610
Principles of town planning : translating concepts to applications 500
Iron‐Sulfur Clusters: Biogenesis and Biochemistry 400
Healable Polymer Systems: Fundamentals, Synthesis and Applications 400
热门求助领域 (近24小时)
化学 材料科学 医学 生物 工程类 纳米技术 有机化学 物理 生物化学 化学工程 计算机科学 复合材料 内科学 催化作用 光电子学 物理化学 电极 冶金 遗传学 细胞生物学
热门帖子
关注 科研通微信公众号,转发送积分 6071453
求助须知:如何正确求助?哪些是违规求助? 7902960
关于积分的说明 16340025
捐赠科研通 5211747
什么是DOI,文献DOI怎么找? 2787567
邀请新用户注册赠送积分活动 1770269
关于科研通互助平台的介绍 1648148