计算机科学
人工智能
分割
深度学习
水准点(测量)
图像融合
机器学习
模式
多模态
图像(数学)
融合
万维网
地理
社会学
哲学
大地测量学
语言学
社会科学
作者
Yifei Zhang,Désiré Sidibé,Olivier Morel,Fabrice Mériaudeau
标识
DOI:10.1016/j.imavis.2020.104042
摘要
Recent advances in deep learning have shown excellent performance in various scene understanding tasks. However, in some complex environments or under challenging conditions, it is necessary to employ multiple modalities that provide complementary information on the same scene. A variety of studies have demonstrated that deep multimodal fusion for semantic image segmentation achieves significant performance improvement. These fusion approaches take the benefits of multiple information sources and generate an optimal joint prediction automatically. This paper describes the essential background concepts of deep multimodal fusion and the relevant applications in computer vision. In particular, we provide a systematic survey of multimodal fusion methodologies, multimodal segmentation datasets, and quantitative evaluations on the benchmark datasets. Existing fusion methods are summarized according to a common taxonomy: early fusion, late fusion, and hybrid fusion. Based on their performance, we analyze the strengths and weaknesses of different fusion strategies. Current challenges and design choices are discussed, aiming to provide the reader with a comprehensive and heuristic view of deep multimodal image segmentation.
科研通智能强力驱动
Strongly Powered by AbleSci AI