计算机科学
卷积神经网络
人工智能
胎头
放大倍数
超声波
图像质量
学习迁移
任务(项目管理)
模式识别(心理学)
计算机视觉
图像(数学)
放射科
医学
生物
经济
胎儿
管理
怀孕
遗传学
作者
Zhiping Lin,Shengli Li,Dong Ni,Yu‐Cai Liao,Huaxuan Wen,Jie Du,Siping Chen,Tianfu Wang,Baiying Lei
标识
DOI:10.1016/j.media.2019.101548
摘要
It is essential to measure anatomical parameters in prenatal ultrasound images for the growth and development of the fetus, which is highly relied on obtaining a standard plane. However, the acquisition of a standard plane is, in turn, highly subjective and depends on the clinical experience of sonographers. In order to deal with this challenge, we propose a new multi-task learning framework using a faster regional convolutional neural network (MF R-CNN) architecture for standard plane detection and quality assessment. MF R-CNN can identify the critical anatomical structure of the fetal head and analyze whether the magnification of the ultrasound image is appropriate, and then performs quality assessment of ultrasound images based on clinical protocols. Specifically, the first five convolution blocks of the MF R-CNN learn the features shared within the input data, which can be associated with the detection and classification tasks, and then extend to the task-specific output streams. In training, in order to speed up the different convergence of different tasks, we devise a section train method based on transfer learning. In addition, our proposed method also uses prior clinical and statistical knowledge to reduce the false detection rate. By identifying the key anatomical structure and magnification of the ultrasound image, we score the ultrasonic plane of fetal head to judge whether it is a standard image or not. Experimental results on our own-collected dataset show that our method can accurately make a quality assessment of an ultrasound plane within half a second. Our method achieves promising performance compared with state-of-the-art methods, which can improve the examination effectiveness and alleviate the measurement error caused by improper ultrasound scanning.
科研通智能强力驱动
Strongly Powered by AbleSci AI