摘要
目的为解决基于深度学习算法在执行胎儿四腔心超声切面图像质量评测时无法准确反映心脏区域中瓣膜与房室间隔及心室心房区域的可见程度问题,提出一种目标检测与两级分割相结合的胎儿四腔心超声切面图像质量评测方法。方法首先利用自行构建的胎儿超声切面数据集训练主流的YOLOv5x(you only look once v5x)模型,实现四腔心区域与胸腔区域的有效定位。当检测到四腔心区域在胸腔区域内时,将其视为感兴趣区域送入训练好的U2-Net模型,进一步分割出包含心房室及瓣膜的部分。然后利用形态学算子去除其外围可能存在的少许心脏外膜区域得到四腔心内区域后,通过直方图修正与最大类间方差法(OTSU)相结合的方法分割出瓣膜连同房室间隔区域,并通过减法操作得到心室心房区域的分割图。最后通过联合胎儿四腔心超声切面图像中瓣膜连同房室间隔与心室心房区域的面积之比、瓣膜与房室间隔区域以及心室心房区域的平均灰度构建评分公式与评分标准,实现胎儿四腔心超声切面图像质量的有效评测。结果在胸腔和四腔心区域的检测任务上的mAP@0.5、mAP@0.5-0.95和召回率分别为99.5%、84.6%和99.9%;在四腔心内部区域分割任务上的灵敏度、特异度和准确度分别为95.0%、95.1%和94.9%;所提质量评测方法在所构建的A、B、C三类评测数据集上分别取得了93.7%、90.3%和99.1%的准确率。结论所提方法的评测结果与医生主观评测结果相近,具有较好的可解释性,拥有良好的实际应用价值。
Objective To diagnose the fetal congenital heart disease(CHD)during screening,clinical-related ultrasound technique is adopted and focused on the captured images in terms of different critical cardiac scanning planes.Ultrasound scanning image quality assessment(QA)is indispensable for its efficiency and effectiveness.Four-chamber(4C)viewrelated multiple fetal cardiac ultrasound scan planes are commonly-used for CHD.The emerging artificial intelligence(AI)technique is beneficial for automatic fetal 4C view QA algorithm research further.In recent years,deep convolutional neural network(DCNN)based AI technique has been widely developing in the context of medical image processing and analy‐sis.However,duo to the lack of relevant data-sets and the 4C region only occupying a small part in the whole fetal 4C view,the confidence of detection bounding-box from general purpose object detection network is challenged to reflect the visibility and clarity of four chambers of the heart and its related crux area well,which includes mitral valve,tricuspid valve,interatrial septum and interventricular septum in cardiac region.In addition,current fetal 4C view QA methods are still challenged for reasonable explainability based on pure deep learning(DL)techniques.To resolve this problem,we proposed a novel fetal 4C view QA algorithm through the integration of object detection and two-stage segmentation operations,which is mutual-benefited for both of DL and traditional image processing techniques to get better accuracy and interpretability.Method A self-built medical data-set of 1696 images is used for fetal 4C view QA research.The data-set offers common objects in context(COCO)format labels for 4C and thorax regions,semantic segmentation labels for 4C inner regions,and also contains QA labels annotated manually.First,object detection network of you only look once v5x(YOLOv5x)is trained to realize effective 4C region detection and extraction.It illustrates that the 4C region’s location is normal and can be treated as the region of interest(ROI)when the detected 4C region locates inner the thorax region.And the ROI will be fed into the semantic segmentation network U2-Net which is well trained based the 4C inner region data-set which is a sub-set of the self-built data-set.The U2-Net considers four chambers of the heart and crux areas as foreground and implements the initial segmentation.And,the U2-Net output is a gray-scale image,in which values of background pixels are restrained effectively and foreground pixels are highlighted as well.Then,the maximum inter-class variance method(OTSU’s method)is adopted to binarize the U2-Net output.The morphological erosion operation is employed to optimize the binary segmentation result further,and a binary mask is produced to isolate the gray-scale 4C region.Next,OTSUintegrated histogram adjustment is used to separate the crux area leveraged from the isolated 4C region.And,the rest part of it can be considered as the four chambers of the heart area.After that,three QA indices are designed,and they can be used to represent area ration of the crux and four chambers of the heart,average gray scale of the crux and four chambers of the heart.Finally,to achieve effective fetal 4C view QA,evaluation formulations and standards are developed based on the three QA indices above.Result The experimental results show that the well trained YOLOv5x model based on the self-built data-set can achieve 99.5%mAP@0.5 and 84.6%mAP@0.5-0.95 in the object detection task for thorax and 4C regions respectively,and the recall rate is as high as 99.9%;and the well trained U2-Net model can achieve 95.0%sensitivity,95.1%specificity and 94.9%accuracy in the segmentation task for 4C inner region.The proposed fetal 4C view QA method can get 93.7%,90.3%and 99.1%accuracy on each evaluation data-set of class A,B and C.Conclusion To solve the problem that DL based image classification network cannot consider the location relationships of anatomy parts,and the object detection network cannot reflect their visibility and clarity,which leads to unreliable evaluation results,a fetal 4C view QA algorithm is proposed based on an object detection incorporated with two-stage segmentation.Evaluation results show that the trained object detection network has its potentials for 4C and thorax region detection,and the trained semantic segmentation network is optimized for 4C inner region extraction as well.The adopted two-stage segmentation strategy which combines DL and traditional image processing can not only shrink the costs of data annotation and network training greatly,but also strengthen optimal explainable results as well.The designed QA standards can be developed farther in terms of the three key indices.
作者
徐光柱
钱奕凡
王阳
刘蓉
周军
雷帮军
Xu Guangzhu;Qian Yifan;Wang Yang;Liu Rong;Zhou Jun;Lei Bangjun(Hubei Key Laboratory of Intelligent Vision Based Monitoring for Hydroelectric Engineering,China Three Gorges University,Yichang 443002,China;College of Computer and Information Technology,China Three Gorges University,Yichang 443002,China;Ultrasound Department,Yichang Central People’s Hospital,Yichang 443003,China)
出处
《中国图象图形学报》
CSCD
北大核心
2023年第8期2476-2490,共15页
Journal of Image and Graphics