To overcome the shortcomings of 1 D and 2 D Otsu’s thresholding techniques, the 3 D Otsu method has been developed.Among all Otsu’s methods, 3 D Otsu technique provides the best threshold values for the multi-level ...To overcome the shortcomings of 1 D and 2 D Otsu’s thresholding techniques, the 3 D Otsu method has been developed.Among all Otsu’s methods, 3 D Otsu technique provides the best threshold values for the multi-level thresholding processes. In this paper, to improve the quality of segmented images, a simple and effective multilevel thresholding method is introduced. The proposed approach focuses on preserving edge detail by computing the 3 D Otsu along the fusion phenomena. The advantages of the presented scheme include higher quality outcomes, better preservation of tiny details and boundaries and reduced execution time with rising threshold levels. The fusion approach depends upon the differences between pixel intensity values within a small local space of an image;it aims to improve localized information after the thresholding process. The fusion of images based on local contrast can improve image segmentation performance by minimizing the loss of local contrast, loss of details and gray-level distributions. Results show that the proposed method yields more promising segmentation results when compared to conventional1 D Otsu, 2 D Otsu and 3 D Otsu methods, as evident from the objective and subjective evaluations.展开更多
With the widespread application of deep learning in the field of computer vision,gradually allowing medical image technology to assist doctors in making diagnoses has great practical and research significance.Aiming a...With the widespread application of deep learning in the field of computer vision,gradually allowing medical image technology to assist doctors in making diagnoses has great practical and research significance.Aiming at the shortcomings of the traditional U-Net model in 3D spatial information extraction,model over-fitting,and low degree of semantic information fusion,an improved medical image segmentation model has been used to achieve more accurate segmentation of medical images.In this model,we make full use of the residual network(ResNet)to solve the over-fitting problem.In order to process and aggregate data at different scales,the inception network is used instead of the traditional convolutional layer,and the dilated convolution is used to increase the receptive field.The conditional random field(CRF)can complete the contour refinement work.Compared with the traditional 3D U-Net network,the segmentation accuracy of the improved liver and tumor images increases by 2.89%and 7.66%,respectively.As a part of the image processing process,the method in this paper not only can be used for medical image segmentation,but also can lay the foundation for subsequent image 3D reconstruction work.展开更多
In order to effectively improve the pathological diagnosis capability and feature resolution of 3D human brain CT images,a threshold segmentation method of multi-resolution 3D human brain CT image based on edge pixel ...In order to effectively improve the pathological diagnosis capability and feature resolution of 3D human brain CT images,a threshold segmentation method of multi-resolution 3D human brain CT image based on edge pixel grayscale feature decomposition is proposed in this paper.In this method,first,original 3D human brain image information is collected,and CT image filtering is performed to the collected information through the gradient value decomposition method,and edge contour features of the 3D human brain CT image are extracted.Then,the threshold segmentation method is adopted to segment the regional pixel feature block of the 3D human brain CT image to segment the image into block vectors with high-resolution feature points,and the 3D human brain CT image is reconstructed with the salient feature point as center.Simulation results show that the method proposed in this paper can provide accuracy up to 100%when the signal-to-noise ratio is 0,and with the increase of signal-to-noise ratio,the accuracy provided by this method is stable at 100%.Comparison results show that the threshold segmentation method of multi-resolution 3D human brain CT image based on edge pixel grayscale feature decomposition is signicantly better than traditional methods in pathological feature estimation accuracy,and it effectively improves the rapid pathological diagnosis and positioning recognition abilities to CT images.展开更多
Tumour segmentation in medical images(especially 3D tumour segmentation)is highly challenging due to the possible similarity between tumours and adjacent tissues,occurrence of multiple tumours and variable tumour shap...Tumour segmentation in medical images(especially 3D tumour segmentation)is highly challenging due to the possible similarity between tumours and adjacent tissues,occurrence of multiple tumours and variable tumour shapes and sizes.The popular deep learning‐based segmentation algorithms generally rely on the convolutional neural network(CNN)and Transformer.The former cannot extract the global image features effectively while the latter lacks the inductive bias and involves the complicated computation for 3D volume data.The existing hybrid CNN‐Transformer network can only provide the limited performance improvement or even poorer segmentation performance than the pure CNN.To address these issues,a short‐term and long‐term memory self‐attention network is proposed.Firstly,a distinctive self‐attention block uses the Transformer to explore the correlation among the region features at different levels extracted by the CNN.Then,the memory structure filters and combines the above information to exclude the similar regions and detect the multiple tumours.Finally,the multi‐layer reconstruction blocks will predict the tumour boundaries.Experimental results demonstrate that our method outperforms other methods in terms of subjective visual and quantitative evaluation.Compared with the most competitive method,the proposed method provides Dice(82.4%vs.76.6%)and Hausdorff distance 95%(HD95)(10.66 vs.11.54 mm)on the KiTS19 as well as Dice(80.2%vs.78.4%)and HD95(9.632 vs.12.17 mm)on the LiTS.展开更多
Currently,deep learning is widely used in medical image segmentation and has achieved good results.However,3D medical image segmentation tasks with diverse lesion characters,blurred edges,and unstable positions requir...Currently,deep learning is widely used in medical image segmentation and has achieved good results.However,3D medical image segmentation tasks with diverse lesion characters,blurred edges,and unstable positions require complex networks with a large number of parameters.It is computationally expensive and results in high requirements on equipment,making it hard to deploy the network in hospitals.In this work,we propose a method for network lightweighting and applied it to a 3D CNN based network.We experimented on a COVID-19 lesion segmentation dataset.Specifically,we use three cascaded one-dimensional convolutions to replace a 3D convolution,and integrate instance normalization with the previous layer of one-dimensional convolutions to accelerate network inference.In addition,we simplify test-time augmentation and deep supervision of the network.Experiments show that the lightweight network can reduce the prediction time of each sample and the memory usage by 50%and reduce the number of parameters by 60%compared with the original network.The training time of one epoch is also reduced by 50%with the segmentation accuracy dropped within the acceptable range.展开更多
This paper presents a method which uses multiple types of expert knowledge together in 3D medical image segmentation based on rough set theory. The focus of this paper is how to approximate a ROI(region of interest) w...This paper presents a method which uses multiple types of expert knowledge together in 3D medical image segmentation based on rough set theory. The focus of this paper is how to approximate a ROI(region of interest) when there are multiple types of expert knowledge. Based on rough set theory, the image can be split into three regions: positive regions; negative regions; boundary regions. With multiple knowledge we refine ROI as an intersection of all of the expected shapes with single knowledge. At last we show the results of implementing a rough 3D image segmentation and visualization system.展开更多
In this paper we first determine three phases of cell images: background, cytoplasm and nucleolus according to the general physical characteristics of cell images, and then develop a variational model, based on these...In this paper we first determine three phases of cell images: background, cytoplasm and nucleolus according to the general physical characteristics of cell images, and then develop a variational model, based on these characteristics, to segment nucleolus and cytoplasm from their relatively complicated backgrounds. In the meantime, the preprocessing obtained information of cell images using the OTSU algorithm is used to initialize the level set function in the model, which can speed up the segmentation and present satisfactory results in cell image processing.展开更多
A novel technique of three-dimensional (3D) reconstruction, segmentation, display and analysis of series slices of images including microscopic wide field optical sectioning by deconvolution method, cryo-electron micr...A novel technique of three-dimensional (3D) reconstruction, segmentation, display and analysis of series slices of images including microscopic wide field optical sectioning by deconvolution method, cryo-electron microscope slices by Fou-rier-Bessel synthesis and electron tomography (ET), and a series of computed tomography (CT) was developed to perform si-multaneous measurement on the structure and function of biomedical samples. The paper presents the 3D reconstruction seg-mentation display and analysis results of pollen spore, chaperonin, virus, head, cervical bone, tibia and carpus. At the same time, it also puts forward some potential applications of the new technique in the biomedical realm.展开更多
AIM: To explore a segmentation algorithm based on deep learning to achieve accurate diagnosis and treatment of patients with retinal fluid.METHODS: A two-dimensional(2D) fully convolutional network for retinal segment...AIM: To explore a segmentation algorithm based on deep learning to achieve accurate diagnosis and treatment of patients with retinal fluid.METHODS: A two-dimensional(2D) fully convolutional network for retinal segmentation was employed. In order to solve the category imbalance in retinal optical coherence tomography(OCT) images, the network parameters and loss function based on the 2D fully convolutional network were modified. For this network, the correlations of corresponding positions among adjacent images in space are ignored. Thus, we proposed a three-dimensional(3D) fully convolutional network for segmentation in the retinal OCT images.RESULTS: The algorithm was evaluated according to segmentation accuracy, Kappa coefficient, and F1 score. For the 3D fully convolutional network proposed in this paper, the overall segmentation accuracy rate is 99.56%, Kappa coefficient is 98.47%, and F1 score of retinal fluid is 95.50%. CONCLUSION: The OCT image segmentation algorithm based on deep learning is primarily founded on the 2D convolutional network. The 3D network architecture proposed in this paper reduces the influence of category imbalance, realizes end-to-end segmentation of volume images, and achieves optimal segmentation results. The segmentation maps are practically the same as the manual annotations of doctors, and can provide doctors with more accurate diagnostic data.展开更多
As a kind of special material in geotechnical engineering, the mudded weak interlayer plays a crucial part in slope stability. In this paper, we presented a method to determine the threshold value of section micrograp...As a kind of special material in geotechnical engineering, the mudded weak interlayer plays a crucial part in slope stability. In this paper, we presented a method to determine the threshold value of section micrographs of the mudded weak interlayer in slope during its meso-structure qualification process. Some soil tests, scanning electron microscopy(SEM) and image segmentation technology were performed to fulfill our purpose. Specifically, the relation between 3 D-porosity and the threshold was obtained by least square fitting of the threshold-porosity curves and a simplified pore equivalent model. Using this relation and the 3 D-porosity determined by soil experiments, we can figure out the polynomial equation of the threshold value. The threshold values obtained by the other existing methods in literature were employed to validate our present results.展开更多
Deep learning (DL) has experienced an exponential development in recent years, with major impact in many medical fields, especially in the field of medical image and, respectively, as a specific task, in the segmentat...Deep learning (DL) has experienced an exponential development in recent years, with major impact in many medical fields, especially in the field of medical image and, respectively, as a specific task, in the segmentation of the medical image. We aim to create a computer assisted diagnostic method, optimized by the use of deep learning (DL) and validated by a randomized controlled clinical trial, is a highly automated tool for diagnosing and staging precancerous and cervical cancer and thyroid cancers. We aim to design a high-performance deep learning model, combined from convolutional neural network (U-Net)-based architectures, for segmentation of the medical image that is independent of the type of organs/tissues, dimensions or type of image (2D/3D) and to validate the DL model in a randomized, controlled clinical trial. We used as a methodology primarily the analysis of U-Net-based architectures to identify the key elements that we considered important in the design and optimization of the combined DL model, from the U-Net-based architectures, imagined by us. Secondly, we will validate the performance of the DL model through a randomized controlled clinical trial. The DL model designed by us will be a highly automated tool for diagnosing and staging precancers and cervical cancer and thyroid cancers. The combined model we designed takes into account the key features of each of the architectures Overcomplete Convolutional Network Kite-Net (Kite-Net), Attention gate mechanism is an improvement added on convolutional network architecture for fast and precise segmentation of images (Attention U-Net), Harmony Densely Connected Network-Medical image Segmentation (HarDNet-MSEG). In this regard, we will create a comprehensive computer assisted diagnostic methodology validated by a randomized controlled clinical trial. The model will be a highly automated tool for diagnosing and staging precancers and cervical cancer and thyroid cancers. This would help drastically minimize the time and effort that specialists put into analyzing medical images, help to achieve a better therapeutic plan, and can provide a “second opinion” of computer assisted diagnosis.展开更多
Cancer is one of the leading causes of death in the world,with radiotherapy as one of the treatment options.Radiotherapy planning starts with delineating the affected area from healthy organs,called organs at risk(OAR...Cancer is one of the leading causes of death in the world,with radiotherapy as one of the treatment options.Radiotherapy planning starts with delineating the affected area from healthy organs,called organs at risk(OAR).A new approach to automatic OAR seg-mentation in the chest cavity in Computed Tomography(CT)images is presented.The proposed approach is based on the modified U‐Net architecture with the ResNet‐34 encoder,which is the baseline adopted in this work.The new two‐branch CS‐SA U‐Net architecture is proposed,which consists of two parallel U‐Net models in which self‐attention blocks with cosine similarity as query‐key similarity function(CS‐SA)blocks are inserted between the encoder and decoder,which enabled the use of con-sistency regularisation.The proposed solution demonstrates state‐of‐the‐art performance for the problem of OAR segmentation in CT images on the publicly available SegTHOR benchmark dataset in terms of a Dice coefficient(oesophagus-0.8714,heart-0.9516,trachea-0.9286,aorta-0.9510)and Hausdorff distance(oesophagus-0.2541,heart-0.1514,trachea-0.1722,aorta-0.1114)and significantly outperforms the baseline.The current approach is demonstrated to be viable for improving the quality of OAR segmentation for radiotherapy planning.展开更多
This paper proposes an improved high-precision 3D semantic mapping method for indoor scenes using RGB-D images.The current semantic mapping algorithms suffer from low semantic annotation accuracy and insufficient real...This paper proposes an improved high-precision 3D semantic mapping method for indoor scenes using RGB-D images.The current semantic mapping algorithms suffer from low semantic annotation accuracy and insufficient real-time performance.To address these issues,we first adopt the Elastic Fusion algorithm to select key frames from indoor environment image sequences captured by the Kinect sensor and construct the indoor environment space model.Then,an indoor RGB-D image semantic segmentation network is proposed,which uses multi-scale feature fusion to quickly and accurately obtain object labeling information at the pixel level of the spatial point cloud model.Finally,Bayesian updating is used to conduct incremental semantic label fusion on the established spatial point cloud model.We also employ dense conditional random fields(CRF)to optimize the 3D semantic map model,resulting in a high-precision spatial semantic map of indoor scenes.Experimental results show that the proposed semantic mapping system can process image sequences collected by RGB-D sensors in real-time and output accurate semantic segmentation results of indoor scene images and the current local spatial semantic map.Finally,it constructs a globally consistent high-precision indoor scenes 3D semantic map.展开更多
Unmanned aerial vehicle(UAV)array InSAR is a new type of single-flight 3D SAR imaging system with the advantages of high coherence and resolution.However,due to the low altitude of the platform,the elevation ambiguity...Unmanned aerial vehicle(UAV)array InSAR is a new type of single-flight 3D SAR imaging system with the advantages of high coherence and resolution.However,due to the low altitude of the platform,the elevation ambiguity of the system is smaller than the maximal terrain elevation.Since the spatial phase unwrapping is conducted based on the assumption of phase continuity,the inappropriate ambiguity will cause significant unwrapping errors.In this work,a 3D phase unwrapping algorithm assisted by image segmentation is proposed to address this issue.The Markov random field(MRF)is utilized for image segmentation.The optimal spatial phase unwrapping network is achieved based on the segmentation results.The asymptotic 3D phase unwrapping is adopted to get the refined 3D reconstruction.Results based on the real airborne array-InSAR data show that the proposed method effectively improves the elevation ambiguity.展开更多
An infant male presented with the rare anatomy consisting of situs solitus,concordant atrioventricular connections to L-looped ventricles,double outlet right ventricle(DORV),and hypoplastic aortic arch.6 months after ...An infant male presented with the rare anatomy consisting of situs solitus,concordant atrioventricular connections to L-looped ventricles,double outlet right ventricle(DORV),and hypoplastic aortic arch.6 months after neonatal aortic arch repair,the morphologic right ventricle function deteriorated,and surgical evaluation was undertaken to determine if either biventricular repair with a systemic morphologic left ventricle or right ventricular exclusion was possible.After initial echocardiography,magnetic resonance imaging(MRI)was used to create detailed axial and 4-dimensional(4D)images and 3-dimensional(3D)printed models.The detailed anatomy of this rare,complex case and its use in pre-surgical planning is presented.展开更多
Coronavirus has infected more than 753 million people,ranging in severity from one person to another,where more than six million infected people died worldwide.Computer-aided diagnostic(CAD)with artificial intelligenc...Coronavirus has infected more than 753 million people,ranging in severity from one person to another,where more than six million infected people died worldwide.Computer-aided diagnostic(CAD)with artificial intelligence(AI)showed outstanding performance in effectively diagnosing this virus in real-time.Computed tomography is a complementary diagnostic tool to clarify the damage of COVID-19 in the lungs even before symptoms appear in patients.This paper conducts a systematic literature review of deep learning methods for classifying the segmentation of COVID-19 infection in the lungs.We used the methodology of systematic reviews and meta-analyses(PRISMA)flow method.This research aims to systematically analyze the supervised deep learning methods,open resource datasets,data augmentation methods,and loss functions used for various segment shapes of COVID-19 infection from computerized tomography(CT)chest images.We have selected 56 primary studies relevant to the topic of the paper.We have compared different aspects of the algorithms used to segment infected areas in the CT images.Limitations to deep learning in the segmentation of infected areas still need to be developed to predict smaller regions of infection at the beginning of their appearance.展开更多
Visual attention mechanisms allow humans to extract relevant and important information from raw input percepts. Many applications in robotics and computer vision have modeled human visual attention mechanisms using a ...Visual attention mechanisms allow humans to extract relevant and important information from raw input percepts. Many applications in robotics and computer vision have modeled human visual attention mechanisms using a bottom-up data centric approach. In contrast, recent studies in cognitive science highlight advantages of a top-down approach to the attention mechanisms, especially in applications involving goal-directed search. In this paper, we propose a top-down approach for extracting salient objects/regions of space. The top-down methodology first isolates different objects in an unorganized point cloud, and compares each object for uniqueness. A measure of saliency using the properties of geodesic distance on the object’s surface is defined. Our method works on 3D point cloud data, and identifies salient objects of high curvature and unique silhouette. These being the most unique features of a scene, are robust to clutter, occlusions and view point changes. We provide the details of the proposed method and initial experimental results.展开更多
In civil aviation security screening, laptops, with their intricate structural composition, provide the potential for criminals to conceal dangerous items. Presently, the security process necessitates passengers to in...In civil aviation security screening, laptops, with their intricate structural composition, provide the potential for criminals to conceal dangerous items. Presently, the security process necessitates passengers to individually present their laptops for inspection. The paper introduced a method for laptop removal. By combining projection algorithms with the YOLOv7-Seg model, a laptop’s three views were generated through projection, and instance segmentation of these views was achieved using YOLOv7-Seg. The resulting 2D masks from instance segmentation at different angles were employed to reconstruct a 3D mask through angle restoration. Ultimately, the intersection of this 3D mask with the original 3D data enabled the successful extraction of the laptop’s 3D information. Experimental results demonstrated that the fusion of projection and instance segmentation facilitated the automatic removal of laptops from CT data. Moreover, higher instance segmentation model accuracy leads to more precise removal outcomes. By implementing the laptop removal functionality, the civil aviation security screening process becomes more efficient and convenient. Passengers will no longer be required to individually handle their laptops, effectively enhancing the efficiency and accuracy of security screening.展开更多
Semantic segmentation of remote sensing images is an important but unsolved problem in the remote sensing society.Advanced image semantic segmentation models,such as DeepLabv3+,have achieved astonishing performance fo...Semantic segmentation of remote sensing images is an important but unsolved problem in the remote sensing society.Advanced image semantic segmentation models,such as DeepLabv3+,have achieved astonishing performance for semantically labeling very high resolution(VHR)remote sensing images.However,it is difficult for these models to capture the precise outlines of ground objects and explore the context information that revealing relationships among image objects for optimizing segmentation results.Consequently,this study proposes a semantic segmentation method for VHR images by incorporating deep learning semantic segmentation model(DeepLabv3+)and objectbased image analysis(OBIA),wherein DSM is employed to provide geometric information to enhance the interpretation of VHR images.The proposed method first obtains two initial probabilistic labeling predictions using a DeepLabv3+network on spectral image and a random forest(RF)classifier on hand-crafted features,respectively.These two predictions are then integrated by Dempster-Shafer(D-S)evidence theory to be fed into an object-constrained higher-order conditional random field(CRF)framework to estimate the final semantic labeling results with the consideration of the spatial contextual information.The proposed method is applied to the ISPRS 2D semantic labeling benchmark,and competitive overall accuracies of 90.6%and 85.0%are achieved for Vaihingen and Potsdam datasets,respectively.展开更多
As a deep learning network with an encoder-decoder architecture,UNet and its series of improved versions have been widely used in medical image segmentation with great applications.However,when used to segment targets...As a deep learning network with an encoder-decoder architecture,UNet and its series of improved versions have been widely used in medical image segmentation with great applications.However,when used to segment targets in 3D medical images such as magnetic resonance imaging(MRI),computed tomography(CT),these models do not model the relevance of images in vertical space,resulting in poor accurate analysis of consecutive slices of the same patient.On the other hand,the large amount of detail lost during the encoding process makes these models incapable of segmenting small-scale tumor targets.Aiming at the scene of small-scale target segmentation in 3D medical images,a fully new neural network model SUNet++is proposed on the basis of UNet and UNet++.SUNet++improves the existing models mainly in three aspects:1)the modeling strategy of slice superposition is used to thoroughly excavate the three dimensional information of the data;2)by adding an attention mechanism during the decoding process,small scale targets in the picture are retained and amplified;3)in the up-sampling process,the transposed convolution operation is used to further enhance the effect of the model.In order to verify the effect of the model,we collected and produced a dataset of hyperintensity MRI liver-stage images containing over 400 cases of liver nodules.Experimental results on both public and proprietary datasets demonstrate the superiority of SUNet++in small-scale target segmentation of three-dimensional medical images.展开更多
文摘To overcome the shortcomings of 1 D and 2 D Otsu’s thresholding techniques, the 3 D Otsu method has been developed.Among all Otsu’s methods, 3 D Otsu technique provides the best threshold values for the multi-level thresholding processes. In this paper, to improve the quality of segmented images, a simple and effective multilevel thresholding method is introduced. The proposed approach focuses on preserving edge detail by computing the 3 D Otsu along the fusion phenomena. The advantages of the presented scheme include higher quality outcomes, better preservation of tiny details and boundaries and reduced execution time with rising threshold levels. The fusion approach depends upon the differences between pixel intensity values within a small local space of an image;it aims to improve localized information after the thresholding process. The fusion of images based on local contrast can improve image segmentation performance by minimizing the loss of local contrast, loss of details and gray-level distributions. Results show that the proposed method yields more promising segmentation results when compared to conventional1 D Otsu, 2 D Otsu and 3 D Otsu methods, as evident from the objective and subjective evaluations.
文摘With the widespread application of deep learning in the field of computer vision,gradually allowing medical image technology to assist doctors in making diagnoses has great practical and research significance.Aiming at the shortcomings of the traditional U-Net model in 3D spatial information extraction,model over-fitting,and low degree of semantic information fusion,an improved medical image segmentation model has been used to achieve more accurate segmentation of medical images.In this model,we make full use of the residual network(ResNet)to solve the over-fitting problem.In order to process and aggregate data at different scales,the inception network is used instead of the traditional convolutional layer,and the dilated convolution is used to increase the receptive field.The conditional random field(CRF)can complete the contour refinement work.Compared with the traditional 3D U-Net network,the segmentation accuracy of the improved liver and tumor images increases by 2.89%and 7.66%,respectively.As a part of the image processing process,the method in this paper not only can be used for medical image segmentation,but also can lay the foundation for subsequent image 3D reconstruction work.
文摘In order to effectively improve the pathological diagnosis capability and feature resolution of 3D human brain CT images,a threshold segmentation method of multi-resolution 3D human brain CT image based on edge pixel grayscale feature decomposition is proposed in this paper.In this method,first,original 3D human brain image information is collected,and CT image filtering is performed to the collected information through the gradient value decomposition method,and edge contour features of the 3D human brain CT image are extracted.Then,the threshold segmentation method is adopted to segment the regional pixel feature block of the 3D human brain CT image to segment the image into block vectors with high-resolution feature points,and the 3D human brain CT image is reconstructed with the salient feature point as center.Simulation results show that the method proposed in this paper can provide accuracy up to 100%when the signal-to-noise ratio is 0,and with the increase of signal-to-noise ratio,the accuracy provided by this method is stable at 100%.Comparison results show that the threshold segmentation method of multi-resolution 3D human brain CT image based on edge pixel grayscale feature decomposition is signicantly better than traditional methods in pathological feature estimation accuracy,and it effectively improves the rapid pathological diagnosis and positioning recognition abilities to CT images.
基金supported by the National Key Research and Development Program of China under Grant No.2018YFE0206900the National Natural Science Foundation of China under Grant No.61871440 and CAAI‐Huawei Mind-Spore Open Fund.
文摘Tumour segmentation in medical images(especially 3D tumour segmentation)is highly challenging due to the possible similarity between tumours and adjacent tissues,occurrence of multiple tumours and variable tumour shapes and sizes.The popular deep learning‐based segmentation algorithms generally rely on the convolutional neural network(CNN)and Transformer.The former cannot extract the global image features effectively while the latter lacks the inductive bias and involves the complicated computation for 3D volume data.The existing hybrid CNN‐Transformer network can only provide the limited performance improvement or even poorer segmentation performance than the pure CNN.To address these issues,a short‐term and long‐term memory self‐attention network is proposed.Firstly,a distinctive self‐attention block uses the Transformer to explore the correlation among the region features at different levels extracted by the CNN.Then,the memory structure filters and combines the above information to exclude the similar regions and detect the multiple tumours.Finally,the multi‐layer reconstruction blocks will predict the tumour boundaries.Experimental results demonstrate that our method outperforms other methods in terms of subjective visual and quantitative evaluation.Compared with the most competitive method,the proposed method provides Dice(82.4%vs.76.6%)and Hausdorff distance 95%(HD95)(10.66 vs.11.54 mm)on the KiTS19 as well as Dice(80.2%vs.78.4%)and HD95(9.632 vs.12.17 mm)on the LiTS.
文摘Currently,deep learning is widely used in medical image segmentation and has achieved good results.However,3D medical image segmentation tasks with diverse lesion characters,blurred edges,and unstable positions require complex networks with a large number of parameters.It is computationally expensive and results in high requirements on equipment,making it hard to deploy the network in hospitals.In this work,we propose a method for network lightweighting and applied it to a 3D CNN based network.We experimented on a COVID-19 lesion segmentation dataset.Specifically,we use three cascaded one-dimensional convolutions to replace a 3D convolution,and integrate instance normalization with the previous layer of one-dimensional convolutions to accelerate network inference.In addition,we simplify test-time augmentation and deep supervision of the network.Experiments show that the lightweight network can reduce the prediction time of each sample and the memory usage by 50%and reduce the number of parameters by 60%compared with the original network.The training time of one epoch is also reduced by 50%with the segmentation accuracy dropped within the acceptable range.
基金PHD Site from Chinese Educational Department,Grant number:20040699015
文摘This paper presents a method which uses multiple types of expert knowledge together in 3D medical image segmentation based on rough set theory. The focus of this paper is how to approximate a ROI(region of interest) when there are multiple types of expert knowledge. Based on rough set theory, the image can be split into three regions: positive regions; negative regions; boundary regions. With multiple knowledge we refine ROI as an intersection of all of the expected shapes with single knowledge. At last we show the results of implementing a rough 3D image segmentation and visualization system.
基金supported by the National Basic Research Program of China (Grant No. 2011CB707701)the National Natural Science Foundation of China (Grant No. 60873124)+2 种基金the Joint Research Foundation of Beijing Education Committee (GrantNo. JD100010607)the International Science and Technology Supporting Programme (Grant No. 2008BAH26B00)the Zhejiang Service Robot Key Laboratory (Grant No. 2008E10004)
文摘In this paper we first determine three phases of cell images: background, cytoplasm and nucleolus according to the general physical characteristics of cell images, and then develop a variational model, based on these characteristics, to segment nucleolus and cytoplasm from their relatively complicated backgrounds. In the meantime, the preprocessing obtained information of cell images using the OTSU algorithm is used to initialize the level set function in the model, which can speed up the segmentation and present satisfactory results in cell image processing.
文摘A novel technique of three-dimensional (3D) reconstruction, segmentation, display and analysis of series slices of images including microscopic wide field optical sectioning by deconvolution method, cryo-electron microscope slices by Fou-rier-Bessel synthesis and electron tomography (ET), and a series of computed tomography (CT) was developed to perform si-multaneous measurement on the structure and function of biomedical samples. The paper presents the 3D reconstruction seg-mentation display and analysis results of pollen spore, chaperonin, virus, head, cervical bone, tibia and carpus. At the same time, it also puts forward some potential applications of the new technique in the biomedical realm.
基金Supported by National Science Foundation of China(No.81800878)Interdisciplinary Program of Shanghai Jiao Tong University(No.YG2017QN24)+1 种基金Key Technological Research Projects of Songjiang District(No.18sjkjgg24)Bethune Langmu Ophthalmological Research Fund for Young and Middle-aged People(No.BJ-LM2018002J)
文摘AIM: To explore a segmentation algorithm based on deep learning to achieve accurate diagnosis and treatment of patients with retinal fluid.METHODS: A two-dimensional(2D) fully convolutional network for retinal segmentation was employed. In order to solve the category imbalance in retinal optical coherence tomography(OCT) images, the network parameters and loss function based on the 2D fully convolutional network were modified. For this network, the correlations of corresponding positions among adjacent images in space are ignored. Thus, we proposed a three-dimensional(3D) fully convolutional network for segmentation in the retinal OCT images.RESULTS: The algorithm was evaluated according to segmentation accuracy, Kappa coefficient, and F1 score. For the 3D fully convolutional network proposed in this paper, the overall segmentation accuracy rate is 99.56%, Kappa coefficient is 98.47%, and F1 score of retinal fluid is 95.50%. CONCLUSION: The OCT image segmentation algorithm based on deep learning is primarily founded on the 2D convolutional network. The 3D network architecture proposed in this paper reduces the influence of category imbalance, realizes end-to-end segmentation of volume images, and achieves optimal segmentation results. The segmentation maps are practically the same as the manual annotations of doctors, and can provide doctors with more accurate diagnostic data.
基金Funded by the National Natural Science Foundation of China(No.51574201)the State Key Laboratory of Geohazard Prevention and Geoenvironment Protection Foundation(Chengdu University of Technology)(No.SKLGP2015K006)the Scientific&Technical Youth Innovation Group(Southwest Petroleum University)(No.2015CXTD05)
文摘As a kind of special material in geotechnical engineering, the mudded weak interlayer plays a crucial part in slope stability. In this paper, we presented a method to determine the threshold value of section micrographs of the mudded weak interlayer in slope during its meso-structure qualification process. Some soil tests, scanning electron microscopy(SEM) and image segmentation technology were performed to fulfill our purpose. Specifically, the relation between 3 D-porosity and the threshold was obtained by least square fitting of the threshold-porosity curves and a simplified pore equivalent model. Using this relation and the 3 D-porosity determined by soil experiments, we can figure out the polynomial equation of the threshold value. The threshold values obtained by the other existing methods in literature were employed to validate our present results.
文摘Deep learning (DL) has experienced an exponential development in recent years, with major impact in many medical fields, especially in the field of medical image and, respectively, as a specific task, in the segmentation of the medical image. We aim to create a computer assisted diagnostic method, optimized by the use of deep learning (DL) and validated by a randomized controlled clinical trial, is a highly automated tool for diagnosing and staging precancerous and cervical cancer and thyroid cancers. We aim to design a high-performance deep learning model, combined from convolutional neural network (U-Net)-based architectures, for segmentation of the medical image that is independent of the type of organs/tissues, dimensions or type of image (2D/3D) and to validate the DL model in a randomized, controlled clinical trial. We used as a methodology primarily the analysis of U-Net-based architectures to identify the key elements that we considered important in the design and optimization of the combined DL model, from the U-Net-based architectures, imagined by us. Secondly, we will validate the performance of the DL model through a randomized controlled clinical trial. The DL model designed by us will be a highly automated tool for diagnosing and staging precancers and cervical cancer and thyroid cancers. The combined model we designed takes into account the key features of each of the architectures Overcomplete Convolutional Network Kite-Net (Kite-Net), Attention gate mechanism is an improvement added on convolutional network architecture for fast and precise segmentation of images (Attention U-Net), Harmony Densely Connected Network-Medical image Segmentation (HarDNet-MSEG). In this regard, we will create a comprehensive computer assisted diagnostic methodology validated by a randomized controlled clinical trial. The model will be a highly automated tool for diagnosing and staging precancers and cervical cancer and thyroid cancers. This would help drastically minimize the time and effort that specialists put into analyzing medical images, help to achieve a better therapeutic plan, and can provide a “second opinion” of computer assisted diagnosis.
基金the PID2022‐137451OB‐I00 and PID2022‐137629OA‐I00 projects funded by the MICIU/AEIAEI/10.13039/501100011033 and by ERDF/EU.
文摘Cancer is one of the leading causes of death in the world,with radiotherapy as one of the treatment options.Radiotherapy planning starts with delineating the affected area from healthy organs,called organs at risk(OAR).A new approach to automatic OAR seg-mentation in the chest cavity in Computed Tomography(CT)images is presented.The proposed approach is based on the modified U‐Net architecture with the ResNet‐34 encoder,which is the baseline adopted in this work.The new two‐branch CS‐SA U‐Net architecture is proposed,which consists of two parallel U‐Net models in which self‐attention blocks with cosine similarity as query‐key similarity function(CS‐SA)blocks are inserted between the encoder and decoder,which enabled the use of con-sistency regularisation.The proposed solution demonstrates state‐of‐the‐art performance for the problem of OAR segmentation in CT images on the publicly available SegTHOR benchmark dataset in terms of a Dice coefficient(oesophagus-0.8714,heart-0.9516,trachea-0.9286,aorta-0.9510)and Hausdorff distance(oesophagus-0.2541,heart-0.1514,trachea-0.1722,aorta-0.1114)and significantly outperforms the baseline.The current approach is demonstrated to be viable for improving the quality of OAR segmentation for radiotherapy planning.
基金This work was supported in part by the National Natural Science Foundation of China under Grant U20A20225,61833013in part by Shaanxi Provincial Key Research and Development Program under Grant 2022-GY111.
文摘This paper proposes an improved high-precision 3D semantic mapping method for indoor scenes using RGB-D images.The current semantic mapping algorithms suffer from low semantic annotation accuracy and insufficient real-time performance.To address these issues,we first adopt the Elastic Fusion algorithm to select key frames from indoor environment image sequences captured by the Kinect sensor and construct the indoor environment space model.Then,an indoor RGB-D image semantic segmentation network is proposed,which uses multi-scale feature fusion to quickly and accurately obtain object labeling information at the pixel level of the spatial point cloud model.Finally,Bayesian updating is used to conduct incremental semantic label fusion on the established spatial point cloud model.We also employ dense conditional random fields(CRF)to optimize the 3D semantic map model,resulting in a high-precision spatial semantic map of indoor scenes.Experimental results show that the proposed semantic mapping system can process image sequences collected by RGB-D sensors in real-time and output accurate semantic segmentation results of indoor scene images and the current local spatial semantic map.Finally,it constructs a globally consistent high-precision indoor scenes 3D semantic map.
文摘Unmanned aerial vehicle(UAV)array InSAR is a new type of single-flight 3D SAR imaging system with the advantages of high coherence and resolution.However,due to the low altitude of the platform,the elevation ambiguity of the system is smaller than the maximal terrain elevation.Since the spatial phase unwrapping is conducted based on the assumption of phase continuity,the inappropriate ambiguity will cause significant unwrapping errors.In this work,a 3D phase unwrapping algorithm assisted by image segmentation is proposed to address this issue.The Markov random field(MRF)is utilized for image segmentation.The optimal spatial phase unwrapping network is achieved based on the segmentation results.The asymptotic 3D phase unwrapping is adopted to get the refined 3D reconstruction.Results based on the real airborne array-InSAR data show that the proposed method effectively improves the elevation ambiguity.
文摘An infant male presented with the rare anatomy consisting of situs solitus,concordant atrioventricular connections to L-looped ventricles,double outlet right ventricle(DORV),and hypoplastic aortic arch.6 months after neonatal aortic arch repair,the morphologic right ventricle function deteriorated,and surgical evaluation was undertaken to determine if either biventricular repair with a systemic morphologic left ventricle or right ventricular exclusion was possible.After initial echocardiography,magnetic resonance imaging(MRI)was used to create detailed axial and 4-dimensional(4D)images and 3-dimensional(3D)printed models.The detailed anatomy of this rare,complex case and its use in pre-surgical planning is presented.
文摘Coronavirus has infected more than 753 million people,ranging in severity from one person to another,where more than six million infected people died worldwide.Computer-aided diagnostic(CAD)with artificial intelligence(AI)showed outstanding performance in effectively diagnosing this virus in real-time.Computed tomography is a complementary diagnostic tool to clarify the damage of COVID-19 in the lungs even before symptoms appear in patients.This paper conducts a systematic literature review of deep learning methods for classifying the segmentation of COVID-19 infection in the lungs.We used the methodology of systematic reviews and meta-analyses(PRISMA)flow method.This research aims to systematically analyze the supervised deep learning methods,open resource datasets,data augmentation methods,and loss functions used for various segment shapes of COVID-19 infection from computerized tomography(CT)chest images.We have selected 56 primary studies relevant to the topic of the paper.We have compared different aspects of the algorithms used to segment infected areas in the CT images.Limitations to deep learning in the segmentation of infected areas still need to be developed to predict smaller regions of infection at the beginning of their appearance.
文摘Visual attention mechanisms allow humans to extract relevant and important information from raw input percepts. Many applications in robotics and computer vision have modeled human visual attention mechanisms using a bottom-up data centric approach. In contrast, recent studies in cognitive science highlight advantages of a top-down approach to the attention mechanisms, especially in applications involving goal-directed search. In this paper, we propose a top-down approach for extracting salient objects/regions of space. The top-down methodology first isolates different objects in an unorganized point cloud, and compares each object for uniqueness. A measure of saliency using the properties of geodesic distance on the object’s surface is defined. Our method works on 3D point cloud data, and identifies salient objects of high curvature and unique silhouette. These being the most unique features of a scene, are robust to clutter, occlusions and view point changes. We provide the details of the proposed method and initial experimental results.
文摘In civil aviation security screening, laptops, with their intricate structural composition, provide the potential for criminals to conceal dangerous items. Presently, the security process necessitates passengers to individually present their laptops for inspection. The paper introduced a method for laptop removal. By combining projection algorithms with the YOLOv7-Seg model, a laptop’s three views were generated through projection, and instance segmentation of these views was achieved using YOLOv7-Seg. The resulting 2D masks from instance segmentation at different angles were employed to reconstruct a 3D mask through angle restoration. Ultimately, the intersection of this 3D mask with the original 3D data enabled the successful extraction of the laptop’s 3D information. Experimental results demonstrated that the fusion of projection and instance segmentation facilitated the automatic removal of laptops from CT data. Moreover, higher instance segmentation model accuracy leads to more precise removal outcomes. By implementing the laptop removal functionality, the civil aviation security screening process becomes more efficient and convenient. Passengers will no longer be required to individually handle their laptops, effectively enhancing the efficiency and accuracy of security screening.
基金was funded by the Key Laboratory of Surveying and Mapping Science and Geospatial Information Technology of Ministry of Natural Resources[grant number 2020-2-1]the National Natural Science Foundation of China[grant number 41871372].
文摘Semantic segmentation of remote sensing images is an important but unsolved problem in the remote sensing society.Advanced image semantic segmentation models,such as DeepLabv3+,have achieved astonishing performance for semantically labeling very high resolution(VHR)remote sensing images.However,it is difficult for these models to capture the precise outlines of ground objects and explore the context information that revealing relationships among image objects for optimizing segmentation results.Consequently,this study proposes a semantic segmentation method for VHR images by incorporating deep learning semantic segmentation model(DeepLabv3+)and objectbased image analysis(OBIA),wherein DSM is employed to provide geometric information to enhance the interpretation of VHR images.The proposed method first obtains two initial probabilistic labeling predictions using a DeepLabv3+network on spectral image and a random forest(RF)classifier on hand-crafted features,respectively.These two predictions are then integrated by Dempster-Shafer(D-S)evidence theory to be fed into an object-constrained higher-order conditional random field(CRF)framework to estimate the final semantic labeling results with the consideration of the spatial contextual information.The proposed method is applied to the ISPRS 2D semantic labeling benchmark,and competitive overall accuracies of 90.6%and 85.0%are achieved for Vaihingen and Potsdam datasets,respectively.
基金This work was supported by the National Natural Science Foundation of China(No.62072135)Natural Science Foundation of Ningxia Hui Autonomous Region(No.2022AAC03346)Fundamental Research Funds for the Central Universities(No.3072020CF0602).
文摘As a deep learning network with an encoder-decoder architecture,UNet and its series of improved versions have been widely used in medical image segmentation with great applications.However,when used to segment targets in 3D medical images such as magnetic resonance imaging(MRI),computed tomography(CT),these models do not model the relevance of images in vertical space,resulting in poor accurate analysis of consecutive slices of the same patient.On the other hand,the large amount of detail lost during the encoding process makes these models incapable of segmenting small-scale tumor targets.Aiming at the scene of small-scale target segmentation in 3D medical images,a fully new neural network model SUNet++is proposed on the basis of UNet and UNet++.SUNet++improves the existing models mainly in three aspects:1)the modeling strategy of slice superposition is used to thoroughly excavate the three dimensional information of the data;2)by adding an attention mechanism during the decoding process,small scale targets in the picture are retained and amplified;3)in the up-sampling process,the transposed convolution operation is used to further enhance the effect of the model.In order to verify the effect of the model,we collected and produced a dataset of hyperintensity MRI liver-stage images containing over 400 cases of liver nodules.Experimental results on both public and proprietary datasets demonstrate the superiority of SUNet++in small-scale target segmentation of three-dimensional medical images.