Feature extraction of signals plays an important role in classification problems because of data dimension reduction property and potential improvement of a classification accuracy rate. Principal component analysis (...Feature extraction of signals plays an important role in classification problems because of data dimension reduction property and potential improvement of a classification accuracy rate. Principal component analysis (PCA), wavelets transform or Fourier transform methods are often used for feature extraction. In this paper, we propose a multi-scale PCA, which combines discrete wavelet transform, and PCA for feature extraction of signals in both the spatial and temporal domains. Our study shows that the multi-scale PCA combined with the proposed new classification methods leads to high classification accuracy for the considered signals.展开更多
Tea leaf picking is a crucial stage in tea production that directly influences the quality and value of the tea.Traditional tea-picking machines may compromise the quality of the tea leaves.High-quality teas are often...Tea leaf picking is a crucial stage in tea production that directly influences the quality and value of the tea.Traditional tea-picking machines may compromise the quality of the tea leaves.High-quality teas are often handpicked and need more delicate operations in intelligent picking machines.Compared with traditional image processing techniques,deep learning models have stronger feature extraction capabilities,and better generalization and are more suitable for practical tea shoot harvesting.However,current research mostly focuses on shoot detection and cannot directly accomplish end-to-end shoot segmentation tasks.We propose a tea shoot instance segmentation model based on multi-scale mixed attention(Mask2FusionNet)using a dataset from the tea garden in Hangzhou.We further analyzed the characteristics of the tea shoot dataset,where the proportion of small to medium-sized targets is 89.9%.Our algorithm is compared with several mainstream object segmentation algorithms,and the results demonstrate that our model achieves an accuracy of 82%in recognizing the tea shoots,showing a better performance compared to other models.Through ablation experiments,we found that ResNet50,PointRend strategy,and the Feature Pyramid Network(FPN)architecture can improve performance by 1.6%,1.4%,and 2.4%,respectively.These experiments demonstrated that our proposed multi-scale and point selection strategy optimizes the feature extraction capability for overlapping small targets.The results indicate that the proposed Mask2FusionNet model can perform the shoot segmentation in unstructured environments,realizing the individual distinction of tea shoots,and complete extraction of the shoot edge contours with a segmentation accuracy of 82.0%.The research results can provide algorithmic support for the segmentation and intelligent harvesting of premium tea shoots at different scales.展开更多
In order to extract the richer feature information of ship targets from sea clutter, and address the high dimensional data problem, a method termed as multi-scale fusion kernel sparse preserving projection(MSFKSPP) ba...In order to extract the richer feature information of ship targets from sea clutter, and address the high dimensional data problem, a method termed as multi-scale fusion kernel sparse preserving projection(MSFKSPP) based on the maximum margin criterion(MMC) is proposed for recognizing the class of ship targets utilizing the high-resolution range profile(HRRP). Multi-scale fusion is introduced to capture the local and detailed information in small-scale features, and the global and contour information in large-scale features, offering help to extract the edge information from sea clutter and further improving the target recognition accuracy. The proposed method can maximally preserve the multi-scale fusion sparse of data and maximize the class separability in the reduced dimensionality by reproducing kernel Hilbert space. Experimental results on the measured radar data show that the proposed method can effectively extract the features of ship target from sea clutter, further reduce the feature dimensionality, and improve target recognition performance.展开更多
The success of intelligent transportation systems relies heavily on accurate traffic prediction,in which how to model the underlying spatial-temporal information from traffic data has come under the spotlight.Most exi...The success of intelligent transportation systems relies heavily on accurate traffic prediction,in which how to model the underlying spatial-temporal information from traffic data has come under the spotlight.Most existing frameworks typically utilize separate modules for spatial and temporal correlations modeling.However,this stepwise pattern may limit the effectiveness and efficiency in spatial-temporal feature extraction and cause the overlook of important information in some steps.Furthermore,it is lacking sufficient guidance from prior information while modeling based on a given spatial adjacency graph(e.g.,deriving from the geodesic distance or approximate connectivity),and may not reflect the actual interaction between nodes.To overcome those limitations,our paper proposes a spatial-temporal graph synchronous aggregation(STGSA)model to extract the localized and long-term spatial-temporal dependencies simultaneously.Specifically,a tailored graph aggregation method in the vertex domain is designed to extract spatial and temporal features in one graph convolution process.In each STGSA block,we devise a directed temporal correlation graph to represent the localized and long-term dependencies between nodes,and the potential temporal dependence is further fine-tuned by an adaptive weighting operation.Meanwhile,we construct an elaborated spatial adjacency matrix to represent the road sensor graph by considering both physical distance and node similarity in a datadriven manner.Then,inspired by the multi-head attention mechanism which can jointly emphasize information from different r epresentation subspaces,we construct a multi-stream module based on the STGSA blocks to capture global information.It projects the embedding input repeatedly with multiple different channels.Finally,the predicted values are generated by stacking several multi-stream modules.Extensive experiments are constructed on six real-world datasets,and numerical results show that the proposed STGSA model significantly outperforms the benchmarks.展开更多
Background Recurrent recovery is a common method for video super-resolution(VSR)that models the correlation between frames via hidden states.However,the application of this structure in real-world scenarios can lead t...Background Recurrent recovery is a common method for video super-resolution(VSR)that models the correlation between frames via hidden states.However,the application of this structure in real-world scenarios can lead to unsatisfactory artifacts.We found that in real-world VSR training,the use of unknown and complex degradation can better simulate the degradation process in the real world.Methods Based on this,we propose the RealFuVSR model,which simulates real-world degradation and mitigates artifacts caused by the VSR.Specifically,we propose a multiscale feature extraction module(MSF)module that extracts and fuses features from multiple scales,thereby facilitating the elimination of hidden state artifacts.To improve the accuracy of the hidden state alignment information,RealFuVSR uses an advanced optical flow-guided deformable convolution.Moreover,a cascaded residual upsampling module was used to eliminate noise caused by the upsampling process.Results The experiment demonstrates that RealFuVSR model can not only recover high-quality videos but also outperforms the state-of-the-art RealBasicVSR and RealESRGAN models.展开更多
Aiming at the difficulty of fault identification caused by manual extraction of fault features of rotating machinery,a one-dimensional multi-scale convolutional auto-encoder fault diagnosis model is proposed,based on ...Aiming at the difficulty of fault identification caused by manual extraction of fault features of rotating machinery,a one-dimensional multi-scale convolutional auto-encoder fault diagnosis model is proposed,based on the standard convolutional auto-encoder.In this model,the parallel convolutional and deconvolutional kernels of different scales are used to extract the features from the input signal and reconstruct the input signal;then the feature map extracted by multi-scale convolutional kernels is used as the input of the classifier;and finally the parameters of the whole model are fine-tuned using labeled data.Experiments on one set of simulation fault data and two sets of rolling bearing fault data are conducted to validate the proposed method.The results show that the model can achieve 99.75%,99.3%and 100%diagnostic accuracy,respectively.In addition,the diagnostic accuracy and reconstruction error of the one-dimensional multi-scale convolutional auto-encoder are compared with traditional machine learning,convolutional neural networks and a traditional convolutional auto-encoder.The final results show that the proposed model has a better recognition effect for rolling bearing fault data.展开更多
Along with the progression of Internet of Things(IoT)technology,network terminals are becoming continuously more intelligent.IoT has been widely applied in various scenarios,including urban infrastructure,transportati...Along with the progression of Internet of Things(IoT)technology,network terminals are becoming continuously more intelligent.IoT has been widely applied in various scenarios,including urban infrastructure,transportation,industry,personal life,and other socio-economic fields.The introduction of deep learning has brought new security challenges,like an increment in abnormal traffic,which threatens network security.Insufficient feature extraction leads to less accurate classification results.In abnormal traffic detection,the data of network traffic is high-dimensional and complex.This data not only increases the computational burden of model training but also makes information extraction more difficult.To address these issues,this paper proposes an MD-MRD-ResNeXt model for abnormal network traffic detection.To fully utilize the multi-scale information in network traffic,a Multi-scale Dilated feature extraction(MD)block is introduced.This module can effectively understand and process information at various scales and uses dilated convolution technology to significantly broaden the model’s receptive field.The proposed Max-feature-map Residual with Dual-channel pooling(MRD)block integrates the maximum feature map with the residual block.This module ensures the model focuses on key information,thereby optimizing computational efficiency and reducing unnecessary information redundancy.Experimental results show that compared to the latest methods,the proposed abnormal traffic detection model improves accuracy by about 2%.展开更多
Diseases in tea trees can result in significant losses in both the quality and quantity of tea production.Regular monitoring can help to prevent the occurrence of large-scale diseases in tea plantations.However,existi...Diseases in tea trees can result in significant losses in both the quality and quantity of tea production.Regular monitoring can help to prevent the occurrence of large-scale diseases in tea plantations.However,existingmethods face challenges such as a high number of parameters and low recognition accuracy,which hinders their application in tea plantation monitoring equipment.This paper presents a lightweight I-MobileNetV2 model for identifying diseases in tea leaves,to address these challenges.The proposed method first embeds a Coordinate Attention(CA)module into the originalMobileNetV2 network,enabling the model to locate disease regions accurately.Secondly,a Multi-branch Parallel Convolution(MPC)module is employed to extract disease features across multiple scales,improving themodel’s adaptability to different disease scales.Finally,the AutoML for Model Compression(AMC)is used to compress themodel and reduce computational complexity.Experimental results indicate that our proposed algorithm attains an average accuracy of 96.12%on our self-built tea leaf disease dataset,surpassing the original MobileNetV2 by 1.91%.Furthermore,the number of model parameters have been reduced by 40%,making itmore suitable for practical application in tea plantation environments.展开更多
Breast cancer is a significant threat to the global population,affecting not only women but also a threat to the entire population.With recent advancements in digital pathology,Eosin and hematoxylin images provide enh...Breast cancer is a significant threat to the global population,affecting not only women but also a threat to the entire population.With recent advancements in digital pathology,Eosin and hematoxylin images provide enhanced clarity in examiningmicroscopic features of breast tissues based on their staining properties.Early cancer detection facilitates the quickening of the therapeutic process,thereby increasing survival rates.The analysis made by medical professionals,especially pathologists,is time-consuming and challenging,and there arises a need for automated breast cancer detection systems.The upcoming artificial intelligence platforms,especially deep learning models,play an important role in image diagnosis and prediction.Initially,the histopathology biopsy images are taken from standard data sources.Further,the gathered images are given as input to the Multi-Scale Dilated Vision Transformer,where the essential features are acquired.Subsequently,the features are subjected to the Bidirectional Long Short-Term Memory(Bi-LSTM)for classifying the breast cancer disorder.The efficacy of the model is evaluated using divergent metrics.When compared with other methods,the proposed work reveals that it offers impressive results for detection.展开更多
Recently,with the urgent demand for data-driven approaches in practical industrial scenarios,the deep learning diagnosis model in noise environments has attracted increasing attention.However,the existing research has...Recently,with the urgent demand for data-driven approaches in practical industrial scenarios,the deep learning diagnosis model in noise environments has attracted increasing attention.However,the existing research has two limitations:(1)the complex and changeable environmental noise,which cannot ensure the high-performance diagnosis of the model in different noise domains and(2)the possibility of multiple faults occurring simultaneously,which brings challenges to the model diagnosis.This paper presents a novel anti-noise multi-scale convolutional neural network(AM-CNN)for solving the issue of compound fault diagnosis under different intensity noises.First,we propose a residual pre-processing block according to the principle of noise superposition to process the input information and present the residual loss to construct a new loss function.Additionally,considering the strong coupling of input information,we design a multi-scale convolution block to realize multi-scale feature extraction for enhancing the proposed model’s robustness and effectiveness.Finally,a multi-label classifier is utilized to simultaneously distinguish multiple bearing faults.The proposed AM-CNN is verified under our collected compound fault dataset.On average,AM-CNN improves 39.93%accuracy and 25.84%F1-macro under the no-noise working condition and 45.67%accuracy and 27.72%F1-macro under different intensity noise working conditions compared with the existing methods.Furthermore,the experimental results show that AM-CNN can achieve good cross-domain performance with 100%accuracy and 100%F1-macro.Thus,AM-CNN has the potential to be an accurate and stable fault diagnosis tool.展开更多
As a highly vascular eye part,the choroid is crucial in various eye disease diagnoses.However,limited research has focused on the inner structure of the choroid due to the challenges in obtaining sufficient accurate l...As a highly vascular eye part,the choroid is crucial in various eye disease diagnoses.However,limited research has focused on the inner structure of the choroid due to the challenges in obtaining sufficient accurate label data,particularly for the choroidal vessels.Meanwhile,the existing direct choroidal vessel segmentation methods for the intelligent diagnosis of vascular assisted ophthalmic diseases are still unsatisfactory due to noise data,while the synergistic segmentation methods compromise vessel segmentation performance for the choroid layer segmentation tasks.Common cascaded structures grapple with error propagation during training.To address these challenges,we propose a cascade learning segmentation method for the inner vessel structures of the choroid in this paper.Specifically,we propose TransformerAssisted Cascade Learning Network(TACLNet)for choroidal vessel segmentation,which comprises a two-stage training strategy:pre-training for choroid layer segmentation and joint training for choroid layer and choroidal vessel segmentation.We also enhance the skip connection structures by introducing a multi-scale subtraction connection module designated as MSC,capturing differential and detailed information simultaneously.Additionally,we implement an auxiliary Transformer branch named ATB to integrate global features into the segmentation process.Experimental results exhibit that our method achieves the state-of-the-art performance for choroidal vessel segmentation.Besides,we further validate the significant superiority of the proposed method for retinal fluid segmentation in optical coherence tomography(OCT)scans on a publicly available dataset.All these fully prove that our TACLNet contributes to the advancement of choroidal vessel segmentation and is of great significance for ophthalmic research and clinical application.展开更多
文摘Feature extraction of signals plays an important role in classification problems because of data dimension reduction property and potential improvement of a classification accuracy rate. Principal component analysis (PCA), wavelets transform or Fourier transform methods are often used for feature extraction. In this paper, we propose a multi-scale PCA, which combines discrete wavelet transform, and PCA for feature extraction of signals in both the spatial and temporal domains. Our study shows that the multi-scale PCA combined with the proposed new classification methods leads to high classification accuracy for the considered signals.
基金This research was supported by the National Natural Science Foundation of China No.62276086the National Key R&D Program of China No.2022YFD2000100Zhejiang Provincial Natural Science Foundation of China under Grant No.LTGN23D010002.
文摘Tea leaf picking is a crucial stage in tea production that directly influences the quality and value of the tea.Traditional tea-picking machines may compromise the quality of the tea leaves.High-quality teas are often handpicked and need more delicate operations in intelligent picking machines.Compared with traditional image processing techniques,deep learning models have stronger feature extraction capabilities,and better generalization and are more suitable for practical tea shoot harvesting.However,current research mostly focuses on shoot detection and cannot directly accomplish end-to-end shoot segmentation tasks.We propose a tea shoot instance segmentation model based on multi-scale mixed attention(Mask2FusionNet)using a dataset from the tea garden in Hangzhou.We further analyzed the characteristics of the tea shoot dataset,where the proportion of small to medium-sized targets is 89.9%.Our algorithm is compared with several mainstream object segmentation algorithms,and the results demonstrate that our model achieves an accuracy of 82%in recognizing the tea shoots,showing a better performance compared to other models.Through ablation experiments,we found that ResNet50,PointRend strategy,and the Feature Pyramid Network(FPN)architecture can improve performance by 1.6%,1.4%,and 2.4%,respectively.These experiments demonstrated that our proposed multi-scale and point selection strategy optimizes the feature extraction capability for overlapping small targets.The results indicate that the proposed Mask2FusionNet model can perform the shoot segmentation in unstructured environments,realizing the individual distinction of tea shoots,and complete extraction of the shoot edge contours with a segmentation accuracy of 82.0%.The research results can provide algorithmic support for the segmentation and intelligent harvesting of premium tea shoots at different scales.
基金supported by the National Natural Science Foundation of China (62271255,61871218)the Fundamental Research Funds for the Central University (3082019NC2019002)+1 种基金the Aeronautical Science Foundation (ASFC-201920007002)the Program of Remote Sensing Intelligent Monitoring and Emergency Services for Regional Security Elements。
文摘In order to extract the richer feature information of ship targets from sea clutter, and address the high dimensional data problem, a method termed as multi-scale fusion kernel sparse preserving projection(MSFKSPP) based on the maximum margin criterion(MMC) is proposed for recognizing the class of ship targets utilizing the high-resolution range profile(HRRP). Multi-scale fusion is introduced to capture the local and detailed information in small-scale features, and the global and contour information in large-scale features, offering help to extract the edge information from sea clutter and further improving the target recognition accuracy. The proposed method can maximally preserve the multi-scale fusion sparse of data and maximize the class separability in the reduced dimensionality by reproducing kernel Hilbert space. Experimental results on the measured radar data show that the proposed method can effectively extract the features of ship target from sea clutter, further reduce the feature dimensionality, and improve target recognition performance.
基金partially supported by the National Key Research and Development Program of China(2020YFB2104001)。
文摘The success of intelligent transportation systems relies heavily on accurate traffic prediction,in which how to model the underlying spatial-temporal information from traffic data has come under the spotlight.Most existing frameworks typically utilize separate modules for spatial and temporal correlations modeling.However,this stepwise pattern may limit the effectiveness and efficiency in spatial-temporal feature extraction and cause the overlook of important information in some steps.Furthermore,it is lacking sufficient guidance from prior information while modeling based on a given spatial adjacency graph(e.g.,deriving from the geodesic distance or approximate connectivity),and may not reflect the actual interaction between nodes.To overcome those limitations,our paper proposes a spatial-temporal graph synchronous aggregation(STGSA)model to extract the localized and long-term spatial-temporal dependencies simultaneously.Specifically,a tailored graph aggregation method in the vertex domain is designed to extract spatial and temporal features in one graph convolution process.In each STGSA block,we devise a directed temporal correlation graph to represent the localized and long-term dependencies between nodes,and the potential temporal dependence is further fine-tuned by an adaptive weighting operation.Meanwhile,we construct an elaborated spatial adjacency matrix to represent the road sensor graph by considering both physical distance and node similarity in a datadriven manner.Then,inspired by the multi-head attention mechanism which can jointly emphasize information from different r epresentation subspaces,we construct a multi-stream module based on the STGSA blocks to capture global information.It projects the embedding input repeatedly with multiple different channels.Finally,the predicted values are generated by stacking several multi-stream modules.Extensive experiments are constructed on six real-world datasets,and numerical results show that the proposed STGSA model significantly outperforms the benchmarks.
基金Supported by Open Project of the Ministry of Industry and Information Technology Key Laboratory of Performance and Reliability Testing and Evaluation for Basic Software and Hardware。
文摘Background Recurrent recovery is a common method for video super-resolution(VSR)that models the correlation between frames via hidden states.However,the application of this structure in real-world scenarios can lead to unsatisfactory artifacts.We found that in real-world VSR training,the use of unknown and complex degradation can better simulate the degradation process in the real world.Methods Based on this,we propose the RealFuVSR model,which simulates real-world degradation and mitigates artifacts caused by the VSR.Specifically,we propose a multiscale feature extraction module(MSF)module that extracts and fuses features from multiple scales,thereby facilitating the elimination of hidden state artifacts.To improve the accuracy of the hidden state alignment information,RealFuVSR uses an advanced optical flow-guided deformable convolution.Moreover,a cascaded residual upsampling module was used to eliminate noise caused by the upsampling process.Results The experiment demonstrates that RealFuVSR model can not only recover high-quality videos but also outperforms the state-of-the-art RealBasicVSR and RealESRGAN models.
基金The National Natural Science Foundation of China(No.51675098)
文摘Aiming at the difficulty of fault identification caused by manual extraction of fault features of rotating machinery,a one-dimensional multi-scale convolutional auto-encoder fault diagnosis model is proposed,based on the standard convolutional auto-encoder.In this model,the parallel convolutional and deconvolutional kernels of different scales are used to extract the features from the input signal and reconstruct the input signal;then the feature map extracted by multi-scale convolutional kernels is used as the input of the classifier;and finally the parameters of the whole model are fine-tuned using labeled data.Experiments on one set of simulation fault data and two sets of rolling bearing fault data are conducted to validate the proposed method.The results show that the model can achieve 99.75%,99.3%and 100%diagnostic accuracy,respectively.In addition,the diagnostic accuracy and reconstruction error of the one-dimensional multi-scale convolutional auto-encoder are compared with traditional machine learning,convolutional neural networks and a traditional convolutional auto-encoder.The final results show that the proposed model has a better recognition effect for rolling bearing fault data.
基金supported by the Key Research and Development Program of Xinjiang Uygur Autonomous Region(No.2022B01008)the National Natural Science Foundation of China(No.62363032)+4 种基金the Natural Science Foundation of Xinjiang Uygur Autonomous Region(No.2023D01C20)the Scientific Research Foundation of Higher Education(No.XJEDU2022P011)National Science and Technology Major Project(No.2022ZD0115803)Tianshan Innovation Team Program of Xinjiang Uygur Autonomous Region(No.2023D14012)the“Heaven Lake Doctor”Project(No.202104120018).
文摘Along with the progression of Internet of Things(IoT)technology,network terminals are becoming continuously more intelligent.IoT has been widely applied in various scenarios,including urban infrastructure,transportation,industry,personal life,and other socio-economic fields.The introduction of deep learning has brought new security challenges,like an increment in abnormal traffic,which threatens network security.Insufficient feature extraction leads to less accurate classification results.In abnormal traffic detection,the data of network traffic is high-dimensional and complex.This data not only increases the computational burden of model training but also makes information extraction more difficult.To address these issues,this paper proposes an MD-MRD-ResNeXt model for abnormal network traffic detection.To fully utilize the multi-scale information in network traffic,a Multi-scale Dilated feature extraction(MD)block is introduced.This module can effectively understand and process information at various scales and uses dilated convolution technology to significantly broaden the model’s receptive field.The proposed Max-feature-map Residual with Dual-channel pooling(MRD)block integrates the maximum feature map with the residual block.This module ensures the model focuses on key information,thereby optimizing computational efficiency and reducing unnecessary information redundancy.Experimental results show that compared to the latest methods,the proposed abnormal traffic detection model improves accuracy by about 2%.
基金supported by National Key Research and Development Program(No.2016YFD0201305-07)Guizhou Provincial Basic Research Program(Natural Science)(No.ZK[2023]060)Open Fund Project in Semiconductor Power Device Reliability Engineering Center of Ministry of Education(No.ERCMEKFJJ2019-06).
文摘Diseases in tea trees can result in significant losses in both the quality and quantity of tea production.Regular monitoring can help to prevent the occurrence of large-scale diseases in tea plantations.However,existingmethods face challenges such as a high number of parameters and low recognition accuracy,which hinders their application in tea plantation monitoring equipment.This paper presents a lightweight I-MobileNetV2 model for identifying diseases in tea leaves,to address these challenges.The proposed method first embeds a Coordinate Attention(CA)module into the originalMobileNetV2 network,enabling the model to locate disease regions accurately.Secondly,a Multi-branch Parallel Convolution(MPC)module is employed to extract disease features across multiple scales,improving themodel’s adaptability to different disease scales.Finally,the AutoML for Model Compression(AMC)is used to compress themodel and reduce computational complexity.Experimental results indicate that our proposed algorithm attains an average accuracy of 96.12%on our self-built tea leaf disease dataset,surpassing the original MobileNetV2 by 1.91%.Furthermore,the number of model parameters have been reduced by 40%,making itmore suitable for practical application in tea plantation environments.
基金Deanship of Research and Graduate Studies at King Khalid University for funding this work through Small Group Research Project under Grant Number RGP1/261/45.
文摘Breast cancer is a significant threat to the global population,affecting not only women but also a threat to the entire population.With recent advancements in digital pathology,Eosin and hematoxylin images provide enhanced clarity in examiningmicroscopic features of breast tissues based on their staining properties.Early cancer detection facilitates the quickening of the therapeutic process,thereby increasing survival rates.The analysis made by medical professionals,especially pathologists,is time-consuming and challenging,and there arises a need for automated breast cancer detection systems.The upcoming artificial intelligence platforms,especially deep learning models,play an important role in image diagnosis and prediction.Initially,the histopathology biopsy images are taken from standard data sources.Further,the gathered images are given as input to the Multi-Scale Dilated Vision Transformer,where the essential features are acquired.Subsequently,the features are subjected to the Bidirectional Long Short-Term Memory(Bi-LSTM)for classifying the breast cancer disorder.The efficacy of the model is evaluated using divergent metrics.When compared with other methods,the proposed work reveals that it offers impressive results for detection.
基金supported by the National Key R&D Program of China(Grant No.2020YFB1709604)the State Key Laboratory of Mechanical System and Vibration(Grant No.MSVZD202103)+1 种基金the Shanghai Municipal Science and Technology Major Project(Grant No.2021SHZDZX0102)。
文摘Recently,with the urgent demand for data-driven approaches in practical industrial scenarios,the deep learning diagnosis model in noise environments has attracted increasing attention.However,the existing research has two limitations:(1)the complex and changeable environmental noise,which cannot ensure the high-performance diagnosis of the model in different noise domains and(2)the possibility of multiple faults occurring simultaneously,which brings challenges to the model diagnosis.This paper presents a novel anti-noise multi-scale convolutional neural network(AM-CNN)for solving the issue of compound fault diagnosis under different intensity noises.First,we propose a residual pre-processing block according to the principle of noise superposition to process the input information and present the residual loss to construct a new loss function.Additionally,considering the strong coupling of input information,we design a multi-scale convolution block to realize multi-scale feature extraction for enhancing the proposed model’s robustness and effectiveness.Finally,a multi-label classifier is utilized to simultaneously distinguish multiple bearing faults.The proposed AM-CNN is verified under our collected compound fault dataset.On average,AM-CNN improves 39.93%accuracy and 25.84%F1-macro under the no-noise working condition and 45.67%accuracy and 27.72%F1-macro under different intensity noise working conditions compared with the existing methods.Furthermore,the experimental results show that AM-CNN can achieve good cross-domain performance with 100%accuracy and 100%F1-macro.Thus,AM-CNN has the potential to be an accurate and stable fault diagnosis tool.
基金supported by the National Natural Science Foundation of China under Grant Nos.62301330 and 62101346the Guangdong Basic and Applied Basic Research Foundation under Grant Nos.20231121103807001,2022A1515110101the Guangdong Provincial Key Laboratory under Grant No.2023B1212060076.
文摘As a highly vascular eye part,the choroid is crucial in various eye disease diagnoses.However,limited research has focused on the inner structure of the choroid due to the challenges in obtaining sufficient accurate label data,particularly for the choroidal vessels.Meanwhile,the existing direct choroidal vessel segmentation methods for the intelligent diagnosis of vascular assisted ophthalmic diseases are still unsatisfactory due to noise data,while the synergistic segmentation methods compromise vessel segmentation performance for the choroid layer segmentation tasks.Common cascaded structures grapple with error propagation during training.To address these challenges,we propose a cascade learning segmentation method for the inner vessel structures of the choroid in this paper.Specifically,we propose TransformerAssisted Cascade Learning Network(TACLNet)for choroidal vessel segmentation,which comprises a two-stage training strategy:pre-training for choroid layer segmentation and joint training for choroid layer and choroidal vessel segmentation.We also enhance the skip connection structures by introducing a multi-scale subtraction connection module designated as MSC,capturing differential and detailed information simultaneously.Additionally,we implement an auxiliary Transformer branch named ATB to integrate global features into the segmentation process.Experimental results exhibit that our method achieves the state-of-the-art performance for choroidal vessel segmentation.Besides,we further validate the significant superiority of the proposed method for retinal fluid segmentation in optical coherence tomography(OCT)scans on a publicly available dataset.All these fully prove that our TACLNet contributes to the advancement of choroidal vessel segmentation and is of great significance for ophthalmic research and clinical application.