Oscillation detection has been a hot research topic in industries due to the high incidence of oscillation loops and their negative impact on plant profitability.Although numerous automatic detection techniques have b...Oscillation detection has been a hot research topic in industries due to the high incidence of oscillation loops and their negative impact on plant profitability.Although numerous automatic detection techniques have been proposed,most of them can only address part of the practical difficulties.An oscillation is heuristically defined as a visually apparent periodic variation.However,manual visual inspection is labor-intensive and prone to missed detection.Convolutional neural networks(CNNs),inspired by animal visual systems,have been raised with powerful feature extraction capabilities.In this work,an exploration of the typical CNN models for visual oscillation detection is performed.Specifically,we tested MobileNet-V1,ShuffleNet-V2,Efficient Net-B0,and GhostNet models,and found that such a visual framework is well-suited for oscillation detection.The feasibility and validity of this framework are verified utilizing extensive numerical and industrial cases.Compared with state-of-theart oscillation detectors,the suggested framework is more straightforward and more robust to noise and mean-nonstationarity.In addition,this framework generalizes well and is capable of handling features that are not present in the training data,such as multiple oscillations and outliers.展开更多
Image processing networks have gained great success in many fields,and thus the issue of copyright protection for image processing networks hasbecome a focus of attention. Model watermarking techniques are widely used...Image processing networks have gained great success in many fields,and thus the issue of copyright protection for image processing networks hasbecome a focus of attention. Model watermarking techniques are widely usedin model copyright protection, but there are two challenges: (1) designinguniversal trigger sample watermarking for different network models is stilla challenge;(2) existing methods of copyright protection based on trigger swatermarking are difficult to resist forgery attacks. In this work, we propose adual model watermarking framework for copyright protection in image processingnetworks. The trigger sample watermark is embedded in the trainingprocess of the model, which can effectively verify the model copyright. And wedesign a common method for generating trigger sample watermarks based ongenerative adversarial networks, adaptively generating trigger sample watermarksaccording to different models. The spatial watermark is embedded intothe model output. When an attacker steals model copyright using a forgedtrigger sample watermark, which can be correctly extracted to distinguishbetween the piratical and the protected model. The experiments show that theproposed framework has good performance in different image segmentationnetworks of UNET, UNET++, and FCN (fully convolutional network), andeffectively resists forgery attacks.展开更多
We redesign the parameterized quantum circuit in the quantum deep neural network, construct a three-layer structure as the hidden layer, and then use classical optimization algorithms to train the parameterized quantu...We redesign the parameterized quantum circuit in the quantum deep neural network, construct a three-layer structure as the hidden layer, and then use classical optimization algorithms to train the parameterized quantum circuit, thereby propose a novel hybrid quantum deep neural network(HQDNN) used for image classification. After bilinear interpolation reduces the original image to a suitable size, an improved novel enhanced quantum representation(INEQR) is used to encode it into quantum states as the input of the HQDNN. Multi-layer parameterized quantum circuits are used as the main structure to implement feature extraction and classification. The output results of parameterized quantum circuits are converted into classical data through quantum measurements and then optimized on a classical computer. To verify the performance of the HQDNN, we conduct binary classification and three classification experiments on the MNIST(Modified National Institute of Standards and Technology) data set. In the first binary classification, the accuracy of 0 and 4 exceeds98%. Then we compare the performance of three classification with other algorithms, the results on two datasets show that the classification accuracy is higher than that of quantum deep neural network and general quantum convolutional neural network.展开更多
Watermarks can provide reliable and secure copyright protection for optical coherence tomography(OCT)fundus images.The effective image segmentation is helpful for promoting OCT image watermarking.However,OCT images ha...Watermarks can provide reliable and secure copyright protection for optical coherence tomography(OCT)fundus images.The effective image segmentation is helpful for promoting OCT image watermarking.However,OCT images have a large amount of low-quality data,which seriously affects the performance of segmentationmethods.Therefore,this paper proposes an effective segmentation method for OCT fundus image watermarking using a rough convolutional neural network(RCNN).First,the rough-set-based feature discretization module is designed to preprocess the input data.Second,a dual attention mechanism for feature channels and spatial regions in the CNN is added to enable the model to adaptively select important information for fusion.Finally,the refinement module for enhancing the extraction power of multi-scale information is added to improve the edge accuracy in segmentation.RCNN is compared with CE-Net and MultiResUNet on 83 gold standard 3D retinal OCT data samples.The average dice similarly coefficient(DSC)obtained by RCNN is 6%higher than that of CE-Net.The average 95 percent Hausdorff distance(95HD)and average symmetric surface distance(ASD)obtained by RCNN are 32.4%and 33.3%lower than those of MultiResUNet,respectively.We also evaluate the effect of feature discretization,as well as analyze the initial learning rate of RCNN and conduct ablation experiments with the four different models.The experimental results indicate that our method can improve the segmentation accuracy of OCT fundus images,providing strong support for its application in medical image watermarking.展开更多
The neuron model has been widely employed in neural-morphic computing systems and chaotic circuits.This study aims to develop a novel circuit simulation of a three-neuron Hopfield neural network(HNN)with coupled hyper...The neuron model has been widely employed in neural-morphic computing systems and chaotic circuits.This study aims to develop a novel circuit simulation of a three-neuron Hopfield neural network(HNN)with coupled hyperbolic memristors through the modification of a single coupling connection weight.The bistable mode of the hyperbolic memristive HNN(mHNN),characterized by the coexistence of asymmetric chaos and periodic attractors,is effectively demonstrated through the utilization of conventional nonlinear analysis techniques.These techniques include bifurcation diagrams,two-parameter maximum Lyapunov exponent plots,local attractor basins,and phase trajectory diagrams.Moreover,an encryption technique for color images is devised by leveraging the mHNN model and asymmetric structural attractors.This method demonstrates significant benefits in correlation,information entropy,and resistance to differential attacks,providing strong evidence for its effectiveness in encryption.Additionally,an improved modular circuit design method is employed to create the analog equivalent circuit of the memristive HNN.The correctness of the circuit design is confirmed through Multisim simulations,which align with numerical simulations conducted in Matlab.展开更多
Convolutional neural networks (CNNs) are widely used in image classification tasks, but their increasing model size and computation make them challenging to implement on embedded systems with constrained hardware reso...Convolutional neural networks (CNNs) are widely used in image classification tasks, but their increasing model size and computation make them challenging to implement on embedded systems with constrained hardware resources. To address this issue, the MobileNetV1 network was developed, which employs depthwise convolution to reduce network complexity. MobileNetV1 employs a stride of 2 in several convolutional layers to decrease the spatial resolution of feature maps, thereby lowering computational costs. However, this stride setting can lead to a loss of spatial information, particularly affecting the detection and representation of smaller objects or finer details in images. To maintain the trade-off between complexity and model performance, a lightweight convolutional neural network with hierarchical multi-scale feature fusion based on the MobileNetV1 network is proposed. The network consists of two main subnetworks. The first subnetwork uses a depthwise dilated separable convolution (DDSC) layer to learn imaging features with fewer parameters, which results in a lightweight and computationally inexpensive network. Furthermore, depthwise dilated convolution in DDSC layer effectively expands the field of view of filters, allowing them to incorporate a larger context. The second subnetwork is a hierarchical multi-scale feature fusion (HMFF) module that uses parallel multi-resolution branches architecture to process the input feature map in order to extract the multi-scale feature information of the input image. Experimental results on the CIFAR-10, Malaria, and KvasirV1 datasets demonstrate that the proposed method is efficient, reducing the network parameters and computational cost by 65.02% and 39.78%, respectively, while maintaining the network performance compared to the MobileNetV1 baseline.展开更多
Blind image quality assessment(BIQA)is of fundamental importance in low-level computer vision community.Increasing interest has been drawn in exploiting deep neural networks for BIQA.Despite of the notable success ach...Blind image quality assessment(BIQA)is of fundamental importance in low-level computer vision community.Increasing interest has been drawn in exploiting deep neural networks for BIQA.Despite of the notable success achieved,there is a broad consensus that training deep convolutional neural networks(DCNN)heavily relies on massive annotated data.Unfortunately,BIQA is typically a small sample problem,resulting the generalization ability of BIQA severely restricted.In order to improve the accuracy and generalization ability of BIQA metrics,this work proposed a totally opinion-unaware BIQA in which no subjective annotations are involved in the training stage.Multiple full-reference image quality assessment(FR-IQA)metrics are employed to label the distorted image as a substitution of subjective quality annotation.A deep neural network(DNN)is trained to blindly predict the multiple FR-IQA score in absence of corresponding pristine image.In the end,a selfsupervised FR-IQA score aggregator implemented by adversarial auto-encoder pools the predictions of multiple FR-IQA scores into the final quality predicting score.Even though none of subjective scores are involved in the training stage,experimental results indicate that our proposed full reference induced BIQA framework is as competitive as state-of-the-art BIQA metrics.展开更多
This paper presented an online quality inspection system based on artificial neural networks. Chromatism classification and edge detection are two difficult problems in glass steel surface quality inspection. Two arti...This paper presented an online quality inspection system based on artificial neural networks. Chromatism classification and edge detection are two difficult problems in glass steel surface quality inspection. Two artificial neural networks were made and the two problems were solved. The one solved chromatism classification. Hue, saturation and their probability of three colors, whose appearing probabilities were maximum in color histogram, were selected as input parameters, and the number of output node could be adjusted with the change of requirement. The other solved edge detection. In this neutral network, edge detection of gray scale image was able to be tested with trained neural networks for a binary image. It prevent the difficulty that the number of needed training samples was too large if gray scale images were directly regarded as training samples. This system is able to be applied to not only glass steel fault inspection but also other product online quality inspection and classification.展开更多
Automatic segmentation of medical images provides a reliable scientific basis for disease diagnosis and analysis.Notably,most existing methods that combine the strengths of convolutional neural networks(CNNs)and Trans...Automatic segmentation of medical images provides a reliable scientific basis for disease diagnosis and analysis.Notably,most existing methods that combine the strengths of convolutional neural networks(CNNs)and Transformers have made significant progress.However,there are some limitations in the current integration of CNN and Transformer technology in two key aspects.Firstly,most methods either overlook or fail to fully incorporate the complementary nature between local and global features.Secondly,the significance of integrating the multiscale encoder features from the dual-branch network to enhance the decoding features is often disregarded in methods that combine CNN and Transformer.To address this issue,we present a groundbreaking dual-branch cross-attention fusion network(DCFNet),which efficiently combines the power of Swin Transformer and CNN to generate complementary global and local features.We then designed the Feature Cross-Fusion(FCF)module to efficiently fuse local and global features.In the FCF,the utilization of the Channel-wise Cross-fusion Transformer(CCT)serves the purpose of aggregatingmulti-scale features,and the Feature FusionModule(FFM)is employed to effectively aggregate dual-branch prominent feature regions from the spatial perspective.Furthermore,within the decoding phase of the dual-branch network,our proposed Channel Attention Block(CAB)aims to emphasize the significance of the channel features between the up-sampled features and the features generated by the FCFmodule to enhance the details of the decoding.Experimental results demonstrate that DCFNet exhibits enhanced accuracy in segmentation performance.Compared to other state-of-the-art(SOTA)methods,our segmentation framework exhibits a superior level of competitiveness.DCFNet’s accurate segmentation of medical images can greatly assist medical professionals in making crucial diagnoses of lesion areas in advance.展开更多
In intelligent perception and diagnosis of medical equipment,the visual and morphological changes in retinal vessels are closely related to the severity of cardiovascular diseases(e.g.,diabetes and hypertension).Intel...In intelligent perception and diagnosis of medical equipment,the visual and morphological changes in retinal vessels are closely related to the severity of cardiovascular diseases(e.g.,diabetes and hypertension).Intelligent auxiliary diagnosis of these diseases depends on the accuracy of the retinal vascular segmentation results.To address this challenge,we design a Dual-Branch-UNet framework,which comprises a Dual-Branch encoder structure for feature extraction based on the traditional U-Net model for medical image segmentation.To be more explicit,we utilize a novel parallel encoder made up of various convolutional modules to enhance the encoder portion of the original U-Net.Then,image features are combined at each layer to produce richer semantic data and the model’s capacity is adjusted to various input images.Meanwhile,in the lower sampling section,we give up pooling and conduct the lower sampling by convolution operation to control step size for information fusion.We also employ an attentionmodule in the decoder stage to filter the image noises so as to lessen the response of irrelevant features.Experiments are verified and compared on the DRIVE and ARIA datasets for retinal vessels segmentation.The proposed Dual-Branch-UNet has proved to be superior to other five typical state-of-the-art methods.展开更多
With limited number of labeled samples,hyperspectral image(HSI)classification is a difficult Problem in current research.The graph neural network(GNN)has emerged as an approach to semi-supervised classification,and th...With limited number of labeled samples,hyperspectral image(HSI)classification is a difficult Problem in current research.The graph neural network(GNN)has emerged as an approach to semi-supervised classification,and the application of GNN to hyperspectral images has attracted much attention.However,in the existing GNN-based methods a single graph neural network or graph filter is mainly used to extract HSI features,which does not take full advantage of various graph neural networks(graph filters).Moreover,the traditional GNNs have the problem of oversmoothing.To alleviate these shortcomings,we introduce a deep hybrid multi-graph neural network(DHMG),where two different graph filters,i.e.,the spectral filter and the autoregressive moving average(ARMA)filter,are utilized in two branches.The former can well extract the spectral features of the nodes,and the latter has a good suppression effect on graph noise.The network realizes information interaction between the two branches and takes good advantage of different graph filters.In addition,to address the problem of oversmoothing,a dense network is proposed,where the local graph features are preserved.The dense structure satisfies the needs of different classification targets presenting different features.Finally,we introduce a GraphSAGEbased network to refine the graph features produced by the deep hybrid network.Extensive experiments on three public HSI datasets strongly demonstrate that the DHMG dramatically outperforms the state-ofthe-art models.展开更多
Image processing technique was employed to analyze pitting corrosion morphologies of 304 stainless steel exposed to FeCl3 environments. BP neural network models were developed for the prediction of pitting corrosion m...Image processing technique was employed to analyze pitting corrosion morphologies of 304 stainless steel exposed to FeCl3 environments. BP neural network models were developed for the prediction of pitting corrosion mass loss using the obtained data of the total and the average pit areas which were extracted from pitting binary image. The results showed that the predicted results obtained by the 2-5-1 type BP neural network model are in good agreement with the experimental data of pitting corrosion mass loss. The maximum relative error of prediction is 6.78%.展开更多
Considering the problem that the scattering echo images of airborne Doppler weather radar are often reduced by ground clutters,the accuracy and confidence of meteorology target detection are reduced.In this paper,a de...Considering the problem that the scattering echo images of airborne Doppler weather radar are often reduced by ground clutters,the accuracy and confidence of meteorology target detection are reduced.In this paper,a deep convolutional neural network(DCNN)is proposed for meteorology target detection and ground clutter suppression with a large collection of airborne weather radar images as network input.For each weather radar image,the corresponding digital elevation model(DEM)image is extracted on basis of the radar antenna scan-ning parameters and plane position,and is further fed to the net-work as a supplement for ground clutter suppression.The fea-tures of actual meteorology targets are learned in each bottle-neck module of the proposed network and convolved into deeper iterations in the forward propagation process.Then the network parameters are updated by the back propagation itera-tion of the training error.Experimental results on the real mea-sured images show that our proposed DCNN outperforms the counterparts in terms of six evaluation factors.Meanwhile,the network outputs are in good agreement with the expected mete-orology detection results(labels).It is demonstrated that the pro-posed network would have a promising meteorology observa-tion application with minimal effort on network variables or parameter changes.展开更多
Observing and analyzing surface images is critical for studying the interaction between plasma and irradiated plasma-facing materials.This paper presents a method for the automatic recognition of bubbles in transmissi...Observing and analyzing surface images is critical for studying the interaction between plasma and irradiated plasma-facing materials.This paper presents a method for the automatic recognition of bubbles in transmission electron microscope(TEM)images of W nanofibers using image processing techniques and convolutional neural network(CNN).We employ a three-stage approach consisting of Otsu,local-threshold,and watershed segmentation to extract bubbles from noisy images.To address over-segmentation,we propose a combination of area factor and radial pixel intensity scanning.A CNN is used to recognize bubbles,outperforming traditional neural network models such as Alex Net and Google Net with an accuracy of 97.1%and recall of 98.6%.Our method is tested on both clear and blurred TEM images,and demonstrates humanlike performance in recognizing bubbles.This work contributes to the development of quantitative image analysis in the field of plasma-material interactions,offering a scalable solution for analyzing material defects.Overall,this study's findings establish the potential for automatic defect recognition and its applications in the assessment of plasma-material interactions.This method can be employed in a variety of specialties,including plasma physics and materials science.展开更多
Deep learning has been widely used in the field of mammographic image classification owing to its superiority in automatic feature extraction.However,general deep learning models cannot achieve very satisfactory class...Deep learning has been widely used in the field of mammographic image classification owing to its superiority in automatic feature extraction.However,general deep learning models cannot achieve very satisfactory classification results on mammographic images because these models are not specifically designed for mammographic images and do not take the specific traits of these images into account.To exploit the essential discriminant information of mammographic images,we propose a novel classification method based on a convolutional neural network.Specifically,the proposed method designs two branches to extract the discriminative features from mammographic images from the mediolateral oblique and craniocaudal(CC)mammographic views.The features extracted from the two-view mammographic images contain complementary information that enables breast cancer to be more easily distinguished.Moreover,the attention block is introduced to capture the channel-wise information by adjusting the weight of each feature map,which is beneficial to emphasising the important features of mammographic images.Furthermore,we add a penalty term based on the fuzzy cluster algorithm to the cross-entropy function,which improves the generalisation ability of the classification model by maximising the interclass distance and minimising the intraclass distance of the samples.The experimental results on The Digital database for Screening Mammography INbreast and MIAS mammography databases illustrate that the proposed method achieves the best classification performance and is more robust than the compared state-ofthe-art classification methods.展开更多
With the rapid development of sports,the number of sports images has increased dramatically.Intelligent and automatic processing and analysis of moving images are significant,which can not only facilitate users to qui...With the rapid development of sports,the number of sports images has increased dramatically.Intelligent and automatic processing and analysis of moving images are significant,which can not only facilitate users to quickly search and access moving images but also facilitate staff to store and manage moving image data and contribute to the intellectual development of the sports industry.In this paper,a method of table tennis identification and positioning based on a convolutional neural network is proposed,which solves the problem that the identification and positioning method based on color features and contour features is not adaptable in various environments.At the same time,the learning methods and techniques of table tennis detection,positioning,and trajectory prediction are studied.A deep learning framework for recognition learning of rotating flying table tennis is put forward.The mechanism and methods of positioning,trajectory prediction,and intelligent automatic processing of moving images are studied,and the self-built data sets are trained and verified.展开更多
With the development of artificial intelligence-related technologies such as deep learning,various organizations,including the government,are making various efforts to generate and manage big data for use in artificia...With the development of artificial intelligence-related technologies such as deep learning,various organizations,including the government,are making various efforts to generate and manage big data for use in artificial intelligence.However,it is difficult to acquire big data due to various social problems and restrictions such as personal information leakage.There are many problems in introducing technology in fields that do not have enough training data necessary to apply deep learning technology.Therefore,this study proposes a mixed contour data augmentation technique,which is a data augmentation technique using contour images,to solve a problem caused by a lack of data.ResNet,a famous convolutional neural network(CNN)architecture,and CIFAR-10,a benchmark data set,are used for experimental performance evaluation to prove the superiority of the proposed method.And to prove that high performance improvement can be achieved even with a small training dataset,the ratio of the training dataset was divided into 70%,50%,and 30%for comparative analysis.As a result of applying the mixed contour data augmentation technique,it was possible to achieve a classification accuracy improvement of up to 4.64%and high accuracy even with a small amount of data set.In addition,it is expected that the mixed contour data augmentation technique can be applied in various fields by proving the excellence of the proposed data augmentation technique using benchmark datasets.展开更多
Time series classification(TSC)has attracted a lot of attention for time series data mining tasks and has been applied in various fields.With the success of deep learning(DL)in computer vision recognition,people are s...Time series classification(TSC)has attracted a lot of attention for time series data mining tasks and has been applied in various fields.With the success of deep learning(DL)in computer vision recognition,people are starting to use deep learning to tackle TSC tasks.Quantum neural networks(QNN)have recently demonstrated their superiority over traditional machine learning in methods such as image processing and natural language processing,but research using quantum neural networks to handle TSC tasks has not received enough attention.Therefore,we proposed a learning framework based on multiple imaging and hybrid QNN(MIHQNN)for TSC tasks.We investigate the possibility of converting 1D time series to 2D images and classifying the converted images using hybrid QNN.We explored the differences between MIHQNN based on single time series imaging and MIHQNN based on the fusion of multiple time series imaging.Four quantum circuits were also selected and designed to study the impact of quantum circuits on TSC tasks.We tested our method on several standard datasets and achieved significant results compared to several current TSC methods,demonstrating the effectiveness of MIHQNN.This research highlights the potential of applying quantum computing to TSC and provides the theoretical and experimental background for future research.展开更多
A variety of neural networks have been presented to deal with issues in deep learning in the last decades.Despite the prominent success achieved by the neural network,it still lacks theoretical guidance to design an e...A variety of neural networks have been presented to deal with issues in deep learning in the last decades.Despite the prominent success achieved by the neural network,it still lacks theoretical guidance to design an efficient neural network model,and verifying the performance of a model needs excessive resources.Previous research studies have demonstrated that many existing models can be regarded as different numerical discretizations of differential equations.This connection sheds light on designing an effective recurrent neural network(RNN)by resorting to numerical analysis.Simple RNN is regarded as a discretisation of the forward Euler scheme.Considering the limited solution accuracy of the forward Euler methods,a Taylor‐type discrete scheme is presented with lower truncation error and a Taylor‐type RNN(T‐RNN)is designed with its guidance.Extensive experiments are conducted to evaluate its performance on statistical language models and emotion analysis tasks.The noticeable gains obtained by T‐RNN present its superiority and the feasibility of designing the neural network model using numerical methods.展开更多
COVID-19 disease caused by the SARS-CoV-2 virus has created social and economic disruption across theworld.The ability of the COVID-19 virus to quickly mutate and transfer has created serious concerns across the world...COVID-19 disease caused by the SARS-CoV-2 virus has created social and economic disruption across theworld.The ability of the COVID-19 virus to quickly mutate and transfer has created serious concerns across the world.It is essential to detectCOVID-19 infection caused by different variants to take preventive measures accordingly.The existing method of detection of infections caused by COVID-19 and its variants is costly and time-consuming.The impacts of theCOVID-19 pandemic in developing countries are very drastic due to the unavailability of medical facilities and infrastructure to handle the pandemic.Pneumonia is the major symptom of COVID-19 infection.The radiology of the lungs in varies in the case of bacterial pneumonia as compared to COVID-19-caused pneumonia.The pattern of pneumonia in lungs in radiology images can also be used to identify the cause associated with pneumonia.In this paper,we propose the methodology of identifying the cause(either due to COVID-19 or other types of infections)of pneumonia from radiology images.Furthermore,because different variants of COVID-19 lead to different patterns of pneumonia,the proposed methodology identifies pneumonia,the COVID-19 caused pneumonia,and Omicron caused pneumonia from the radiology images.To fulfill the above-mentioned tasks,we have used three Convolution Neural Networks(CNNs)at each stage of the proposed methodology.The results unveil that the proposed step-by-step solution enhances the accuracy of pneumonia detection along with finding its cause,despite having a limited dataset.展开更多
基金the National Natural Science Foundation of China(62003298,62163036)the Major Project of Science and Technology of Yunnan Province(202202AD080005,202202AH080009)the Yunnan University Professional Degree Graduate Practice Innovation Fund Project(ZC-22222770)。
文摘Oscillation detection has been a hot research topic in industries due to the high incidence of oscillation loops and their negative impact on plant profitability.Although numerous automatic detection techniques have been proposed,most of them can only address part of the practical difficulties.An oscillation is heuristically defined as a visually apparent periodic variation.However,manual visual inspection is labor-intensive and prone to missed detection.Convolutional neural networks(CNNs),inspired by animal visual systems,have been raised with powerful feature extraction capabilities.In this work,an exploration of the typical CNN models for visual oscillation detection is performed.Specifically,we tested MobileNet-V1,ShuffleNet-V2,Efficient Net-B0,and GhostNet models,and found that such a visual framework is well-suited for oscillation detection.The feasibility and validity of this framework are verified utilizing extensive numerical and industrial cases.Compared with state-of-theart oscillation detectors,the suggested framework is more straightforward and more robust to noise and mean-nonstationarity.In addition,this framework generalizes well and is capable of handling features that are not present in the training data,such as multiple oscillations and outliers.
基金supported by the National Natural Science Foundation of China under grants U1836208,by the Priority Academic Program Development of Jiangsu Higher Education Institutions (PAPD)fundby the Collaborative Innovation Center of Atmospheric Environment and Equipment Technology (CICAEET)fund,China.
文摘Image processing networks have gained great success in many fields,and thus the issue of copyright protection for image processing networks hasbecome a focus of attention. Model watermarking techniques are widely usedin model copyright protection, but there are two challenges: (1) designinguniversal trigger sample watermarking for different network models is stilla challenge;(2) existing methods of copyright protection based on trigger swatermarking are difficult to resist forgery attacks. In this work, we propose adual model watermarking framework for copyright protection in image processingnetworks. The trigger sample watermark is embedded in the trainingprocess of the model, which can effectively verify the model copyright. And wedesign a common method for generating trigger sample watermarks based ongenerative adversarial networks, adaptively generating trigger sample watermarksaccording to different models. The spatial watermark is embedded intothe model output. When an attacker steals model copyright using a forgedtrigger sample watermark, which can be correctly extracted to distinguishbetween the piratical and the protected model. The experiments show that theproposed framework has good performance in different image segmentationnetworks of UNET, UNET++, and FCN (fully convolutional network), andeffectively resists forgery attacks.
基金Project supported by the Natural Science Foundation of Shandong Province,China (Grant No. ZR2021MF049)the Joint Fund of Natural Science Foundation of Shandong Province (Grant Nos. ZR2022LLZ012 and ZR2021LLZ001)。
文摘We redesign the parameterized quantum circuit in the quantum deep neural network, construct a three-layer structure as the hidden layer, and then use classical optimization algorithms to train the parameterized quantum circuit, thereby propose a novel hybrid quantum deep neural network(HQDNN) used for image classification. After bilinear interpolation reduces the original image to a suitable size, an improved novel enhanced quantum representation(INEQR) is used to encode it into quantum states as the input of the HQDNN. Multi-layer parameterized quantum circuits are used as the main structure to implement feature extraction and classification. The output results of parameterized quantum circuits are converted into classical data through quantum measurements and then optimized on a classical computer. To verify the performance of the HQDNN, we conduct binary classification and three classification experiments on the MNIST(Modified National Institute of Standards and Technology) data set. In the first binary classification, the accuracy of 0 and 4 exceeds98%. Then we compare the performance of three classification with other algorithms, the results on two datasets show that the classification accuracy is higher than that of quantum deep neural network and general quantum convolutional neural network.
基金the China Postdoctoral Science Foundation under Grant 2021M701838the Natural Science Foundation of Hainan Province of China under Grants 621MS042 and 622MS067the Hainan Medical University Teaching Achievement Award Cultivation under Grant HYjcpx202209.
文摘Watermarks can provide reliable and secure copyright protection for optical coherence tomography(OCT)fundus images.The effective image segmentation is helpful for promoting OCT image watermarking.However,OCT images have a large amount of low-quality data,which seriously affects the performance of segmentationmethods.Therefore,this paper proposes an effective segmentation method for OCT fundus image watermarking using a rough convolutional neural network(RCNN).First,the rough-set-based feature discretization module is designed to preprocess the input data.Second,a dual attention mechanism for feature channels and spatial regions in the CNN is added to enable the model to adaptively select important information for fusion.Finally,the refinement module for enhancing the extraction power of multi-scale information is added to improve the edge accuracy in segmentation.RCNN is compared with CE-Net and MultiResUNet on 83 gold standard 3D retinal OCT data samples.The average dice similarly coefficient(DSC)obtained by RCNN is 6%higher than that of CE-Net.The average 95 percent Hausdorff distance(95HD)and average symmetric surface distance(ASD)obtained by RCNN are 32.4%and 33.3%lower than those of MultiResUNet,respectively.We also evaluate the effect of feature discretization,as well as analyze the initial learning rate of RCNN and conduct ablation experiments with the four different models.The experimental results indicate that our method can improve the segmentation accuracy of OCT fundus images,providing strong support for its application in medical image watermarking.
基金Project supported by the National Nature Science Foundation of China(Grant Nos.51737003 and 51977060)the Natural Science Foundation of Hebei Province(Grant No.E2011202051).
文摘The neuron model has been widely employed in neural-morphic computing systems and chaotic circuits.This study aims to develop a novel circuit simulation of a three-neuron Hopfield neural network(HNN)with coupled hyperbolic memristors through the modification of a single coupling connection weight.The bistable mode of the hyperbolic memristive HNN(mHNN),characterized by the coexistence of asymmetric chaos and periodic attractors,is effectively demonstrated through the utilization of conventional nonlinear analysis techniques.These techniques include bifurcation diagrams,two-parameter maximum Lyapunov exponent plots,local attractor basins,and phase trajectory diagrams.Moreover,an encryption technique for color images is devised by leveraging the mHNN model and asymmetric structural attractors.This method demonstrates significant benefits in correlation,information entropy,and resistance to differential attacks,providing strong evidence for its effectiveness in encryption.Additionally,an improved modular circuit design method is employed to create the analog equivalent circuit of the memristive HNN.The correctness of the circuit design is confirmed through Multisim simulations,which align with numerical simulations conducted in Matlab.
文摘Convolutional neural networks (CNNs) are widely used in image classification tasks, but their increasing model size and computation make them challenging to implement on embedded systems with constrained hardware resources. To address this issue, the MobileNetV1 network was developed, which employs depthwise convolution to reduce network complexity. MobileNetV1 employs a stride of 2 in several convolutional layers to decrease the spatial resolution of feature maps, thereby lowering computational costs. However, this stride setting can lead to a loss of spatial information, particularly affecting the detection and representation of smaller objects or finer details in images. To maintain the trade-off between complexity and model performance, a lightweight convolutional neural network with hierarchical multi-scale feature fusion based on the MobileNetV1 network is proposed. The network consists of two main subnetworks. The first subnetwork uses a depthwise dilated separable convolution (DDSC) layer to learn imaging features with fewer parameters, which results in a lightweight and computationally inexpensive network. Furthermore, depthwise dilated convolution in DDSC layer effectively expands the field of view of filters, allowing them to incorporate a larger context. The second subnetwork is a hierarchical multi-scale feature fusion (HMFF) module that uses parallel multi-resolution branches architecture to process the input feature map in order to extract the multi-scale feature information of the input image. Experimental results on the CIFAR-10, Malaria, and KvasirV1 datasets demonstrate that the proposed method is efficient, reducing the network parameters and computational cost by 65.02% and 39.78%, respectively, while maintaining the network performance compared to the MobileNetV1 baseline.
基金supported by the Public Welfare Technology Application Research Project of Zhejiang Province,China(No.LGF21F010001)the Key Research and Development Program of Zhejiang Province,China(Grant No.2019C01002)the Key Research and Development Program of Zhejiang Province,China(Grant No.2021C03138)。
文摘Blind image quality assessment(BIQA)is of fundamental importance in low-level computer vision community.Increasing interest has been drawn in exploiting deep neural networks for BIQA.Despite of the notable success achieved,there is a broad consensus that training deep convolutional neural networks(DCNN)heavily relies on massive annotated data.Unfortunately,BIQA is typically a small sample problem,resulting the generalization ability of BIQA severely restricted.In order to improve the accuracy and generalization ability of BIQA metrics,this work proposed a totally opinion-unaware BIQA in which no subjective annotations are involved in the training stage.Multiple full-reference image quality assessment(FR-IQA)metrics are employed to label the distorted image as a substitution of subjective quality annotation.A deep neural network(DNN)is trained to blindly predict the multiple FR-IQA score in absence of corresponding pristine image.In the end,a selfsupervised FR-IQA score aggregator implemented by adversarial auto-encoder pools the predictions of multiple FR-IQA scores into the final quality predicting score.Even though none of subjective scores are involved in the training stage,experimental results indicate that our proposed full reference induced BIQA framework is as competitive as state-of-the-art BIQA metrics.
基金Supported by Science and Technology Fundation (China University of Geosciences) (No.200520)
文摘This paper presented an online quality inspection system based on artificial neural networks. Chromatism classification and edge detection are two difficult problems in glass steel surface quality inspection. Two artificial neural networks were made and the two problems were solved. The one solved chromatism classification. Hue, saturation and their probability of three colors, whose appearing probabilities were maximum in color histogram, were selected as input parameters, and the number of output node could be adjusted with the change of requirement. The other solved edge detection. In this neutral network, edge detection of gray scale image was able to be tested with trained neural networks for a binary image. It prevent the difficulty that the number of needed training samples was too large if gray scale images were directly regarded as training samples. This system is able to be applied to not only glass steel fault inspection but also other product online quality inspection and classification.
基金supported by the National Key R&D Program of China(2018AAA0102100)the National Natural Science Foundation of China(No.62376287)+3 种基金the International Science and Technology Innovation Joint Base of Machine Vision and Medical Image Processing in Hunan Province(2021CB1013)the Key Research and Development Program of Hunan Province(2022SK2054)the Natural Science Foundation of Hunan Province(No.2022JJ30762,2023JJ70016)the 111 Project under Grant(No.B18059).
文摘Automatic segmentation of medical images provides a reliable scientific basis for disease diagnosis and analysis.Notably,most existing methods that combine the strengths of convolutional neural networks(CNNs)and Transformers have made significant progress.However,there are some limitations in the current integration of CNN and Transformer technology in two key aspects.Firstly,most methods either overlook or fail to fully incorporate the complementary nature between local and global features.Secondly,the significance of integrating the multiscale encoder features from the dual-branch network to enhance the decoding features is often disregarded in methods that combine CNN and Transformer.To address this issue,we present a groundbreaking dual-branch cross-attention fusion network(DCFNet),which efficiently combines the power of Swin Transformer and CNN to generate complementary global and local features.We then designed the Feature Cross-Fusion(FCF)module to efficiently fuse local and global features.In the FCF,the utilization of the Channel-wise Cross-fusion Transformer(CCT)serves the purpose of aggregatingmulti-scale features,and the Feature FusionModule(FFM)is employed to effectively aggregate dual-branch prominent feature regions from the spatial perspective.Furthermore,within the decoding phase of the dual-branch network,our proposed Channel Attention Block(CAB)aims to emphasize the significance of the channel features between the up-sampled features and the features generated by the FCFmodule to enhance the details of the decoding.Experimental results demonstrate that DCFNet exhibits enhanced accuracy in segmentation performance.Compared to other state-of-the-art(SOTA)methods,our segmentation framework exhibits a superior level of competitiveness.DCFNet’s accurate segmentation of medical images can greatly assist medical professionals in making crucial diagnoses of lesion areas in advance.
基金supported by National Natural Science Foundation of China(NSFC)(61976123,62072213)Taishan Young Scholars Program of Shandong Provinceand Key Development Program for Basic Research of Shandong Province(ZR2020ZD44).
文摘In intelligent perception and diagnosis of medical equipment,the visual and morphological changes in retinal vessels are closely related to the severity of cardiovascular diseases(e.g.,diabetes and hypertension).Intelligent auxiliary diagnosis of these diseases depends on the accuracy of the retinal vascular segmentation results.To address this challenge,we design a Dual-Branch-UNet framework,which comprises a Dual-Branch encoder structure for feature extraction based on the traditional U-Net model for medical image segmentation.To be more explicit,we utilize a novel parallel encoder made up of various convolutional modules to enhance the encoder portion of the original U-Net.Then,image features are combined at each layer to produce richer semantic data and the model’s capacity is adjusted to various input images.Meanwhile,in the lower sampling section,we give up pooling and conduct the lower sampling by convolution operation to control step size for information fusion.We also employ an attentionmodule in the decoder stage to filter the image noises so as to lessen the response of irrelevant features.Experiments are verified and compared on the DRIVE and ARIA datasets for retinal vessels segmentation.The proposed Dual-Branch-UNet has proved to be superior to other five typical state-of-the-art methods.
文摘With limited number of labeled samples,hyperspectral image(HSI)classification is a difficult Problem in current research.The graph neural network(GNN)has emerged as an approach to semi-supervised classification,and the application of GNN to hyperspectral images has attracted much attention.However,in the existing GNN-based methods a single graph neural network or graph filter is mainly used to extract HSI features,which does not take full advantage of various graph neural networks(graph filters).Moreover,the traditional GNNs have the problem of oversmoothing.To alleviate these shortcomings,we introduce a deep hybrid multi-graph neural network(DHMG),where two different graph filters,i.e.,the spectral filter and the autoregressive moving average(ARMA)filter,are utilized in two branches.The former can well extract the spectral features of the nodes,and the latter has a good suppression effect on graph noise.The network realizes information interaction between the two branches and takes good advantage of different graph filters.In addition,to address the problem of oversmoothing,a dense network is proposed,where the local graph features are preserved.The dense structure satisfies the needs of different classification targets presenting different features.Finally,we introduce a GraphSAGEbased network to refine the graph features produced by the deep hybrid network.Extensive experiments on three public HSI datasets strongly demonstrate that the DHMG dramatically outperforms the state-ofthe-art models.
文摘Image processing technique was employed to analyze pitting corrosion morphologies of 304 stainless steel exposed to FeCl3 environments. BP neural network models were developed for the prediction of pitting corrosion mass loss using the obtained data of the total and the average pit areas which were extracted from pitting binary image. The results showed that the predicted results obtained by the 2-5-1 type BP neural network model are in good agreement with the experimental data of pitting corrosion mass loss. The maximum relative error of prediction is 6.78%.
基金supported by the China Ministry of Industry and Information Technology Foundation and Aeronautical Science Foundation of China(ASFC-201920007002)the National Key Research and Development Plan(2021YFB1600603)the Open Fund of Key Laboratory of Civil Aircraft Airworthiness Technology,Civil Aviation University of China.
文摘Considering the problem that the scattering echo images of airborne Doppler weather radar are often reduced by ground clutters,the accuracy and confidence of meteorology target detection are reduced.In this paper,a deep convolutional neural network(DCNN)is proposed for meteorology target detection and ground clutter suppression with a large collection of airborne weather radar images as network input.For each weather radar image,the corresponding digital elevation model(DEM)image is extracted on basis of the radar antenna scan-ning parameters and plane position,and is further fed to the net-work as a supplement for ground clutter suppression.The fea-tures of actual meteorology targets are learned in each bottle-neck module of the proposed network and convolved into deeper iterations in the forward propagation process.Then the network parameters are updated by the back propagation itera-tion of the training error.Experimental results on the real mea-sured images show that our proposed DCNN outperforms the counterparts in terms of six evaluation factors.Meanwhile,the network outputs are in good agreement with the expected mete-orology detection results(labels).It is demonstrated that the pro-posed network would have a promising meteorology observa-tion application with minimal effort on network variables or parameter changes.
基金supported by the National Key R&D Program of China(No.2017YFE0300106)Dalian Science and Technology Star Project(No.2020RQ136)+1 种基金the Central Guidance on Local Science and Technology Development Fund of Liaoning Province(No.2022010055-JH6/100)the Fundamental Research Funds for the Central Universities(No.DUT21RC(3)066)。
文摘Observing and analyzing surface images is critical for studying the interaction between plasma and irradiated plasma-facing materials.This paper presents a method for the automatic recognition of bubbles in transmission electron microscope(TEM)images of W nanofibers using image processing techniques and convolutional neural network(CNN).We employ a three-stage approach consisting of Otsu,local-threshold,and watershed segmentation to extract bubbles from noisy images.To address over-segmentation,we propose a combination of area factor and radial pixel intensity scanning.A CNN is used to recognize bubbles,outperforming traditional neural network models such as Alex Net and Google Net with an accuracy of 97.1%and recall of 98.6%.Our method is tested on both clear and blurred TEM images,and demonstrates humanlike performance in recognizing bubbles.This work contributes to the development of quantitative image analysis in the field of plasma-material interactions,offering a scalable solution for analyzing material defects.Overall,this study's findings establish the potential for automatic defect recognition and its applications in the assessment of plasma-material interactions.This method can be employed in a variety of specialties,including plasma physics and materials science.
基金Guangdong Basic and Applied Basic Research Foundation,Grant/Award Number:2019A1515110582Shenzhen Key Laboratory of Visual Object Detection and Recognition,Grant/Award Number:ZDSYS20190902093015527National Natural Science Foundation of China,Grant/Award Number:61876051。
文摘Deep learning has been widely used in the field of mammographic image classification owing to its superiority in automatic feature extraction.However,general deep learning models cannot achieve very satisfactory classification results on mammographic images because these models are not specifically designed for mammographic images and do not take the specific traits of these images into account.To exploit the essential discriminant information of mammographic images,we propose a novel classification method based on a convolutional neural network.Specifically,the proposed method designs two branches to extract the discriminative features from mammographic images from the mediolateral oblique and craniocaudal(CC)mammographic views.The features extracted from the two-view mammographic images contain complementary information that enables breast cancer to be more easily distinguished.Moreover,the attention block is introduced to capture the channel-wise information by adjusting the weight of each feature map,which is beneficial to emphasising the important features of mammographic images.Furthermore,we add a penalty term based on the fuzzy cluster algorithm to the cross-entropy function,which improves the generalisation ability of the classification model by maximising the interclass distance and minimising the intraclass distance of the samples.The experimental results on The Digital database for Screening Mammography INbreast and MIAS mammography databases illustrate that the proposed method achieves the best classification performance and is more robust than the compared state-ofthe-art classification methods.
文摘With the rapid development of sports,the number of sports images has increased dramatically.Intelligent and automatic processing and analysis of moving images are significant,which can not only facilitate users to quickly search and access moving images but also facilitate staff to store and manage moving image data and contribute to the intellectual development of the sports industry.In this paper,a method of table tennis identification and positioning based on a convolutional neural network is proposed,which solves the problem that the identification and positioning method based on color features and contour features is not adaptable in various environments.At the same time,the learning methods and techniques of table tennis detection,positioning,and trajectory prediction are studied.A deep learning framework for recognition learning of rotating flying table tennis is put forward.The mechanism and methods of positioning,trajectory prediction,and intelligent automatic processing of moving images are studied,and the self-built data sets are trained and verified.
文摘With the development of artificial intelligence-related technologies such as deep learning,various organizations,including the government,are making various efforts to generate and manage big data for use in artificial intelligence.However,it is difficult to acquire big data due to various social problems and restrictions such as personal information leakage.There are many problems in introducing technology in fields that do not have enough training data necessary to apply deep learning technology.Therefore,this study proposes a mixed contour data augmentation technique,which is a data augmentation technique using contour images,to solve a problem caused by a lack of data.ResNet,a famous convolutional neural network(CNN)architecture,and CIFAR-10,a benchmark data set,are used for experimental performance evaluation to prove the superiority of the proposed method.And to prove that high performance improvement can be achieved even with a small training dataset,the ratio of the training dataset was divided into 70%,50%,and 30%for comparative analysis.As a result of applying the mixed contour data augmentation technique,it was possible to achieve a classification accuracy improvement of up to 4.64%and high accuracy even with a small amount of data set.In addition,it is expected that the mixed contour data augmentation technique can be applied in various fields by proving the excellence of the proposed data augmentation technique using benchmark datasets.
基金Project supported by the National Natural Science Foundation of China (Grant Nos.61772295 and 61572270)the PHD foundation of Chongqing Normal University (Grant No.19XLB003)Chongqing Technology Foresight and Institutional Innovation Project (Grant No.cstc2021jsyjyzysbAX0011)。
文摘Time series classification(TSC)has attracted a lot of attention for time series data mining tasks and has been applied in various fields.With the success of deep learning(DL)in computer vision recognition,people are starting to use deep learning to tackle TSC tasks.Quantum neural networks(QNN)have recently demonstrated their superiority over traditional machine learning in methods such as image processing and natural language processing,but research using quantum neural networks to handle TSC tasks has not received enough attention.Therefore,we proposed a learning framework based on multiple imaging and hybrid QNN(MIHQNN)for TSC tasks.We investigate the possibility of converting 1D time series to 2D images and classifying the converted images using hybrid QNN.We explored the differences between MIHQNN based on single time series imaging and MIHQNN based on the fusion of multiple time series imaging.Four quantum circuits were also selected and designed to study the impact of quantum circuits on TSC tasks.We tested our method on several standard datasets and achieved significant results compared to several current TSC methods,demonstrating the effectiveness of MIHQNN.This research highlights the potential of applying quantum computing to TSC and provides the theoretical and experimental background for future research.
基金supported in part by the National Natural Science Foundation of China under Grant 62176109in part by the Tibetan Information Processing and Machine Translation Key Laboratory of Qinghai Province under Grant 2021‐Z‐003+3 种基金in part by the Natural Science Foundation of Gansu Province under Grant 21JR7RA531 and Grant 22JR5RA487in part by the Fundamental Research Funds for the Central Universities under Grant lzujbky‐2022‐23in part by the CAAI‐Huawei MindSpore Open Fund under Grant CAAIXSJLJJ‐2022‐020Ain part by the Supercomputing Center of Lanzhou University,in part by Sichuan Science and Technology Program No.2022nsfsc0916.
文摘A variety of neural networks have been presented to deal with issues in deep learning in the last decades.Despite the prominent success achieved by the neural network,it still lacks theoretical guidance to design an efficient neural network model,and verifying the performance of a model needs excessive resources.Previous research studies have demonstrated that many existing models can be regarded as different numerical discretizations of differential equations.This connection sheds light on designing an effective recurrent neural network(RNN)by resorting to numerical analysis.Simple RNN is regarded as a discretisation of the forward Euler scheme.Considering the limited solution accuracy of the forward Euler methods,a Taylor‐type discrete scheme is presented with lower truncation error and a Taylor‐type RNN(T‐RNN)is designed with its guidance.Extensive experiments are conducted to evaluate its performance on statistical language models and emotion analysis tasks.The noticeable gains obtained by T‐RNN present its superiority and the feasibility of designing the neural network model using numerical methods.
文摘COVID-19 disease caused by the SARS-CoV-2 virus has created social and economic disruption across theworld.The ability of the COVID-19 virus to quickly mutate and transfer has created serious concerns across the world.It is essential to detectCOVID-19 infection caused by different variants to take preventive measures accordingly.The existing method of detection of infections caused by COVID-19 and its variants is costly and time-consuming.The impacts of theCOVID-19 pandemic in developing countries are very drastic due to the unavailability of medical facilities and infrastructure to handle the pandemic.Pneumonia is the major symptom of COVID-19 infection.The radiology of the lungs in varies in the case of bacterial pneumonia as compared to COVID-19-caused pneumonia.The pattern of pneumonia in lungs in radiology images can also be used to identify the cause associated with pneumonia.In this paper,we propose the methodology of identifying the cause(either due to COVID-19 or other types of infections)of pneumonia from radiology images.Furthermore,because different variants of COVID-19 lead to different patterns of pneumonia,the proposed methodology identifies pneumonia,the COVID-19 caused pneumonia,and Omicron caused pneumonia from the radiology images.To fulfill the above-mentioned tasks,we have used three Convolution Neural Networks(CNNs)at each stage of the proposed methodology.The results unveil that the proposed step-by-step solution enhances the accuracy of pneumonia detection along with finding its cause,despite having a limited dataset.