In view of the composition analysis and identification of ancient glass products, L1 regularization, K-Means cluster analysis, elbow rule and other methods were comprehensively used to build logical regression, cluste...In view of the composition analysis and identification of ancient glass products, L1 regularization, K-Means cluster analysis, elbow rule and other methods were comprehensively used to build logical regression, cluster analysis, hyper-parameter test and other models, and SPSS, Python and other tools were used to obtain the classification rules of glass products under different fluxes, sub classification under different chemical compositions, hyper-parameter K value test and rationality analysis. Research can provide theoretical support for the protection and restoration of ancient glass relics.展开更多
By defining fuzzy valued simple functions and giving L1(μ) approximations of fuzzy valued integrably bounded functions by such simple functions, the paper analyses by L1(μ)-norm the approximation capability of four-...By defining fuzzy valued simple functions and giving L1(μ) approximations of fuzzy valued integrably bounded functions by such simple functions, the paper analyses by L1(μ)-norm the approximation capability of four-layer feedforward regular fuzzy neural networks to the fuzzy valued integrably bounded function F : Rn → FcO(R). That is, if the transfer functionσ: R→R is non-polynomial and integrable function on each finite interval, F may be innorm approximated by fuzzy valued functions defined as to anydegree of accuracy. Finally some real examples demonstrate the conclusions.展开更多
In this paper, polynomial fuzzy neural network classifiers (PFNNCs) is proposed by means of density fuzzy c-means and L2-norm regularization. The overall design of PFNNCs was realized by means of fuzzy rules that come...In this paper, polynomial fuzzy neural network classifiers (PFNNCs) is proposed by means of density fuzzy c-means and L2-norm regularization. The overall design of PFNNCs was realized by means of fuzzy rules that come in form of three parts, namely premise part, consequence part and aggregation part. The premise part was developed by density fuzzy c-means that helps determine the apex parameters of membership functions, while the consequence part was realized by means of two types of polynomials including linear and quadratic. L2-norm regularization that can alleviate the overfitting problem was exploited to estimate the parameters of polynomials, which constructed the aggregation part. Experimental results of several data sets demonstrate that the proposed classifiers show higher classification accuracy in comparison with some other classifiers reported in the literature.展开更多
In this paper, we built upon the estimating primaries by sparse inversion (EPSI) method. We use the 3D curvelet transform and modify the EPSI method to the sparse inversion of the biconvex optimization and Ll-norm r...In this paper, we built upon the estimating primaries by sparse inversion (EPSI) method. We use the 3D curvelet transform and modify the EPSI method to the sparse inversion of the biconvex optimization and Ll-norm regularization, and use alternating optimization to directly estimate the primary reflection coefficients and source wavelet. The 3D curvelet transform is used as a sparseness constraint when inverting the primary reflection coefficients, which results in avoiding the prediction subtraction process in the surface-related multiples elimination (SRME) method. The proposed method not only reduces the damage to the effective waves but also improves the elimination of multiples. It is also a wave equation- based method for elimination of surface multiple reflections, which effectively removes surface multiples under complex submarine conditions.展开更多
Bayesian empirical likelihood is a semiparametric method that combines parametric priors and nonparametric likelihoods, that is, replacing the parametric likelihood function in Bayes theorem with a nonparametric empir...Bayesian empirical likelihood is a semiparametric method that combines parametric priors and nonparametric likelihoods, that is, replacing the parametric likelihood function in Bayes theorem with a nonparametric empirical likelihood function, which can be used without assuming the distribution of the data. It can effectively avoid the problems caused by the wrong setting of the model. In the variable selection based on Bayesian empirical likelihood, the penalty term is introduced into the model in the form of parameter prior. In this paper, we propose a novel variable selection method, L<sub>1/2</sub> regularization based on Bayesian empirical likelihood. The L<sub>1/2</sub> penalty is introduced into the model through a scale mixture of uniform representation of generalized Gaussian prior, and the posterior distribution is then sampled using MCMC method. Simulations demonstrate that the proposed method can have better predictive ability when the error violates the zero-mean normality assumption of the standard parameter model, and can perform variable selection.展开更多
The Lt-norm method is one of the widely used matching filters for adaptive multiple subtraction. When the primaries and multiples are mixed together, the L1-norm method might damage the primaries, leading to poor late...The Lt-norm method is one of the widely used matching filters for adaptive multiple subtraction. When the primaries and multiples are mixed together, the L1-norm method might damage the primaries, leading to poor lateral continuity. In this paper, we propose a constrained L1-norm method for adaptive multiple subtraction by introducing the lateral continuity constraint for the estimated primaries. We measure the lateral continuity using prediction-error filters (PEF). We illustrate our method with the synthetic Pluto dataset. The results show that the constrained L1-norm method can simultaneously attenuate the multiples and preserve the primaries.展开更多
Based on exact penalty function, a new neural network for solving the L1-norm optimization problem is proposed. In comparison with Kennedy and Chua’s network(1988), it has better properties.Based on Bandler’s fault ...Based on exact penalty function, a new neural network for solving the L1-norm optimization problem is proposed. In comparison with Kennedy and Chua’s network(1988), it has better properties.Based on Bandler’s fault location method(1982), a new nonlinearly constrained L1-norm problem is developed. It can be solved with less computing time through only one optimization processing. The proposed neural network can be used to solve the analog diagnosis L1 problem. The validity of the proposed neural networks and the fault location L1 method are illustrated by extensive computer simulations.展开更多
Seismic data regularization is an important preprocessing step in seismic signal processing. Traditional seismic acquisition methods follow the Shannon–Nyquist sampling theorem, whereas compressive sensing(CS) prov...Seismic data regularization is an important preprocessing step in seismic signal processing. Traditional seismic acquisition methods follow the Shannon–Nyquist sampling theorem, whereas compressive sensing(CS) provides a fundamentally new paradigm to overcome limitations in data acquisition. Besides the sparse representation of seismic signal in some transform domain and the 1-norm reconstruction algorithm, the seismic data regularization quality of CS-based techniques strongly depends on random undersampling schemes. For 2D seismic data, discrete uniform-based methods have been investigated, where some seismic traces are randomly sampled with an equal probability. However, in theory and practice, some seismic traces with different probability are required to be sampled for satisfying the assumptions in CS. Therefore, designing new undersampling schemes is imperative. We propose a Bernoulli-based random undersampling scheme and its jittered version to determine the regular traces that are randomly sampled with different probability, while both schemes comply with the Bernoulli process distribution. We performed experiments using the Fourier and curvelet transforms and the spectral projected gradient reconstruction algorithm for 1-norm(SPGL1), and ten different random seeds. According to the signal-to-noise ratio(SNR) between the original and reconstructed seismic data, the detailed experimental results from 2D numerical and physical simulation data show that the proposed novel schemes perform overall better than the discrete uniform schemes.展开更多
L(d, 1)-labeling is a kind of graph coloring problem from frequency assignment in radio networks, in which adjacent nodes must receive colors that are at least d apart while nodes at distance two from each other must ...L(d, 1)-labeling is a kind of graph coloring problem from frequency assignment in radio networks, in which adjacent nodes must receive colors that are at least d apart while nodes at distance two from each other must receive different colors. We focus on L(d, 1)-labeling of regular tilings for d≥3 since the cases d=0, 1 or 2 have been researched by Calamoneri and Petreschi. For all three kinds of regular tilings, we give their L (d, 1)-labeling numbers for any integer d≥3. Therefore, combined with the results given by Calamoneri and Petreschi, the L(d, 1)-labeling numbers of regular tilings for any nonnegative integer d may be determined completely.展开更多
The paper discusses the core parameters of the 3 D and 4 D variational merging based on L1 norm regularization,namely optimization characteristic correlation length of background error covariance matrix and regulariza...The paper discusses the core parameters of the 3 D and 4 D variational merging based on L1 norm regularization,namely optimization characteristic correlation length of background error covariance matrix and regularization parameter. Classical 3 D/4 D variational merging is based on the theory that error follows Gaussian distribution. It involves the solution of the objective functional gradient in minimization iteration,which requires the data to have continuity and differentiability. Classic 3 D/4 D-dimensional variational merging method was extended,and L1 norm was used as the constraint coupling to the classical variational merged model. Experiment was carried out by using linear advection-diffusion equation as four-dimensional prediction model,and parameter optimization of this method is discussed. Considering the strong temporal and spatial variation of water vapor,this method is further applied to the precipitable water vapor( PWV) merging by calculating reanalysis data and GNSS retrieval.Parameters were adjusted gradually to analyze the influence of background field on the merging result,and the experiment results show that the mathematical algorithm adopted in this paper is feasible.展开更多
Radial functions have become a useful tool in numerical mathematics. On the sphere they have to be identified with the zonal functions. We investigate zonal polynomials with mass concentration at the pole, in the sens...Radial functions have become a useful tool in numerical mathematics. On the sphere they have to be identified with the zonal functions. We investigate zonal polynomials with mass concentration at the pole, in the sense of their L1-norm is attaining the minimum value. Such polynomials satisfy a complicated system of nonlinear e-quations (algebraic if the space dimension is odd, only) and also a singular differential equation of third order. The exact order of decay of the minimum value with respect to the polynomial degree is determined. By our results we can prove that some nodal systems on the sphere, which are defined by a minimum-property, are providing fundamental matrices which are diagonal-dominant or bounded with respect to the ∞-norm, at least, as the polynomial degree tends to infinity.展开更多
Least-squares migration (LSM) is applied to image subsurface structures and lithology by minimizing the objective function of the observed seismic and reverse-time migration residual data of various underground refl...Least-squares migration (LSM) is applied to image subsurface structures and lithology by minimizing the objective function of the observed seismic and reverse-time migration residual data of various underground reflectivity models. LSM reduces the migration artifacts, enhances the spatial resolution of the migrated images, and yields a more accurate subsurface reflectivity distribution than that of standard migration. The introduction of regularization constraints effectively improves the stability of the least-squares offset. The commonly used regularization terms are based on the L2-norm, which smooths the migration results, e.g., by smearing the reflectivities, while providing stability. However, in exploration geophysics, reflection structures based on velocity and density are generally observed to be discontinuous in depth, illustrating sparse reflectance. To obtain a sparse migration profile, we propose the super-resolution least-squares Kirchhoff prestack depth migration by solving the L0-norm-constrained optimization problem. Additionally, we introduce a two-stage iterative soft and hard thresholding algorithm to retrieve the super-resolution reflectivity distribution. Further, the proposed algorithm is applied to complex synthetic data. Furthermore, the sensitivity of the proposed algorithm to noise and the dominant frequency of the source wavelet was evaluated. Finally, we conclude that the proposed method improves the spatial resolution and achieves impulse-like reflectivity distribution and can be applied to structural interpretations and complex subsurface imaging.展开更多
In this work, we consider a homotopic principle for solving large-scale and dense l1underdetermined problems and its applications in image processing and classification. We solve the face recognition problem where the...In this work, we consider a homotopic principle for solving large-scale and dense l1underdetermined problems and its applications in image processing and classification. We solve the face recognition problem where the input image contains corrupted and/or lost pixels. The approach involves two steps: first, the incomplete or corrupted image is subject to an inpainting process, and secondly, the restored image is used to carry out the classification or recognition task. Addressing these two steps involves solving large scale l1minimization problems. To that end, we propose to solve a sequence of linear equality constrained multiquadric problems that depends on a regularization parameter that converges to zero. The procedure generates a central path that converges to a point on the solution set of the l1underdetermined problem. In order to solve each subproblem, a conjugate gradient algorithm is formulated. When noise is present in the model, inexact directions are taken so that an approximate solution is computed faster. This prevents the ill conditioning produced when the conjugate gradient is required to iterate until a zero residual is attained.展开更多
In the network security system,intrusion detection plays a significant role.The network security system detects the malicious actions in the network and also conforms the availability,integrity and confidentiality of da...In the network security system,intrusion detection plays a significant role.The network security system detects the malicious actions in the network and also conforms the availability,integrity and confidentiality of data informa-tion resources.Intrusion identification system can easily detect the false positive alerts.If large number of false positive alerts are created then it makes intrusion detection system as difficult to differentiate the false positive alerts from genuine attacks.Many research works have been done.The issues in the existing algo-rithms are more memory space and need more time to execute the transactions of records.This paper proposes a novel framework of network security Intrusion Detection System(IDS)using Modified Frequent Pattern(MFP-Tree)via K-means algorithm.The accuracy rate of Modified Frequent Pattern Tree(MFPT)-K means method infinding the various attacks are Normal 94.89%,for DoS based attack 98.34%,for User to Root(U2R)attacks got 96.73%,Remote to Local(R2L)got 95.89%and Probe attack got 92.67%and is optimal when it is compared with other existing algorithms of K-Means and APRIORI.展开更多
Integrated with sensors,processors,and radio frequency(RF)communication modules,intelligent bearing could achieve the autonomous perception and autonomous decision-making,guarantying the safety and reliability during ...Integrated with sensors,processors,and radio frequency(RF)communication modules,intelligent bearing could achieve the autonomous perception and autonomous decision-making,guarantying the safety and reliability during their use.However,because of the resource limitations of the end device,processors in the intelligent bearing are unable to carry the computational load of deep learning models like convolutional neural network(CNN),which involves a great amount of multiplicative operations.To minimize the computation cost of the conventional CNN,based on the idea of AdderNet,a 1-D adder neural network with a wide first-layer kernel(WAddNN)suitable for bearing fault diagnosis is proposed in this paper.The proposed method uses the l1-norm distance between filters and input features as the output response,thus making the whole network almost free of multiplicative operations.The whole model takes the original signal as the input,uses a wide kernel in the first adder layer to extract features and suppress the high frequency noise,and then uses two layers of small kernels for nonlinear mapping.Through experimental comparison with CNN models of the same structure,WAddNN is able to achieve a similar accuracy as CNN models with significantly reduced computational cost.The proposed model provides a new fault diagnosis method for intelligent bearings with limited resources.展开更多
文摘In view of the composition analysis and identification of ancient glass products, L1 regularization, K-Means cluster analysis, elbow rule and other methods were comprehensively used to build logical regression, cluster analysis, hyper-parameter test and other models, and SPSS, Python and other tools were used to obtain the classification rules of glass products under different fluxes, sub classification under different chemical compositions, hyper-parameter K value test and rationality analysis. Research can provide theoretical support for the protection and restoration of ancient glass relics.
基金Supported by the National Natural Science Foundation of China(No:69872039)
文摘By defining fuzzy valued simple functions and giving L1(μ) approximations of fuzzy valued integrably bounded functions by such simple functions, the paper analyses by L1(μ)-norm the approximation capability of four-layer feedforward regular fuzzy neural networks to the fuzzy valued integrably bounded function F : Rn → FcO(R). That is, if the transfer functionσ: R→R is non-polynomial and integrable function on each finite interval, F may be innorm approximated by fuzzy valued functions defined as to anydegree of accuracy. Finally some real examples demonstrate the conclusions.
基金This work was supported in part by the National Natural Science Foundation of China under Grant 61673295the Natural Science Foundation of Tianjin under Grant 18JCYBJC85200by the National College Students’ innovation and entrepreneurship project under Grant 201710060041.
文摘In this paper, polynomial fuzzy neural network classifiers (PFNNCs) is proposed by means of density fuzzy c-means and L2-norm regularization. The overall design of PFNNCs was realized by means of fuzzy rules that come in form of three parts, namely premise part, consequence part and aggregation part. The premise part was developed by density fuzzy c-means that helps determine the apex parameters of membership functions, while the consequence part was realized by means of two types of polynomials including linear and quadratic. L2-norm regularization that can alleviate the overfitting problem was exploited to estimate the parameters of polynomials, which constructed the aggregation part. Experimental results of several data sets demonstrate that the proposed classifiers show higher classification accuracy in comparison with some other classifiers reported in the literature.
基金supported by the National Science and Technology Major Project (No.2011ZX05023-005-008)
文摘In this paper, we built upon the estimating primaries by sparse inversion (EPSI) method. We use the 3D curvelet transform and modify the EPSI method to the sparse inversion of the biconvex optimization and Ll-norm regularization, and use alternating optimization to directly estimate the primary reflection coefficients and source wavelet. The 3D curvelet transform is used as a sparseness constraint when inverting the primary reflection coefficients, which results in avoiding the prediction subtraction process in the surface-related multiples elimination (SRME) method. The proposed method not only reduces the damage to the effective waves but also improves the elimination of multiples. It is also a wave equation- based method for elimination of surface multiple reflections, which effectively removes surface multiples under complex submarine conditions.
文摘Bayesian empirical likelihood is a semiparametric method that combines parametric priors and nonparametric likelihoods, that is, replacing the parametric likelihood function in Bayes theorem with a nonparametric empirical likelihood function, which can be used without assuming the distribution of the data. It can effectively avoid the problems caused by the wrong setting of the model. In the variable selection based on Bayesian empirical likelihood, the penalty term is introduced into the model in the form of parameter prior. In this paper, we propose a novel variable selection method, L<sub>1/2</sub> regularization based on Bayesian empirical likelihood. The L<sub>1/2</sub> penalty is introduced into the model through a scale mixture of uniform representation of generalized Gaussian prior, and the posterior distribution is then sampled using MCMC method. Simulations demonstrate that the proposed method can have better predictive ability when the error violates the zero-mean normality assumption of the standard parameter model, and can perform variable selection.
基金This work is sponsored by National Natural Science Foundation of China (No. 40874056), Important National Science & Technology Specific Projects 2008ZX05023-005-004, and the NCET Fund.Acknowledgements The authors are grateful to Liu Yang, and Zhu Sheng-wang for their constructive remarks on this manuscript.
文摘The Lt-norm method is one of the widely used matching filters for adaptive multiple subtraction. When the primaries and multiples are mixed together, the L1-norm method might damage the primaries, leading to poor lateral continuity. In this paper, we propose a constrained L1-norm method for adaptive multiple subtraction by introducing the lateral continuity constraint for the estimated primaries. We measure the lateral continuity using prediction-error filters (PEF). We illustrate our method with the synthetic Pluto dataset. The results show that the constrained L1-norm method can simultaneously attenuate the multiples and preserve the primaries.
基金Supported by Doctoral Special Fund of State Education Commissionthe National Natural Science Foundation of China,Grant No.59477001 and No.59707002
文摘Based on exact penalty function, a new neural network for solving the L1-norm optimization problem is proposed. In comparison with Kennedy and Chua’s network(1988), it has better properties.Based on Bandler’s fault location method(1982), a new nonlinearly constrained L1-norm problem is developed. It can be solved with less computing time through only one optimization processing. The proposed neural network can be used to solve the analog diagnosis L1 problem. The validity of the proposed neural networks and the fault location L1 method are illustrated by extensive computer simulations.
基金financially supported by The 2011 Prospective Research Project of SINOPEC(P11096)
文摘Seismic data regularization is an important preprocessing step in seismic signal processing. Traditional seismic acquisition methods follow the Shannon–Nyquist sampling theorem, whereas compressive sensing(CS) provides a fundamentally new paradigm to overcome limitations in data acquisition. Besides the sparse representation of seismic signal in some transform domain and the 1-norm reconstruction algorithm, the seismic data regularization quality of CS-based techniques strongly depends on random undersampling schemes. For 2D seismic data, discrete uniform-based methods have been investigated, where some seismic traces are randomly sampled with an equal probability. However, in theory and practice, some seismic traces with different probability are required to be sampled for satisfying the assumptions in CS. Therefore, designing new undersampling schemes is imperative. We propose a Bernoulli-based random undersampling scheme and its jittered version to determine the regular traces that are randomly sampled with different probability, while both schemes comply with the Bernoulli process distribution. We performed experiments using the Fourier and curvelet transforms and the spectral projected gradient reconstruction algorithm for 1-norm(SPGL1), and ten different random seeds. According to the signal-to-noise ratio(SNR) between the original and reconstructed seismic data, the detailed experimental results from 2D numerical and physical simulation data show that the proposed novel schemes perform overall better than the discrete uniform schemes.
文摘L(d, 1)-labeling is a kind of graph coloring problem from frequency assignment in radio networks, in which adjacent nodes must receive colors that are at least d apart while nodes at distance two from each other must receive different colors. We focus on L(d, 1)-labeling of regular tilings for d≥3 since the cases d=0, 1 or 2 have been researched by Calamoneri and Petreschi. For all three kinds of regular tilings, we give their L (d, 1)-labeling numbers for any integer d≥3. Therefore, combined with the results given by Calamoneri and Petreschi, the L(d, 1)-labeling numbers of regular tilings for any nonnegative integer d may be determined completely.
基金Supported by Open Foundation Project of Shenyang Institute of Atmospheric Environment,China Meteorological Administration(2016SYIAE14)Natural Science Foundation of Anhui Province,China(1708085QD89)National Natural Science Foundation of China(41805080)
文摘The paper discusses the core parameters of the 3 D and 4 D variational merging based on L1 norm regularization,namely optimization characteristic correlation length of background error covariance matrix and regularization parameter. Classical 3 D/4 D variational merging is based on the theory that error follows Gaussian distribution. It involves the solution of the objective functional gradient in minimization iteration,which requires the data to have continuity and differentiability. Classic 3 D/4 D-dimensional variational merging method was extended,and L1 norm was used as the constraint coupling to the classical variational merged model. Experiment was carried out by using linear advection-diffusion equation as four-dimensional prediction model,and parameter optimization of this method is discussed. Considering the strong temporal and spatial variation of water vapor,this method is further applied to the precipitable water vapor( PWV) merging by calculating reanalysis data and GNSS retrieval.Parameters were adjusted gradually to analyze the influence of background field on the merging result,and the experiment results show that the mathematical algorithm adopted in this paper is feasible.
文摘Radial functions have become a useful tool in numerical mathematics. On the sphere they have to be identified with the zonal functions. We investigate zonal polynomials with mass concentration at the pole, in the sense of their L1-norm is attaining the minimum value. Such polynomials satisfy a complicated system of nonlinear e-quations (algebraic if the space dimension is odd, only) and also a singular differential equation of third order. The exact order of decay of the minimum value with respect to the polynomial degree is determined. By our results we can prove that some nodal systems on the sphere, which are defined by a minimum-property, are providing fundamental matrices which are diagonal-dominant or bounded with respect to the ∞-norm, at least, as the polynomial degree tends to infinity.
基金supported by the National Natural Science Foundation of China(No.41422403)
文摘Least-squares migration (LSM) is applied to image subsurface structures and lithology by minimizing the objective function of the observed seismic and reverse-time migration residual data of various underground reflectivity models. LSM reduces the migration artifacts, enhances the spatial resolution of the migrated images, and yields a more accurate subsurface reflectivity distribution than that of standard migration. The introduction of regularization constraints effectively improves the stability of the least-squares offset. The commonly used regularization terms are based on the L2-norm, which smooths the migration results, e.g., by smearing the reflectivities, while providing stability. However, in exploration geophysics, reflection structures based on velocity and density are generally observed to be discontinuous in depth, illustrating sparse reflectance. To obtain a sparse migration profile, we propose the super-resolution least-squares Kirchhoff prestack depth migration by solving the L0-norm-constrained optimization problem. Additionally, we introduce a two-stage iterative soft and hard thresholding algorithm to retrieve the super-resolution reflectivity distribution. Further, the proposed algorithm is applied to complex synthetic data. Furthermore, the sensitivity of the proposed algorithm to noise and the dominant frequency of the source wavelet was evaluated. Finally, we conclude that the proposed method improves the spatial resolution and achieves impulse-like reflectivity distribution and can be applied to structural interpretations and complex subsurface imaging.
文摘In this work, we consider a homotopic principle for solving large-scale and dense l1underdetermined problems and its applications in image processing and classification. We solve the face recognition problem where the input image contains corrupted and/or lost pixels. The approach involves two steps: first, the incomplete or corrupted image is subject to an inpainting process, and secondly, the restored image is used to carry out the classification or recognition task. Addressing these two steps involves solving large scale l1minimization problems. To that end, we propose to solve a sequence of linear equality constrained multiquadric problems that depends on a regularization parameter that converges to zero. The procedure generates a central path that converges to a point on the solution set of the l1underdetermined problem. In order to solve each subproblem, a conjugate gradient algorithm is formulated. When noise is present in the model, inexact directions are taken so that an approximate solution is computed faster. This prevents the ill conditioning produced when the conjugate gradient is required to iterate until a zero residual is attained.
文摘In the network security system,intrusion detection plays a significant role.The network security system detects the malicious actions in the network and also conforms the availability,integrity and confidentiality of data informa-tion resources.Intrusion identification system can easily detect the false positive alerts.If large number of false positive alerts are created then it makes intrusion detection system as difficult to differentiate the false positive alerts from genuine attacks.Many research works have been done.The issues in the existing algo-rithms are more memory space and need more time to execute the transactions of records.This paper proposes a novel framework of network security Intrusion Detection System(IDS)using Modified Frequent Pattern(MFP-Tree)via K-means algorithm.The accuracy rate of Modified Frequent Pattern Tree(MFPT)-K means method infinding the various attacks are Normal 94.89%,for DoS based attack 98.34%,for User to Root(U2R)attacks got 96.73%,Remote to Local(R2L)got 95.89%and Probe attack got 92.67%and is optimal when it is compared with other existing algorithms of K-Means and APRIORI.
基金support provided by the China National Key Research and Development Program of China under Grant 2019YFB2004300the National Natural Science Foundation of China under Grant 51975065 and 51805051.
文摘Integrated with sensors,processors,and radio frequency(RF)communication modules,intelligent bearing could achieve the autonomous perception and autonomous decision-making,guarantying the safety and reliability during their use.However,because of the resource limitations of the end device,processors in the intelligent bearing are unable to carry the computational load of deep learning models like convolutional neural network(CNN),which involves a great amount of multiplicative operations.To minimize the computation cost of the conventional CNN,based on the idea of AdderNet,a 1-D adder neural network with a wide first-layer kernel(WAddNN)suitable for bearing fault diagnosis is proposed in this paper.The proposed method uses the l1-norm distance between filters and input features as the output response,thus making the whole network almost free of multiplicative operations.The whole model takes the original signal as the input,uses a wide kernel in the first adder layer to extract features and suppress the high frequency noise,and then uses two layers of small kernels for nonlinear mapping.Through experimental comparison with CNN models of the same structure,WAddNN is able to achieve a similar accuracy as CNN models with significantly reduced computational cost.The proposed model provides a new fault diagnosis method for intelligent bearings with limited resources.