Fraud of credit cards is a major issue for financial organizations and individuals.As fraudulent actions become more complex,a demand for better fraud detection systems is rising.Deep learning approaches have shown pr...Fraud of credit cards is a major issue for financial organizations and individuals.As fraudulent actions become more complex,a demand for better fraud detection systems is rising.Deep learning approaches have shown promise in several fields,including detecting credit card fraud.However,the efficacy of these models is heavily dependent on the careful selection of appropriate hyperparameters.This paper introduces models that integrate deep learning models with hyperparameter tuning techniques to learn the patterns and relationships within credit card transaction data,thereby improving fraud detection.Three deep learning models:AutoEncoder(AE),Convolution Neural Network(CNN),and Long Short-Term Memory(LSTM)are proposed to investigate how hyperparameter adjustment impacts the efficacy of deep learning models used to identify credit card fraud.The experiments conducted on a European credit card fraud dataset using different hyperparameters and three deep learning models demonstrate that the proposed models achieve a tradeoff between detection rate and precision,leading these models to be effective in accurately predicting credit card fraud.The results demonstrate that LSTM significantly outperformed AE and CNN in terms of accuracy(99.2%),detection rate(93.3%),and area under the curve(96.3%).These proposed models have surpassed those of existing studies and are expected to make a significant contribution to the field of credit card fraud detection.展开更多
Current spectral analysis for evaluating the rail ride quality effectively is based on the sampling data at a uniform time interval, but the train is of fluctuation velocity in motion, which results in a non-uniform i...Current spectral analysis for evaluating the rail ride quality effectively is based on the sampling data at a uniform time interval, but the train is of fluctuation velocity in motion, which results in a non-uniform interval between consecutive sampling data. Therefore the accuracy of routine spectral analysis is in doubt when applying it in evaluating the rail ride quality. This paper presents a new way, namely, re-sampling with variable frequency to eliminate the influence of the train's uneven velocity. Its feature is that there is no need for precision measurement of the train's moving speed. Experiment results from the test-bed of rolling stock vibration show that it is valid.展开更多
In order to address the issues of traditional resampling algorithms involving computational accuracy and efficiency in rolling element bearing fault diagnosis, an equal division impulse-based(EDI-based) resampling a...In order to address the issues of traditional resampling algorithms involving computational accuracy and efficiency in rolling element bearing fault diagnosis, an equal division impulse-based(EDI-based) resampling algorithm is proposed. First, the time marks of every rising edge of the rotating speed pulse and the corresponding amplitudes of faulty bearing vibration signal are determined. Then, every adjacent the rotating pulse is divided equally, and the time marks in every adjacent rotating speed pulses and the corresponding amplitudes of vibration signal are obtained by the interpolation algorithm. Finally, all the time marks and the corresponding amplitudes of vibration signal are arranged and the time marks are transformed into the angle domain to obtain the resampling signal. Speed-up and speed-down faulty bearing signals are employed to verify the validity of the proposed method, and experimental results show that the proposed method is effective for diagnosing faulty bearings. Furthermore, the traditional order tracking techniques are applied to the experimental bearing signals, and the results show that the proposed method produces higher accurate outcomes in less computation time.展开更多
Order analysis is regarded as one of the most significant method for monitoring and analyzing rotational machinery for the phenomenon of " frequency smear".However,the order analysis based on resampling is a...Order analysis is regarded as one of the most significant method for monitoring and analyzing rotational machinery for the phenomenon of " frequency smear".However,the order analysis based on resampling is a signal processingwhich converts the constant time interval sampling into constant angle interval sampling,while with the variety of the rotational speed.The superiority of the order analysis is investigatedon implement of order analysis.Andthrough comparing the advantage and disadvantage between spectrum and order analysis,the paper will discuss the order analysis form a deep perspective.展开更多
Gravity-aided inertial navigation is a hot issue in the applications of underwater autonomous vehicle(UAV). Since the matching process is conducted with a gravity anomaly database tabulated in the form of a digital mo...Gravity-aided inertial navigation is a hot issue in the applications of underwater autonomous vehicle(UAV). Since the matching process is conducted with a gravity anomaly database tabulated in the form of a digital model and the resolution is 2’ × 2’,a filter model based on vehicle position is derived and the particularity of inertial navigation system(INS) output is employed to estimate a parameter in the system model. Meanwhile, the matching algorithm based on point mass filter(PMF) is applied and several optimal selection strategies are discussed. It is obtained that the point mass filter algorithm based on the deterministic resampling method has better practicability. The reliability and the accuracy of the algorithm are verified via simulation tests.展开更多
Various uncertainties arising during acquisition process of geoscience data may result in anomalous data instances(i.e.,outliers)that do not conform with the expected pattern of regular data instances.With sparse mult...Various uncertainties arising during acquisition process of geoscience data may result in anomalous data instances(i.e.,outliers)that do not conform with the expected pattern of regular data instances.With sparse multivariate data obtained from geotechnical site investigation,it is impossible to identify outliers with certainty due to the distortion of statistics of geotechnical parameters caused by outliers and their associated statistical uncertainty resulted from data sparsity.This paper develops a probabilistic outlier detection method for sparse multivariate data obtained from geotechnical site investigation.The proposed approach quantifies the outlying probability of each data instance based on Mahalanobis distance and determines outliers as those data instances with outlying probabilities greater than 0.5.It tackles the distortion issue of statistics estimated from the dataset with outliers by a re-sampling technique and accounts,rationally,for the statistical uncertainty by Bayesian machine learning.Moreover,the proposed approach also suggests an exclusive method to determine outlying components of each outlier.The proposed approach is illustrated and verified using simulated and real-life dataset.It showed that the proposed approach properly identifies outliers among sparse multivariate data and their corresponding outlying components in a probabilistic manner.It can significantly reduce the masking effect(i.e.,missing some actual outliers due to the distortion of statistics by the outliers and statistical uncertainty).It also found that outliers among sparse multivariate data instances affect significantly the construction of multivariate distribution of geotechnical parameters for uncertainty quantification.This emphasizes the necessity of data cleaning process(e.g.,outlier detection)for uncertainty quantification based on geoscience data.展开更多
An efficient resampling reliability approach was developed to consider the effect of statistical uncertainties in input properties arising due to insufficient data when estimating the reliability of rock slopes and tu...An efficient resampling reliability approach was developed to consider the effect of statistical uncertainties in input properties arising due to insufficient data when estimating the reliability of rock slopes and tunnels.This approach considers the effect of uncertainties in both distribution parameters(mean and standard deviation)and types of input properties.Further,the approach was generalized to make it capable of analyzing complex problems with explicit/implicit performance functions(PFs),single/multiple PFs,and correlated/non-correlated input properties.It couples resampling statistical tool,i.e.jackknife,with advanced reliability tools like Latin hypercube sampling(LHS),Sobol’s global sensitivity,moving least square-response surface method(MLS-RSM),and Nataf’s transformation.The developed approach was demonstrated for four cases encompassing different types.Results were compared with a recently developed bootstrap-based resampling reliability approach.The results show that the approach is accurate and significantly efficient compared with the bootstrap-based approach.The proposed approach reflects the effect of statistical uncertainties of input properties by estimating distributions/confidence intervals of reliability index/probability of failure(s)instead of their fixed-point estimates.Further,sufficiently accurate results were obtained by considering uncertainties in distribution parameters only and ignoring those in distribution types.展开更多
The merging of a panchromatic (PAN) image with a multispectral satellite image (MSI) to increase the spatial resolution of the MSI, while simultaneously preserving its spectral information is classically referred as P...The merging of a panchromatic (PAN) image with a multispectral satellite image (MSI) to increase the spatial resolution of the MSI, while simultaneously preserving its spectral information is classically referred as PAN-sharpening. We employed a recent dataset derived from very high resolution of WorldView-2 satellite (PAN and MSI) for two test sites (one over an urban area and the other over Antarctica), to comprehensively evaluate the performance of six existing PAN-sharpening algorithms. The algorithms under consideration were the Gram-Schmidt (GS), Ehlers fusion (EF), modified hue-intensity-saturation (Mod-HIS), high pass filtering (HPF), the Brovey transform (BT), and wavelet-based principal component analysis (W-PC). Quality assessment of the sharpened images was carried out by using 20 quality indices. We also analyzed the performance of nearest neighbour (NN), bilinear interpolation (BI), and cubic convolution (CC) resampling methods to test their practicability in the PAN-sharpening process. Our results indicate that the comprehensive performance of PAN-sharpening methods decreased in the following order: GS > W-PC > EF > HPF > Mod-HIS > BT, while resampling methods followed the order: NN > BI > CC.展开更多
A new method to detect multiple outliers in multivariate data is proposed. It is a combination of minimum subsets, resampling and self-organizing map (SOM) algorithm introduced by Kohonen,which provides a robust way w...A new method to detect multiple outliers in multivariate data is proposed. It is a combination of minimum subsets, resampling and self-organizing map (SOM) algorithm introduced by Kohonen,which provides a robust way with neural network. In this method, the number and organization of the neurons are selected by the characteristics of the spectra, e.g., the spectra data are often changed linearly with the concentration of the components and are often measured repeatedly, etc. So the spatial distribution of the neurons can be arranged by this characteristic. With this method, all the outliers in the spectra can be detected, which cannot be solved by the traditional method, and the speed of computation is higher than that of the traditional neural network method. The results of the simulation and the experiment show that this method is simple, effective, intuitionistic and all the outliers in the spectra can be detected in a short time. It is useful when associated with the regression model in the near infra-red research.展开更多
In this paper, we describe resourceefficient hardware architectures for softwaredefined radio (SDR) frontends. These architectures are made efficient by using a polyphase channelizer that performs arbitrary sample r...In this paper, we describe resourceefficient hardware architectures for softwaredefined radio (SDR) frontends. These architectures are made efficient by using a polyphase channelizer that performs arbitrary sample rate changes, frequency selection, and bandwidth control. We discuss area, time, and power optimization for field programmable gate array (FPGA) based architectures in an Mpath polyphase filter bank with modified Npath polyphase filter. Such systems allow resampling by arbitrary ratios while simultaneously performing baseband aliasing from center frequencies at Nyquist zones that are not multiples of the output sample rate. A nonmaximally decimated polyphase filter bank, where the number of data loads is not equal to the number of M subfilters, processes M subfilters in a time period that is either less than or greater than the Mdataload ' s time period. We present a loadprocess architecture (LPA) and a runtime architecture (RA) (based on serial polyphase structure) which have different scheduling. In LPA, Nsubfilters are loaded, and then M subfilters are processed at a clock rate that is a multiple of the input data rate. This is necessary to meet the output time constraint of the down-sampled data. In RA, Msubfilters processes are efficiently scheduled within Ndataload time while simultaneously loading N subfilters. This requires reduced clock rates compared with LPA, and potentially less power is consumed. A polyphase filter bank that uses different resampling factors for maximally decimated, underdecimated, overdecimated, and combined upand downsampled scenarios is used as a case study, and an analysis of area, time, and power for their FPGA architectures is given. For resourceoptimized SDR frontends, RA is superior for reducing operating clock rates and dynamic power consumption. RA is also superior for reducing area resources, except when indices are prestored in LUTs.展开更多
This paper presents efficient processing engines for software-defined radio (SDR) front-ends. These engines, based on a polyphase channelizer, perform arbitrary sample-rate changes, frequency selection, and bandwidt...This paper presents efficient processing engines for software-defined radio (SDR) front-ends. These engines, based on a polyphase channelizer, perform arbitrary sample-rate changes, frequency selection, and bandwidth control. This paper presents an M-path polyphase filter bank based on a modified N-path polyphase filter. Such a system allows resampling by arbitrary ratios while performing baseband aliasing from center frequencies at Nyquist zones that are not multiples of the output sample rate. This resampling technique is based on sliding cyclic data load interacting with cyclic-shifted coefficients. A non-maximally-decimated polyphase filterbank (where the number of data loads is not equal to the number of M subfilters) processes M subfilters in a time period that is less than or greater than the M data loads. A polyphase filter bank with five different resampling modes is used as a case study for embedded resamp/ing in SDR front-ends. These modes are (i) maximally decimated, (ii) Under-decimated, (iii) over-decimated, and combined up- and down-sampling with (iv) single stride length, and (v) multiple stride lengths. These modes can be used to obtain any required rational sampling rate change in an SDR front-end based on a polyphase channelizer. They can also be used for translation to and from arbitrary center frequencies that are unrelated to the output sample rates.展开更多
Previous studies have proposed higher requirements for the transient characteristics of a DC transformer used in a flexible high-voltage direct current(HVDC) system to achieve faster sampling speed and meet wider band...Previous studies have proposed higher requirements for the transient characteristics of a DC transformer used in a flexible high-voltage direct current(HVDC) system to achieve faster sampling speed and meet wider bandwidth requirements of the control and protection signal, and to eventually suppress the large transient fault current. In this study, a transient characteristics verification method is proposed for transient characteristics verification of a DC transformer used in a flexible HVDC system based on resampling technology and LabVIEW measurement technology after analyzing the key technology for transient characteristics verification of a DC transformer. A laboratory experiment for the transient characteristics of a full-fiber electronic DC transformer is conducted, and experimental results show that such verification method can be employed for frequency response and step response verification of a DC transformer at 10% of the rated voltage and current, and can eventually improve the screening of a DC transformer.展开更多
The wide-field electromagnetic method is widely used in hydrocarbon exploration,mineral deposit detection,and geological disaster prediction.However,apparent resistivity and normalized field amplitude exceeding 2048 H...The wide-field electromagnetic method is widely used in hydrocarbon exploration,mineral deposit detection,and geological disaster prediction.However,apparent resistivity and normalized field amplitude exceeding 2048 Hz often exhibit upward warping in data,making geophysical inversion and interpretation challenging.The cumulative error of the crystal oscillator in signal transmission and acquisition contributes to an upturned apparent resistivity curve.To address this,a high-frequency information extraction method is proposed based on time-domain signal reconstruction,which helps to record a complete current data sequence;moreover,it helps estimate the crystal oscillator error for the transmitted signal.Considering the recorded error,a received signal was corrected using a set of reconstruction algorithms.After processing,the high-frequency component of the wide-field electromagnetic data was not upturned,while accurate high-frequency information was extracted from the signal.Therefore,the proposed method helped effectively extract high-frequency components of all wide-field electromagnetic data.展开更多
The design, analysis and parallel implementation of particle filter(PF) were investigated. Firstly, to tackle the particle degeneracy problem in the PF, an iterated importance density function(IIDF) was proposed, wher...The design, analysis and parallel implementation of particle filter(PF) were investigated. Firstly, to tackle the particle degeneracy problem in the PF, an iterated importance density function(IIDF) was proposed, where a new term associating with the current measurement information(CMI) was introduced into the expression of the sampled particles. Through the repeated use of the least squares estimate, the CMI can be integrated into the sampling stage in an iterative manner, conducing to the greatly improved sampling quality. By running the IIDF, an iterated PF(IPF) can be obtained. Subsequently, a parallel resampling(PR) was proposed for the purpose of parallel implementation of IPF, whose main idea was the same as systematic resampling(SR) but performed differently. The PR directly used the integral part of the product of the particle weight and particle number as the number of times that a particle was replicated, and it simultaneously eliminated the particles with the smallest weights, which are the two key differences from the SR. The detailed implementation procedures on the graphics processing unit of IPF based on the PR were presented at last. The performance of the IPF, PR and their parallel implementations are illustrated via one-dimensional numerical simulation and practical application of passive radar target tracking.展开更多
Imbalanced data classification is one of the major problems in machine learning.This imbalanced dataset typically has significant differences in the number of data samples between its classes.In most cases,the perform...Imbalanced data classification is one of the major problems in machine learning.This imbalanced dataset typically has significant differences in the number of data samples between its classes.In most cases,the performance of the machine learning algorithm such as Support Vector Machine(SVM)is affected when dealing with an imbalanced dataset.The classification accuracy is mostly skewed toward the majority class and poor results are exhibited in the prediction of minority-class samples.In this paper,a hybrid approach combining data pre-processing technique andSVMalgorithm based on improved Simulated Annealing(SA)was proposed.Firstly,the data preprocessing technique which primarily aims at solving the resampling strategy of handling imbalanced datasets was proposed.In this technique,the data were first synthetically generated to equalize the number of samples between classes and followed by a reduction step to remove redundancy and duplicated data.Next is the training of a balanced dataset using SVM.Since this algorithm requires an iterative process to search for the best penalty parameter during training,an improved SA algorithm was proposed for this task.In this proposed improvement,a new acceptance criterion for the solution to be accepted in the SA algorithm was introduced to enhance the accuracy of the optimization process.Experimental works based on ten publicly available imbalanced datasets have demonstrated higher accuracy in the classification tasks using the proposed approach in comparison with the conventional implementation of SVM.Registering at an average of 89.65%of accuracy for the binary class classification has demonstrated the good performance of the proposed works.展开更多
A method for determining medium quality factor is developed on the basis of analyzing the attenuation dispersion of the arrived first period P wave. In order to enhance signal to noise ratio, improve the resolution in...A method for determining medium quality factor is developed on the basis of analyzing the attenuation dispersion of the arrived first period P wave. In order to enhance signal to noise ratio, improve the resolution in measurement and reduce systematic error we applied the data resampling technique. The group velocity delay of P wave was derived by using an improved multi-filtering method. Based on a linear viscoelastic relaxation model we deduced the medium quality factor Qm, and associated error with 95% confidence level. Applying the method to the seismic record of the Xiuyan M=5.4 earthquake sequences we obtained the following result: 1 High Qm started to appear from Nov. 9, 1999. The events giving the deduced high Qm value clustered in a region with their epicenter dis- tances being between 32 and 46 km to the Yingkou station. This Qm versus distance observation obviously deviates from the normal trend of Qm linearly increasing with distance. 2 The average Qm before the 29 Dec. 1999 M=5.4 earthquake is 460, while the average Qm between the M=5.4 event and the 12 Jan. 2000 M=5.1 earthquake is 391, and the average Qm after the M=5.1 event is 204.展开更多
The estimation of image resampling factors is an important problem in image forensics.Among all the resampling factor estimation methods,spectrumbased methods are one of the most widely used methods and have attracted...The estimation of image resampling factors is an important problem in image forensics.Among all the resampling factor estimation methods,spectrumbased methods are one of the most widely used methods and have attracted a lot of research interest.However,because of inherent ambiguity,spectrum-based methods fail to discriminate upscale and downscale operations without any prior information.In general,the application of resampling leaves detectable traces in both spatial domain and frequency domain of a resampled image.Firstly,the resampling process will introduce correlations between neighboring pixels.In this case,a set of periodic pixels that are correlated to their neighbors can be found in a resampled image.Secondly,the resampled image has distinct and strong peaks on spectrum while the spectrum of original image has no clear peaks.Hence,in this paper,we propose a dual-stream convolutional neural network for image resampling factors estimation.One of the two streams is gray stream whose purpose is to extract resampling traces features directly from the rescaled images.The other is frequency stream that discovers the differences of spectrum between rescaled and original images.The features from two streams are then fused to construct a feature representation including the resampling traces left in spatial and frequency domain,which is later fed into softmax layer for resampling factor estimation.Experimental results show that the proposed method is effective on resampling factor estimation and outperforms some CNN-based methods.展开更多
In the mechanical fault detection and diagnosis field, it is more and more important to analyze the instantaneous frequency (IF) character of complex vibration signal. The improved IF estimation method is put forwar...In the mechanical fault detection and diagnosis field, it is more and more important to analyze the instantaneous frequency (IF) character of complex vibration signal. The improved IF estimation method is put forward aiming at the shortage of traditional Hilbert transform. It is based on Hilbert transform in wavelet domain. With the help of relationship between the real part and the imaginary part obtained from the complex coefficient of continuous wavelet transform or the analyti- cal signal reconstructed in wavelet packet decomposition, the instantaneous phase function of the subcomponent is extracted. In order to improve the precise of IF estimated out, some means such as Linear regression, adaptive filtering, resampling are applied into the instantaneous phase obtained, then, the central differencing operator is used to get desired IF. Simulation results with synthetic and gearbox fault signals are included to illustrate the proposed method.展开更多
Object tracking with abrupt motion is an important research topic and has attracted wide attention.To obtain accurate tracking results,an improved particle filter tracking algorithm based on sparse representation and ...Object tracking with abrupt motion is an important research topic and has attracted wide attention.To obtain accurate tracking results,an improved particle filter tracking algorithm based on sparse representation and nonlinear resampling is proposed in this paper. First,the sparse representation is used to compute particle weights by considering the fact that the weights are sparse when the object moves abruptly,so the potential object region can be predicted more precisely. Then,a nonlinear resampling process is proposed by utilizing the nonlinear sorting strategy,which can solve the problem of particle diversity impoverishment caused by traditional resampling methods. Experimental results based on videos containing objects with various abrupt motions have demonstrated the effectiveness of the proposed algorithm.展开更多
文摘Fraud of credit cards is a major issue for financial organizations and individuals.As fraudulent actions become more complex,a demand for better fraud detection systems is rising.Deep learning approaches have shown promise in several fields,including detecting credit card fraud.However,the efficacy of these models is heavily dependent on the careful selection of appropriate hyperparameters.This paper introduces models that integrate deep learning models with hyperparameter tuning techniques to learn the patterns and relationships within credit card transaction data,thereby improving fraud detection.Three deep learning models:AutoEncoder(AE),Convolution Neural Network(CNN),and Long Short-Term Memory(LSTM)are proposed to investigate how hyperparameter adjustment impacts the efficacy of deep learning models used to identify credit card fraud.The experiments conducted on a European credit card fraud dataset using different hyperparameters and three deep learning models demonstrate that the proposed models achieve a tradeoff between detection rate and precision,leading these models to be effective in accurately predicting credit card fraud.The results demonstrate that LSTM significantly outperformed AE and CNN in terms of accuracy(99.2%),detection rate(93.3%),and area under the curve(96.3%).These proposed models have surpassed those of existing studies and are expected to make a significant contribution to the field of credit card fraud detection.
基金Supported by Major Engineering Project Culture Foundation of Science and Technology's New Ideas in Colleges and Universities (No.705044)
文摘Current spectral analysis for evaluating the rail ride quality effectively is based on the sampling data at a uniform time interval, but the train is of fluctuation velocity in motion, which results in a non-uniform interval between consecutive sampling data. Therefore the accuracy of routine spectral analysis is in doubt when applying it in evaluating the rail ride quality. This paper presents a new way, namely, re-sampling with variable frequency to eliminate the influence of the train's uneven velocity. Its feature is that there is no need for precision measurement of the train's moving speed. Experiment results from the test-bed of rolling stock vibration show that it is valid.
基金Fundamental Research Funds for the Central Universities(No.2016JBM051)
文摘In order to address the issues of traditional resampling algorithms involving computational accuracy and efficiency in rolling element bearing fault diagnosis, an equal division impulse-based(EDI-based) resampling algorithm is proposed. First, the time marks of every rising edge of the rotating speed pulse and the corresponding amplitudes of faulty bearing vibration signal are determined. Then, every adjacent the rotating pulse is divided equally, and the time marks in every adjacent rotating speed pulses and the corresponding amplitudes of vibration signal are obtained by the interpolation algorithm. Finally, all the time marks and the corresponding amplitudes of vibration signal are arranged and the time marks are transformed into the angle domain to obtain the resampling signal. Speed-up and speed-down faulty bearing signals are employed to verify the validity of the proposed method, and experimental results show that the proposed method is effective for diagnosing faulty bearings. Furthermore, the traditional order tracking techniques are applied to the experimental bearing signals, and the results show that the proposed method produces higher accurate outcomes in less computation time.
文摘Order analysis is regarded as one of the most significant method for monitoring and analyzing rotational machinery for the phenomenon of " frequency smear".However,the order analysis based on resampling is a signal processingwhich converts the constant time interval sampling into constant angle interval sampling,while with the variety of the rotational speed.The superiority of the order analysis is investigatedon implement of order analysis.Andthrough comparing the advantage and disadvantage between spectrum and order analysis,the paper will discuss the order analysis form a deep perspective.
基金supported by the National Natural Science Foundation of China(61673060)the National Key R&D Plan(2016YFB0501700)
文摘Gravity-aided inertial navigation is a hot issue in the applications of underwater autonomous vehicle(UAV). Since the matching process is conducted with a gravity anomaly database tabulated in the form of a digital model and the resolution is 2’ × 2’,a filter model based on vehicle position is derived and the particularity of inertial navigation system(INS) output is employed to estimate a parameter in the system model. Meanwhile, the matching algorithm based on point mass filter(PMF) is applied and several optimal selection strategies are discussed. It is obtained that the point mass filter algorithm based on the deterministic resampling method has better practicability. The reliability and the accuracy of the algorithm are verified via simulation tests.
基金supported by the National Key R&D Program of China(Project No.2016YFC0800200)the NRF-NSFC 3rd Joint Research Grant(Earth Science)(Project No.41861144022)+2 种基金the National Natural Science Foundation of China(Project Nos.51679174,and 51779189)the Shenzhen Key Technology R&D Program(Project No.20170324)The financial support is grateful acknowledged。
文摘Various uncertainties arising during acquisition process of geoscience data may result in anomalous data instances(i.e.,outliers)that do not conform with the expected pattern of regular data instances.With sparse multivariate data obtained from geotechnical site investigation,it is impossible to identify outliers with certainty due to the distortion of statistics of geotechnical parameters caused by outliers and their associated statistical uncertainty resulted from data sparsity.This paper develops a probabilistic outlier detection method for sparse multivariate data obtained from geotechnical site investigation.The proposed approach quantifies the outlying probability of each data instance based on Mahalanobis distance and determines outliers as those data instances with outlying probabilities greater than 0.5.It tackles the distortion issue of statistics estimated from the dataset with outliers by a re-sampling technique and accounts,rationally,for the statistical uncertainty by Bayesian machine learning.Moreover,the proposed approach also suggests an exclusive method to determine outlying components of each outlier.The proposed approach is illustrated and verified using simulated and real-life dataset.It showed that the proposed approach properly identifies outliers among sparse multivariate data and their corresponding outlying components in a probabilistic manner.It can significantly reduce the masking effect(i.e.,missing some actual outliers due to the distortion of statistics by the outliers and statistical uncertainty).It also found that outliers among sparse multivariate data instances affect significantly the construction of multivariate distribution of geotechnical parameters for uncertainty quantification.This emphasizes the necessity of data cleaning process(e.g.,outlier detection)for uncertainty quantification based on geoscience data.
文摘An efficient resampling reliability approach was developed to consider the effect of statistical uncertainties in input properties arising due to insufficient data when estimating the reliability of rock slopes and tunnels.This approach considers the effect of uncertainties in both distribution parameters(mean and standard deviation)and types of input properties.Further,the approach was generalized to make it capable of analyzing complex problems with explicit/implicit performance functions(PFs),single/multiple PFs,and correlated/non-correlated input properties.It couples resampling statistical tool,i.e.jackknife,with advanced reliability tools like Latin hypercube sampling(LHS),Sobol’s global sensitivity,moving least square-response surface method(MLS-RSM),and Nataf’s transformation.The developed approach was demonstrated for four cases encompassing different types.Results were compared with a recently developed bootstrap-based resampling reliability approach.The results show that the approach is accurate and significantly efficient compared with the bootstrap-based approach.The proposed approach reflects the effect of statistical uncertainties of input properties by estimating distributions/confidence intervals of reliability index/probability of failure(s)instead of their fixed-point estimates.Further,sufficiently accurate results were obtained by considering uncertainties in distribution parameters only and ignoring those in distribution types.
文摘The merging of a panchromatic (PAN) image with a multispectral satellite image (MSI) to increase the spatial resolution of the MSI, while simultaneously preserving its spectral information is classically referred as PAN-sharpening. We employed a recent dataset derived from very high resolution of WorldView-2 satellite (PAN and MSI) for two test sites (one over an urban area and the other over Antarctica), to comprehensively evaluate the performance of six existing PAN-sharpening algorithms. The algorithms under consideration were the Gram-Schmidt (GS), Ehlers fusion (EF), modified hue-intensity-saturation (Mod-HIS), high pass filtering (HPF), the Brovey transform (BT), and wavelet-based principal component analysis (W-PC). Quality assessment of the sharpened images was carried out by using 20 quality indices. We also analyzed the performance of nearest neighbour (NN), bilinear interpolation (BI), and cubic convolution (CC) resampling methods to test their practicability in the PAN-sharpening process. Our results indicate that the comprehensive performance of PAN-sharpening methods decreased in the following order: GS > W-PC > EF > HPF > Mod-HIS > BT, while resampling methods followed the order: NN > BI > CC.
文摘A new method to detect multiple outliers in multivariate data is proposed. It is a combination of minimum subsets, resampling and self-organizing map (SOM) algorithm introduced by Kohonen,which provides a robust way with neural network. In this method, the number and organization of the neurons are selected by the characteristics of the spectra, e.g., the spectra data are often changed linearly with the concentration of the components and are often measured repeatedly, etc. So the spatial distribution of the neurons can be arranged by this characteristic. With this method, all the outliers in the spectra can be detected, which cannot be solved by the traditional method, and the speed of computation is higher than that of the traditional neural network method. The results of the simulation and the experiment show that this method is simple, effective, intuitionistic and all the outliers in the spectra can be detected in a short time. It is useful when associated with the regression model in the near infra-red research.
文摘In this paper, we describe resourceefficient hardware architectures for softwaredefined radio (SDR) frontends. These architectures are made efficient by using a polyphase channelizer that performs arbitrary sample rate changes, frequency selection, and bandwidth control. We discuss area, time, and power optimization for field programmable gate array (FPGA) based architectures in an Mpath polyphase filter bank with modified Npath polyphase filter. Such systems allow resampling by arbitrary ratios while simultaneously performing baseband aliasing from center frequencies at Nyquist zones that are not multiples of the output sample rate. A nonmaximally decimated polyphase filter bank, where the number of data loads is not equal to the number of M subfilters, processes M subfilters in a time period that is either less than or greater than the Mdataload ' s time period. We present a loadprocess architecture (LPA) and a runtime architecture (RA) (based on serial polyphase structure) which have different scheduling. In LPA, Nsubfilters are loaded, and then M subfilters are processed at a clock rate that is a multiple of the input data rate. This is necessary to meet the output time constraint of the down-sampled data. In RA, Msubfilters processes are efficiently scheduled within Ndataload time while simultaneously loading N subfilters. This requires reduced clock rates compared with LPA, and potentially less power is consumed. A polyphase filter bank that uses different resampling factors for maximally decimated, underdecimated, overdecimated, and combined upand downsampled scenarios is used as a case study, and an analysis of area, time, and power for their FPGA architectures is given. For resourceoptimized SDR frontends, RA is superior for reducing operating clock rates and dynamic power consumption. RA is also superior for reducing area resources, except when indices are prestored in LUTs.
文摘This paper presents efficient processing engines for software-defined radio (SDR) front-ends. These engines, based on a polyphase channelizer, perform arbitrary sample-rate changes, frequency selection, and bandwidth control. This paper presents an M-path polyphase filter bank based on a modified N-path polyphase filter. Such a system allows resampling by arbitrary ratios while performing baseband aliasing from center frequencies at Nyquist zones that are not multiples of the output sample rate. This resampling technique is based on sliding cyclic data load interacting with cyclic-shifted coefficients. A non-maximally-decimated polyphase filterbank (where the number of data loads is not equal to the number of M subfilters) processes M subfilters in a time period that is less than or greater than the M data loads. A polyphase filter bank with five different resampling modes is used as a case study for embedded resamp/ing in SDR front-ends. These modes are (i) maximally decimated, (ii) Under-decimated, (iii) over-decimated, and combined up- and down-sampling with (iv) single stride length, and (v) multiple stride lengths. These modes can be used to obtain any required rational sampling rate change in an SDR front-end based on a polyphase channelizer. They can also be used for translation to and from arbitrary center frequencies that are unrelated to the output sample rates.
基金supported by the State Grid Corporation Science and Technology Project(No.JL71-15-039)
文摘Previous studies have proposed higher requirements for the transient characteristics of a DC transformer used in a flexible high-voltage direct current(HVDC) system to achieve faster sampling speed and meet wider bandwidth requirements of the control and protection signal, and to eventually suppress the large transient fault current. In this study, a transient characteristics verification method is proposed for transient characteristics verification of a DC transformer used in a flexible HVDC system based on resampling technology and LabVIEW measurement technology after analyzing the key technology for transient characteristics verification of a DC transformer. A laboratory experiment for the transient characteristics of a full-fiber electronic DC transformer is conducted, and experimental results show that such verification method can be employed for frequency response and step response verification of a DC transformer at 10% of the rated voltage and current, and can eventually improve the screening of a DC transformer.
基金Project(42004056)supported by the National Natural Science Foundation of ChinaProject(ZR2020QD052)supported by the Natural Science Foundation of Shandong Province,ChinaProject(2019YFC0604902)supported by the National Key Research and Development Program of China。
文摘The wide-field electromagnetic method is widely used in hydrocarbon exploration,mineral deposit detection,and geological disaster prediction.However,apparent resistivity and normalized field amplitude exceeding 2048 Hz often exhibit upward warping in data,making geophysical inversion and interpretation challenging.The cumulative error of the crystal oscillator in signal transmission and acquisition contributes to an upturned apparent resistivity curve.To address this,a high-frequency information extraction method is proposed based on time-domain signal reconstruction,which helps to record a complete current data sequence;moreover,it helps estimate the crystal oscillator error for the transmitted signal.Considering the recorded error,a received signal was corrected using a set of reconstruction algorithms.After processing,the high-frequency component of the wide-field electromagnetic data was not upturned,while accurate high-frequency information was extracted from the signal.Therefore,the proposed method helped effectively extract high-frequency components of all wide-field electromagnetic data.
基金Project(61372136) supported by the National Natural Science Foundation of China
文摘The design, analysis and parallel implementation of particle filter(PF) were investigated. Firstly, to tackle the particle degeneracy problem in the PF, an iterated importance density function(IIDF) was proposed, where a new term associating with the current measurement information(CMI) was introduced into the expression of the sampled particles. Through the repeated use of the least squares estimate, the CMI can be integrated into the sampling stage in an iterative manner, conducing to the greatly improved sampling quality. By running the IIDF, an iterated PF(IPF) can be obtained. Subsequently, a parallel resampling(PR) was proposed for the purpose of parallel implementation of IPF, whose main idea was the same as systematic resampling(SR) but performed differently. The PR directly used the integral part of the product of the particle weight and particle number as the number of times that a particle was replicated, and it simultaneously eliminated the particles with the smallest weights, which are the two key differences from the SR. The detailed implementation procedures on the graphics processing unit of IPF based on the PR were presented at last. The performance of the IPF, PR and their parallel implementations are illustrated via one-dimensional numerical simulation and practical application of passive radar target tracking.
文摘Imbalanced data classification is one of the major problems in machine learning.This imbalanced dataset typically has significant differences in the number of data samples between its classes.In most cases,the performance of the machine learning algorithm such as Support Vector Machine(SVM)is affected when dealing with an imbalanced dataset.The classification accuracy is mostly skewed toward the majority class and poor results are exhibited in the prediction of minority-class samples.In this paper,a hybrid approach combining data pre-processing technique andSVMalgorithm based on improved Simulated Annealing(SA)was proposed.Firstly,the data preprocessing technique which primarily aims at solving the resampling strategy of handling imbalanced datasets was proposed.In this technique,the data were first synthetically generated to equalize the number of samples between classes and followed by a reduction step to remove redundancy and duplicated data.Next is the training of a balanced dataset using SVM.Since this algorithm requires an iterative process to search for the best penalty parameter during training,an improved SA algorithm was proposed for this task.In this proposed improvement,a new acceptance criterion for the solution to be accepted in the SA algorithm was introduced to enhance the accuracy of the optimization process.Experimental works based on ten publicly available imbalanced datasets have demonstrated higher accuracy in the classification tasks using the proposed approach in comparison with the conventional implementation of SVM.Registering at an average of 89.65%of accuracy for the binary class classification has demonstrated the good performance of the proposed works.
基金State Key Project of Science and Technology during the Tenth Five-year Plan (2004BA601B01-03-01).
文摘A method for determining medium quality factor is developed on the basis of analyzing the attenuation dispersion of the arrived first period P wave. In order to enhance signal to noise ratio, improve the resolution in measurement and reduce systematic error we applied the data resampling technique. The group velocity delay of P wave was derived by using an improved multi-filtering method. Based on a linear viscoelastic relaxation model we deduced the medium quality factor Qm, and associated error with 95% confidence level. Applying the method to the seismic record of the Xiuyan M=5.4 earthquake sequences we obtained the following result: 1 High Qm started to appear from Nov. 9, 1999. The events giving the deduced high Qm value clustered in a region with their epicenter dis- tances being between 32 and 46 km to the Yingkou station. This Qm versus distance observation obviously deviates from the normal trend of Qm linearly increasing with distance. 2 The average Qm before the 29 Dec. 1999 M=5.4 earthquake is 460, while the average Qm between the M=5.4 event and the 12 Jan. 2000 M=5.1 earthquake is 391, and the average Qm after the M=5.1 event is 204.
基金the National Natural Science Foundation of China(No.62072480)the Key Areas R&D Program of Guangdong(No.2019B010136002)the Key ScientificResearch Program of Guangzhou(No.201804020068).
文摘The estimation of image resampling factors is an important problem in image forensics.Among all the resampling factor estimation methods,spectrumbased methods are one of the most widely used methods and have attracted a lot of research interest.However,because of inherent ambiguity,spectrum-based methods fail to discriminate upscale and downscale operations without any prior information.In general,the application of resampling leaves detectable traces in both spatial domain and frequency domain of a resampled image.Firstly,the resampling process will introduce correlations between neighboring pixels.In this case,a set of periodic pixels that are correlated to their neighbors can be found in a resampled image.Secondly,the resampled image has distinct and strong peaks on spectrum while the spectrum of original image has no clear peaks.Hence,in this paper,we propose a dual-stream convolutional neural network for image resampling factors estimation.One of the two streams is gray stream whose purpose is to extract resampling traces features directly from the rescaled images.The other is frequency stream that discovers the differences of spectrum between rescaled and original images.The features from two streams are then fused to construct a feature representation including the resampling traces left in spatial and frequency domain,which is later fed into softmax layer for resampling factor estimation.Experimental results show that the proposed method is effective on resampling factor estimation and outperforms some CNN-based methods.
基金This project is supported by National Natural Science Foundation of China (No.50605065)Natural Science Foundation Project of CQ CSTC(No.2007BB2142).
文摘In the mechanical fault detection and diagnosis field, it is more and more important to analyze the instantaneous frequency (IF) character of complex vibration signal. The improved IF estimation method is put forward aiming at the shortage of traditional Hilbert transform. It is based on Hilbert transform in wavelet domain. With the help of relationship between the real part and the imaginary part obtained from the complex coefficient of continuous wavelet transform or the analyti- cal signal reconstructed in wavelet packet decomposition, the instantaneous phase function of the subcomponent is extracted. In order to improve the precise of IF estimated out, some means such as Linear regression, adaptive filtering, resampling are applied into the instantaneous phase obtained, then, the central differencing operator is used to get desired IF. Simulation results with synthetic and gearbox fault signals are included to illustrate the proposed method.
基金Supported by the National Natural Science Foundation of China(61701029)
文摘Object tracking with abrupt motion is an important research topic and has attracted wide attention.To obtain accurate tracking results,an improved particle filter tracking algorithm based on sparse representation and nonlinear resampling is proposed in this paper. First,the sparse representation is used to compute particle weights by considering the fact that the weights are sparse when the object moves abruptly,so the potential object region can be predicted more precisely. Then,a nonlinear resampling process is proposed by utilizing the nonlinear sorting strategy,which can solve the problem of particle diversity impoverishment caused by traditional resampling methods. Experimental results based on videos containing objects with various abrupt motions have demonstrated the effectiveness of the proposed algorithm.