Finite rate of innovation sampling is a novel sub-Nyquist sampling method that can reconstruct a signal from sparse sampling data.The application of this method in ultrasonic testing greatly reduces the signal samplin...Finite rate of innovation sampling is a novel sub-Nyquist sampling method that can reconstruct a signal from sparse sampling data.The application of this method in ultrasonic testing greatly reduces the signal sampling rate and the quantity of sampling data.However,the pulse number of the signal must be known beforehand for the signal reconstruction procedure.The accuracy of this prior information directly affects the accuracy of the estimated parameters of the signal and influences the assessment of flaws,leading to a lower defect detection ratio.Although the pulse number can be pre-given by theoretical analysis,the process is still unable to assess actual complex random orientation defects.Therefore,this paper proposes a new method that uses singular value decomposition(SVD) for estimating the pulse number from sparse sampling data and avoids the shortcoming of providing the pulse number in advance for signal reconstruction.When the sparse sampling data have been acquired from the ultrasonic signal,these data are transformed to discrete Fourier coefficients.A Hankel matrix is then constructed from these coefficients,and SVD is performed on the matrix.The decomposition coefficients reserve the information of the pulse number.When the decomposition coefficients generated by noise according to noise level are removed,the number of the remaining decomposition coefficients is the signal pulse number.The feasibility of the proposed method was verified through simulation experiments.The applicability was tested in ultrasonic experiments by using sample flawed pipelines.Results from simulations and real experiments demonstrated the efficiency of this method.展开更多
In the current paper,we considered the Fisher information matrix from the generalized Rayleigh distribution(GR)distribution in ranked set sampling(RSS).The numerical results show that the ranked set sample carries mor...In the current paper,we considered the Fisher information matrix from the generalized Rayleigh distribution(GR)distribution in ranked set sampling(RSS).The numerical results show that the ranked set sample carries more information about λ and α than a simple random sample of equivalent size.In order to give more insight into the performance of RSS with respect to(w.r.t.)simple random sampling(SRS),a modified unbiased estimator and a modified best linear unbiased estimator(BLUE)of scale and shape λ and α from GR distribution in SRS and RSS are studied.The numerical results show that the modified unbiased estimator and the modified BLUE of λ and α in RSS are significantly more efficient than the ones in SRS.展开更多
In stratified survey sampling, sometimes we have complete auxiliary information. One of the fundamental questions is how to effectively use the complete auxiliary information at the estimation stage. In this paper, we...In stratified survey sampling, sometimes we have complete auxiliary information. One of the fundamental questions is how to effectively use the complete auxiliary information at the estimation stage. In this paper, we extend the model-calibration method to obtain estimators of the finite population mean by using complete auxiliary information from stratified sampling survey data. We show that the resulting estimators effectively use auxiliary information at the estimation stage and possess a number of attractive features such as asymptotically design-unbiased irrespective of the working model and approximately model-unbiased under the model. When a linear working-model is used, the resulting estimators reduce to the usual calibration estimator(or GREG).展开更多
Data processing of small samples is an important and valuable research problem in the electronic equipment test. Because it is difficult and complex to determine the probability distribution of small samples, it is di...Data processing of small samples is an important and valuable research problem in the electronic equipment test. Because it is difficult and complex to determine the probability distribution of small samples, it is difficult to use the traditional probability theory to process the samples and assess the degree of uncertainty. Using the grey relational theory and the norm theory, the grey distance information approach, which is based on the grey distance information quantity of a sample and the average grey distance information quantity of the samples, is proposed in this article. The definitions of the grey distance information quantity of a sample and the average grey distance information quantity of the samples, with their characteristics and algorithms, are introduced. The correlative problems, including the algorithm of estimated value, the standard deviation, and the acceptance and rejection criteria of the samples and estimated results, are also proposed. Moreover, the information whitening ratio is introduced to select the weight algorithm and to compare the different samples. Several examples are given to demonstrate the application of the proposed approach. The examples show that the proposed approach, which has no demand for the probability distribution of small samples, is feasible and effective.展开更多
Background Functional mapping, despite its proven efficiency, suffers from a “chicken or egg” scenario, in that, poor spatial features lead to inadequate spectral alignment and vice versa during training, often resu...Background Functional mapping, despite its proven efficiency, suffers from a “chicken or egg” scenario, in that, poor spatial features lead to inadequate spectral alignment and vice versa during training, often resulting in slow convergence, high computational costs, and learning failures, particularly when small datasets are used. Methods A novel method is presented for dense-shape correspondence, whereby the spatial information transformed by neural networks is combined with the projections onto spectral maps to overcome the “chicken or egg” challenge by selectively sampling only points with high confidence in their alignment. These points then contribute to the alignment and spectral loss terms, boosting training, and accelerating convergence by a factor of five. To ensure full unsupervised learning, the Gromov–Hausdorff distance metric was used to select the points with the maximal alignment score displaying most confidence. Results The effectiveness of the proposed approach was demonstrated on several benchmark datasets, whereby results were reported as superior to those of spectral and spatial-based methods. Conclusions The proposed method provides a promising new approach to dense-shape correspondence, addressing the key challenges in the field and offering significant advantages over the current methods, including faster convergence, improved accuracy, and reduced computational costs.展开更多
When analyzing and evaluating risks in insurance, people are often confronted with the situation of incomplete information and insufficient data, which is known as a small-sample problem. In this paper, a one-dimensio...When analyzing and evaluating risks in insurance, people are often confronted with the situation of incomplete information and insufficient data, which is known as a small-sample problem. In this paper, a one-dimensional small-sample problem in insurance was investigated using the kernel density estimation method (KerM) and general limited information diffusion method (GIDM). In particular, MacCormack technique was applied to get the solutions of GIDM equations and then the optimal diffusion solution was acquired based on the two optimization principles. Finally, the analysis introduced in this paper was verified by treating some examples and satisfying results were obtained.展开更多
The prediction process often runs with small samples and under-sufficient information.To target this problem,we propose a performance comparison study that combines prediction and optimization algorithms based on expe...The prediction process often runs with small samples and under-sufficient information.To target this problem,we propose a performance comparison study that combines prediction and optimization algorithms based on experimental data analysis.Through a large number of prediction and optimization experiments,the accuracy and stability of the prediction method and the correction ability of the optimization method are studied.First,five traditional single-item prediction methods are used to process small samples with under-sufficient information,and the standard deviation method is used to assign weights on the five methods for combined forecasting.The accuracy of the prediction results is ranked.The mean and variance of the rankings reflect the accuracy and stability of the prediction method.Second,the error elimination prediction optimization method is proposed.To make,the prediction results are corrected by error elimination optimization method(EEOM),Markov optimization and two-layer optimization separately to obtain more accurate prediction results.The degree improvement and decline are used to reflect the correction ability of the optimization method.The results show that the accuracy and stability of combined prediction are the best in the prediction methods,and the correction ability of error elimination optimization is the best in the optimization methods.The combination of the two methods can well solve the problem of prediction with small samples and under-sufficient information.Finally,the accuracy of the combination of the combined prediction and the error elimination optimization is verified by predicting the number of unsafe events in civil aviation in a certain year.展开更多
In order to meet the demand of testability analysis and evaluation for complex equipment under a small sample test in the equipment life cycle, the hierarchical hybrid testability model- ing and evaluation method (HH...In order to meet the demand of testability analysis and evaluation for complex equipment under a small sample test in the equipment life cycle, the hierarchical hybrid testability model- ing and evaluation method (HHTME), which combines the testabi- lity structure model (TSM) with the testability Bayesian networks model (TBNM), is presented. Firstly, the testability network topo- logy of complex equipment is built by using the hierarchical hybrid testability modeling method. Secondly, the prior conditional prob- ability distribution between network nodes is determined through expert experience. Then the Bayesian method is used to update the conditional probability distribution, according to history test information, virtual simulation information and similar product in- formation. Finally, the learned hierarchical hybrid testability model (HHTM) is used to estimate the testability of equipment. Compared with the results of other modeling methods, the relative deviation of the HHTM is only 0.52%, and the evaluation result is the most accu rate.展开更多
Abstract--This paper conducts a survey on iterative learn- ing control (ILC) with incomplete information and associated control system design, which is a frontier of the ILC field. The incomplete information, includ...Abstract--This paper conducts a survey on iterative learn- ing control (ILC) with incomplete information and associated control system design, which is a frontier of the ILC field. The incomplete information, including passive and active types, can cause data loss or fragment due to various factors. Passive incomplete information refers to incomplete data and information caused by practical system limitations during data collection, storage, transmission, and processing, such as data dropouts, delays, disordering, and limited transmission bandwidth. Active incomplete information refers to incomplete data and information caused by man-made reduction of data quantity and quality on the premise that the given objective is satisfied, such as sampling and quantization. This survey emphasizes two aspects: the first one is how to guarantee good learning performance and tracking performance with passive incomplete data, and the second is how to balance the control performance index and data demand by active means. The promising research directions along this topic are also addressed, where data robustness is highly emphasized. This survey is expected to improve understanding of the restrictive relationship and trade-off between incomplete data and tracking performance, quantitatively, and promote further developments of ILC theory. Index Terms--Data dropout, data robustness, incomplete in- formation, iterative learning controi(ILC), quantized control, sampled control, varying lengths.展开更多
Coupling Bayes’Theorem with a two-dimensional(2D)groundwater solute advection-diffusion transport equation allows an inverse model to be established to identify a set of contamination source parameters including sour...Coupling Bayes’Theorem with a two-dimensional(2D)groundwater solute advection-diffusion transport equation allows an inverse model to be established to identify a set of contamination source parameters including source intensity(M),release location(0 X,0 Y)and release time(0 T),based on monitoring well data.To address the issues of insufficient monitoring wells or weak correlation between monitoring data and model parameters,a monitoring well design optimization approach was developed based on the Bayesian formula and information entropy.To demonstrate how the model works,an exemplar problem with an instantaneous release of a contaminant in a confined groundwater aquifer was employed.The information entropy of the model parameters posterior distribution was used as a criterion to evaluate the monitoring data quantity index.The optimal monitoring well position and monitoring frequency were solved by the two-step Monte Carlo method and differential evolution algorithm given a known well monitoring locations and monitoring events.Based on the optimized monitoring well position and sampling frequency,the contamination source was identified by an improved Metropolis algorithm using the Latin hypercube sampling approach.The case study results show that the following parameters were obtained:1)the optimal monitoring well position(D)is at(445,200);and 2)the optimal monitoring frequency(Δt)is 7,providing that the monitoring events is set as 5 times.Employing the optimized monitoring well position and frequency,the mean errors of inverse modeling results in source parameters(M,X0,Y0,T0)were 9.20%,0.25%,0.0061%,and 0.33%,respectively.The optimized monitoring well position and sampling frequency canIt was also learnt that the improved Metropolis-Hastings algorithm(a Markov chain Monte Carlo method)can make the inverse modeling result independent of the initial sampling points and achieves an overall optimization,which significantly improved the accuracy and numerical stability of the inverse modeling results.展开更多
Characterizing spatial distribution of soil liquefaction potential is critical for assessing liquefactionrelated hazards(e.g.building damages caused by liquefaction-induced differential settlement).However,in engineer...Characterizing spatial distribution of soil liquefaction potential is critical for assessing liquefactionrelated hazards(e.g.building damages caused by liquefaction-induced differential settlement).However,in engineering practice,soil liquefaction potential is usually measured at limited locations in a specific site using in situ tests,e.g.cone penetration tests(CPTs),due to the restrictions of time,cost and access to subsurface space.In these cases,liquefaction potential of soil at untested locations requires to be interpreted from limited measured data points using proper interpolation method,leading to remarkable statistical uncertainty in liquefaction assessment.This underlines an important question of how to optimize the locations of CPT soundings and determine the minimum number of CPTs for achieving a target reliability level of liquefaction assessment.To tackle this issue,this study proposes a smart sampling strategy for determining the minimum number of CPTs and their optimal locations in a selfadaptive and data-driven manner.The proposed sampling strategy leverages on information entropy and Bayesian compressive sampling(BCS).Both simulated and real CPT data are used to demonstrate the proposed method.Illustrative examples indicate that the proposed method can adaptively and sequentially select the required number and optimal locations of CPTs.展开更多
There are many kinds of fires occurring under different conditions. For a specific site, it is difficult to collect sufficient data for analyzing the fire risk. In this paper, we suggest an information diffusion techn...There are many kinds of fires occurring under different conditions. For a specific site, it is difficult to collect sufficient data for analyzing the fire risk. In this paper, we suggest an information diffusion technique to analyze fire risk with a small sample. The information distribution method is applied to change crisp observations into fuzzy sets, and then to effectively construct a fuzzy relationship between fire and surroundings. With the data of Shanghai in winter, we show how to use the technique to analyze the fire risk.展开更多
There exists some discord or contradiction of information during the process of fault diagnosis for rotary machine. But the traditional methods used in fault diagnosis can not dispose of the information. A model of fa...There exists some discord or contradiction of information during the process of fault diagnosis for rotary machine. But the traditional methods used in fault diagnosis can not dispose of the information. A model of fault diagnosis for a rotary machine based on information entropy theory and rough set theory is presented in this paper. The model has clear mathematical definition and can dispose both complete unification information and complete inconsistent information of vibration faults. By using the model, decision rules of six typical vibration faults of a steam turbine and electric generating set are deduced from experiment samples. Finally, the decision rules are validated by selected samples and good identification results are acquired.展开更多
In order to understand the security conditions of the incomplete interval-valued information system (IllS) and acquire the corresponding solution of security problems, this paper proposes a multi-attribute group dec...In order to understand the security conditions of the incomplete interval-valued information system (IllS) and acquire the corresponding solution of security problems, this paper proposes a multi-attribute group decision- making (MAGDM) security assessment method based on the technique for order performance by similarity to ideal solution (TOPSIS). For IllS with preference information, combining with dominance-based rough set approach (DRSA), the effect of incomplete interval-valued information on decision results is discussed. For the imprecise judgment matrices, the security attribute weight can be obtained using Gibbs sampling. A numerical example shows that the proposed method can acquire some valuable knowledge hidden in the incomplete interval-valued information. The effectiveness of the proposed method in the synthetic security assessment for IIIS is verified.展开更多
Background: Remote sensing-based inventories are essential in estimating forest cover in tropical and subtropical countries, where ground inventories cannot be performed periodically at a large scale owing to high cos...Background: Remote sensing-based inventories are essential in estimating forest cover in tropical and subtropical countries, where ground inventories cannot be performed periodically at a large scale owing to high costs and forest inaccessibility(e.g. REDD projects) and are mandatory for constructing historical records that can be used as forest cover baselines. Given the conditions of such inventories, the survey area is partitioned into a grid of imagery segments of pre-fixed size where the proportion of forest cover can be measured within segments using a combination of unsupervised(automated or semi-automated) classification of satellite imagery and manual(i.e. visual on-screen)enhancements. Because visual on-screen operations are time expensive procedures, manual classification can be performed only for a sample of imagery segments selected at a first stage, while forest cover within each selected segment is estimated at a second stage from a sample of pixels selected within the segment. Because forest cover data arising from unsupervised satellite imagery classification may be freely available(e.g. Landsat imagery)over the entire survey area(wall-to-wall data) and are likely to be good proxies of manually classified cover data(sample data), they can be adopted as suitable auxiliary information.Methods: The question is how to choose the sample areas where manual classification is carried out. We have investigated the efficiency of one-per-stratum stratified sampling for selecting segments and pixels, where to carry out manual classification and to determine the efficiency of the difference estimator for exploiting auxiliary information at the estimation level. The performance of this strategy is compared with simple random sampling without replacement.Results: Our results were obtained theoretically from three artificial populations constructed from the Landsat classification(forest/non forest) available at pixel level for a study area located in central Italy, assuming three levels of error rates of the unsupervised classification of satellite imagery. The exploitation of map data as auxiliary information in the difference estimator proves to be highly effective with respect to the Horvitz-Thompson estimator,in which no auxiliary information is exploited. The use of one-per-stratum stratified sampling provides relevant improvement with respect to the use of simple random sampling without replacement.Conclusions: The use of one-per-stratum stratified sampling with many imagery segments selected at the first stage and few pixels within at the second stage- jointly with a difference estimator- proves to be a suitable strategy to estimate forest cover by remote sensing-based inventories.展开更多
In this paper, auxiliary information is used to determine an estimator of finite population total using nonparametric regression under stratified random sampling. To achieve this, a model-based approach is adopted by ...In this paper, auxiliary information is used to determine an estimator of finite population total using nonparametric regression under stratified random sampling. To achieve this, a model-based approach is adopted by making use of the local polynomial regression estimation to predict the nonsampled values of the survey variable y. The performance of the proposed estimator is investigated against some design-based and model-based regression estimators. The simulation experiments show that the resulting estimator exhibits good properties. Generally, good confidence intervals are seen for the nonparametric regression estimators, and use of the proposed estimator leads to relatively smaller values of RE compared to other estimators.展开更多
基金Supported by the National Natural Science Foundation of China(Grant No.51375217)
文摘Finite rate of innovation sampling is a novel sub-Nyquist sampling method that can reconstruct a signal from sparse sampling data.The application of this method in ultrasonic testing greatly reduces the signal sampling rate and the quantity of sampling data.However,the pulse number of the signal must be known beforehand for the signal reconstruction procedure.The accuracy of this prior information directly affects the accuracy of the estimated parameters of the signal and influences the assessment of flaws,leading to a lower defect detection ratio.Although the pulse number can be pre-given by theoretical analysis,the process is still unable to assess actual complex random orientation defects.Therefore,this paper proposes a new method that uses singular value decomposition(SVD) for estimating the pulse number from sparse sampling data and avoids the shortcoming of providing the pulse number in advance for signal reconstruction.When the sparse sampling data have been acquired from the ultrasonic signal,these data are transformed to discrete Fourier coefficients.A Hankel matrix is then constructed from these coefficients,and SVD is performed on the matrix.The decomposition coefficients reserve the information of the pulse number.When the decomposition coefficients generated by noise according to noise level are removed,the number of the remaining decomposition coefficients is the signal pulse number.The feasibility of the proposed method was verified through simulation experiments.The applicability was tested in ultrasonic experiments by using sample flawed pipelines.Results from simulations and real experiments demonstrated the efficiency of this method.
基金Supported by National Science Foundation of China(11901236,12261036),Scientific Research Fund of Hunan Provincial Education Department(21A0328)Provincial Natural Science Foundation of Hunan(2022JJ30469)+1 种基金Young Core Teacher Foundation of Hunan Province([2020]43)Jishou University Laboratory Program(JDDL2017001,JDLF2021024).
文摘In the current paper,we considered the Fisher information matrix from the generalized Rayleigh distribution(GR)distribution in ranked set sampling(RSS).The numerical results show that the ranked set sample carries more information about λ and α than a simple random sample of equivalent size.In order to give more insight into the performance of RSS with respect to(w.r.t.)simple random sampling(SRS),a modified unbiased estimator and a modified best linear unbiased estimator(BLUE)of scale and shape λ and α from GR distribution in SRS and RSS are studied.The numerical results show that the modified unbiased estimator and the modified BLUE of λ and α in RSS are significantly more efficient than the ones in SRS.
基金Supported by the National Natural Science Foundation of China(10571093)
文摘In stratified survey sampling, sometimes we have complete auxiliary information. One of the fundamental questions is how to effectively use the complete auxiliary information at the estimation stage. In this paper, we extend the model-calibration method to obtain estimators of the finite population mean by using complete auxiliary information from stratified sampling survey data. We show that the resulting estimators effectively use auxiliary information at the estimation stage and possess a number of attractive features such as asymptotically design-unbiased irrespective of the working model and approximately model-unbiased under the model. When a linear working-model is used, the resulting estimators reduce to the usual calibration estimator(or GREG).
文摘Data processing of small samples is an important and valuable research problem in the electronic equipment test. Because it is difficult and complex to determine the probability distribution of small samples, it is difficult to use the traditional probability theory to process the samples and assess the degree of uncertainty. Using the grey relational theory and the norm theory, the grey distance information approach, which is based on the grey distance information quantity of a sample and the average grey distance information quantity of the samples, is proposed in this article. The definitions of the grey distance information quantity of a sample and the average grey distance information quantity of the samples, with their characteristics and algorithms, are introduced. The correlative problems, including the algorithm of estimated value, the standard deviation, and the acceptance and rejection criteria of the samples and estimated results, are also proposed. Moreover, the information whitening ratio is introduced to select the weight algorithm and to compare the different samples. Several examples are given to demonstrate the application of the proposed approach. The examples show that the proposed approach, which has no demand for the probability distribution of small samples, is feasible and effective.
基金Supported by the Zimin Institute for Engineering Solutions Advancing Better Lives。
文摘Background Functional mapping, despite its proven efficiency, suffers from a “chicken or egg” scenario, in that, poor spatial features lead to inadequate spectral alignment and vice versa during training, often resulting in slow convergence, high computational costs, and learning failures, particularly when small datasets are used. Methods A novel method is presented for dense-shape correspondence, whereby the spatial information transformed by neural networks is combined with the projections onto spectral maps to overcome the “chicken or egg” challenge by selectively sampling only points with high confidence in their alignment. These points then contribute to the alignment and spectral loss terms, boosting training, and accelerating convergence by a factor of five. To ensure full unsupervised learning, the Gromov–Hausdorff distance metric was used to select the points with the maximal alignment score displaying most confidence. Results The effectiveness of the proposed approach was demonstrated on several benchmark datasets, whereby results were reported as superior to those of spectral and spatial-based methods. Conclusions The proposed method provides a promising new approach to dense-shape correspondence, addressing the key challenges in the field and offering significant advantages over the current methods, including faster convergence, improved accuracy, and reduced computational costs.
基金Project supported by the National Natural Science Foundation of China (Grant No.10271072)
文摘When analyzing and evaluating risks in insurance, people are often confronted with the situation of incomplete information and insufficient data, which is known as a small-sample problem. In this paper, a one-dimensional small-sample problem in insurance was investigated using the kernel density estimation method (KerM) and general limited information diffusion method (GIDM). In particular, MacCormack technique was applied to get the solutions of GIDM equations and then the optimal diffusion solution was acquired based on the two optimization principles. Finally, the analysis introduced in this paper was verified by treating some examples and satisfying results were obtained.
基金This work was supported by the Scientific Research Projects of Tianjin Educational Committee(No.2020KJ029)。
文摘The prediction process often runs with small samples and under-sufficient information.To target this problem,we propose a performance comparison study that combines prediction and optimization algorithms based on experimental data analysis.Through a large number of prediction and optimization experiments,the accuracy and stability of the prediction method and the correction ability of the optimization method are studied.First,five traditional single-item prediction methods are used to process small samples with under-sufficient information,and the standard deviation method is used to assign weights on the five methods for combined forecasting.The accuracy of the prediction results is ranked.The mean and variance of the rankings reflect the accuracy and stability of the prediction method.Second,the error elimination prediction optimization method is proposed.To make,the prediction results are corrected by error elimination optimization method(EEOM),Markov optimization and two-layer optimization separately to obtain more accurate prediction results.The degree improvement and decline are used to reflect the correction ability of the optimization method.The results show that the accuracy and stability of combined prediction are the best in the prediction methods,and the correction ability of error elimination optimization is the best in the optimization methods.The combination of the two methods can well solve the problem of prediction with small samples and under-sufficient information.Finally,the accuracy of the combination of the combined prediction and the error elimination optimization is verified by predicting the number of unsafe events in civil aviation in a certain year.
基金supported by the National Defense Pre-research Foundation of China(51327030104)
文摘In order to meet the demand of testability analysis and evaluation for complex equipment under a small sample test in the equipment life cycle, the hierarchical hybrid testability model- ing and evaluation method (HHTME), which combines the testabi- lity structure model (TSM) with the testability Bayesian networks model (TBNM), is presented. Firstly, the testability network topo- logy of complex equipment is built by using the hierarchical hybrid testability modeling method. Secondly, the prior conditional prob- ability distribution between network nodes is determined through expert experience. Then the Bayesian method is used to update the conditional probability distribution, according to history test information, virtual simulation information and similar product in- formation. Finally, the learned hierarchical hybrid testability model (HHTM) is used to estimate the testability of equipment. Compared with the results of other modeling methods, the relative deviation of the HHTM is only 0.52%, and the evaluation result is the most accu rate.
基金supported by the National Natural Science Foundation of China(61673045)Beijing Natural Science Foundation(4152040)
文摘Abstract--This paper conducts a survey on iterative learn- ing control (ILC) with incomplete information and associated control system design, which is a frontier of the ILC field. The incomplete information, including passive and active types, can cause data loss or fragment due to various factors. Passive incomplete information refers to incomplete data and information caused by practical system limitations during data collection, storage, transmission, and processing, such as data dropouts, delays, disordering, and limited transmission bandwidth. Active incomplete information refers to incomplete data and information caused by man-made reduction of data quantity and quality on the premise that the given objective is satisfied, such as sampling and quantization. This survey emphasizes two aspects: the first one is how to guarantee good learning performance and tracking performance with passive incomplete data, and the second is how to balance the control performance index and data demand by active means. The promising research directions along this topic are also addressed, where data robustness is highly emphasized. This survey is expected to improve understanding of the restrictive relationship and trade-off between incomplete data and tracking performance, quantitatively, and promote further developments of ILC theory. Index Terms--Data dropout, data robustness, incomplete in- formation, iterative learning controi(ILC), quantized control, sampled control, varying lengths.
基金This work was supported by Major Science and Technology Program for Water Pollution Control and Treatment(No.2015ZX07406005)Also thanks to the National Natural Science Foundation of China(No.41430643 and No.51774270)the National Key Research&Development Plan(No.2016YFC0501109).
文摘Coupling Bayes’Theorem with a two-dimensional(2D)groundwater solute advection-diffusion transport equation allows an inverse model to be established to identify a set of contamination source parameters including source intensity(M),release location(0 X,0 Y)and release time(0 T),based on monitoring well data.To address the issues of insufficient monitoring wells or weak correlation between monitoring data and model parameters,a monitoring well design optimization approach was developed based on the Bayesian formula and information entropy.To demonstrate how the model works,an exemplar problem with an instantaneous release of a contaminant in a confined groundwater aquifer was employed.The information entropy of the model parameters posterior distribution was used as a criterion to evaluate the monitoring data quantity index.The optimal monitoring well position and monitoring frequency were solved by the two-step Monte Carlo method and differential evolution algorithm given a known well monitoring locations and monitoring events.Based on the optimized monitoring well position and sampling frequency,the contamination source was identified by an improved Metropolis algorithm using the Latin hypercube sampling approach.The case study results show that the following parameters were obtained:1)the optimal monitoring well position(D)is at(445,200);and 2)the optimal monitoring frequency(Δt)is 7,providing that the monitoring events is set as 5 times.Employing the optimized monitoring well position and frequency,the mean errors of inverse modeling results in source parameters(M,X0,Y0,T0)were 9.20%,0.25%,0.0061%,and 0.33%,respectively.The optimized monitoring well position and sampling frequency canIt was also learnt that the improved Metropolis-Hastings algorithm(a Markov chain Monte Carlo method)can make the inverse modeling result independent of the initial sampling points and achieves an overall optimization,which significantly improved the accuracy and numerical stability of the inverse modeling results.
基金supported by grants from the Research Grant Council of Hong Kong Special Administrative Region,China(Project Nos.CityU 11202121 and CityU 11213119).
文摘Characterizing spatial distribution of soil liquefaction potential is critical for assessing liquefactionrelated hazards(e.g.building damages caused by liquefaction-induced differential settlement).However,in engineering practice,soil liquefaction potential is usually measured at limited locations in a specific site using in situ tests,e.g.cone penetration tests(CPTs),due to the restrictions of time,cost and access to subsurface space.In these cases,liquefaction potential of soil at untested locations requires to be interpreted from limited measured data points using proper interpolation method,leading to remarkable statistical uncertainty in liquefaction assessment.This underlines an important question of how to optimize the locations of CPT soundings and determine the minimum number of CPTs for achieving a target reliability level of liquefaction assessment.To tackle this issue,this study proposes a smart sampling strategy for determining the minimum number of CPTs and their optimal locations in a selfadaptive and data-driven manner.The proposed sampling strategy leverages on information entropy and Bayesian compressive sampling(BCS).Both simulated and real CPT data are used to demonstrate the proposed method.Illustrative examples indicate that the proposed method can adaptively and sequentially select the required number and optimal locations of CPTs.
文摘There are many kinds of fires occurring under different conditions. For a specific site, it is difficult to collect sufficient data for analyzing the fire risk. In this paper, we suggest an information diffusion technique to analyze fire risk with a small sample. The information distribution method is applied to change crisp observations into fuzzy sets, and then to effectively construct a fuzzy relationship between fire and surroundings. With the data of Shanghai in winter, we show how to use the technique to analyze the fire risk.
基金The paper is supported by the 863 Program of China under Grant No 2006AA04A110
文摘There exists some discord or contradiction of information during the process of fault diagnosis for rotary machine. But the traditional methods used in fault diagnosis can not dispose of the information. A model of fault diagnosis for a rotary machine based on information entropy theory and rough set theory is presented in this paper. The model has clear mathematical definition and can dispose both complete unification information and complete inconsistent information of vibration faults. By using the model, decision rules of six typical vibration faults of a steam turbine and electric generating set are deduced from experiment samples. Finally, the decision rules are validated by selected samples and good identification results are acquired.
基金Supported by the National Natural Science Foundation of China(No.60605019)
文摘In order to understand the security conditions of the incomplete interval-valued information system (IllS) and acquire the corresponding solution of security problems, this paper proposes a multi-attribute group decision- making (MAGDM) security assessment method based on the technique for order performance by similarity to ideal solution (TOPSIS). For IllS with preference information, combining with dominance-based rough set approach (DRSA), the effect of incomplete interval-valued information on decision results is discussed. For the imprecise judgment matrices, the security attribute weight can be obtained using Gibbs sampling. A numerical example shows that the proposed method can acquire some valuable knowledge hidden in the incomplete interval-valued information. The effectiveness of the proposed method in the synthetic security assessment for IIIS is verified.
文摘Background: Remote sensing-based inventories are essential in estimating forest cover in tropical and subtropical countries, where ground inventories cannot be performed periodically at a large scale owing to high costs and forest inaccessibility(e.g. REDD projects) and are mandatory for constructing historical records that can be used as forest cover baselines. Given the conditions of such inventories, the survey area is partitioned into a grid of imagery segments of pre-fixed size where the proportion of forest cover can be measured within segments using a combination of unsupervised(automated or semi-automated) classification of satellite imagery and manual(i.e. visual on-screen)enhancements. Because visual on-screen operations are time expensive procedures, manual classification can be performed only for a sample of imagery segments selected at a first stage, while forest cover within each selected segment is estimated at a second stage from a sample of pixels selected within the segment. Because forest cover data arising from unsupervised satellite imagery classification may be freely available(e.g. Landsat imagery)over the entire survey area(wall-to-wall data) and are likely to be good proxies of manually classified cover data(sample data), they can be adopted as suitable auxiliary information.Methods: The question is how to choose the sample areas where manual classification is carried out. We have investigated the efficiency of one-per-stratum stratified sampling for selecting segments and pixels, where to carry out manual classification and to determine the efficiency of the difference estimator for exploiting auxiliary information at the estimation level. The performance of this strategy is compared with simple random sampling without replacement.Results: Our results were obtained theoretically from three artificial populations constructed from the Landsat classification(forest/non forest) available at pixel level for a study area located in central Italy, assuming three levels of error rates of the unsupervised classification of satellite imagery. The exploitation of map data as auxiliary information in the difference estimator proves to be highly effective with respect to the Horvitz-Thompson estimator,in which no auxiliary information is exploited. The use of one-per-stratum stratified sampling provides relevant improvement with respect to the use of simple random sampling without replacement.Conclusions: The use of one-per-stratum stratified sampling with many imagery segments selected at the first stage and few pixels within at the second stage- jointly with a difference estimator- proves to be a suitable strategy to estimate forest cover by remote sensing-based inventories.
文摘In this paper, auxiliary information is used to determine an estimator of finite population total using nonparametric regression under stratified random sampling. To achieve this, a model-based approach is adopted by making use of the local polynomial regression estimation to predict the nonsampled values of the survey variable y. The performance of the proposed estimator is investigated against some design-based and model-based regression estimators. The simulation experiments show that the resulting estimator exhibits good properties. Generally, good confidence intervals are seen for the nonparametric regression estimators, and use of the proposed estimator leads to relatively smaller values of RE compared to other estimators.