This study presents a kinematic calibration method for exoskeletal inertial motion capture (EI-MoCap) system with considering the random colored noise such as gyroscopic drift.In this method, the geometric parameters ...This study presents a kinematic calibration method for exoskeletal inertial motion capture (EI-MoCap) system with considering the random colored noise such as gyroscopic drift.In this method, the geometric parameters are calibrated by the traditional calibration method at first. Then, in order to calibrate the parameters affected by the random colored noise, the expectation maximization (EM) algorithm is introduced. Through the use of geometric parameters calibrated by the traditional calibration method, the iterations under the EM framework are decreased and the efficiency of the proposed method on embedded system is improved. The performance of the proposed kinematic calibration method is compared to the traditional calibration method. Furthermore, the feasibility of the proposed method is verified on the EI-MoCap system. The simulation and experiment demonstrate that the motion capture precision is significantly improved by 16.79%and 7.16%respectively in comparison to the traditional calibration method.展开更多
The neutron spectrum unfolding by Bonner sphere spectrometer(BSS) is considered a complex multidimensional model,which requires complex mathematical methods to solve the first kind of Fredholm integral equation. In or...The neutron spectrum unfolding by Bonner sphere spectrometer(BSS) is considered a complex multidimensional model,which requires complex mathematical methods to solve the first kind of Fredholm integral equation. In order to solve the problem of the maximum likelihood expectation maximization(MLEM) algorithm which is easy to suffer the pitfalls of local optima and the particle swarm optimization(PSO) algorithm which is easy to get unreasonable flight direction and step length of particles, which leads to the invalid iteration and affect efficiency and accuracy, an improved PSO-MLEM algorithm, combined of PSO and MLEM algorithm, is proposed for neutron spectrum unfolding. The dynamic acceleration factor is used to balance the ability of global and local search, and improves the convergence speed and accuracy of the algorithm. Firstly, the Monte Carlo method was used to simulated the BSS to obtain the response function and count rates of BSS. In the simulation of count rate, four reference spectra from the IAEA Technical Report Series No. 403 were used as input parameters of the Monte Carlo method. The PSO-MLEM algorithm was used to unfold the neutron spectrum of the simulated data and was verified by the difference of the unfolded spectrum to the reference spectrum. Finally, the 252Cf neutron source was measured by BSS, and the PSO-MLEM algorithm was used to unfold the experimental neutron spectrum.Compared with maximum entropy deconvolution(MAXED), PSO and MLEM algorithm, the PSO-MLEM algorithm has fewer parameters and automatically adjusts the dynamic acceleration factor to solve the problem of local optima. The convergence speed of the PSO-MLEM algorithm is 1.4 times and 3.1 times that of the MLEM and PSO algorithms. Compared with PSO, MLEM and MAXED, the correlation coefficients of PSO-MLEM algorithm are increased by 33.1%, 33.5% and 1.9%, and the relative mean errors are decreased by 98.2%, 97.8% and 67.4%.展开更多
Energy spectra of neutrons are important for identification of unknown neutron sources and for determination of the equivalent dose. Although standard energy spectra of neutrons are available in some situations, e.g.,...Energy spectra of neutrons are important for identification of unknown neutron sources and for determination of the equivalent dose. Although standard energy spectra of neutrons are available in some situations, e.g., for some radiotherapy treatment machines, they are unknown in other cases, e.g., for photoneutrons created in radiotherapy rooms and neutrons generated in nuclear reactors. In situations where neutron energy spectra need to be determined, unfolding the required neutron energy spectra using the Bonner sphere spectrometer (BSS) and nested neutron spectrometer (NNS) has been found promising. However, without any prior knowledge on the spectra, the unfolding process has remained a tedious task. In this work, a standalone numerical tool named ‘‘NRUunfold’’ was developed which could satisfactorily unfold neutron spectra for BSS or NNS, or any other systems using similar detection methodology. A generic and versatile algorithm based on maximum-likelihood expectation– maximization method was developed and benchmarked against the widely used STAY’SL algorithm which was based on the least squares method. The present method could output decent results in the absence of precisely calculated initial guess, although it was also remarked that employment of exceptionally bizarre initial spectra could lead to some unreasonable output spectra. The neutron count rates computed using the manufacturer’s response functions were used for sensitivity studies. The present NRUunfold code could be useful for neutron energy spectrum unfolding for BSS or NNS applications in the absence of a precisely calculated initial guess.展开更多
A new parallel expectation-maximization (EM) algorithm is proposed for large databases. The purpose of the algorithm is to accelerate the operation of the EM algorithm. As a well-known algorithm for estimation in ge...A new parallel expectation-maximization (EM) algorithm is proposed for large databases. The purpose of the algorithm is to accelerate the operation of the EM algorithm. As a well-known algorithm for estimation in generic statistical problems, the EM algorithm has been widely used in many domains. But it often requires significant computational resources. So it is needed to develop more elaborate methods to adapt the databases to a large number of records or large dimensionality. The parallel EM algorithm is based on partial Esteps which has the standard convergence guarantee of EM. The algorithm utilizes fully the advantage of parallel computation. It was confirmed that the algorithm obtains about 2.6 speedups in contrast with the standard EM algorithm through its application to large databases. The running time will decrease near linearly when the number of processors increasing.展开更多
Intrusion detection is the investigation process of information about the system activities or its data to detect any malicious behavior or unauthorized activity.Most of the IDS implement K-means clustering technique ...Intrusion detection is the investigation process of information about the system activities or its data to detect any malicious behavior or unauthorized activity.Most of the IDS implement K-means clustering technique due to its linear complexity and fast computing ability.Nonetheless,it is Naïve use of the mean data value for the cluster core that presents a major drawback.The chances of two circular clusters having different radius and centering at the same mean will occur.This condition cannot be addressed by the K-means algorithm because the mean value of the various clusters is very similar together.However,if the clusters are not spherical,it fails.To overcome this issue,a new integrated hybrid model by integrating expectation maximizing(EM)clustering using a Gaussian mixture model(GMM)and naïve Bays classifier have been proposed.In this model,GMM give more flexibility than K-Means in terms of cluster covariance.Also,they use probabilities function and soft clustering,that’s why they can have multiple cluster for a single data.In GMM,we can define the cluster form in GMM by two parameters:the mean and the standard deviation.This means that by using these two parameters,the cluster can take any kind of elliptical shape.EM-GMM will be used to cluster data based on data activity into the corresponding category.展开更多
The quality of synthetic aperture radar(SAR)image degrades in the case of multiple imaging projection planes(IPPs)and multiple overlapping ship targets,and then the performance of target classification and recognition...The quality of synthetic aperture radar(SAR)image degrades in the case of multiple imaging projection planes(IPPs)and multiple overlapping ship targets,and then the performance of target classification and recognition can be influenced.For addressing this issue,a method for extracting ship targets with overlaps via the expectation maximization(EM)algorithm is pro-posed.First,the scatterers of ship targets are obtained via the target detection technique.Then,the EM algorithm is applied to extract the scatterers of a single ship target with a single IPP.Afterwards,a novel image amplitude estimation approach is pro-posed,with which the radar image of a single target with a sin-gle IPP can be generated.The proposed method can accom-plish IPP selection and targets separation in the image domain,which can improve the image quality and reserve the target information most possibly.Results of simulated and real mea-sured data demonstrate the effectiveness of the proposed method.展开更多
Classical survival analysis assumes all subjects will experience the event of interest, but in some cases, a portion of the population may never encounter the event. These survival methods further assume independent s...Classical survival analysis assumes all subjects will experience the event of interest, but in some cases, a portion of the population may never encounter the event. These survival methods further assume independent survival times, which is not valid for honey bees, which live in nests. The study introduces a semi-parametric marginal proportional hazards mixture cure (PHMC) model with exchangeable correlation structure, using generalized estimating equations for survival data analysis. The model was tested on clustered right-censored bees survival data with a cured fraction, where two bee species were subjected to different entomopathogens to test the effect of the entomopathogens on the survival of the bee species. The Expectation-Solution algorithm is used to estimate the parameters. The study notes a weak positive association between cure statuses (ρ1=0.0007) and survival times for uncured bees (ρ2=0.0890), emphasizing their importance. The odds of being uncured for A. mellifera is higher than the odds for species M. ferruginea. The bee species, A. mellifera are more susceptible to entomopathogens icipe 7, icipe 20, and icipe 69. The Cox-Snell residuals show that the proposed semiparametric PH model generally fits the data well as compared to model that assume independent correlation structure. Thus, the semi parametric marginal proportional hazards mixture cure is parsimonious model for correlated bees survival data.展开更多
Compositional data, such as relative information, is a crucial aspect of machine learning and other related fields. It is typically recorded as closed data or sums to a constant, like 100%. The statistical linear mode...Compositional data, such as relative information, is a crucial aspect of machine learning and other related fields. It is typically recorded as closed data or sums to a constant, like 100%. The statistical linear model is the most used technique for identifying hidden relationships between underlying random variables of interest. However, data quality is a significant challenge in machine learning, especially when missing data is present. The linear regression model is a commonly used statistical modeling technique used in various applications to find relationships between variables of interest. When estimating linear regression parameters which are useful for things like future prediction and partial effects analysis of independent variables, maximum likelihood estimation (MLE) is the method of choice. However, many datasets contain missing observations, which can lead to costly and time-consuming data recovery. To address this issue, the expectation-maximization (EM) algorithm has been suggested as a solution for situations including missing data. The EM algorithm repeatedly finds the best estimates of parameters in statistical models that depend on variables or data that have not been observed. This is called maximum likelihood or maximum a posteriori (MAP). Using the present estimate as input, the expectation (E) step constructs a log-likelihood function. Finding the parameters that maximize the anticipated log-likelihood, as determined in the E step, is the job of the maximization (M) phase. This study looked at how well the EM algorithm worked on a made-up compositional dataset with missing observations. It used both the robust least square version and ordinary least square regression techniques. The efficacy of the EM algorithm was compared with two alternative imputation techniques, k-Nearest Neighbor (k-NN) and mean imputation (), in terms of Aitchison distances and covariance.展开更多
In this article,we study strong limit theorems for weighted sums of extended negatively dependent random variables under the sub-linear expectations.We establish general strong law and complete convergence theorems fo...In this article,we study strong limit theorems for weighted sums of extended negatively dependent random variables under the sub-linear expectations.We establish general strong law and complete convergence theorems for weighted sums of extended negatively dependent random variables under the sub-linear expectations.Our results of strong limit theorems are more general than some related results previously obtained by Thrum(1987),Li et al.(1995)and Wu(2010)in classical probability space.展开更多
Clustering is an unsupervised learning method used to organize raw data in such a way that those with the same (similar) characteristics are found in the same class and those that are dissimilar are found in different...Clustering is an unsupervised learning method used to organize raw data in such a way that those with the same (similar) characteristics are found in the same class and those that are dissimilar are found in different classes. In this day and age, the very rapid increase in the amount of data being produced brings new challenges in the analysis and storage of this data. Recently, there is a growing interest in key areas such as real-time data mining, which reveal an urgent need to process very large data under strict performance constraints. The objective of this paper is to survey four algorithms including K-Means algorithm, FCM algorithm, EM algorithm and BIRCH, used for data clustering and then show their strengths and weaknesses. Another task is to compare the results obtained by applying each of these algorithms to the same data and to give a conclusion based on these results.展开更多
针对当前广义频分复用(Generalized Frequency Division Multiplexing,GFDM)系统时变信道估计精度低的问题,提出基于稀疏贝叶斯学习的GFDM系统联合信道估计与符号检测算法.具体地,采用无干扰导频插入的GFDM多重响应信号模型,在稀疏贝叶...针对当前广义频分复用(Generalized Frequency Division Multiplexing,GFDM)系统时变信道估计精度低的问题,提出基于稀疏贝叶斯学习的GFDM系统联合信道估计与符号检测算法.具体地,采用无干扰导频插入的GFDM多重响应信号模型,在稀疏贝叶斯学习框架下,结合期望最大化算法(Expectation-Maximization,EM)和卡尔曼滤波与平滑算法实现块时变信道的最大似然估计;基于信道状态信息的估计值进行GFDM符号检测,并通过信道估计与符号检测的迭代处理逐步提高信道估计与符号检测的精度.仿真结果表明,所提算法能够获得接近完美信道状态信息条件下的误码率性能,且具有收敛速度快、对多普勒频移鲁棒性高等优点.展开更多
基金supported by the National Natural Science Foundation of China (61503392)。
文摘This study presents a kinematic calibration method for exoskeletal inertial motion capture (EI-MoCap) system with considering the random colored noise such as gyroscopic drift.In this method, the geometric parameters are calibrated by the traditional calibration method at first. Then, in order to calibrate the parameters affected by the random colored noise, the expectation maximization (EM) algorithm is introduced. Through the use of geometric parameters calibrated by the traditional calibration method, the iterations under the EM framework are decreased and the efficiency of the proposed method on embedded system is improved. The performance of the proposed kinematic calibration method is compared to the traditional calibration method. Furthermore, the feasibility of the proposed method is verified on the EI-MoCap system. The simulation and experiment demonstrate that the motion capture precision is significantly improved by 16.79%and 7.16%respectively in comparison to the traditional calibration method.
基金supported by the National Natural science Foundation of China (No. 42127807)the Sichuan Science and Technology Program (No. 2020YJ0334)the Sichuan Science and Technology Breeding Program (No. 2022041)。
文摘The neutron spectrum unfolding by Bonner sphere spectrometer(BSS) is considered a complex multidimensional model,which requires complex mathematical methods to solve the first kind of Fredholm integral equation. In order to solve the problem of the maximum likelihood expectation maximization(MLEM) algorithm which is easy to suffer the pitfalls of local optima and the particle swarm optimization(PSO) algorithm which is easy to get unreasonable flight direction and step length of particles, which leads to the invalid iteration and affect efficiency and accuracy, an improved PSO-MLEM algorithm, combined of PSO and MLEM algorithm, is proposed for neutron spectrum unfolding. The dynamic acceleration factor is used to balance the ability of global and local search, and improves the convergence speed and accuracy of the algorithm. Firstly, the Monte Carlo method was used to simulated the BSS to obtain the response function and count rates of BSS. In the simulation of count rate, four reference spectra from the IAEA Technical Report Series No. 403 were used as input parameters of the Monte Carlo method. The PSO-MLEM algorithm was used to unfold the neutron spectrum of the simulated data and was verified by the difference of the unfolded spectrum to the reference spectrum. Finally, the 252Cf neutron source was measured by BSS, and the PSO-MLEM algorithm was used to unfold the experimental neutron spectrum.Compared with maximum entropy deconvolution(MAXED), PSO and MLEM algorithm, the PSO-MLEM algorithm has fewer parameters and automatically adjusts the dynamic acceleration factor to solve the problem of local optima. The convergence speed of the PSO-MLEM algorithm is 1.4 times and 3.1 times that of the MLEM and PSO algorithms. Compared with PSO, MLEM and MAXED, the correlation coefficients of PSO-MLEM algorithm are increased by 33.1%, 33.5% and 1.9%, and the relative mean errors are decreased by 98.2%, 97.8% and 67.4%.
基金support from the Neutron computer cluster from the Department of Physics, City University of Hong Kong
文摘Energy spectra of neutrons are important for identification of unknown neutron sources and for determination of the equivalent dose. Although standard energy spectra of neutrons are available in some situations, e.g., for some radiotherapy treatment machines, they are unknown in other cases, e.g., for photoneutrons created in radiotherapy rooms and neutrons generated in nuclear reactors. In situations where neutron energy spectra need to be determined, unfolding the required neutron energy spectra using the Bonner sphere spectrometer (BSS) and nested neutron spectrometer (NNS) has been found promising. However, without any prior knowledge on the spectra, the unfolding process has remained a tedious task. In this work, a standalone numerical tool named ‘‘NRUunfold’’ was developed which could satisfactorily unfold neutron spectra for BSS or NNS, or any other systems using similar detection methodology. A generic and versatile algorithm based on maximum-likelihood expectation– maximization method was developed and benchmarked against the widely used STAY’SL algorithm which was based on the least squares method. The present method could output decent results in the absence of precisely calculated initial guess, although it was also remarked that employment of exceptionally bizarre initial spectra could lead to some unreasonable output spectra. The neutron count rates computed using the manufacturer’s response functions were used for sensitivity studies. The present NRUunfold code could be useful for neutron energy spectrum unfolding for BSS or NNS applications in the absence of a precisely calculated initial guess.
基金the National Natural Science Foundation of China(79990584)
文摘A new parallel expectation-maximization (EM) algorithm is proposed for large databases. The purpose of the algorithm is to accelerate the operation of the EM algorithm. As a well-known algorithm for estimation in generic statistical problems, the EM algorithm has been widely used in many domains. But it often requires significant computational resources. So it is needed to develop more elaborate methods to adapt the databases to a large number of records or large dimensionality. The parallel EM algorithm is based on partial Esteps which has the standard convergence guarantee of EM. The algorithm utilizes fully the advantage of parallel computation. It was confirmed that the algorithm obtains about 2.6 speedups in contrast with the standard EM algorithm through its application to large databases. The running time will decrease near linearly when the number of processors increasing.
文摘Intrusion detection is the investigation process of information about the system activities or its data to detect any malicious behavior or unauthorized activity.Most of the IDS implement K-means clustering technique due to its linear complexity and fast computing ability.Nonetheless,it is Naïve use of the mean data value for the cluster core that presents a major drawback.The chances of two circular clusters having different radius and centering at the same mean will occur.This condition cannot be addressed by the K-means algorithm because the mean value of the various clusters is very similar together.However,if the clusters are not spherical,it fails.To overcome this issue,a new integrated hybrid model by integrating expectation maximizing(EM)clustering using a Gaussian mixture model(GMM)and naïve Bays classifier have been proposed.In this model,GMM give more flexibility than K-Means in terms of cluster covariance.Also,they use probabilities function and soft clustering,that’s why they can have multiple cluster for a single data.In GMM,we can define the cluster form in GMM by two parameters:the mean and the standard deviation.This means that by using these two parameters,the cluster can take any kind of elliptical shape.EM-GMM will be used to cluster data based on data activity into the corresponding category.
基金This work was supported by the National Science Fund for Distinguished Young Scholars(62325104).
文摘The quality of synthetic aperture radar(SAR)image degrades in the case of multiple imaging projection planes(IPPs)and multiple overlapping ship targets,and then the performance of target classification and recognition can be influenced.For addressing this issue,a method for extracting ship targets with overlaps via the expectation maximization(EM)algorithm is pro-posed.First,the scatterers of ship targets are obtained via the target detection technique.Then,the EM algorithm is applied to extract the scatterers of a single ship target with a single IPP.Afterwards,a novel image amplitude estimation approach is pro-posed,with which the radar image of a single target with a sin-gle IPP can be generated.The proposed method can accom-plish IPP selection and targets separation in the image domain,which can improve the image quality and reserve the target information most possibly.Results of simulated and real mea-sured data demonstrate the effectiveness of the proposed method.
基金supported by National Natural Science Foundation of China(No.61806006)Priority Academic Program Development of Jiangsu Higher Education Institutions111 Project(No.B12018)。
文摘Classical survival analysis assumes all subjects will experience the event of interest, but in some cases, a portion of the population may never encounter the event. These survival methods further assume independent survival times, which is not valid for honey bees, which live in nests. The study introduces a semi-parametric marginal proportional hazards mixture cure (PHMC) model with exchangeable correlation structure, using generalized estimating equations for survival data analysis. The model was tested on clustered right-censored bees survival data with a cured fraction, where two bee species were subjected to different entomopathogens to test the effect of the entomopathogens on the survival of the bee species. The Expectation-Solution algorithm is used to estimate the parameters. The study notes a weak positive association between cure statuses (ρ1=0.0007) and survival times for uncured bees (ρ2=0.0890), emphasizing their importance. The odds of being uncured for A. mellifera is higher than the odds for species M. ferruginea. The bee species, A. mellifera are more susceptible to entomopathogens icipe 7, icipe 20, and icipe 69. The Cox-Snell residuals show that the proposed semiparametric PH model generally fits the data well as compared to model that assume independent correlation structure. Thus, the semi parametric marginal proportional hazards mixture cure is parsimonious model for correlated bees survival data.
文摘Compositional data, such as relative information, is a crucial aspect of machine learning and other related fields. It is typically recorded as closed data or sums to a constant, like 100%. The statistical linear model is the most used technique for identifying hidden relationships between underlying random variables of interest. However, data quality is a significant challenge in machine learning, especially when missing data is present. The linear regression model is a commonly used statistical modeling technique used in various applications to find relationships between variables of interest. When estimating linear regression parameters which are useful for things like future prediction and partial effects analysis of independent variables, maximum likelihood estimation (MLE) is the method of choice. However, many datasets contain missing observations, which can lead to costly and time-consuming data recovery. To address this issue, the expectation-maximization (EM) algorithm has been suggested as a solution for situations including missing data. The EM algorithm repeatedly finds the best estimates of parameters in statistical models that depend on variables or data that have not been observed. This is called maximum likelihood or maximum a posteriori (MAP). Using the present estimate as input, the expectation (E) step constructs a log-likelihood function. Finding the parameters that maximize the anticipated log-likelihood, as determined in the E step, is the job of the maximization (M) phase. This study looked at how well the EM algorithm worked on a made-up compositional dataset with missing observations. It used both the robust least square version and ordinary least square regression techniques. The efficacy of the EM algorithm was compared with two alternative imputation techniques, k-Nearest Neighbor (k-NN) and mean imputation (), in terms of Aitchison distances and covariance.
基金supported by the Natural Science Foundation of Guangxi(Grant No.2024GXNSFAA010476)the National Natural Science Foundation of China(Grant No.12361031)。
文摘In this article,we study strong limit theorems for weighted sums of extended negatively dependent random variables under the sub-linear expectations.We establish general strong law and complete convergence theorems for weighted sums of extended negatively dependent random variables under the sub-linear expectations.Our results of strong limit theorems are more general than some related results previously obtained by Thrum(1987),Li et al.(1995)and Wu(2010)in classical probability space.
文摘Clustering is an unsupervised learning method used to organize raw data in such a way that those with the same (similar) characteristics are found in the same class and those that are dissimilar are found in different classes. In this day and age, the very rapid increase in the amount of data being produced brings new challenges in the analysis and storage of this data. Recently, there is a growing interest in key areas such as real-time data mining, which reveal an urgent need to process very large data under strict performance constraints. The objective of this paper is to survey four algorithms including K-Means algorithm, FCM algorithm, EM algorithm and BIRCH, used for data clustering and then show their strengths and weaknesses. Another task is to compare the results obtained by applying each of these algorithms to the same data and to give a conclusion based on these results.
文摘针对当前广义频分复用(Generalized Frequency Division Multiplexing,GFDM)系统时变信道估计精度低的问题,提出基于稀疏贝叶斯学习的GFDM系统联合信道估计与符号检测算法.具体地,采用无干扰导频插入的GFDM多重响应信号模型,在稀疏贝叶斯学习框架下,结合期望最大化算法(Expectation-Maximization,EM)和卡尔曼滤波与平滑算法实现块时变信道的最大似然估计;基于信道状态信息的估计值进行GFDM符号检测,并通过信道估计与符号检测的迭代处理逐步提高信道估计与符号检测的精度.仿真结果表明,所提算法能够获得接近完美信道状态信息条件下的误码率性能,且具有收敛速度快、对多普勒频移鲁棒性高等优点.