The EM algorithm is a very popular maximum likelihood estimation method, the iterative algorithm for solving the maximum likelihood estimator when the observation data is the incomplete data, but also is very effectiv...The EM algorithm is a very popular maximum likelihood estimation method, the iterative algorithm for solving the maximum likelihood estimator when the observation data is the incomplete data, but also is very effective algorithm to estimate the finite mixture model parameters. However, EM algorithm can not guarantee to find the global optimal solution, and often easy to fall into local optimal solution, so it is sensitive to the determination of initial value to iteration. Traditional EM algorithm select the initial value at random, we propose an improved method of selection of initial value. First, we use the k-nearest-neighbor method to delete outliers. Second, use the k-means to initialize the EM algorithm. Compare this method with the original random initial value method, numerical experiments show that the parameter estimation effect of the initialization of the EM algorithm is significantly better than the effect of the original EM algorithm.展开更多
Since the joint probabilistic data association(JPDA)algorithm results in calculation explosion with the increasing number of targets,a multi-target tracking algorithm based on Gaussian mixture model(GMM)clustering is ...Since the joint probabilistic data association(JPDA)algorithm results in calculation explosion with the increasing number of targets,a multi-target tracking algorithm based on Gaussian mixture model(GMM)clustering is proposed.The algorithm is used to cluster the measurements,and the association matrix between measurements and tracks is constructed by the posterior probability.Compared with the traditional data association algorithm,this algorithm has better tracking performance and less computational complexity.Simulation results demonstrate the effectiveness of the proposed algorithm.展开更多
An improved Gaussian mixture model (GMM)- based clustering method is proposed for the difficult case where the true distribution of data is against the assumed GMM. First, an improved model selection criterion, the ...An improved Gaussian mixture model (GMM)- based clustering method is proposed for the difficult case where the true distribution of data is against the assumed GMM. First, an improved model selection criterion, the completed likelihood minimum message length criterion, is derived. It can measure both the goodness-of-fit of the candidate GMM to the data and the goodness-of-partition of the data. Secondly, by utilizing the proposed criterion as the clustering objective function, an improved expectation- maximization (EM) algorithm is developed, which can avoid poor local optimal solutions compared to the standard EM algorithm for estimating the model parameters. The experimental results demonstrate that the proposed method can rectify the over-fitting tendency of representative GMM-based clustering approaches and can robustly provide more accurate clustering results.展开更多
To solve the problem of color distortion after dehazing in the sky region by using the classical dark channel prior method to process the hazy images with large regions of sky,an improved dark channel image dehazing m...To solve the problem of color distortion after dehazing in the sky region by using the classical dark channel prior method to process the hazy images with large regions of sky,an improved dark channel image dehazing method based on Gaussian mixture model is proposed.Firstly,we use the Gaussian mixture model to model the hazy image,and then use the expectation maximization(EM)algorithm to optimize the parameters,so that the hazy image can be divided into the sky region and the non-sky region.Secondly,the sky region is divided into a light haze region,a medium haze region and a heavy haze region according to the different dark channel values to estimate the transmission respectively.Thirdly,the restored image is obtained by combining the atmospheric scattering model.Finally,adaptive local tone mapping for high dynamic range images is used to adjust the brightness of the restored image.The experimental results show that the proposed method can effectively eliminate the color distortion in the sky region,and the restored image is clearer and has better visual effect.展开更多
An efficient approach was proposed for discriminating shadows from moving objects. In the background subtraction stage, moving objects were extracted. Then, the initial classification for moving shadow pixels and fore...An efficient approach was proposed for discriminating shadows from moving objects. In the background subtraction stage, moving objects were extracted. Then, the initial classification for moving shadow pixels and foreground object pixels was performed by using color invariant features. In the shadow model learning stage, instead of a single Gaussian distribution, it was assumed that the density function computed on the values of chromaticity difference or bright difference, can be modeled as a mixture of Gaussian consisting of two density functions. Meanwhile, the Gaussian parameter estimation was performed by using EM algorithm. The estimates were used to obtain shadow mask according to two constraints. Finally, experiments were carried out. The visual experiment results confirm the effectiveness of proposed method. Quantitative results in terms of the shadow detection rate and the shadow discrimination rate(the maximum values are 85.79% and 97.56%, respectively) show that the proposed approach achieves a satisfying result with post-processing step.展开更多
In this paper, a Gaussian mixture model (GMM) based classifier is described to tell whether precipitation events will happen on a certain day at a certain time from historical meteorological data. The classifier deals...In this paper, a Gaussian mixture model (GMM) based classifier is described to tell whether precipitation events will happen on a certain day at a certain time from historical meteorological data. The classifier deals with a two-class classification problem where one class represents precipitation events and the other represents non-precipitation events. The concept of ambiguity is introduced to represent cases where weather conditions between the two classes like drizzles, intermittent or overcast are more likely to happen. Six groups of experiments are carried out to evaluate the performance of the classifier using different configurations based on the observation data released by Shanghai Baoshan weather station. Specifically, a typical classification performance of about 75% accuracy, 30% precision and 80% recall is achieved for prediction tasks with a time span of 12 hours.展开更多
Although there are many papers on variable selection methods based on mean model in the nite mixture of regression models,little work has been done on how to select signi cant explanatory variables in the modeling of ...Although there are many papers on variable selection methods based on mean model in the nite mixture of regression models,little work has been done on how to select signi cant explanatory variables in the modeling of the variance parameter.In this paper,we propose and study a novel class of models:a skew-normal mixture of joint location and scale models to analyze the heteroscedastic skew-normal data coming from a heterogeneous population.The problem of variable selection for the proposed models is considered.In particular,a modi ed Expectation-Maximization(EM)algorithm for estimating the model parameters is developed.The consistency and the oracle property of the penalized estimators is established.Simulation studies are conducted to investigate the nite sample performance of the proposed methodolo-gies.An example is illustrated by the proposed methodologies.展开更多
In this paper, we propose a robust mixture regression model based on the skew scale mixtures of normal distributions (RMR-SSMN) which can accommodate asymmetric, heavy-tailed and contaminated data better. For the vari...In this paper, we propose a robust mixture regression model based on the skew scale mixtures of normal distributions (RMR-SSMN) which can accommodate asymmetric, heavy-tailed and contaminated data better. For the variable selection problem, the penalized likelihood approach with a new combined penalty function which balances the SCAD and l<sub>2</sub> penalty is proposed. The adjusted EM algorithm is presented to get parameter estimates of RMR-SSMN models at a faster convergence rate. As simulations show, our mixture models are more robust than general FMR models and the new combined penalty function outperforms SCAD for variable selection. Finally, the proposed methodology and algorithm are applied to a real data set and achieve reasonable results.展开更多
<p> <span style="color:#000000;"><span style="color:#000000;">Normal Variance-Mean Mixture (NVMM) provide</span></span><span style="color:#000000;"><...<p> <span style="color:#000000;"><span style="color:#000000;">Normal Variance-Mean Mixture (NVMM) provide</span></span><span style="color:#000000;"><span style="color:#000000;"><span style="color:#000000;">s</span></span></span><span><span><span><span style="color:#000000;"> a general framework for deriving models with desirable properties for modelling financial market variables such as exchange rates, equity prices, and interest rates measured over short time intervals, </span><i><span style="color:#000000;">i.e.</span></i><span style="color:#000000;"> daily or weekly. Such data sets are characterized by non-normality and are usually skewed, fat-tailed and exhibit excess kurtosis. </span><span style="color:#000000;">The Generalised Hyperbolic distribution (GHD) introduced by Barndorff-</span><span style="color:#000000;">Nielsen </span></span></span></span><span style="color:#000000;"><span style="color:#000000;"><span style="color:#000000;">(1977)</span></span></span><span><span><span><span style="color:#000000;"> which act as Normal variance-mean mixtures with Generalised Inverse Gaussian (GIG) mixing distribution nest a number of special and limiting case distributions. The Normal Inverse Gaussian (NIG) distribution is obtained when the Inverse Gaussian is the mixing distribution, </span><i><span style="color:#000000;">i.e</span></i></span></span></span><span style="color:#000000;"><span style="color:#000000;"><i><span style="color:#000000;">.</span></i></span></span><span><span><span><span style="color:#000000;">, the index parameter of the GIG is</span><span style="color:red;"> <img src="Edit_721a4317-7ef5-4796-9713-b9057bc426fc.bmp" alt="" /></span><span style="color:#000000;">. The NIG is very popular because of its analytical tractability. In the mixing mechanism</span></span></span></span><span style="color:#000000;"><span style="color:#000000;"><span style="color:#000000;">,</span></span></span><span><span><span><span><span style="color:#000000;"> the mixing distribution characterizes the prior information of the random variable of the conditional distribution. Therefore, considering finite mixture models is one way of extending the work. The GIG is a three parameter distribution denoted by </span><img src="Edit_d21f2e1e-d426-401e-bf8b-f56d268dddb6.bmp" alt="" /></span><span><span style="color:#000000;"> and nest several special and limiting cases. When </span><img src="Edit_ffee9824-2b75-4ea6-a3d2-e048d49b553f.bmp" alt="" /></span><span><span style="color:#000000;">, we have </span><img src="Edit_654ea565-9798-4435-9a59-a0a1a7c282df.bmp" alt="" /></span><span style="color:#000000;"> which is called an Inverse Gaussian (IG) distribution. </span><span><span><span style="color:#000000;">When </span><img src="Edit_b15daf3d-849f-440a-9e4f-7b0c78d519e5.bmp" alt="" /></span><span style="color:red;"><span style="color:#000000;">, </span><img src="Edit_08a2088c-f57e-401c-8fb9-9974eec5947a.bmp" alt="" /><span style="color:#000000;">, </span><img src="Edit_130f4d7c-3e27-4937-b60f-6bf6e41f1f52.bmp" alt="" /><span style="color:#000000;">,</span></span><span><span style="color:#000000;"> we have </span><img src="Edit_215e67cb-b0d9-44e1-88d1-a2598dea05af.bmp" alt="" /></span><span style="color:red;"><span style="color:#000000;">, </span><img src="Edit_6bf9602b-a9c9-4a9d-aed0-049c47fe8dfe.bmp" alt="" /></span></span><span style="color:red;"><span style="color:#000000;"> </span><span><span style="color:#000000;">and </span><img src="Edit_d642ba7f-8b63-4830-aea1-d6e5fba31cc8.bmp" alt="" /></span></span><span><span style="color:#000000;"> distributions respectively. These distributions are related to </span><img src="Edit_0ca6658e-54cb-4d4d-87fa-25eb3a0a8934.bmp" alt="" /></span><span style="color:#000000;"> and are called weighted inverse Gaussian distributions. In this</span> <span style="color:#000000;">work</span></span></span></span><span style="color:#000000;"><span style="color:#000000;"><span style="color:#000000;">,</span></span></span><span><span><span><span style="color:#000000;"> we consider a finite mixture of </span><img src="Edit_30ee74b7-0bfc-413d-b4d6-43902ec6c69d.bmp" alt="" /></span></span></span><span><span><span><span><span style="color:#000000;"> and </span><img src="Edit_ba62dff8-eb11-48f9-8388-68f5ee954c00.bmp" alt="" /></span></span></span></span><span style="color:#000000;"><span style="color:#000000;"><span style="color:#000000;"> and show that the mixture is also a weighted Inverse Gaussian distribution and use it to construct a NVMM. Due to the complexity of the likelihood, direct maximization is difficult. An EM type algorithm is provided for the Maximum Likelihood estimation of the parameters of the proposed model. We adopt an iterative scheme which is not based on explicit solution to the normal equations. This subtle approach reduces the computational difficulty of solving the complicated quantities involved directly to designing an iterative scheme based on a representation of the normal equation. The algorithm is easily programmable and we obtained a monotonic convergence for the data sets used.</span></span></span> </p>展开更多
Mixture models have become more popular in modelling compared to standard distributions. The mixing distributions play a role in capturing the variability of the random variable in the conditional distribution. Studie...Mixture models have become more popular in modelling compared to standard distributions. The mixing distributions play a role in capturing the variability of the random variable in the conditional distribution. Studies have lately focused on finite mixture models as mixing distributions in the mixing mechanism. In the present work, we consider a Normal Variance Mean mix<span>ture model. The mixing distribution is a finite mixture of two special cases of</span><span> Generalised Inverse Gaussian distribution with indexes <span style="white-space:nowrap;">-1/2 and -3/2</span>. The </span><span>parameters of the mixed model are obtained via the Expectation-Maximization</span><span> (EM) algorithm. The iterative scheme is based on a presentation of the normal equations. An application to some financial data has been done.展开更多
文摘The EM algorithm is a very popular maximum likelihood estimation method, the iterative algorithm for solving the maximum likelihood estimator when the observation data is the incomplete data, but also is very effective algorithm to estimate the finite mixture model parameters. However, EM algorithm can not guarantee to find the global optimal solution, and often easy to fall into local optimal solution, so it is sensitive to the determination of initial value to iteration. Traditional EM algorithm select the initial value at random, we propose an improved method of selection of initial value. First, we use the k-nearest-neighbor method to delete outliers. Second, use the k-means to initialize the EM algorithm. Compare this method with the original random initial value method, numerical experiments show that the parameter estimation effect of the initialization of the EM algorithm is significantly better than the effect of the original EM algorithm.
基金the National Natural Science Foundation of China(61771367)the Science and Technology on Communication Networks Laboratory(HHS19641X003).
文摘Since the joint probabilistic data association(JPDA)algorithm results in calculation explosion with the increasing number of targets,a multi-target tracking algorithm based on Gaussian mixture model(GMM)clustering is proposed.The algorithm is used to cluster the measurements,and the association matrix between measurements and tracks is constructed by the posterior probability.Compared with the traditional data association algorithm,this algorithm has better tracking performance and less computational complexity.Simulation results demonstrate the effectiveness of the proposed algorithm.
基金The National Natural Science Foundation of China(No.61105048,60972165)the Doctoral Fund of Ministry of Education of China(No.20110092120034)+2 种基金the Natural Science Foundation of Jiangsu Province(No.BK2010240)the Technology Foundation for Selected Overseas Chinese Scholar,Ministry of Human Resources and Social Security of China(No.6722000008)the Open Fund of Jiangsu Province Key Laboratory for Remote Measuring and Control(No.YCCK201005)
文摘An improved Gaussian mixture model (GMM)- based clustering method is proposed for the difficult case where the true distribution of data is against the assumed GMM. First, an improved model selection criterion, the completed likelihood minimum message length criterion, is derived. It can measure both the goodness-of-fit of the candidate GMM to the data and the goodness-of-partition of the data. Secondly, by utilizing the proposed criterion as the clustering objective function, an improved expectation- maximization (EM) algorithm is developed, which can avoid poor local optimal solutions compared to the standard EM algorithm for estimating the model parameters. The experimental results demonstrate that the proposed method can rectify the over-fitting tendency of representative GMM-based clustering approaches and can robustly provide more accurate clustering results.
基金National Natural Science Foundation of China(Nos.61841303,61963023)Project of Humanities and Social Sciences of Ministry of Education in China(No.19YJC760012)。
文摘To solve the problem of color distortion after dehazing in the sky region by using the classical dark channel prior method to process the hazy images with large regions of sky,an improved dark channel image dehazing method based on Gaussian mixture model is proposed.Firstly,we use the Gaussian mixture model to model the hazy image,and then use the expectation maximization(EM)algorithm to optimize the parameters,so that the hazy image can be divided into the sky region and the non-sky region.Secondly,the sky region is divided into a light haze region,a medium haze region and a heavy haze region according to the different dark channel values to estimate the transmission respectively.Thirdly,the restored image is obtained by combining the atmospheric scattering model.Finally,adaptive local tone mapping for high dynamic range images is used to adjust the brightness of the restored image.The experimental results show that the proposed method can effectively eliminate the color distortion in the sky region,and the restored image is clearer and has better visual effect.
基金Project(50805023)supported by the National Natural Science Foundation of ChinaProject(BA2010093)supported by the Special Fund of Jiangsu Province for the Transformation of Scientific and Technological Achievements,ChinaProject(2008144)supported by the Hexa-type Elites Peak Program of Jiangsu Province,China
文摘An efficient approach was proposed for discriminating shadows from moving objects. In the background subtraction stage, moving objects were extracted. Then, the initial classification for moving shadow pixels and foreground object pixels was performed by using color invariant features. In the shadow model learning stage, instead of a single Gaussian distribution, it was assumed that the density function computed on the values of chromaticity difference or bright difference, can be modeled as a mixture of Gaussian consisting of two density functions. Meanwhile, the Gaussian parameter estimation was performed by using EM algorithm. The estimates were used to obtain shadow mask according to two constraints. Finally, experiments were carried out. The visual experiment results confirm the effectiveness of proposed method. Quantitative results in terms of the shadow detection rate and the shadow discrimination rate(the maximum values are 85.79% and 97.56%, respectively) show that the proposed approach achieves a satisfying result with post-processing step.
文摘In this paper, a Gaussian mixture model (GMM) based classifier is described to tell whether precipitation events will happen on a certain day at a certain time from historical meteorological data. The classifier deals with a two-class classification problem where one class represents precipitation events and the other represents non-precipitation events. The concept of ambiguity is introduced to represent cases where weather conditions between the two classes like drizzles, intermittent or overcast are more likely to happen. Six groups of experiments are carried out to evaluate the performance of the classifier using different configurations based on the observation data released by Shanghai Baoshan weather station. Specifically, a typical classification performance of about 75% accuracy, 30% precision and 80% recall is achieved for prediction tasks with a time span of 12 hours.
基金Supported by the National Natural Science Foundation of China(11861041).
文摘Although there are many papers on variable selection methods based on mean model in the nite mixture of regression models,little work has been done on how to select signi cant explanatory variables in the modeling of the variance parameter.In this paper,we propose and study a novel class of models:a skew-normal mixture of joint location and scale models to analyze the heteroscedastic skew-normal data coming from a heterogeneous population.The problem of variable selection for the proposed models is considered.In particular,a modi ed Expectation-Maximization(EM)algorithm for estimating the model parameters is developed.The consistency and the oracle property of the penalized estimators is established.Simulation studies are conducted to investigate the nite sample performance of the proposed methodolo-gies.An example is illustrated by the proposed methodologies.
文摘In this paper, we propose a robust mixture regression model based on the skew scale mixtures of normal distributions (RMR-SSMN) which can accommodate asymmetric, heavy-tailed and contaminated data better. For the variable selection problem, the penalized likelihood approach with a new combined penalty function which balances the SCAD and l<sub>2</sub> penalty is proposed. The adjusted EM algorithm is presented to get parameter estimates of RMR-SSMN models at a faster convergence rate. As simulations show, our mixture models are more robust than general FMR models and the new combined penalty function outperforms SCAD for variable selection. Finally, the proposed methodology and algorithm are applied to a real data set and achieve reasonable results.
文摘<p> <span style="color:#000000;"><span style="color:#000000;">Normal Variance-Mean Mixture (NVMM) provide</span></span><span style="color:#000000;"><span style="color:#000000;"><span style="color:#000000;">s</span></span></span><span><span><span><span style="color:#000000;"> a general framework for deriving models with desirable properties for modelling financial market variables such as exchange rates, equity prices, and interest rates measured over short time intervals, </span><i><span style="color:#000000;">i.e.</span></i><span style="color:#000000;"> daily or weekly. Such data sets are characterized by non-normality and are usually skewed, fat-tailed and exhibit excess kurtosis. </span><span style="color:#000000;">The Generalised Hyperbolic distribution (GHD) introduced by Barndorff-</span><span style="color:#000000;">Nielsen </span></span></span></span><span style="color:#000000;"><span style="color:#000000;"><span style="color:#000000;">(1977)</span></span></span><span><span><span><span style="color:#000000;"> which act as Normal variance-mean mixtures with Generalised Inverse Gaussian (GIG) mixing distribution nest a number of special and limiting case distributions. The Normal Inverse Gaussian (NIG) distribution is obtained when the Inverse Gaussian is the mixing distribution, </span><i><span style="color:#000000;">i.e</span></i></span></span></span><span style="color:#000000;"><span style="color:#000000;"><i><span style="color:#000000;">.</span></i></span></span><span><span><span><span style="color:#000000;">, the index parameter of the GIG is</span><span style="color:red;"> <img src="Edit_721a4317-7ef5-4796-9713-b9057bc426fc.bmp" alt="" /></span><span style="color:#000000;">. The NIG is very popular because of its analytical tractability. In the mixing mechanism</span></span></span></span><span style="color:#000000;"><span style="color:#000000;"><span style="color:#000000;">,</span></span></span><span><span><span><span><span style="color:#000000;"> the mixing distribution characterizes the prior information of the random variable of the conditional distribution. Therefore, considering finite mixture models is one way of extending the work. The GIG is a three parameter distribution denoted by </span><img src="Edit_d21f2e1e-d426-401e-bf8b-f56d268dddb6.bmp" alt="" /></span><span><span style="color:#000000;"> and nest several special and limiting cases. When </span><img src="Edit_ffee9824-2b75-4ea6-a3d2-e048d49b553f.bmp" alt="" /></span><span><span style="color:#000000;">, we have </span><img src="Edit_654ea565-9798-4435-9a59-a0a1a7c282df.bmp" alt="" /></span><span style="color:#000000;"> which is called an Inverse Gaussian (IG) distribution. </span><span><span><span style="color:#000000;">When </span><img src="Edit_b15daf3d-849f-440a-9e4f-7b0c78d519e5.bmp" alt="" /></span><span style="color:red;"><span style="color:#000000;">, </span><img src="Edit_08a2088c-f57e-401c-8fb9-9974eec5947a.bmp" alt="" /><span style="color:#000000;">, </span><img src="Edit_130f4d7c-3e27-4937-b60f-6bf6e41f1f52.bmp" alt="" /><span style="color:#000000;">,</span></span><span><span style="color:#000000;"> we have </span><img src="Edit_215e67cb-b0d9-44e1-88d1-a2598dea05af.bmp" alt="" /></span><span style="color:red;"><span style="color:#000000;">, </span><img src="Edit_6bf9602b-a9c9-4a9d-aed0-049c47fe8dfe.bmp" alt="" /></span></span><span style="color:red;"><span style="color:#000000;"> </span><span><span style="color:#000000;">and </span><img src="Edit_d642ba7f-8b63-4830-aea1-d6e5fba31cc8.bmp" alt="" /></span></span><span><span style="color:#000000;"> distributions respectively. These distributions are related to </span><img src="Edit_0ca6658e-54cb-4d4d-87fa-25eb3a0a8934.bmp" alt="" /></span><span style="color:#000000;"> and are called weighted inverse Gaussian distributions. In this</span> <span style="color:#000000;">work</span></span></span></span><span style="color:#000000;"><span style="color:#000000;"><span style="color:#000000;">,</span></span></span><span><span><span><span style="color:#000000;"> we consider a finite mixture of </span><img src="Edit_30ee74b7-0bfc-413d-b4d6-43902ec6c69d.bmp" alt="" /></span></span></span><span><span><span><span><span style="color:#000000;"> and </span><img src="Edit_ba62dff8-eb11-48f9-8388-68f5ee954c00.bmp" alt="" /></span></span></span></span><span style="color:#000000;"><span style="color:#000000;"><span style="color:#000000;"> and show that the mixture is also a weighted Inverse Gaussian distribution and use it to construct a NVMM. Due to the complexity of the likelihood, direct maximization is difficult. An EM type algorithm is provided for the Maximum Likelihood estimation of the parameters of the proposed model. We adopt an iterative scheme which is not based on explicit solution to the normal equations. This subtle approach reduces the computational difficulty of solving the complicated quantities involved directly to designing an iterative scheme based on a representation of the normal equation. The algorithm is easily programmable and we obtained a monotonic convergence for the data sets used.</span></span></span> </p>
文摘Mixture models have become more popular in modelling compared to standard distributions. The mixing distributions play a role in capturing the variability of the random variable in the conditional distribution. Studies have lately focused on finite mixture models as mixing distributions in the mixing mechanism. In the present work, we consider a Normal Variance Mean mix<span>ture model. The mixing distribution is a finite mixture of two special cases of</span><span> Generalised Inverse Gaussian distribution with indexes <span style="white-space:nowrap;">-1/2 and -3/2</span>. The </span><span>parameters of the mixed model are obtained via the Expectation-Maximization</span><span> (EM) algorithm. The iterative scheme is based on a presentation of the normal equations. An application to some financial data has been done.