The validity measurement of fuzzy clustering is a key problem. If clustering is formed, it needs a kind of machine to verify its validity. To make mining more accountable, comprehensible and with a usable spatial patt...The validity measurement of fuzzy clustering is a key problem. If clustering is formed, it needs a kind of machine to verify its validity. To make mining more accountable, comprehensible and with a usable spatial pattern, it is necessary to first detect whether the data set has a clustered structure or not before clustering. This paper discusses a detection method for clustered patterns and a fuzzy clustering algorithm, and studies the validity function of the result produced by fuzzy clustering based on two aspects, which reflect the un-certainty of classification during fuzzy partition and spatial location features of spatial data, and proposes a new validity function of fuzzy clustering for spatial data. The experimental result indicates that the new validity function can accurately measure the validity of the results of fuzzy clustering. Especially, for the result of fuzzy clustering of spatial data, it is robust and its classification result is better when compared to other indices.展开更多
Evolution of the Arctic sea ice and its snow cover during the SHEBA year were simulated by applying a high-resolution thermodynamic snow/ice model (HIGHTSI). Attention was paid to the impact of albedo on snow and se...Evolution of the Arctic sea ice and its snow cover during the SHEBA year were simulated by applying a high-resolution thermodynamic snow/ice model (HIGHTSI). Attention was paid to the impact of albedo on snow and sea ice mass balance, effect of snow on total ice mass balance, and the model vertical resolution. The SHEBA annual simulation was made applying the best possible external forcing data set created by the Sea Ice Model Intercomparison Project. The HIGHTSI control run reasonably reproduced the observed snow and ice thickness. A number of albedo schemes were incorporated into HIGHTSI to study the feedback processes between the albedo and snow and ice thickness. The snow thickness turned out to be an essential variable in the albedo parameterization. Albedo schemes dependent on the surface temperature were liable to excessive positive feedback effects generated by errors in the modelled surface temperature. The superimposed ice formation should be taken into account for the annual Arctic sea ice mass balance.展开更多
In this article we study the estimation method of nonparametric regression measurement error model based on a validation data. The estimation procedures are based on orthogonal series estimation and truncated series a...In this article we study the estimation method of nonparametric regression measurement error model based on a validation data. The estimation procedures are based on orthogonal series estimation and truncated series approximation methods without specifying any structure equation and the distribution assumption. The convergence rates of the proposed estimator are derived. By example and through simulation, the method is robust against the misspecification of a measurement error model.展开更多
In this article, we develop estimation approaches for nonparametric multiple regression measurement error models when both independent validation data on covariables and primary data on the response variable and surro...In this article, we develop estimation approaches for nonparametric multiple regression measurement error models when both independent validation data on covariables and primary data on the response variable and surrogate covariables are available. An estimator which integrates Fourier series estimation and truncated series approximation methods is derived without any error model structure assumption between the true covariables and surrogate variables. Most importantly, our proposed methodology can be readily extended to the case that only some of covariates are measured with errors with the assistance of validation data. Under mild conditions, we derive the convergence rates of the proposed estimators. The finite-sample properties of the estimators are investigated through simulation studies.展开更多
The basis of accurate mineral resource estimates is to have a geological model which replicates the nature and style of the orebody. Key inputs into the generation of a good geological model are the sample data and ma...The basis of accurate mineral resource estimates is to have a geological model which replicates the nature and style of the orebody. Key inputs into the generation of a good geological model are the sample data and mapping information. The Obuasi Mine sample data with a lot of legacy issues were subjected to a robust validation process and integrated with mapping information to generate an accurate geological orebody model for mineral resource estimation in Block 8 Lower. Validation of the sample data focused on replacing missing collar coordinates, missing assays, and correcting magnetic declination that was used to convert the downhole surveys from true to magnetic, fix missing lithology and finally assign confidence numbers to all the sample data. The missing coordinates which were replaced ensured that the sample data plotted at their correct location in space as intended from the planning stage. Magnetic declination data, which was maintained constant throughout all the years even though it changes every year, was also corrected in the validation project. The corrected magnetic declination ensured that the drillholes were plotted on their accurate trajectory as per the planned azimuth and also reflected the true position of the intercepted mineralized fissure(s) which was previously not the case and marked a major blot in the modelling of the Obuasi orebody. The incorporation of mapped data with the validated sample data in the wireframes resulted in a better interpretation of the orebody. The updated mineral resource generated by domaining quartz from the sulphides and compared with the old resource showed that the sulphide tonnes in the old resource estimates were overestimated by 1% and the grade overestimated by 8.5%.展开更多
Two statistical validation methods were used to evaluate the confidence level of the Total Column Ozone (TCO) measurements recorded by satellite systems measuring simultaneously, one using the normal distribution and ...Two statistical validation methods were used to evaluate the confidence level of the Total Column Ozone (TCO) measurements recorded by satellite systems measuring simultaneously, one using the normal distribution and another using the Mann-Whitney test. First, the reliability of the TCO measurements was studied hemispherically. While similar coincidences and levels of significance > 0.05 were found with the two statistical tests, an enormous variability in the levels of significance throughout the year was also exposed. Then, using the same statistical comparison methods, a latitudinal study was carried out in order to elucidate the geographical distribution that gave rise to this variability. Our study reveals that between the TOMS and OMI measurements in 2005 there was only a coincidence in 50% of the latitudes, which explained the variability. This implies that for 2005, the TOMS measurements are not completely reliable, except between the -50° and -15° latitude band in the southern hemisphere and between +15° and +50° latitude band in the northern hemisphere. In the case of OMI-OMPS, we observe that between 2011 and 2016 the measurements of both satellite systems are reasonably similar with a confidence level higher than 95%. However, in 2017 a band with a width of 20° latitude centered on the equator appeared, in which the significance levels were much less than 0.05, indicating that one of the measurement systems had begun to fail. In 2018, the fault was not only located in the equator, but was also replicated in various bands in the Southern Hemisphere. We interpret this as evidence of irreversible failure in one of the measurement systems.展开更多
近年来,业务流程的精确管理受到越来越多的关注,作为业务流程管理内容之一的一致性检查技术正变得越来越重要。现有的一致性检查技术主要从模型的控制流角度出发,并未考虑业务流程中的数据或数据的变化对业务流程产生的影响,为此提出了...近年来,业务流程的精确管理受到越来越多的关注,作为业务流程管理内容之一的一致性检查技术正变得越来越重要。现有的一致性检查技术主要从模型的控制流角度出发,并未考虑业务流程中的数据或数据的变化对业务流程产生的影响,为此提出了一种基于数据影响的业务流程一致性检查方法。首先,通过业务流程建模符号(business process modeling notation,BPMN)模型中数据和行为之间的依赖关系来分析数据对偏差活动预期行为的影响,进而获取偏差活动的预期行为集;其次,通过在BPMN模型中引入决策模型和符号(decision model and notation,DMN)决策表来充分捕获当前实例执行的数据上下文与行为上下文之间的关系,以区分有效数据更改和异常数据更改,找到偏差活动的有效预期行为集。最后,通过设计数据对偏差活动影响的各类成本函数提出了有效一致性检查方法。实验结果表明,相比已有工作,该方法在进行业务流程一致性检查时提高了业务流程的一致性,可以成功捕获偏差活动作出反应的适应行为,使得业务流程在复杂多变的环境下表现得更加准确、合理。展开更多
校核、验证与确认(verification,validation,and accreditation,VV&A)是保证仿真模型可信度的关键手段,其中模型验证是核心环节。针对导弹飞行仿真模型结果验证存在的参考数据不可获得、参考数据来源多样、专家验证主观性强等问题,...校核、验证与确认(verification,validation,and accreditation,VV&A)是保证仿真模型可信度的关键手段,其中模型验证是核心环节。针对导弹飞行仿真模型结果验证存在的参考数据不可获得、参考数据来源多样、专家验证主观性强等问题,提出一种基于时间序列分段特征提取的导弹飞行仿真模型结果验证方法。提出了一种综合的时间序列分段线性方法,由基于二阶导数提取趋势边缘点的线性分段算法和基于极值点优化的Top-Down线性分段算法两部分组成,以实现对导弹飞行仿真数据和参考数据进行有效的线性分段表示。基于上述分段结果,对各段时间序列的均值、方差、斜率等特征进行提取,以辅助专家进行验证,从而降低验证中的主观性;或者直接利用TIC系数法、动态时间规整(dynamic time warping,DTW)等方法进行客观的相似性分析。通过充分利用时间序列的分段特征,可实现各种情形下的导弹飞行仿真模型结果验证。通过一个导弹模型结果验证案例演示了所提方法的可行性和有效性。展开更多
文摘The validity measurement of fuzzy clustering is a key problem. If clustering is formed, it needs a kind of machine to verify its validity. To make mining more accountable, comprehensible and with a usable spatial pattern, it is necessary to first detect whether the data set has a clustered structure or not before clustering. This paper discusses a detection method for clustered patterns and a fuzzy clustering algorithm, and studies the validity function of the result produced by fuzzy clustering based on two aspects, which reflect the un-certainty of classification during fuzzy partition and spatial location features of spatial data, and proposes a new validity function of fuzzy clustering for spatial data. The experimental result indicates that the new validity function can accurately measure the validity of the results of fuzzy clustering. Especially, for the result of fuzzy clustering of spatial data, it is robust and its classification result is better when compared to other indices.
基金supported by the EC-funded project DAMOCLES (grant 18509)which is part of the Sixth Framework Program of DFG(grant LU 818/1-1)Natural Science Foundation of China(grants No.40233032,40376006).
文摘Evolution of the Arctic sea ice and its snow cover during the SHEBA year were simulated by applying a high-resolution thermodynamic snow/ice model (HIGHTSI). Attention was paid to the impact of albedo on snow and sea ice mass balance, effect of snow on total ice mass balance, and the model vertical resolution. The SHEBA annual simulation was made applying the best possible external forcing data set created by the Sea Ice Model Intercomparison Project. The HIGHTSI control run reasonably reproduced the observed snow and ice thickness. A number of albedo schemes were incorporated into HIGHTSI to study the feedback processes between the albedo and snow and ice thickness. The snow thickness turned out to be an essential variable in the albedo parameterization. Albedo schemes dependent on the surface temperature were liable to excessive positive feedback effects generated by errors in the modelled surface temperature. The superimposed ice formation should be taken into account for the annual Arctic sea ice mass balance.
文摘In this article we study the estimation method of nonparametric regression measurement error model based on a validation data. The estimation procedures are based on orthogonal series estimation and truncated series approximation methods without specifying any structure equation and the distribution assumption. The convergence rates of the proposed estimator are derived. By example and through simulation, the method is robust against the misspecification of a measurement error model.
文摘In this article, we develop estimation approaches for nonparametric multiple regression measurement error models when both independent validation data on covariables and primary data on the response variable and surrogate covariables are available. An estimator which integrates Fourier series estimation and truncated series approximation methods is derived without any error model structure assumption between the true covariables and surrogate variables. Most importantly, our proposed methodology can be readily extended to the case that only some of covariates are measured with errors with the assistance of validation data. Under mild conditions, we derive the convergence rates of the proposed estimators. The finite-sample properties of the estimators are investigated through simulation studies.
文摘The basis of accurate mineral resource estimates is to have a geological model which replicates the nature and style of the orebody. Key inputs into the generation of a good geological model are the sample data and mapping information. The Obuasi Mine sample data with a lot of legacy issues were subjected to a robust validation process and integrated with mapping information to generate an accurate geological orebody model for mineral resource estimation in Block 8 Lower. Validation of the sample data focused on replacing missing collar coordinates, missing assays, and correcting magnetic declination that was used to convert the downhole surveys from true to magnetic, fix missing lithology and finally assign confidence numbers to all the sample data. The missing coordinates which were replaced ensured that the sample data plotted at their correct location in space as intended from the planning stage. Magnetic declination data, which was maintained constant throughout all the years even though it changes every year, was also corrected in the validation project. The corrected magnetic declination ensured that the drillholes were plotted on their accurate trajectory as per the planned azimuth and also reflected the true position of the intercepted mineralized fissure(s) which was previously not the case and marked a major blot in the modelling of the Obuasi orebody. The incorporation of mapped data with the validated sample data in the wireframes resulted in a better interpretation of the orebody. The updated mineral resource generated by domaining quartz from the sulphides and compared with the old resource showed that the sulphide tonnes in the old resource estimates were overestimated by 1% and the grade overestimated by 8.5%.
文摘Two statistical validation methods were used to evaluate the confidence level of the Total Column Ozone (TCO) measurements recorded by satellite systems measuring simultaneously, one using the normal distribution and another using the Mann-Whitney test. First, the reliability of the TCO measurements was studied hemispherically. While similar coincidences and levels of significance > 0.05 were found with the two statistical tests, an enormous variability in the levels of significance throughout the year was also exposed. Then, using the same statistical comparison methods, a latitudinal study was carried out in order to elucidate the geographical distribution that gave rise to this variability. Our study reveals that between the TOMS and OMI measurements in 2005 there was only a coincidence in 50% of the latitudes, which explained the variability. This implies that for 2005, the TOMS measurements are not completely reliable, except between the -50° and -15° latitude band in the southern hemisphere and between +15° and +50° latitude band in the northern hemisphere. In the case of OMI-OMPS, we observe that between 2011 and 2016 the measurements of both satellite systems are reasonably similar with a confidence level higher than 95%. However, in 2017 a band with a width of 20° latitude centered on the equator appeared, in which the significance levels were much less than 0.05, indicating that one of the measurement systems had begun to fail. In 2018, the fault was not only located in the equator, but was also replicated in various bands in the Southern Hemisphere. We interpret this as evidence of irreversible failure in one of the measurement systems.
文摘近年来,业务流程的精确管理受到越来越多的关注,作为业务流程管理内容之一的一致性检查技术正变得越来越重要。现有的一致性检查技术主要从模型的控制流角度出发,并未考虑业务流程中的数据或数据的变化对业务流程产生的影响,为此提出了一种基于数据影响的业务流程一致性检查方法。首先,通过业务流程建模符号(business process modeling notation,BPMN)模型中数据和行为之间的依赖关系来分析数据对偏差活动预期行为的影响,进而获取偏差活动的预期行为集;其次,通过在BPMN模型中引入决策模型和符号(decision model and notation,DMN)决策表来充分捕获当前实例执行的数据上下文与行为上下文之间的关系,以区分有效数据更改和异常数据更改,找到偏差活动的有效预期行为集。最后,通过设计数据对偏差活动影响的各类成本函数提出了有效一致性检查方法。实验结果表明,相比已有工作,该方法在进行业务流程一致性检查时提高了业务流程的一致性,可以成功捕获偏差活动作出反应的适应行为,使得业务流程在复杂多变的环境下表现得更加准确、合理。
文摘校核、验证与确认(verification,validation,and accreditation,VV&A)是保证仿真模型可信度的关键手段,其中模型验证是核心环节。针对导弹飞行仿真模型结果验证存在的参考数据不可获得、参考数据来源多样、专家验证主观性强等问题,提出一种基于时间序列分段特征提取的导弹飞行仿真模型结果验证方法。提出了一种综合的时间序列分段线性方法,由基于二阶导数提取趋势边缘点的线性分段算法和基于极值点优化的Top-Down线性分段算法两部分组成,以实现对导弹飞行仿真数据和参考数据进行有效的线性分段表示。基于上述分段结果,对各段时间序列的均值、方差、斜率等特征进行提取,以辅助专家进行验证,从而降低验证中的主观性;或者直接利用TIC系数法、动态时间规整(dynamic time warping,DTW)等方法进行客观的相似性分析。通过充分利用时间序列的分段特征,可实现各种情形下的导弹飞行仿真模型结果验证。通过一个导弹模型结果验证案例演示了所提方法的可行性和有效性。