Although quality assurance and quality control procedures are routinely applied in most air quality networks, outliers can still occur due to instrument malfunctions, the influence of harsh environments and the limita...Although quality assurance and quality control procedures are routinely applied in most air quality networks, outliers can still occur due to instrument malfunctions, the influence of harsh environments and the limitation of measuring methods. Such outliers pose challenges for data-powered applications such as data assimilation, statistical analysis of pollution characteristics and ensemble forecasting. Here, a fully automatic outlier detection method was developed based on the probability of residuals, which are the discrepancies between the observed and the estimated concentration values. The estimation can be conducted using filtering—or regressions when appropriate—to discriminate four types of outliers characterized by temporal and spatial inconsistency, instrument-induced low variances, periodic calibration exceptions, and less PM_(10) than PM_(2.5) in concentration observations, respectively. This probabilistic method was applied to detect all four types of outliers in hourly surface measurements of six pollutants(PM_(2.5), PM_(10),SO_2,NO_2,CO and O_3) from 1436 stations of the China National Environmental Monitoring Network during 2014-16. Among the measurements, 0.65%-5.68% are marked as outliers. with PM_(10) and CO more prone to outliers. Our method successfully identifies a trend of decreasing outliers from 2014 to 2016,which corresponds to known improvements in the quality assurance and quality control procedures of the China National Environmental Monitoring Network. The outliers can have a significant impact on the annual mean concentrations of PM_(2.5),with differences exceeding 10 μg m^(-3) at 66 sites.展开更多
Sea fog is a disastrous weather phenomenon,posing a risk to the safety of maritime transportation.Dense sea fogs reduce visibility at sea and have frequently caused ship collisions.This study used a geographically wei...Sea fog is a disastrous weather phenomenon,posing a risk to the safety of maritime transportation.Dense sea fogs reduce visibility at sea and have frequently caused ship collisions.This study used a geographically weighted regression(GWR)model to explore the spatial non-stationarity of near-miss collision risk,as detected by a vessel conflict ranking operator(VCRO)model from automatic identification system(AIS)data under the influence of sea fog in the Bohai Sea.Sea fog was identified by a machine learning method that was derived from Himawari-8 satellite data.The spatial distributions of near-miss collision risk,sea fog,and the parameters of GWR were mapped.The results showed that sea fog and near-miss collision risk have specific spatial distribution patterns in the Bohai Sea,in which near-miss collision risk in the fog season is significantly higher than that outside the fog season,especially in the northeast(the sea area near Yingkou Port and Bayuquan Port)and the southeast(the sea area near Yantai Port).GWR outputs further indicated a significant correlation between near-miss collision risk and sea fog in fog season,with higher R-squared(0.890 in fog season,2018),than outside the fog season(0.723 in non-fog season,2018).GWR results revealed spatial non-stationarity in the relationships between-near miss collision risk and sea fog and that the significance of these relationships varied locally.Dividing the specific navigation area made it possible to verify that sea fog has a positive impact on near-miss collision risk.展开更多
Many preterm infants suffer from neural disorders caused by early birth complications. The detection of children with neurological risk is an important challenge. The electroencephalogram is an important technique for...Many preterm infants suffer from neural disorders caused by early birth complications. The detection of children with neurological risk is an important challenge. The electroencephalogram is an important technique for establishing long-term neurological prognosis. Within this scope, the goal of this study is to propose an automatic detection of abnormal preterm babies’ electroencephalograms (EEG). A corpus of 316 neonatal EEG recordings of 100 infants born after less than 35 weeks of gestation were preprocessed and a time series of standard deviation was computed. This time series was thresholded to detect Inter Burst Intervals (IBI). Temporal features were extracted from bursts and IBI. Feature selection was carried out with classification in one step so as to select the best combination of features in terms of classification performance. Two classifiers were tested: Multiple Linear Regressions and Support Vector Machines (SVM). Performance was computed using cross validations. Methods were validated on a corpus of 100 infants with no serious brain damage. The Multiple Linear Regression method shows the best results with a sensitivity of 86.11% ± 10.01%, a specificity of 77.44% ± 7.62% and an AUC (Area under the ROC curves) of 0.82 ± 0.04. An accurate detection of abnormal EEG for preterm infants is feasible. This study is a first step towards an automatic analysis of the premature brain, making it possible to lighten the physician’s workload in the future.展开更多
The supreme goal of the Automatic Test case selection techniques is to guarantee systematic coverage, to recognize the usual error forms and to lessen the test of redundancy. It is unfeasible to carry out all the test...The supreme goal of the Automatic Test case selection techniques is to guarantee systematic coverage, to recognize the usual error forms and to lessen the test of redundancy. It is unfeasible to carry out all the test cases consistently. For this reason, the test cases are picked and prioritize it. The major goal of test case prioritization is to prioritize the test case sequence and finds faults as early as possible to improve the efficiency. Regression testing is used to ensure the validity and the enhancement part of the changed software. In this paper, we propose a new path compression technique (PCUA) for both old version and new version of BPEL dataset. In order to analyze the enhancement part of an application and to find an error in an enhancement part of an application, center of the tree has been calculated. Moreover in the comparative analysis, our proposed PCUA- COT technique is compared with the existing XPFG technique in terms of time consuming and error detection in the path of an enhancement part of BPEL dataset. The experimental results have been shown that our proposed work is better than the existing technique in terms of time consuming and error detection.展开更多
Scientific computing libraries,whether in-house or open-source,have witnessed enormous progress in both engineering and scientific research.Therefore,it is important to ensure that modifications to the source code,pro...Scientific computing libraries,whether in-house or open-source,have witnessed enormous progress in both engineering and scientific research.Therefore,it is important to ensure that modifications to the source code,prompted by bug fixing or new feature development,do not compromise the accuracy and functionality that have been already validated and verified.This paper introduces a method for establishing and implementing an automatic regression test environment,using the open-source multi-physics library SPHinXsys as an illustrative example.Initially,a reference database for each benchmark test is generated from observed data across multiple executions.This comprehensive database encapsulates the maximum variation range of metrics for different strategies,including the time-averaged,ensemble-averaged,and dynamic time warping methods.It accounts for uncertainties arising from parallel computing,particle relaxation,physical instabilities,and more.Subsequently,new results obtained after source code modifications undergo testing based on a curve-similarity comparison against the reference database.Whenever the source code is updated,the regression test is automatically executed for all test cases,providing a comprehensive assessment of the validity of the current results.This regression test environment has been successfully implemented in all dynamic test cases within SPHinXsys,including fluid dynamics,solid mechanics,fluid-structure interaction,thermal and mass diffusion,reaction-diffusion,and their multi-physics couplings,and demonstrates robust capabilities in testing different problems.It is noted that while the current test environment is built and implemented for a particular scientific computing library,its underlying principles are generic and can be easily adapted for use with other libraries,achieving equal effectiveness.展开更多
The optimization of the rule base of a fuzzy logic system (FLS) based on evolutionary algorithm has achievednotable results. However, due to the diversity of the deep structure in the hierarchical fuzzy system (HFS) a...The optimization of the rule base of a fuzzy logic system (FLS) based on evolutionary algorithm has achievednotable results. However, due to the diversity of the deep structure in the hierarchical fuzzy system (HFS) and thecorrelation of each sub fuzzy system, the uncertainty of the HFS’s deep structure increases. For the HFS, a largenumber of studies mainly use fixed structures, which cannot be selected automatically. To solve this problem, thispaper proposes a novel approach for constructing the incremental HFS. During system design, the deep structureand the rule base of the HFS are encoded separately. Subsequently, the deep structure is adaptively mutated basedon the fitness value, so as to realize the diversity of deep structures while ensuring reasonable competition amongthe structures. Finally, the differential evolution (DE) is used to optimize the deep structure of HFS and theparameters of antecedent and consequent simultaneously. The simulation results confirm the effectiveness of themodel. Specifically, the root mean square errors in the Laser dataset and Friedman dataset are 0.0395 and 0.0725,respectively with rule counts of rules is 8 and 12, respectively.When compared to alternative methods, the resultsindicate that the proposed method offers improvements in accuracy and rule counts.展开更多
轨道交通客流量影响因素是轨道交通方面研究的一个关注点,不同站点客流量的时空非平稳性被认为与站域建成环境有关。通过构建时空地理加权(geographically and temporally weighted regression,GTWR)模型,揭示了土地多样性、密度、站点...轨道交通客流量影响因素是轨道交通方面研究的一个关注点,不同站点客流量的时空非平稳性被认为与站域建成环境有关。通过构建时空地理加权(geographically and temporally weighted regression,GTWR)模型,揭示了土地多样性、密度、站点属性3个方面因素在时间和空间维度上对天津市轨道交通客流量的影响。结果表明:相较于传统的地理加权(geographically weighted regression,GWR)模型和最小二乘法(ordinary least squares,OLS)模型,GTWR具有更好的拟合优度;公交站点密度对轨道交通客流产生促进作用,尤其在工作日的早晚高峰时段和中心城区位置;市中心的商业设施在工作日晚高峰吸引更多的地铁乘客,而在近郊区它们在早高峰吸引更多的地铁乘客;人口密度促进轨道交通的客流量;充足的停车场设施数量可以吸引更多的轨道交通乘客。展开更多
基金supported by the National Natural Science Foundation (Grant Nos.91644216 and 41575128)the CAS Information Technology Program (Grant No.XXH13506-302)Guangdong Provincial Science and Technology Development Special Fund (No.2017B020216007)
文摘Although quality assurance and quality control procedures are routinely applied in most air quality networks, outliers can still occur due to instrument malfunctions, the influence of harsh environments and the limitation of measuring methods. Such outliers pose challenges for data-powered applications such as data assimilation, statistical analysis of pollution characteristics and ensemble forecasting. Here, a fully automatic outlier detection method was developed based on the probability of residuals, which are the discrepancies between the observed and the estimated concentration values. The estimation can be conducted using filtering—or regressions when appropriate—to discriminate four types of outliers characterized by temporal and spatial inconsistency, instrument-induced low variances, periodic calibration exceptions, and less PM_(10) than PM_(2.5) in concentration observations, respectively. This probabilistic method was applied to detect all four types of outliers in hourly surface measurements of six pollutants(PM_(2.5), PM_(10),SO_2,NO_2,CO and O_3) from 1436 stations of the China National Environmental Monitoring Network during 2014-16. Among the measurements, 0.65%-5.68% are marked as outliers. with PM_(10) and CO more prone to outliers. Our method successfully identifies a trend of decreasing outliers from 2014 to 2016,which corresponds to known improvements in the quality assurance and quality control procedures of the China National Environmental Monitoring Network. The outliers can have a significant impact on the annual mean concentrations of PM_(2.5),with differences exceeding 10 μg m^(-3) at 66 sites.
文摘Sea fog is a disastrous weather phenomenon,posing a risk to the safety of maritime transportation.Dense sea fogs reduce visibility at sea and have frequently caused ship collisions.This study used a geographically weighted regression(GWR)model to explore the spatial non-stationarity of near-miss collision risk,as detected by a vessel conflict ranking operator(VCRO)model from automatic identification system(AIS)data under the influence of sea fog in the Bohai Sea.Sea fog was identified by a machine learning method that was derived from Himawari-8 satellite data.The spatial distributions of near-miss collision risk,sea fog,and the parameters of GWR were mapped.The results showed that sea fog and near-miss collision risk have specific spatial distribution patterns in the Bohai Sea,in which near-miss collision risk in the fog season is significantly higher than that outside the fog season,especially in the northeast(the sea area near Yingkou Port and Bayuquan Port)and the southeast(the sea area near Yantai Port).GWR outputs further indicated a significant correlation between near-miss collision risk and sea fog in fog season,with higher R-squared(0.890 in fog season,2018),than outside the fog season(0.723 in non-fog season,2018).GWR results revealed spatial non-stationarity in the relationships between-near miss collision risk and sea fog and that the significance of these relationships varied locally.Dividing the specific navigation area made it possible to verify that sea fog has a positive impact on near-miss collision risk.
文摘Many preterm infants suffer from neural disorders caused by early birth complications. The detection of children with neurological risk is an important challenge. The electroencephalogram is an important technique for establishing long-term neurological prognosis. Within this scope, the goal of this study is to propose an automatic detection of abnormal preterm babies’ electroencephalograms (EEG). A corpus of 316 neonatal EEG recordings of 100 infants born after less than 35 weeks of gestation were preprocessed and a time series of standard deviation was computed. This time series was thresholded to detect Inter Burst Intervals (IBI). Temporal features were extracted from bursts and IBI. Feature selection was carried out with classification in one step so as to select the best combination of features in terms of classification performance. Two classifiers were tested: Multiple Linear Regressions and Support Vector Machines (SVM). Performance was computed using cross validations. Methods were validated on a corpus of 100 infants with no serious brain damage. The Multiple Linear Regression method shows the best results with a sensitivity of 86.11% ± 10.01%, a specificity of 77.44% ± 7.62% and an AUC (Area under the ROC curves) of 0.82 ± 0.04. An accurate detection of abnormal EEG for preterm infants is feasible. This study is a first step towards an automatic analysis of the premature brain, making it possible to lighten the physician’s workload in the future.
文摘The supreme goal of the Automatic Test case selection techniques is to guarantee systematic coverage, to recognize the usual error forms and to lessen the test of redundancy. It is unfeasible to carry out all the test cases consistently. For this reason, the test cases are picked and prioritize it. The major goal of test case prioritization is to prioritize the test case sequence and finds faults as early as possible to improve the efficiency. Regression testing is used to ensure the validity and the enhancement part of the changed software. In this paper, we propose a new path compression technique (PCUA) for both old version and new version of BPEL dataset. In order to analyze the enhancement part of an application and to find an error in an enhancement part of an application, center of the tree has been calculated. Moreover in the comparative analysis, our proposed PCUA- COT technique is compared with the existing XPFG technique in terms of time consuming and error detection in the path of an enhancement part of BPEL dataset. The experimental results have been shown that our proposed work is better than the existing technique in terms of time consuming and error detection.
基金supported by the China Scholarship Council(Grant No.202006230071)the Deutsche Forschungsgemeinschaft(DFG)(Grant No.DFG HU1527/12-4).
文摘Scientific computing libraries,whether in-house or open-source,have witnessed enormous progress in both engineering and scientific research.Therefore,it is important to ensure that modifications to the source code,prompted by bug fixing or new feature development,do not compromise the accuracy and functionality that have been already validated and verified.This paper introduces a method for establishing and implementing an automatic regression test environment,using the open-source multi-physics library SPHinXsys as an illustrative example.Initially,a reference database for each benchmark test is generated from observed data across multiple executions.This comprehensive database encapsulates the maximum variation range of metrics for different strategies,including the time-averaged,ensemble-averaged,and dynamic time warping methods.It accounts for uncertainties arising from parallel computing,particle relaxation,physical instabilities,and more.Subsequently,new results obtained after source code modifications undergo testing based on a curve-similarity comparison against the reference database.Whenever the source code is updated,the regression test is automatically executed for all test cases,providing a comprehensive assessment of the validity of the current results.This regression test environment has been successfully implemented in all dynamic test cases within SPHinXsys,including fluid dynamics,solid mechanics,fluid-structure interaction,thermal and mass diffusion,reaction-diffusion,and their multi-physics couplings,and demonstrates robust capabilities in testing different problems.It is noted that while the current test environment is built and implemented for a particular scientific computing library,its underlying principles are generic and can be easily adapted for use with other libraries,achieving equal effectiveness.
基金the Sichuan Science and Technology Program(2021ZYD0016).
文摘The optimization of the rule base of a fuzzy logic system (FLS) based on evolutionary algorithm has achievednotable results. However, due to the diversity of the deep structure in the hierarchical fuzzy system (HFS) and thecorrelation of each sub fuzzy system, the uncertainty of the HFS’s deep structure increases. For the HFS, a largenumber of studies mainly use fixed structures, which cannot be selected automatically. To solve this problem, thispaper proposes a novel approach for constructing the incremental HFS. During system design, the deep structureand the rule base of the HFS are encoded separately. Subsequently, the deep structure is adaptively mutated basedon the fitness value, so as to realize the diversity of deep structures while ensuring reasonable competition amongthe structures. Finally, the differential evolution (DE) is used to optimize the deep structure of HFS and theparameters of antecedent and consequent simultaneously. The simulation results confirm the effectiveness of themodel. Specifically, the root mean square errors in the Laser dataset and Friedman dataset are 0.0395 and 0.0725,respectively with rule counts of rules is 8 and 12, respectively.When compared to alternative methods, the resultsindicate that the proposed method offers improvements in accuracy and rule counts.
文摘轨道交通客流量影响因素是轨道交通方面研究的一个关注点,不同站点客流量的时空非平稳性被认为与站域建成环境有关。通过构建时空地理加权(geographically and temporally weighted regression,GTWR)模型,揭示了土地多样性、密度、站点属性3个方面因素在时间和空间维度上对天津市轨道交通客流量的影响。结果表明:相较于传统的地理加权(geographically weighted regression,GWR)模型和最小二乘法(ordinary least squares,OLS)模型,GTWR具有更好的拟合优度;公交站点密度对轨道交通客流产生促进作用,尤其在工作日的早晚高峰时段和中心城区位置;市中心的商业设施在工作日晚高峰吸引更多的地铁乘客,而在近郊区它们在早高峰吸引更多的地铁乘客;人口密度促进轨道交通的客流量;充足的停车场设施数量可以吸引更多的轨道交通乘客。