Due to the restricted satellite payloads in LEO mega-constellation networks(LMCNs),remote sensing image analysis,online learning and other big data services desirably need onboard distributed processing(OBDP).In exist...Due to the restricted satellite payloads in LEO mega-constellation networks(LMCNs),remote sensing image analysis,online learning and other big data services desirably need onboard distributed processing(OBDP).In existing technologies,the efficiency of big data applications(BDAs)in distributed systems hinges on the stable-state and low-latency links between worker nodes.However,LMCNs with high-dynamic nodes and long-distance links can not provide the above conditions,which makes the performance of OBDP hard to be intuitively measured.To bridge this gap,a multidimensional simulation platform is indispensable that can simulate the network environment of LMCNs and put BDAs in it for performance testing.Using STK's APIs and parallel computing framework,we achieve real-time simulation for thousands of satellite nodes,which are mapped as application nodes through software defined network(SDN)and container technologies.We elaborate the architecture and mechanism of the simulation platform,and take the Starlink and Hadoop as realistic examples for simulations.The results indicate that LMCNs have dynamic end-to-end latency which fluctuates periodically with the constellation movement.Compared to ground data center networks(GDCNs),LMCNs deteriorate the computing and storage job throughput,which can be alleviated by the utilization of erasure codes and data flow scheduling of worker nodes.展开更多
Background:Missing data are frequently occurred in clinical studies.Due to the development of precision medicine,there is an increased interest in N-of-1 trial.Bayesian models are one of main statistical methods for a...Background:Missing data are frequently occurred in clinical studies.Due to the development of precision medicine,there is an increased interest in N-of-1 trial.Bayesian models are one of main statistical methods for analyzing the data of N-of-1 trials.This simulation study aimed to compare two statistical methods for handling missing values of quantitative data in Bayesian N-of-1 trials.Methods:The simulated data of N-of-1 trials with different coefficients of autocorrelation,effect sizes and missing ratios are obtained by SAS 9.1 system.The missing values are filled with mean filling and regression filling respectively in the condition of different coefficients of autocorrelation,effect sizes and missing ratios by SPSS 25.0 software.Bayesian models are built to estimate the posterior means by Winbugs 14 software.Results:When the missing ratio is relatively small,e.g.5%,missing values have relatively little effect on the results.Therapeutic effects may be underestimated when the coefficient of autocorrelation increases and no filling is used.However,it may be overestimated when mean or regression filling is used,and the results after mean filling are closer to the actual effect than regression filling.In the case of moderate missing ratio,the estimated effect after mean filling is closer to the actual effect compared to regression filling.When a large missing ratio(20%)occurs,data missing can lead to significantly underestimate the effect.In this case,the estimated effect after regression filling is closer to the actual effect compared to mean filling.Conclusion:Data missing can affect the estimated therapeutic effects using Bayesian models in N-of-1 trials.The present study suggests that mean filling can be used under situation of missing ratio≤10%.Otherwise,regression filling may be preferable.展开更多
The monopile is the most common foundation to support offshore wind turbines.In the marine environment,local scour due to combined currents and waves is a significant issue that must be considered in the design of win...The monopile is the most common foundation to support offshore wind turbines.In the marine environment,local scour due to combined currents and waves is a significant issue that must be considered in the design of wind turbine foundations.In this paper,a full-scale numerical model was developed and validated based on field data from Rudong,China.The scour development around monopiles was investigated,and the effects of waves and the Reynolds number Re were analyzed.Several formulas for predicting the scour depth in the literature have been evaluated.It is found that waves can accelerate scour development even if the KC number is small(0.78<KC<1.57).The formula obtained from small-scale model tests may be unsafe or wasteful when it is applied in practical design due to the scale effect.A new equation for predicting the scour depth based on the average pile Reynolds number(Rea)is proposed and validated with field data.The equilibrium scour depth predicted using the proposed equation is evaluated and compared with those from nine equations in the literature.It is demonstrated that the values predicted from the proposed equation and from the S/M(Sheppard/Melville)equation are closer to the field data.展开更多
Recent advances in deep learning have expanded new possibilities for fluid flow simulation in petroleum reservoirs.However,the predominant approach in existing research is to train neural networks using high-fidelity ...Recent advances in deep learning have expanded new possibilities for fluid flow simulation in petroleum reservoirs.However,the predominant approach in existing research is to train neural networks using high-fidelity numerical simulation data.This presents a significant challenge because the sole source of authentic wellbore production data for training is sparse.In response to this challenge,this work introduces a novel architecture called physics-informed neural network based on domain decomposition(PINN-DD),aiming to effectively utilize the sparse production data of wells for reservoir simulation with large-scale systems.To harness the capabilities of physics-informed neural networks(PINNs)in handling small-scale spatial-temporal domain while addressing the challenges of large-scale systems with sparse labeled data,the computational domain is divided into two distinct sub-domains:the well-containing and the well-free sub-domain.Moreover,the two sub-domains and the interface are rigorously constrained by the governing equations,data matching,and boundary conditions.The accuracy of the proposed method is evaluated on two problems,and its performance is compared against state-of-the-art PINNs through numerical analysis as a benchmark.The results demonstrate the superiority of PINN-DD in handling large-scale reservoir simulation with limited data and show its potential to outperform conventional PINNs in such scenarios.展开更多
An attempt has been made to develop a distributed software infrastructure model for onboard data fusion system simulation, which is also applied to netted radar systems, onboard distributed detection systems and advan...An attempt has been made to develop a distributed software infrastructure model for onboard data fusion system simulation, which is also applied to netted radar systems, onboard distributed detection systems and advanced C3I systems. Two architectures are provided and verified: one is based on pure TCP/IP protocol and C/S model, and implemented with Winsock, the other is based on CORBA (common object request broker architecture). The performance of data fusion simulation system, i.e. reliability, flexibility and scalability, is improved and enhanced by two models. The study of them makes valuable explore on incorporating the distributed computation concepts into radar system simulation techniques.展开更多
Imbalanced data classification is one of the major problems in machine learning.This imbalanced dataset typically has significant differences in the number of data samples between its classes.In most cases,the perform...Imbalanced data classification is one of the major problems in machine learning.This imbalanced dataset typically has significant differences in the number of data samples between its classes.In most cases,the performance of the machine learning algorithm such as Support Vector Machine(SVM)is affected when dealing with an imbalanced dataset.The classification accuracy is mostly skewed toward the majority class and poor results are exhibited in the prediction of minority-class samples.In this paper,a hybrid approach combining data pre-processing technique andSVMalgorithm based on improved Simulated Annealing(SA)was proposed.Firstly,the data preprocessing technique which primarily aims at solving the resampling strategy of handling imbalanced datasets was proposed.In this technique,the data were first synthetically generated to equalize the number of samples between classes and followed by a reduction step to remove redundancy and duplicated data.Next is the training of a balanced dataset using SVM.Since this algorithm requires an iterative process to search for the best penalty parameter during training,an improved SA algorithm was proposed for this task.In this proposed improvement,a new acceptance criterion for the solution to be accepted in the SA algorithm was introduced to enhance the accuracy of the optimization process.Experimental works based on ten publicly available imbalanced datasets have demonstrated higher accuracy in the classification tasks using the proposed approach in comparison with the conventional implementation of SVM.Registering at an average of 89.65%of accuracy for the binary class classification has demonstrated the good performance of the proposed works.展开更多
A large amount of data can partly assure good fitting quality for the trained neural networks.When the quantity of experimental or on-site monitoring data is commonly insufficient and the quality is difficult to contr...A large amount of data can partly assure good fitting quality for the trained neural networks.When the quantity of experimental or on-site monitoring data is commonly insufficient and the quality is difficult to control in engineering practice,numerical simulations can provide a large amount of controlled high quality data.Once the neural networks are trained by such data,they can be used for predicting the properties/responses of the engineering objects instantly,saving the further computing efforts of simulation tools.Correspondingly,a strategy for efficiently transferring the input and output data used and obtained in numerical simulations to neural networks is desirable for engineers and programmers.In this work,we proposed a simple image representation strategy of numerical simulations,where the input and output data are all represented by images.The temporal and spatial information is kept and the data are greatly compressed.In addition,the results are readable for not only computers but also human resources.Some examples are given,indicating the effectiveness of the proposed strategy.展开更多
The MM5 and its four dimensional variational data assimilation (4D-Var) system are used in this paper. Based on the National Centers for Environmental Prediction/National Center for Atmospheric Research (NCEP/NCAR) re...The MM5 and its four dimensional variational data assimilation (4D-Var) system are used in this paper. Based on the National Centers for Environmental Prediction/National Center for Atmospheric Research (NCEP/NCAR) reanalysis data, the authors generate an optimal initial condition for a typhoon by using the bogus data assimilation (BDA) scheme. BDA is able to recover many of the structural features of typhoons including a warm-core vertex, the correct center position, and the strong circulation. As a result of BDA using a bogus surface low, dramatic improvement is achieved in the 72 h prediction of typhoon Herb. Through several cases, the initialization by BDA effectively generates the harmonious inner structure of the typhoon, but which is lacking in the original analysis field. Therefore the intensity forecast is improved greatly. Some improvements are made in the track forecast, but more work still needs to be done.展开更多
Background:The universal occurrence of randomly distributed dark holes(i.e.,data pits appearing within the tree crown)in LiDAR-derived canopy height models(CHMs)negatively affects the accuracy of extracted forest inve...Background:The universal occurrence of randomly distributed dark holes(i.e.,data pits appearing within the tree crown)in LiDAR-derived canopy height models(CHMs)negatively affects the accuracy of extracted forest inventory parameters.Methods:We develop an algorithm based on cloth simulation for constructing a pit-free CHM.Results:The proposed algorithm effectively fills data pits of various sizes whilst preserving canopy details.Our pitfree CHMs derived from point clouds at different proportions of data pits are remarkably better than those constructed using other algorithms,as evidenced by the lowest average root mean square error(0.4981 m)between the reference CHMs and the constructed pit-free CHMs.Moreover,our pit-free CHMs show the best performance overall in terms of maximum tree height estimation(average bias=0.9674 m).Conclusion:The proposed algorithm can be adopted when working with different quality LiDAR data and shows high potential in forestry applications.展开更多
This paper describes the historical simulations produced by the Chinese Academy of Meteorological Sciences(CAMS)climate system model(CAMS-CSM),which are contributing to phase 6 of the Coupled Model Intercomparison Pro...This paper describes the historical simulations produced by the Chinese Academy of Meteorological Sciences(CAMS)climate system model(CAMS-CSM),which are contributing to phase 6 of the Coupled Model Intercomparison Project(CMIP6).The model description,experiment design and model outputs are presented.Three members’historical experiments are conducted by CAMS-CSM,with two members starting from different initial conditions,and one excluding the stratospheric aerosol to identify the effect of volcanic eruptions.The outputs of the historical experiments are also validated using observational data.It is found that the model can reproduce the climatological mean states and seasonal cycle of the major climate system quantities,including the surface air temperature,precipitation,and the equatorial thermocline.The long-term trend of air temperature and precipitation is also reasonably captured by CAMS-CSM.There are still some biases in the model that need further improvement.This paper can help the users to better understand the performance and the datasets of CAMS-CSM.展开更多
Basins in many parts of the world are ungauged or poorly gauged, and in some cases existing measurement networks are declining. The purpose of this study was to examine the utility of reanalysis and global precipitati...Basins in many parts of the world are ungauged or poorly gauged, and in some cases existing measurement networks are declining. The purpose of this study was to examine the utility of reanalysis and global precipitation datasets in the river discharge simulation for a data-scarce basin. The White Volta basin of Ghana which is one of international rivers was selected as a study basin. NCEP1, NCEP2, ERA-Interim, and GPCP datasets were compared with corresponding observed precipitation data. Annual variations were not reproduced in NCEP1, NCEP2, and ERA-Interim. However, GPCP data, which is based on satellite and observed data, had good seasonal accuracy and reproduced annual variations well. Moreover, five datasets were used as input data to a hydrologic model with HYMOD, which is a water balance model, and with WTM, which is a river model;thereafter, the hydrologic model was calibrated for each datum set by a global optimization method, and river discharge were simulated. The results were evaluated by the root mean square error, relative error, and water balance error. As a result, the combination of GPCP precipitation and ERA-Interim evaporation data was the best in terms of most evaluations. The relative errors in the calibration and validation periods were 43.1% and 46.6%, respectively. Moreover, the results for the GPCP precipitation and ERA-Interim evaporation were better than those for the combination of observed precipitation and ERA-Interim evaporation. In conclusion, GPCP precipitation data and ERA-Interim evaporation data are very useful in a data-scarce basin water balance analysis.展开更多
When castings become complicated and the demands for precision of numerical simulation become higher,the numerical data of casting numerical simulation become more massive.On a general personal computer,these massive ...When castings become complicated and the demands for precision of numerical simulation become higher,the numerical data of casting numerical simulation become more massive.On a general personal computer,these massive numerical data may probably exceed the capacity of available memory,resulting in failure of rendering.Based on the out-of-core technique,this paper proposes a method to effectively utilize external storage and reduce memory usage dramatically,so as to solve the problem of insufficient memory for massive data rendering on general personal computers.Based on this method,a new postprocessor is developed.It is capable to illustrate filling and solidification processes of casting,as well as thermal stess.The new post-processor also provides fast interaction to simulation results.Theoretical analysis as well as several practical examples prove that the memory usage and loading time of the post-processor are independent of the size of the relevant files,but the proportion of the number of cells on surface.Meanwhile,the speed of rendering and fetching of value from the mouse is appreciable,and the demands of real-time and interaction are satisfied.展开更多
This research involved an exploratory evaluation of the dynamics of vehicular traffic on a road network across two traffic light-controlled junctions. The study uses the case study of a one-kilometer road system model...This research involved an exploratory evaluation of the dynamics of vehicular traffic on a road network across two traffic light-controlled junctions. The study uses the case study of a one-kilometer road system modelled on Anylogic version 8.8.4. Anylogic is a multi-paradigm simulation tool that supports three main simulation methodologies: discrete event simulation, agent-based modeling, and system dynamics modeling. The system is used to evaluate the implication of stochastic time-based vehicle variables on the general efficiency of road use. Road use efficiency as reflected in this model is based on the percentage of entry vehicles to exit the model within a one-hour simulation period. The study deduced that for the model under review, an increase in entry point time delay has a domineering influence on the efficiency of road use far beyond any other consideration. This study therefore presents a novel approach that leverages Discrete Events Simulation to facilitate efficient road management with a focus on optimum road use efficiency. The study also determined that the inclusion of appropriate random parameters to reflect road use activities at critical event points in a simulation can help in the effective representation of authentic traffic models. The Anylogic simulation software leverages the Classic DEVS and Parallel DEVS formalisms to achieve these objectives.展开更多
For the first time, this article introduces a LiDAR Point Clouds Dataset of Ships composed of both collected and simulated data to address the scarcity of LiDAR data in maritime applications. The collected data are ac...For the first time, this article introduces a LiDAR Point Clouds Dataset of Ships composed of both collected and simulated data to address the scarcity of LiDAR data in maritime applications. The collected data are acquired using specialized maritime LiDAR sensors in both inland waterways and wide-open ocean environments. The simulated data is generated by placing a ship in the LiDAR coordinate system and scanning it with a redeveloped Blensor that emulates the operation of a LiDAR sensor equipped with various laser beams. Furthermore,we also render point clouds for foggy and rainy weather conditions. To describe a realistic shipping environment, a dynamic tail wave is modeled by iterating the wave elevation of each point in a time series. Finally, networks serving small objects are migrated to ship applications by feeding our dataset. The positive effect of simulated data is described in object detection experiments, and the negative impact of tail waves as noise is verified in single-object tracking experiments. The Dataset is available at https://github.com/zqy411470859/ship_dataset.展开更多
In response to the lack of reliable physical parameters in the process simulation of the butadiene extraction,a large amount of phase equilibrium data were collected in the context of the actual process of butadiene p...In response to the lack of reliable physical parameters in the process simulation of the butadiene extraction,a large amount of phase equilibrium data were collected in the context of the actual process of butadiene production by acetonitrile.The accuracy of five prediction methods,UNIFAC(UNIQUAC Functional-group Activity Coefficients),UNIFAC-LL,UNIFAC-LBY,UNIFAC-DMD and COSMO-RS,applied to the butadiene extraction process was verified using partial phase equilibrium data.The results showed that the UNIFAC-DMD method had the highest accuracy in predicting phase equilibrium data for the missing system.COSMO-RS-predicted multiple systems showed good accuracy,and a large number of missing phase equilibrium data were estimated using the UNIFAC-DMD method and COSMO-RS method.The predicted phase equilibrium data were checked for consistency.The NRTL-RK(non-Random Two Liquid-Redlich-Kwong Equation of State)and UNIQUAC thermodynamic models were used to correlate the phase equilibrium data.Industrial device simulations were used to verify the accuracy of the thermodynamic model applied to the butadiene extraction process.The simulation results showed that the average deviations of the simulated results using the correlated thermodynamic model from the actual values were less than 2%compared to that using the commercial simulation software,Aspen Plus and its database.The average deviation was much smaller than that of the simulations using the Aspen Plus database(>10%),indicating that the obtained phase equilibrium data are highly accurate and reliable.The best phase equilibrium data and thermodynamic model parameters for butadiene extraction are provided.This improves the accuracy and reliability of the design,optimization and control of the process,and provides a basis and guarantee for developing a more environmentally friendly and economical butadiene extraction process.展开更多
This article introduces a novel variant of the generalized linear exponential(GLE)distribution,known as the sine generalized linear exponential(SGLE)distribution.The SGLE distribution utilizes the sine transformation ...This article introduces a novel variant of the generalized linear exponential(GLE)distribution,known as the sine generalized linear exponential(SGLE)distribution.The SGLE distribution utilizes the sine transformation to enhance its capabilities.The updated distribution is very adaptable and may be efficiently used in the modeling of survival data and dependability issues.The suggested model incorporates a hazard rate function(HRF)that may display a rising,J-shaped,or bathtub form,depending on its unique characteristics.This model includes many well-known lifespan distributions as separate sub-models.The suggested model is accompanied with a range of statistical features.The model parameters are examined using the techniques of maximum likelihood and Bayesian estimation using progressively censored data.In order to evaluate the effectiveness of these techniques,we provide a set of simulated data for testing purposes.The relevance of the newly presented model is shown via two real-world dataset applications,highlighting its superiority over other respected similar models.展开更多
Typhoon Rananim (0414) has been simulated by using the non-hydrostatic Advanced Regional Prediction System (ARPS) from Center of Analysis and Prediction of Storms (CAPS). The prediction of Rananim has generally ...Typhoon Rananim (0414) has been simulated by using the non-hydrostatic Advanced Regional Prediction System (ARPS) from Center of Analysis and Prediction of Storms (CAPS). The prediction of Rananim has generally been improved with ARPS using the new generation CINRAD Doppler radar data. Numerical experiments with or without using the radar data have shown that model initial fields with the assimilated radar radial velocity data in ARPS can change the wind field at the middle and high levels of the troposphere; fine characteristics of the tropical cyclone (TC) are introduced into the initial wind, the x component of wind speed south of the TC is increased and so is the y component west of it. They lead to improved forecasting of TC tracks for the time after landfall. The field of water vapor mixing ratio, temperature, cloud water mixing ratio and rainwater mixing ratio have also been improved by using radar refiectivity data. The model's initial response to the introduction of hydrometeors has been increased. It is shown that horizontal model resolution has a significant impact on intensity forecasts, by greatly improving the forecasting of TC rainfall, and heavy rainstorm of the TC specially, as well as its distribution and variation with time.展开更多
In the last decade,building energy simulation( BES)became a central component in building energy systems’ design and optimization. For each building location,BES requires one year of hourly weather data. Most buildin...In the last decade,building energy simulation( BES)became a central component in building energy systems’ design and optimization. For each building location,BES requires one year of hourly weather data. Most buildings are designed to last50 + years,consequently,the building design phase should include BES with future weather files considering climate change.This paper presents a comparative study of two methods to produce future climate hourly data files for BES: Morphing and typical meteorological year of future climate( F-TMY). The study uses data from a high-resolution( 9 km) regional climate atmospheric model simulation of Iberia,spanning 10 years of historical and future hourly data. This study compares both methods by analyzing anomalies in air temperature,and the impact in BES predictions of annual and peak energy consumption for space heating, cooling and ventilation in 4 buildings.Additionally, this study performs a sensitivity analysis of morphing method. The analysis shows that F-TMY is representative of the multi-year simulation for BES applications.A high-quality Morphed TMY weather file has a similar performance compared to F-TMY( average difference: 8% versus 7%). Morphing based on different baseline climates,low-grid resolution and/or outdated climate projections leads to BES average differences of 16%~20%.展开更多
Dynamic numerical simulation of water conditions is useful for reservoir management. In remote semi-arid areas, however, meteorological and hydrological time-series data needed for computation are not frequently measu...Dynamic numerical simulation of water conditions is useful for reservoir management. In remote semi-arid areas, however, meteorological and hydrological time-series data needed for computation are not frequently measured and must be obtained using other information. This paper presents a case study of data generation for the computation of thermal conditions in the Joumine Reservoir, Tunisia. Data from the Wind Finder web site and daily sunshine duration at the nearest weather stations were utilized to generate cloud cover and solar radiation data based on meteorological correlations obtained in Japan, which is located at the same latitude as Tunisia. A time series of inflow water temperature was estimated from air temperature using a numerical filter expressed as a linear second-order differential equation. A numerical simulation using a vertical 2-D (two-dimensional) turbulent flow model for a stratified water body with generated data successfully reproduced seasonal thermal conditions in the reservoir, which were monitored using a thermistor chain.展开更多
The progress of safety technologies,based on the continuous advances in vehicle crash worthiness,restraint systems and active safety functions made traffic safer than ever before.Latest developments heading from assis...The progress of safety technologies,based on the continuous advances in vehicle crash worthiness,restraint systems and active safety functions made traffic safer than ever before.Latest developments heading from assisted Advanced Driver Assistance System(ADAS)to Automated Driving(AD),lead to more and more complex real-world situations to be handled,going from standard driving tasks up to critical situations,which may cause a collision.Therefore,throughout the development process of such systems,it becomes common to use simulation technologies in order to assess these systems in advance.To gain results out of the simulation,input data are required;typically,from various sources,so the requirements can be covered.Thus,the challenge of scoping with different input sources arises.To come along with that problem,two main kinds of input data will be needed in general:(1)the descriptive parameter covering all border conditions,so called parameter room;(2)the system specifications for estimation.The quality of the results correlates strongly with the quality of inputs given.In case of ADAS systems and AD functions,the second kind of input data is very well known.Major challenges relate to the first kind of input data.Thus,the paper will describe a way to create input data that cover all descriptive parameters needed from normal driving up to the collision by the combination of accident analysis and real-world road traffic observations.The method aims at being applicable to different data sources and to different countries.展开更多
基金supported by National Natural Sciences Foundation of China(No.62271165,62027802,62201307)the Guangdong Basic and Applied Basic Research Foundation(No.2023A1515030297)+2 种基金the Shenzhen Science and Technology Program ZDSYS20210623091808025Stable Support Plan Program GXWD20231129102638002the Major Key Project of PCL(No.PCL2024A01)。
文摘Due to the restricted satellite payloads in LEO mega-constellation networks(LMCNs),remote sensing image analysis,online learning and other big data services desirably need onboard distributed processing(OBDP).In existing technologies,the efficiency of big data applications(BDAs)in distributed systems hinges on the stable-state and low-latency links between worker nodes.However,LMCNs with high-dynamic nodes and long-distance links can not provide the above conditions,which makes the performance of OBDP hard to be intuitively measured.To bridge this gap,a multidimensional simulation platform is indispensable that can simulate the network environment of LMCNs and put BDAs in it for performance testing.Using STK's APIs and parallel computing framework,we achieve real-time simulation for thousands of satellite nodes,which are mapped as application nodes through software defined network(SDN)and container technologies.We elaborate the architecture and mechanism of the simulation platform,and take the Starlink and Hadoop as realistic examples for simulations.The results indicate that LMCNs have dynamic end-to-end latency which fluctuates periodically with the constellation movement.Compared to ground data center networks(GDCNs),LMCNs deteriorate the computing and storage job throughput,which can be alleviated by the utilization of erasure codes and data flow scheduling of worker nodes.
基金supported by the National Natural Science Foundation of China (No.81973705).
文摘Background:Missing data are frequently occurred in clinical studies.Due to the development of precision medicine,there is an increased interest in N-of-1 trial.Bayesian models are one of main statistical methods for analyzing the data of N-of-1 trials.This simulation study aimed to compare two statistical methods for handling missing values of quantitative data in Bayesian N-of-1 trials.Methods:The simulated data of N-of-1 trials with different coefficients of autocorrelation,effect sizes and missing ratios are obtained by SAS 9.1 system.The missing values are filled with mean filling and regression filling respectively in the condition of different coefficients of autocorrelation,effect sizes and missing ratios by SPSS 25.0 software.Bayesian models are built to estimate the posterior means by Winbugs 14 software.Results:When the missing ratio is relatively small,e.g.5%,missing values have relatively little effect on the results.Therapeutic effects may be underestimated when the coefficient of autocorrelation increases and no filling is used.However,it may be overestimated when mean or regression filling is used,and the results after mean filling are closer to the actual effect than regression filling.In the case of moderate missing ratio,the estimated effect after mean filling is closer to the actual effect compared to regression filling.When a large missing ratio(20%)occurs,data missing can lead to significantly underestimate the effect.In this case,the estimated effect after regression filling is closer to the actual effect compared to mean filling.Conclusion:Data missing can affect the estimated therapeutic effects using Bayesian models in N-of-1 trials.The present study suggests that mean filling can be used under situation of missing ratio≤10%.Otherwise,regression filling may be preferable.
基金financially supported by the National Natural Science Foundation of China (Grant No.52378329)。
文摘The monopile is the most common foundation to support offshore wind turbines.In the marine environment,local scour due to combined currents and waves is a significant issue that must be considered in the design of wind turbine foundations.In this paper,a full-scale numerical model was developed and validated based on field data from Rudong,China.The scour development around monopiles was investigated,and the effects of waves and the Reynolds number Re were analyzed.Several formulas for predicting the scour depth in the literature have been evaluated.It is found that waves can accelerate scour development even if the KC number is small(0.78<KC<1.57).The formula obtained from small-scale model tests may be unsafe or wasteful when it is applied in practical design due to the scale effect.A new equation for predicting the scour depth based on the average pile Reynolds number(Rea)is proposed and validated with field data.The equilibrium scour depth predicted using the proposed equation is evaluated and compared with those from nine equations in the literature.It is demonstrated that the values predicted from the proposed equation and from the S/M(Sheppard/Melville)equation are closer to the field data.
基金funded by the National Natural Science Foundation of China(Grant No.52274048)Beijing Natural Science Foundation(Grant No.3222037)+1 种基金the CNPC 14th Five-Year Perspective Fundamental Research Project(Grant No.2021DJ2104)the Science Foundation of China University of Petroleum-Beijing(No.2462021YXZZ010).
文摘Recent advances in deep learning have expanded new possibilities for fluid flow simulation in petroleum reservoirs.However,the predominant approach in existing research is to train neural networks using high-fidelity numerical simulation data.This presents a significant challenge because the sole source of authentic wellbore production data for training is sparse.In response to this challenge,this work introduces a novel architecture called physics-informed neural network based on domain decomposition(PINN-DD),aiming to effectively utilize the sparse production data of wells for reservoir simulation with large-scale systems.To harness the capabilities of physics-informed neural networks(PINNs)in handling small-scale spatial-temporal domain while addressing the challenges of large-scale systems with sparse labeled data,the computational domain is divided into two distinct sub-domains:the well-containing and the well-free sub-domain.Moreover,the two sub-domains and the interface are rigorously constrained by the governing equations,data matching,and boundary conditions.The accuracy of the proposed method is evaluated on two problems,and its performance is compared against state-of-the-art PINNs through numerical analysis as a benchmark.The results demonstrate the superiority of PINN-DD in handling large-scale reservoir simulation with limited data and show its potential to outperform conventional PINNs in such scenarios.
文摘An attempt has been made to develop a distributed software infrastructure model for onboard data fusion system simulation, which is also applied to netted radar systems, onboard distributed detection systems and advanced C3I systems. Two architectures are provided and verified: one is based on pure TCP/IP protocol and C/S model, and implemented with Winsock, the other is based on CORBA (common object request broker architecture). The performance of data fusion simulation system, i.e. reliability, flexibility and scalability, is improved and enhanced by two models. The study of them makes valuable explore on incorporating the distributed computation concepts into radar system simulation techniques.
文摘Imbalanced data classification is one of the major problems in machine learning.This imbalanced dataset typically has significant differences in the number of data samples between its classes.In most cases,the performance of the machine learning algorithm such as Support Vector Machine(SVM)is affected when dealing with an imbalanced dataset.The classification accuracy is mostly skewed toward the majority class and poor results are exhibited in the prediction of minority-class samples.In this paper,a hybrid approach combining data pre-processing technique andSVMalgorithm based on improved Simulated Annealing(SA)was proposed.Firstly,the data preprocessing technique which primarily aims at solving the resampling strategy of handling imbalanced datasets was proposed.In this technique,the data were first synthetically generated to equalize the number of samples between classes and followed by a reduction step to remove redundancy and duplicated data.Next is the training of a balanced dataset using SVM.Since this algorithm requires an iterative process to search for the best penalty parameter during training,an improved SA algorithm was proposed for this task.In this proposed improvement,a new acceptance criterion for the solution to be accepted in the SA algorithm was introduced to enhance the accuracy of the optimization process.Experimental works based on ten publicly available imbalanced datasets have demonstrated higher accuracy in the classification tasks using the proposed approach in comparison with the conventional implementation of SVM.Registering at an average of 89.65%of accuracy for the binary class classification has demonstrated the good performance of the proposed works.
基金support from the National Natural Science Foundation of China(NSFC)(52178324).
文摘A large amount of data can partly assure good fitting quality for the trained neural networks.When the quantity of experimental or on-site monitoring data is commonly insufficient and the quality is difficult to control in engineering practice,numerical simulations can provide a large amount of controlled high quality data.Once the neural networks are trained by such data,they can be used for predicting the properties/responses of the engineering objects instantly,saving the further computing efforts of simulation tools.Correspondingly,a strategy for efficiently transferring the input and output data used and obtained in numerical simulations to neural networks is desirable for engineers and programmers.In this work,we proposed a simple image representation strategy of numerical simulations,where the input and output data are all represented by images.The temporal and spatial information is kept and the data are greatly compressed.In addition,the results are readable for not only computers but also human resources.Some examples are given,indicating the effectiveness of the proposed strategy.
文摘The MM5 and its four dimensional variational data assimilation (4D-Var) system are used in this paper. Based on the National Centers for Environmental Prediction/National Center for Atmospheric Research (NCEP/NCAR) reanalysis data, the authors generate an optimal initial condition for a typhoon by using the bogus data assimilation (BDA) scheme. BDA is able to recover many of the structural features of typhoons including a warm-core vertex, the correct center position, and the strong circulation. As a result of BDA using a bogus surface low, dramatic improvement is achieved in the 72 h prediction of typhoon Herb. Through several cases, the initialization by BDA effectively generates the harmonious inner structure of the typhoon, but which is lacking in the original analysis field. Therefore the intensity forecast is improved greatly. Some improvements are made in the track forecast, but more work still needs to be done.
基金the National Natural Science Foundation of China(Grant Nos.41671414,41971380 and 41171265)the National Key Research and Development Program of China(No.2016YFB0501404).
文摘Background:The universal occurrence of randomly distributed dark holes(i.e.,data pits appearing within the tree crown)in LiDAR-derived canopy height models(CHMs)negatively affects the accuracy of extracted forest inventory parameters.Methods:We develop an algorithm based on cloth simulation for constructing a pit-free CHM.Results:The proposed algorithm effectively fills data pits of various sizes whilst preserving canopy details.Our pitfree CHMs derived from point clouds at different proportions of data pits are remarkably better than those constructed using other algorithms,as evidenced by the lowest average root mean square error(0.4981 m)between the reference CHMs and the constructed pit-free CHMs.Moreover,our pit-free CHMs show the best performance overall in terms of maximum tree height estimation(average bias=0.9674 m).Conclusion:The proposed algorithm can be adopted when working with different quality LiDAR data and shows high potential in forestry applications.
基金supported by the National Key Research and Development Program of China(Grant No.2019YFC1510001)the National Natural Science Foundation of China(Grant No.91637210)+1 种基金the Basic Research Fund of CAMS(Grant No.2018Z007)the Jiangsu Collaborative Innovation Center for Climate Change。
文摘This paper describes the historical simulations produced by the Chinese Academy of Meteorological Sciences(CAMS)climate system model(CAMS-CSM),which are contributing to phase 6 of the Coupled Model Intercomparison Project(CMIP6).The model description,experiment design and model outputs are presented.Three members’historical experiments are conducted by CAMS-CSM,with two members starting from different initial conditions,and one excluding the stratospheric aerosol to identify the effect of volcanic eruptions.The outputs of the historical experiments are also validated using observational data.It is found that the model can reproduce the climatological mean states and seasonal cycle of the major climate system quantities,including the surface air temperature,precipitation,and the equatorial thermocline.The long-term trend of air temperature and precipitation is also reasonably captured by CAMS-CSM.There are still some biases in the model that need further improvement.This paper can help the users to better understand the performance and the datasets of CAMS-CSM.
文摘Basins in many parts of the world are ungauged or poorly gauged, and in some cases existing measurement networks are declining. The purpose of this study was to examine the utility of reanalysis and global precipitation datasets in the river discharge simulation for a data-scarce basin. The White Volta basin of Ghana which is one of international rivers was selected as a study basin. NCEP1, NCEP2, ERA-Interim, and GPCP datasets were compared with corresponding observed precipitation data. Annual variations were not reproduced in NCEP1, NCEP2, and ERA-Interim. However, GPCP data, which is based on satellite and observed data, had good seasonal accuracy and reproduced annual variations well. Moreover, five datasets were used as input data to a hydrologic model with HYMOD, which is a water balance model, and with WTM, which is a river model;thereafter, the hydrologic model was calibrated for each datum set by a global optimization method, and river discharge were simulated. The results were evaluated by the root mean square error, relative error, and water balance error. As a result, the combination of GPCP precipitation and ERA-Interim evaporation data was the best in terms of most evaluations. The relative errors in the calibration and validation periods were 43.1% and 46.6%, respectively. Moreover, the results for the GPCP precipitation and ERA-Interim evaporation were better than those for the combination of observed precipitation and ERA-Interim evaporation. In conclusion, GPCP precipitation data and ERA-Interim evaporation data are very useful in a data-scarce basin water balance analysis.
基金supported by the New Century Excellent Talents in University(NCET-09-0396)the National Science&Technology Key Projects of Numerical Control(2012ZX04014-031)+1 种基金the Natural Science Foundation of Hubei Province(2011CDB279)the Foundation for Innovative Research Groups of the Natural Science Foundation of Hubei Province,China(2010CDA067)
文摘When castings become complicated and the demands for precision of numerical simulation become higher,the numerical data of casting numerical simulation become more massive.On a general personal computer,these massive numerical data may probably exceed the capacity of available memory,resulting in failure of rendering.Based on the out-of-core technique,this paper proposes a method to effectively utilize external storage and reduce memory usage dramatically,so as to solve the problem of insufficient memory for massive data rendering on general personal computers.Based on this method,a new postprocessor is developed.It is capable to illustrate filling and solidification processes of casting,as well as thermal stess.The new post-processor also provides fast interaction to simulation results.Theoretical analysis as well as several practical examples prove that the memory usage and loading time of the post-processor are independent of the size of the relevant files,but the proportion of the number of cells on surface.Meanwhile,the speed of rendering and fetching of value from the mouse is appreciable,and the demands of real-time and interaction are satisfied.
文摘This research involved an exploratory evaluation of the dynamics of vehicular traffic on a road network across two traffic light-controlled junctions. The study uses the case study of a one-kilometer road system modelled on Anylogic version 8.8.4. Anylogic is a multi-paradigm simulation tool that supports three main simulation methodologies: discrete event simulation, agent-based modeling, and system dynamics modeling. The system is used to evaluate the implication of stochastic time-based vehicle variables on the general efficiency of road use. Road use efficiency as reflected in this model is based on the percentage of entry vehicles to exit the model within a one-hour simulation period. The study deduced that for the model under review, an increase in entry point time delay has a domineering influence on the efficiency of road use far beyond any other consideration. This study therefore presents a novel approach that leverages Discrete Events Simulation to facilitate efficient road management with a focus on optimum road use efficiency. The study also determined that the inclusion of appropriate random parameters to reflect road use activities at critical event points in a simulation can help in the effective representation of authentic traffic models. The Anylogic simulation software leverages the Classic DEVS and Parallel DEVS formalisms to achieve these objectives.
基金supported by the National Natural Science Foundation of China (62173103)the Fundamental Research Funds for the Central Universities of China (3072022JC0402,3072022JC0403)。
文摘For the first time, this article introduces a LiDAR Point Clouds Dataset of Ships composed of both collected and simulated data to address the scarcity of LiDAR data in maritime applications. The collected data are acquired using specialized maritime LiDAR sensors in both inland waterways and wide-open ocean environments. The simulated data is generated by placing a ship in the LiDAR coordinate system and scanning it with a redeveloped Blensor that emulates the operation of a LiDAR sensor equipped with various laser beams. Furthermore,we also render point clouds for foggy and rainy weather conditions. To describe a realistic shipping environment, a dynamic tail wave is modeled by iterating the wave elevation of each point in a time series. Finally, networks serving small objects are migrated to ship applications by feeding our dataset. The positive effect of simulated data is described in object detection experiments, and the negative impact of tail waves as noise is verified in single-object tracking experiments. The Dataset is available at https://github.com/zqy411470859/ship_dataset.
基金supported by the National Natural Science Foundation of China(22178190)。
文摘In response to the lack of reliable physical parameters in the process simulation of the butadiene extraction,a large amount of phase equilibrium data were collected in the context of the actual process of butadiene production by acetonitrile.The accuracy of five prediction methods,UNIFAC(UNIQUAC Functional-group Activity Coefficients),UNIFAC-LL,UNIFAC-LBY,UNIFAC-DMD and COSMO-RS,applied to the butadiene extraction process was verified using partial phase equilibrium data.The results showed that the UNIFAC-DMD method had the highest accuracy in predicting phase equilibrium data for the missing system.COSMO-RS-predicted multiple systems showed good accuracy,and a large number of missing phase equilibrium data were estimated using the UNIFAC-DMD method and COSMO-RS method.The predicted phase equilibrium data were checked for consistency.The NRTL-RK(non-Random Two Liquid-Redlich-Kwong Equation of State)and UNIQUAC thermodynamic models were used to correlate the phase equilibrium data.Industrial device simulations were used to verify the accuracy of the thermodynamic model applied to the butadiene extraction process.The simulation results showed that the average deviations of the simulated results using the correlated thermodynamic model from the actual values were less than 2%compared to that using the commercial simulation software,Aspen Plus and its database.The average deviation was much smaller than that of the simulations using the Aspen Plus database(>10%),indicating that the obtained phase equilibrium data are highly accurate and reliable.The best phase equilibrium data and thermodynamic model parameters for butadiene extraction are provided.This improves the accuracy and reliability of the design,optimization and control of the process,and provides a basis and guarantee for developing a more environmentally friendly and economical butadiene extraction process.
基金This work was supported and funded by the Deanship of Scientific Research at Imam Mohammad Ibn Saud Islamic University(IMSIU)(Grant Number IMSIU-RG23142).
文摘This article introduces a novel variant of the generalized linear exponential(GLE)distribution,known as the sine generalized linear exponential(SGLE)distribution.The SGLE distribution utilizes the sine transformation to enhance its capabilities.The updated distribution is very adaptable and may be efficiently used in the modeling of survival data and dependability issues.The suggested model incorporates a hazard rate function(HRF)that may display a rising,J-shaped,or bathtub form,depending on its unique characteristics.This model includes many well-known lifespan distributions as separate sub-models.The suggested model is accompanied with a range of statistical features.The model parameters are examined using the techniques of maximum likelihood and Bayesian estimation using progressively censored data.In order to evaluate the effectiveness of these techniques,we provide a set of simulated data for testing purposes.The relevance of the newly presented model is shown via two real-world dataset applications,highlighting its superiority over other respected similar models.
基金Technical Plan Key Project of Zhejiang Province (2006C13025)Key Subsidiary Project for Meteorological Science of Wenzhou (S200601)Technical Plan Key Project of Wenzhou (S2003A011)
文摘Typhoon Rananim (0414) has been simulated by using the non-hydrostatic Advanced Regional Prediction System (ARPS) from Center of Analysis and Prediction of Storms (CAPS). The prediction of Rananim has generally been improved with ARPS using the new generation CINRAD Doppler radar data. Numerical experiments with or without using the radar data have shown that model initial fields with the assimilated radar radial velocity data in ARPS can change the wind field at the middle and high levels of the troposphere; fine characteristics of the tropical cyclone (TC) are introduced into the initial wind, the x component of wind speed south of the TC is increased and so is the y component west of it. They lead to improved forecasting of TC tracks for the time after landfall. The field of water vapor mixing ratio, temperature, cloud water mixing ratio and rainwater mixing ratio have also been improved by using radar refiectivity data. The model's initial response to the introduction of hydrometeors has been increased. It is shown that horizontal model resolution has a significant impact on intensity forecasts, by greatly improving the forecasting of TC rainfall, and heavy rainstorm of the TC specially, as well as its distribution and variation with time.
文摘In the last decade,building energy simulation( BES)became a central component in building energy systems’ design and optimization. For each building location,BES requires one year of hourly weather data. Most buildings are designed to last50 + years,consequently,the building design phase should include BES with future weather files considering climate change.This paper presents a comparative study of two methods to produce future climate hourly data files for BES: Morphing and typical meteorological year of future climate( F-TMY). The study uses data from a high-resolution( 9 km) regional climate atmospheric model simulation of Iberia,spanning 10 years of historical and future hourly data. This study compares both methods by analyzing anomalies in air temperature,and the impact in BES predictions of annual and peak energy consumption for space heating, cooling and ventilation in 4 buildings.Additionally, this study performs a sensitivity analysis of morphing method. The analysis shows that F-TMY is representative of the multi-year simulation for BES applications.A high-quality Morphed TMY weather file has a similar performance compared to F-TMY( average difference: 8% versus 7%). Morphing based on different baseline climates,low-grid resolution and/or outdated climate projections leads to BES average differences of 16%~20%.
文摘Dynamic numerical simulation of water conditions is useful for reservoir management. In remote semi-arid areas, however, meteorological and hydrological time-series data needed for computation are not frequently measured and must be obtained using other information. This paper presents a case study of data generation for the computation of thermal conditions in the Joumine Reservoir, Tunisia. Data from the Wind Finder web site and daily sunshine duration at the nearest weather stations were utilized to generate cloud cover and solar radiation data based on meteorological correlations obtained in Japan, which is located at the same latitude as Tunisia. A time series of inflow water temperature was estimated from air temperature using a numerical filter expressed as a linear second-order differential equation. A numerical simulation using a vertical 2-D (two-dimensional) turbulent flow model for a stratified water body with generated data successfully reproduced seasonal thermal conditions in the reservoir, which were monitored using a thermistor chain.
文摘The progress of safety technologies,based on the continuous advances in vehicle crash worthiness,restraint systems and active safety functions made traffic safer than ever before.Latest developments heading from assisted Advanced Driver Assistance System(ADAS)to Automated Driving(AD),lead to more and more complex real-world situations to be handled,going from standard driving tasks up to critical situations,which may cause a collision.Therefore,throughout the development process of such systems,it becomes common to use simulation technologies in order to assess these systems in advance.To gain results out of the simulation,input data are required;typically,from various sources,so the requirements can be covered.Thus,the challenge of scoping with different input sources arises.To come along with that problem,two main kinds of input data will be needed in general:(1)the descriptive parameter covering all border conditions,so called parameter room;(2)the system specifications for estimation.The quality of the results correlates strongly with the quality of inputs given.In case of ADAS systems and AD functions,the second kind of input data is very well known.Major challenges relate to the first kind of input data.Thus,the paper will describe a way to create input data that cover all descriptive parameters needed from normal driving up to the collision by the combination of accident analysis and real-world road traffic observations.The method aims at being applicable to different data sources and to different countries.