Normality testing is a fundamental hypothesis test in the statistical analysis of key biological indicators of diabetes.If this assumption is violated,it may cause the test results to deviate from the true value,leadi...Normality testing is a fundamental hypothesis test in the statistical analysis of key biological indicators of diabetes.If this assumption is violated,it may cause the test results to deviate from the true value,leading to incorrect inferences and conclusions,and ultimately affecting the validity and accuracy of statistical inferences.Considering this,the study designs a unified analysis scheme for different data types based on parametric statistical test methods and non-parametric test methods.The data were grouped according to sample type and divided into discrete data and continuous data.To account for differences among subgroups,the conventional chi-squared test was used for discrete data.The normal distribution is the basis of many statistical methods;if the data does not follow a normal distribution,many statistical methods will fail or produce incorrect results.Therefore,before data analysis and modeling,the data were divided into normal and non-normal groups through normality testing.For normally distributed data,parametric statistical methods were used to judge the differences between groups.For non-normal data,non-parametric tests were employed to improve the accuracy of the analysis.Statistically significant indicators were retained according to the significance index P-value of the statistical test or corresponding statistics.These indicators were then combined with relevant medical background to further explore the etiology leading to the occurrence or transformation of diabetes status.展开更多
Choosing appropriate statistical tests is crucial but deciding which tests to use can be challenging. Different tests suit different types of data and research questions, so it is important to choose the right one. Kn...Choosing appropriate statistical tests is crucial but deciding which tests to use can be challenging. Different tests suit different types of data and research questions, so it is important to choose the right one. Knowing how to select an appropriate test can lead to more accurate results. Invalid results and misleading conclusions may be drawn from a study if an incorrect statistical test is used. Therefore, to avoid these it is essential to understand the nature of the data, the research question, and the assumptions of the tests before selecting one. This is because there are a wide variety of tests available. This paper provides a step-by-step approach to selecting the right statistical test for any study, with an explanation of when it is appropriate to use it and relevant examples of each statistical test. Furthermore, this guide provides a comprehensive overview of the assumptions of each test and what to do if these assumptions are violated.展开更多
Choosing appropriate statistical tests is crucial but deciding which tests to use can be challenging. Different tests suit different types of data and research questions, so it is important to choose the right one. Kn...Choosing appropriate statistical tests is crucial but deciding which tests to use can be challenging. Different tests suit different types of data and research questions, so it is important to choose the right one. Knowing how to select an appropriate test can lead to more accurate results. Invalid results and misleading conclusions may be drawn from a study if an incorrect statistical test is used. Therefore, to avoid these it is essential to understand the nature of the data, the research question, and the assumptions of the tests before selecting one. This is because there are a wide variety of tests available. This paper provides a step-by-step approach to selecting the right statistical test for any study, with an explanation of when it is appropriate to use it and relevant examples of each statistical test. Furthermore, this guide provides a comprehensive overview of the assumptions of each test and what to do if these assumptions are violated.展开更多
Many fields,such as neuroscience,are experiencing the vast prolife ration of cellular data,underscoring the need fo r organizing and interpreting large datasets.A popular approach partitions data into manageable subse...Many fields,such as neuroscience,are experiencing the vast prolife ration of cellular data,underscoring the need fo r organizing and interpreting large datasets.A popular approach partitions data into manageable subsets via hierarchical clustering,but objective methods to determine the appropriate classification granularity are missing.We recently introduced a technique to systematically identify when to stop subdividing clusters based on the fundamental principle that cells must differ more between than within clusters.Here we present the corresponding protocol to classify cellular datasets by combining datadriven unsupervised hierarchical clustering with statistical testing.These general-purpose functions are applicable to any cellular dataset that can be organized as two-dimensional matrices of numerical values,including molecula r,physiological,and anatomical datasets.We demonstrate the protocol using cellular data from the Janelia MouseLight project to chara cterize morphological aspects of neurons.展开更多
“Human-elephant conflict(HEC)”,the alarming issue,in present day context has attracted the attention of environmentalists and policy makers.The rising conflict between human beings and wild elephants is common in Bu...“Human-elephant conflict(HEC)”,the alarming issue,in present day context has attracted the attention of environmentalists and policy makers.The rising conflict between human beings and wild elephants is common in Buxa Tiger Reserve(BTR)and its adjoining area in West Bengal State,India,making the area volatile.People’s attitudes towards elephant conservation activity are very crucial to get rid of HEC,because people’s proximity with wild elephants’habitat can trigger the occurrence of HEC.The aim of this study is to conduct an in-depth investigation about the association of people’s attitudes towards HEC with their locational,demographic,and socio-economic characteristics in BTR and its adjoining area by using Pearson’s bivariate chi-square test and binary logistic regression analysis.BTR is one of the constituent parts of Eastern Doors Elephant Reserve(EDER).We interviewed 500 respondents to understand their perceptions to HEC and investigated their locational,demographic,and socio-economic characteristics including location of village,gender,age,ethnicity,religion,caste,poverty level,education level,primary occupation,secondary occupation,household type,and source of firewood.The results indicate that respondents who are living in enclave forest villages(EFVs),peripheral forest villages(PFVs),corridor village(CVs),or forest and corridor villages(FCVs),mainly males,at the age of 18–48 years old,engaged with agriculture occupation,and living in kancha and mixed houses,have more likelihood to witness HEC.Besides,respondents who are illiterate or at primary education level are more likely to regard elephant as a main problematic animal around their villages and refuse to participate in elephant conservation activity.For the sake of a sustainable environment for both human beings and wildlife,people’s attitudes towards elephants must be friendly in a more prudent way,so that the two communities can live in harmony.展开更多
Sunshine duration (S) based empirical equations have been employed in this study to estimate the daily global solar radiation on a horizontal surface (G) for six meteorological stations in Burundi. Those equations inc...Sunshine duration (S) based empirical equations have been employed in this study to estimate the daily global solar radiation on a horizontal surface (G) for six meteorological stations in Burundi. Those equations include the Ångström-Prescott linear model and four amongst its derivatives, i.e. logarithmic, exponential, power and quadratic functions. Monthly mean values of daily global solar radiation and sunshine duration data for a period of 20 to 23 years, from the Geographical Institute of Burundi (IGEBU), have been used. For any of the six stations, ten single or double linear regressions have been developed from the above-said five functions, to relate in terms of monthly mean values, the daily clearness index () to each of the next two kinds of relative sunshine duration (RSD): and . In those ratios, G<sub>0</sub>, S<sub>0 </sub>and stand for the extraterrestrial daily solar radiation on a horizontal surface, the day length and the modified day length taking into account the natural site’s horizon, respectively. According to the calculated mean values of the clearness index and the RSD, each station experiences a high number of fairly clear (or partially cloudy) days. Estimated values of the dependent variable (y) in each developed linear regression, have been compared to measured values in terms of the coefficients of correlation (R) and of determination (R<sub>2</sub>), the mean bias error (MBE), the root mean square error (RMSE) and the t-statistics. Mean values of these statistical indicators have been used to rank, according to decreasing performance level, firstly the ten developed equations per station on account of the overall six stations, secondly the six stations on account of the overall ten equations. Nevertheless, the obtained values of those indicators lay in the next ranges for all the developed sixty equations:;;;, with . These results lead to assert that any of the sixty developed linear regressions (and thus equations in terms of and ), fits very adequately measured data, and should be used to estimate monthly average daily global solar radiation with sunshine duration for the relevant station. It is also found that using as RSD, is slightly more advantageous than using for estimating the monthly average daily clearness index, . Moreover, values of statistical indicators of this study match adequately data from other works on the same kinds of empirical equations.展开更多
As the increasing popularity and complexity of Web applications and the emergence of their new characteristics, the testing and maintenance of large, complex Web applications are becoming more complex and difficult. W...As the increasing popularity and complexity of Web applications and the emergence of their new characteristics, the testing and maintenance of large, complex Web applications are becoming more complex and difficult. Web applications generally contain lots of pages and are used by enormous users. Statistical testing is an effective way of ensuring their quality. Web usage can be accurately described by Markov chain which has been proved to be an ideal model for software statistical testing. The results of unit testing can be utilized in the latter stages, which is an important strategy for bottom-to-top integration testing, and the other improvement of extended Markov chain model (EMM) is to present the error type vector which is treated as a part of page node. this paper also proposes the algorithm for generating test cases of usage paths. Finally, optional usage reliability evaluation methods and an incremental usability regression testing model for testing and evaluation are presented. Key words statistical testing - evaluation for Web usability - extended Markov chain model (EMM) - Web log mining - reliability evaluation CLC number TP311. 5 Foundation item: Supported by the National Defence Research Project (No. 41315. 9. 2) and National Science and Technology Plan (2001BA102A04-02-03)Biography: MAO Cheng-ying (1978-), male, Ph.D. candidate, research direction: software testing. Research direction: advanced database system, software testing, component technology and data mining.展开更多
In this article, the unit root test for AR(p) model with GARCH errors is considered. The Dickey-Fuller test statistics are rewritten in the form of self-normalized sums, and the asymptotic distribution of the test s...In this article, the unit root test for AR(p) model with GARCH errors is considered. The Dickey-Fuller test statistics are rewritten in the form of self-normalized sums, and the asymptotic distribution of the test statistics is derived under the weak conditions.展开更多
This paper describes the statistical methods of the comparison of the incidence or mortality rates in cancer registry and descriptive epidemiology, and the features of microcomputer program (CANTEST) which was designe...This paper describes the statistical methods of the comparison of the incidence or mortality rates in cancer registry and descriptive epidemiology, and the features of microcomputer program (CANTEST) which was designed to perform the methods. The program was written in IBM BASIC language. Using the program CANTEST we presented here the user can do several statistical tests or estimations as follow: 1. the comparison of the adjusted rates which were calculated by directly or indirectly standardized methods, 2. the calculation of the slope of regression line for testing the linear trends of the adjusted rates, 3. the estimation of the 95% or 99%conndence intervals of the directly adjusted rates, of the cumulative rates (0-64 and 0-74), and of the cumulative risk. Several examples are presented for testing the performances of the program.展开更多
We are very grateful for the letter written by Dr Lange,and indeed apologize for the mistakes noted in the word-ing of our text regarding statistical analysis.This wasdue to changes carried out while revising the manu...We are very grateful for the letter written by Dr Lange,and indeed apologize for the mistakes noted in the word-ing of our text regarding statistical analysis.This wasdue to changes carried out while revising the manuscriptat the request of reviewers,whom we thank for,point-ing out several issues that were actually similar to thosenoted by Dr.Lange.Unfortunately,we were unable todescribe and discuss our findings properly in the context展开更多
Search-based statistical structural testing(SBSST)is a promising technique that uses automated search to construct input distributions for statistical structural testing.It has been proved that a simple search algorit...Search-based statistical structural testing(SBSST)is a promising technique that uses automated search to construct input distributions for statistical structural testing.It has been proved that a simple search algorithm,for example,the hill-climber is able to optimize an input distribution.However,due to the noisy fitness estimation of the minimum triggering probability among all cover elements(Tri-Low-Bound),the existing approach does not show a satisfactory efficiency.Constructing input distributions to satisfy the Tri-Low-Bound criterion requires an extensive computation time.Tri-Low-Bound is considered a strong criterion,and it is demonstrated to sustain a high fault-detecting ability.This article tries to answer the following question:if we use a relaxed constraint that significantly reduces the time consumption on search,can the optimized input distribution still be effective in faultdetecting ability?In this article,we propose a type of criterion called fairnessenhanced-sum-of-triggering-probability(p-L1-Max).The criterion utilizes the sum of triggering probabilities as the fitness value and leverages a parameter p to adjust the uniformness of test data generation.We conducted extensive experiments to compare the computation time and the fault-detecting ability between the two criteria.The result shows that the 1.0-L1-Max criterion has the highest efficiency,and it is more practical to use than the Tri-Low-Bound criterion.To measure a criterion’s fault-detecting ability,we introduce a definition of expected faults found in the effective test set size region.To measure the effective test set size region,we present a theoretical analysis of the expected faults found with respect to various test set sizes and use the uniform distribution as a baseline to derive the effective test set size region’s definition.展开更多
We investigate redshift distributions of three long burst samples, with the first sample containing 131 long bursts with observed redshifts, the second including 220 long bursts with pseudo-redshifts calculated by the...We investigate redshift distributions of three long burst samples, with the first sample containing 131 long bursts with observed redshifts, the second including 220 long bursts with pseudo-redshifts calculated by the variability-luminosity relation, and the third including 1194 long bursts with pseudo-redshifls calculated by the lag-luminosity relation, respectively. In the redshift range 0-1 the Kolmogorov-Smirnov probability of the observed redshift distribution and that of the variability-luminosity relation is large. In the redshift ranges 1-2, 2-3, 3-6.3 and 0-37, the Kolmogorov-Smirnov probabilities of the redshift distribution from lag-luminosity relation and the observed redshift distribution are also large. For the GRBs, which appear both in the two pseudo-redshift burst samples, the KS probability of the pseudo-redshift distribution from the lag-luminosity relation and the observed reshift distribution is 0.447, which is very large. Based on these results, some conclusions are drawn: i) the V-Liso relation might be more believable than the τ-Liso relation in low redshift ranges and the τ-Liso relation might be more real than the V-Liso relation in high redshift ranges; ii) if we do not consider the redshift ranges, the τ-Liso relation might be more physical and intrinsical than the V-Liso relation.展开更多
This paper investigates the correlation between tidal stress and earthquakes for periods ranging from hours to months in the limited zone of the Palu region(Central Sulawesi,Indonesia).Through Schuster and binomial te...This paper investigates the correlation between tidal stress and earthquakes for periods ranging from hours to months in the limited zone of the Palu region(Central Sulawesi,Indonesia).Through Schuster and binomial tests,we examined the relation between the seismicity(time density of seismic events)and tidal potential arising from the Moon and Sun,using all tidal components simultaneously and focusing on the estimation of specific terms.The results show significant correlations between the seismicity and tidal potential for S2(0.5 d)and O1(1.075 d)tidal components in the case of solely isolated earthquake events,particularly for shallow earthquakes.Meanwhile,there is a strong relationship between aftershocks and tidal components larger than the Mf period(13.661 d).Finally,the analysis of the temporal variation of the earthquake-tide relation reveals an optimal correlation for about six years before the 2018 great Palu earthquake.The correlation becomes insignificant afterwards.展开更多
This paper develops the modeling of wind speed by Weibull distribution in the intention to evaluate wind energy potential and help for designing small wind energy plant in Batouri in Cameroon. The Weibull distribution...This paper develops the modeling of wind speed by Weibull distribution in the intention to evaluate wind energy potential and help for designing small wind energy plant in Batouri in Cameroon. The Weibull distribution model was developed using wind speed data collected from a metrological station at the small Airport of Batouri. Four numerical methods (Moment method, Graphical method, Empirical method and Energy pattern factor method) were used to estimate weibull parameters K and C. The application of these four methods is effective using a sample wind speed data set. With some statistical analysis, a comparison of the accuracy of each method is also performed. The study helps to determine that Energy pattern factor method is the most effective (K = 3.8262 and C = 2.4659).展开更多
The aim of this study is to establish the prevailing conditions of changing climatic trends and change point dates in four selected meteorological stations of Uyo, Benin, Port Harcourt, and Warri in the Niger Delta re...The aim of this study is to establish the prevailing conditions of changing climatic trends and change point dates in four selected meteorological stations of Uyo, Benin, Port Harcourt, and Warri in the Niger Delta region of Nigeria. Using daily or 24-hourly annual maximum series (AMS) data with the Indian Meteorological Department (IMD) and the modified Chowdury Indian Meteorological Department (MCIMD) models were adopted to downscale the time series data. Mann-Kendall (MK) trend and Sen’s Slope Estimator (SSE) test showed a statistically significant trend for Uyo and Benin, while Port Harcourt and Warri showed mild trends. The Sen’s Slope magnitude and variation rate were 21.6, 10.8, 6.00 and 4.4 mm/decade, respectively. The trend change-point analysis showed the initial rainfall change-point dates as 2002, 2005, 1988, and 2000 for Uyo, Benin, Port Harcourt, and Warri, respectively. These prove positive changing climatic conditions for rainfall in the study area. Erosion and flood control facilities analysis and design in the Niger Delta will require the application of Non-stationary IDF modelling.展开更多
The study focused on the detection of indicators of climate change in 24-hourly annual maximum series (AMS) rainfall data collected for 36 years (1982-2017) for Warri Township, using different statistical methods yiel...The study focused on the detection of indicators of climate change in 24-hourly annual maximum series (AMS) rainfall data collected for 36 years (1982-2017) for Warri Township, using different statistical methods yielded a statistically insignificant positive mild trend. The IMD and MCIMD downscaled model’s time series data respectively produced MK statistics varying from 1.403 to 1.4729, and 1.403 to 1.463 which were less than the critical Z-value of 1.96. Also, the slope magnitude obtained showed a mild increasing trend in variation from 0.0189 to 0.3713, and 0.0175 to 0.5426, with the rate of change in rainfall intensity at 24 hours duration as 0.4536 and 0.42 mm/hr.year (4.536 and 4.2 mm/decade) for the IMD and the MCIMD time series data, respectively. The trend change point date occurred in the year 2000 from the distribution-free CUSUM test with the trend maintaining a significant and steady increase from 2010 to 2015. Thus, this study established the existence of a trend, which is an indication of a changing climate, and satisfied the condition for rainfall Non-stationary intensity-duration-frequency (NS-IDF) modeling required for infrastructural design for combating flooding events.展开更多
For the functional partially linear models including flexible nonparametric part and functional linear part,the estimators of the nonlinear function and the slope function have been studied in existing literature.How ...For the functional partially linear models including flexible nonparametric part and functional linear part,the estimators of the nonlinear function and the slope function have been studied in existing literature.How to test the correlation between response and explanatory variables,however,still seems to be missing.Therefore,a test procedure for testing the linearity in the functional partially linear models will be proposed in this paper.A test statistic is constructed based on the existing estimators of the nonlinear and the slope functions.Further,we prove that the approximately asymptotic distribution of the proposed statistic is a chi-squared distribution under some regularity conditions.Finally,some simulation studies and a real data application are presented to demonstrate the performance of the proposed test statistic.展开更多
Exclusive hypothesis testing is a new and special class of hypothesis testing.This kind of testing can be applied in survival analysis to understand the association between genomics information and clinical informatio...Exclusive hypothesis testing is a new and special class of hypothesis testing.This kind of testing can be applied in survival analysis to understand the association between genomics information and clinical information about the survival time.Besides,it is well known that Cox's proportional hazards model is the most commonly used model for regression analysis of failure time.In this paper,the authors consider doing the exclusive hypothesis testing for Cox's proportional hazards model with right-censored data.The authors propose the comprehensive test statistics to make decision,and show that the corresponding decision rule can control the asymptotic TypeⅠerrors and have good powers in theory.The numerical studies indicate that the proposed approach works well for practical situations and it is applied to a set of real data arising from Rotterdam Breast Cancer Data study that motivated this study.展开更多
In this paper,based on the observation data of air temperature during 1951-2009 in Shenyang,the interannual and interdecadal variation of annual average temperature,maximum and minimum temperature in Shenyang were con...In this paper,based on the observation data of air temperature during 1951-2009 in Shenyang,the interannual and interdecadal variation of annual average temperature,maximum and minimum temperature in Shenyang were conducted the statistical analysis by means of linear trend estimation and mutation detection by using Mann-Kendall method.As was demonstrated in the results,the annual average temperature,maximum and minimum temperature in Shenyang showed an upward trend,whose linear tendency rate was 0.231,0.181 and 0.218 respectively.The increment trend of annual average temperature,maximum and minimum temperature was extremely clear.The increase in minimum temperature was more significant than that in mean temperature and maximum temperature.The abrupt change point of annual mean temperature in Shenyang appeared in 1981;the abrupt change point of annual mean maximum temperature appeared in 1994;the annual mean minimum temperature underwent mutation in 1978.展开更多
According to the need of project design for offshore engineering and coastal engineering, this paper statistically analyses the annual extreme data of waves acquired at 19 observation stations along the coast of China...According to the need of project design for offshore engineering and coastal engineering, this paper statistically analyses the annual extreme data of waves acquired at 19 observation stations along the coast of China. Five kinds of distribution curves are adopted: Pearson III (P-III), Log-Extreme I (LE), Log-Normal(LN), Weibull(W) and Exponential Γ(EΓ) to check the adaptability to the long-term distribution of annual extreme of wave in the China Sea areas. The New Curve Fitting Method (NFIT) method and Probability Weighted Moments (PWM) method are used to estimate the distribution parameters and thereby to derive the design wave parameters with different return periods at 19 observation stations. The test results show that by combining EΓ distribution and NFIT parameter estimation and optimum seeking by computer, the design wave parameters can be estimated with high accuracy, high speed and high efficiency, and the randomness of the estimated results can be avoided.展开更多
基金National Natural Science Foundation of China(No.12271261)Postgraduate Research and Practice Innovation Program of Jiangsu Province,China(Grant No.SJCX230368)。
文摘Normality testing is a fundamental hypothesis test in the statistical analysis of key biological indicators of diabetes.If this assumption is violated,it may cause the test results to deviate from the true value,leading to incorrect inferences and conclusions,and ultimately affecting the validity and accuracy of statistical inferences.Considering this,the study designs a unified analysis scheme for different data types based on parametric statistical test methods and non-parametric test methods.The data were grouped according to sample type and divided into discrete data and continuous data.To account for differences among subgroups,the conventional chi-squared test was used for discrete data.The normal distribution is the basis of many statistical methods;if the data does not follow a normal distribution,many statistical methods will fail or produce incorrect results.Therefore,before data analysis and modeling,the data were divided into normal and non-normal groups through normality testing.For normally distributed data,parametric statistical methods were used to judge the differences between groups.For non-normal data,non-parametric tests were employed to improve the accuracy of the analysis.Statistically significant indicators were retained according to the significance index P-value of the statistical test or corresponding statistics.These indicators were then combined with relevant medical background to further explore the etiology leading to the occurrence or transformation of diabetes status.
文摘Choosing appropriate statistical tests is crucial but deciding which tests to use can be challenging. Different tests suit different types of data and research questions, so it is important to choose the right one. Knowing how to select an appropriate test can lead to more accurate results. Invalid results and misleading conclusions may be drawn from a study if an incorrect statistical test is used. Therefore, to avoid these it is essential to understand the nature of the data, the research question, and the assumptions of the tests before selecting one. This is because there are a wide variety of tests available. This paper provides a step-by-step approach to selecting the right statistical test for any study, with an explanation of when it is appropriate to use it and relevant examples of each statistical test. Furthermore, this guide provides a comprehensive overview of the assumptions of each test and what to do if these assumptions are violated.
文摘Choosing appropriate statistical tests is crucial but deciding which tests to use can be challenging. Different tests suit different types of data and research questions, so it is important to choose the right one. Knowing how to select an appropriate test can lead to more accurate results. Invalid results and misleading conclusions may be drawn from a study if an incorrect statistical test is used. Therefore, to avoid these it is essential to understand the nature of the data, the research question, and the assumptions of the tests before selecting one. This is because there are a wide variety of tests available. This paper provides a step-by-step approach to selecting the right statistical test for any study, with an explanation of when it is appropriate to use it and relevant examples of each statistical test. Furthermore, this guide provides a comprehensive overview of the assumptions of each test and what to do if these assumptions are violated.
基金supported in part by NIH grants R01NS39600,U01MH114829RF1MH128693(to GAA)。
文摘Many fields,such as neuroscience,are experiencing the vast prolife ration of cellular data,underscoring the need fo r organizing and interpreting large datasets.A popular approach partitions data into manageable subsets via hierarchical clustering,but objective methods to determine the appropriate classification granularity are missing.We recently introduced a technique to systematically identify when to stop subdividing clusters based on the fundamental principle that cells must differ more between than within clusters.Here we present the corresponding protocol to classify cellular datasets by combining datadriven unsupervised hierarchical clustering with statistical testing.These general-purpose functions are applicable to any cellular dataset that can be organized as two-dimensional matrices of numerical values,including molecula r,physiological,and anatomical datasets.We demonstrate the protocol using cellular data from the Janelia MouseLight project to chara cterize morphological aspects of neurons.
文摘“Human-elephant conflict(HEC)”,the alarming issue,in present day context has attracted the attention of environmentalists and policy makers.The rising conflict between human beings and wild elephants is common in Buxa Tiger Reserve(BTR)and its adjoining area in West Bengal State,India,making the area volatile.People’s attitudes towards elephant conservation activity are very crucial to get rid of HEC,because people’s proximity with wild elephants’habitat can trigger the occurrence of HEC.The aim of this study is to conduct an in-depth investigation about the association of people’s attitudes towards HEC with their locational,demographic,and socio-economic characteristics in BTR and its adjoining area by using Pearson’s bivariate chi-square test and binary logistic regression analysis.BTR is one of the constituent parts of Eastern Doors Elephant Reserve(EDER).We interviewed 500 respondents to understand their perceptions to HEC and investigated their locational,demographic,and socio-economic characteristics including location of village,gender,age,ethnicity,religion,caste,poverty level,education level,primary occupation,secondary occupation,household type,and source of firewood.The results indicate that respondents who are living in enclave forest villages(EFVs),peripheral forest villages(PFVs),corridor village(CVs),or forest and corridor villages(FCVs),mainly males,at the age of 18–48 years old,engaged with agriculture occupation,and living in kancha and mixed houses,have more likelihood to witness HEC.Besides,respondents who are illiterate or at primary education level are more likely to regard elephant as a main problematic animal around their villages and refuse to participate in elephant conservation activity.For the sake of a sustainable environment for both human beings and wildlife,people’s attitudes towards elephants must be friendly in a more prudent way,so that the two communities can live in harmony.
文摘Sunshine duration (S) based empirical equations have been employed in this study to estimate the daily global solar radiation on a horizontal surface (G) for six meteorological stations in Burundi. Those equations include the Ångström-Prescott linear model and four amongst its derivatives, i.e. logarithmic, exponential, power and quadratic functions. Monthly mean values of daily global solar radiation and sunshine duration data for a period of 20 to 23 years, from the Geographical Institute of Burundi (IGEBU), have been used. For any of the six stations, ten single or double linear regressions have been developed from the above-said five functions, to relate in terms of monthly mean values, the daily clearness index () to each of the next two kinds of relative sunshine duration (RSD): and . In those ratios, G<sub>0</sub>, S<sub>0 </sub>and stand for the extraterrestrial daily solar radiation on a horizontal surface, the day length and the modified day length taking into account the natural site’s horizon, respectively. According to the calculated mean values of the clearness index and the RSD, each station experiences a high number of fairly clear (or partially cloudy) days. Estimated values of the dependent variable (y) in each developed linear regression, have been compared to measured values in terms of the coefficients of correlation (R) and of determination (R<sub>2</sub>), the mean bias error (MBE), the root mean square error (RMSE) and the t-statistics. Mean values of these statistical indicators have been used to rank, according to decreasing performance level, firstly the ten developed equations per station on account of the overall six stations, secondly the six stations on account of the overall ten equations. Nevertheless, the obtained values of those indicators lay in the next ranges for all the developed sixty equations:;;;, with . These results lead to assert that any of the sixty developed linear regressions (and thus equations in terms of and ), fits very adequately measured data, and should be used to estimate monthly average daily global solar radiation with sunshine duration for the relevant station. It is also found that using as RSD, is slightly more advantageous than using for estimating the monthly average daily clearness index, . Moreover, values of statistical indicators of this study match adequately data from other works on the same kinds of empirical equations.
文摘As the increasing popularity and complexity of Web applications and the emergence of their new characteristics, the testing and maintenance of large, complex Web applications are becoming more complex and difficult. Web applications generally contain lots of pages and are used by enormous users. Statistical testing is an effective way of ensuring their quality. Web usage can be accurately described by Markov chain which has been proved to be an ideal model for software statistical testing. The results of unit testing can be utilized in the latter stages, which is an important strategy for bottom-to-top integration testing, and the other improvement of extended Markov chain model (EMM) is to present the error type vector which is treated as a part of page node. this paper also proposes the algorithm for generating test cases of usage paths. Finally, optional usage reliability evaluation methods and an incremental usability regression testing model for testing and evaluation are presented. Key words statistical testing - evaluation for Web usability - extended Markov chain model (EMM) - Web log mining - reliability evaluation CLC number TP311. 5 Foundation item: Supported by the National Defence Research Project (No. 41315. 9. 2) and National Science and Technology Plan (2001BA102A04-02-03)Biography: MAO Cheng-ying (1978-), male, Ph.D. candidate, research direction: software testing. Research direction: advanced database system, software testing, component technology and data mining.
基金National Natural Science Foundation of China(1047112610671176).
文摘In this article, the unit root test for AR(p) model with GARCH errors is considered. The Dickey-Fuller test statistics are rewritten in the form of self-normalized sums, and the asymptotic distribution of the test statistics is derived under the weak conditions.
文摘This paper describes the statistical methods of the comparison of the incidence or mortality rates in cancer registry and descriptive epidemiology, and the features of microcomputer program (CANTEST) which was designed to perform the methods. The program was written in IBM BASIC language. Using the program CANTEST we presented here the user can do several statistical tests or estimations as follow: 1. the comparison of the adjusted rates which were calculated by directly or indirectly standardized methods, 2. the calculation of the slope of regression line for testing the linear trends of the adjusted rates, 3. the estimation of the 95% or 99%conndence intervals of the directly adjusted rates, of the cumulative rates (0-64 and 0-74), and of the cumulative risk. Several examples are presented for testing the performances of the program.
文摘We are very grateful for the letter written by Dr Lange,and indeed apologize for the mistakes noted in the word-ing of our text regarding statistical analysis.This wasdue to changes carried out while revising the manuscriptat the request of reviewers,whom we thank for,point-ing out several issues that were actually similar to thosenoted by Dr.Lange.Unfortunately,we were unable todescribe and discuss our findings properly in the context
基金Publication of this article in an open access journal was funded by the Portland State University Library’s Open Access Fund.
文摘Search-based statistical structural testing(SBSST)is a promising technique that uses automated search to construct input distributions for statistical structural testing.It has been proved that a simple search algorithm,for example,the hill-climber is able to optimize an input distribution.However,due to the noisy fitness estimation of the minimum triggering probability among all cover elements(Tri-Low-Bound),the existing approach does not show a satisfactory efficiency.Constructing input distributions to satisfy the Tri-Low-Bound criterion requires an extensive computation time.Tri-Low-Bound is considered a strong criterion,and it is demonstrated to sustain a high fault-detecting ability.This article tries to answer the following question:if we use a relaxed constraint that significantly reduces the time consumption on search,can the optimized input distribution still be effective in faultdetecting ability?In this article,we propose a type of criterion called fairnessenhanced-sum-of-triggering-probability(p-L1-Max).The criterion utilizes the sum of triggering probabilities as the fitness value and leverages a parameter p to adjust the uniformness of test data generation.We conducted extensive experiments to compare the computation time and the fault-detecting ability between the two criteria.The result shows that the 1.0-L1-Max criterion has the highest efficiency,and it is more practical to use than the Tri-Low-Bound criterion.To measure a criterion’s fault-detecting ability,we introduce a definition of expected faults found in the effective test set size region.To measure the effective test set size region,we present a theoretical analysis of the expected faults found with respect to various test set sizes and use the uniform distribution as a baseline to derive the effective test set size region’s definition.
基金supported by the National Natural Science Foundation of China(NSFC, No. 10473023)Scientific Research Fund of the Sichuan Provincial Education Department,the K. C. Wong Education Foundation (Hong Kong)the Jiangsu Planned Projects for PostdoctoralResearch Funds.
文摘We investigate redshift distributions of three long burst samples, with the first sample containing 131 long bursts with observed redshifts, the second including 220 long bursts with pseudo-redshifts calculated by the variability-luminosity relation, and the third including 1194 long bursts with pseudo-redshifls calculated by the lag-luminosity relation, respectively. In the redshift range 0-1 the Kolmogorov-Smirnov probability of the observed redshift distribution and that of the variability-luminosity relation is large. In the redshift ranges 1-2, 2-3, 3-6.3 and 0-37, the Kolmogorov-Smirnov probabilities of the redshift distribution from lag-luminosity relation and the observed redshift distribution are also large. For the GRBs, which appear both in the two pseudo-redshift burst samples, the KS probability of the pseudo-redshift distribution from the lag-luminosity relation and the observed reshift distribution is 0.447, which is very large. Based on these results, some conclusions are drawn: i) the V-Liso relation might be more believable than the τ-Liso relation in low redshift ranges and the τ-Liso relation might be more real than the V-Liso relation in high redshift ranges; ii) if we do not consider the redshift ranges, the τ-Liso relation might be more physical and intrinsical than the V-Liso relation.
文摘This paper investigates the correlation between tidal stress and earthquakes for periods ranging from hours to months in the limited zone of the Palu region(Central Sulawesi,Indonesia).Through Schuster and binomial tests,we examined the relation between the seismicity(time density of seismic events)and tidal potential arising from the Moon and Sun,using all tidal components simultaneously and focusing on the estimation of specific terms.The results show significant correlations between the seismicity and tidal potential for S2(0.5 d)and O1(1.075 d)tidal components in the case of solely isolated earthquake events,particularly for shallow earthquakes.Meanwhile,there is a strong relationship between aftershocks and tidal components larger than the Mf period(13.661 d).Finally,the analysis of the temporal variation of the earthquake-tide relation reveals an optimal correlation for about six years before the 2018 great Palu earthquake.The correlation becomes insignificant afterwards.
文摘This paper develops the modeling of wind speed by Weibull distribution in the intention to evaluate wind energy potential and help for designing small wind energy plant in Batouri in Cameroon. The Weibull distribution model was developed using wind speed data collected from a metrological station at the small Airport of Batouri. Four numerical methods (Moment method, Graphical method, Empirical method and Energy pattern factor method) were used to estimate weibull parameters K and C. The application of these four methods is effective using a sample wind speed data set. With some statistical analysis, a comparison of the accuracy of each method is also performed. The study helps to determine that Energy pattern factor method is the most effective (K = 3.8262 and C = 2.4659).
文摘The aim of this study is to establish the prevailing conditions of changing climatic trends and change point dates in four selected meteorological stations of Uyo, Benin, Port Harcourt, and Warri in the Niger Delta region of Nigeria. Using daily or 24-hourly annual maximum series (AMS) data with the Indian Meteorological Department (IMD) and the modified Chowdury Indian Meteorological Department (MCIMD) models were adopted to downscale the time series data. Mann-Kendall (MK) trend and Sen’s Slope Estimator (SSE) test showed a statistically significant trend for Uyo and Benin, while Port Harcourt and Warri showed mild trends. The Sen’s Slope magnitude and variation rate were 21.6, 10.8, 6.00 and 4.4 mm/decade, respectively. The trend change-point analysis showed the initial rainfall change-point dates as 2002, 2005, 1988, and 2000 for Uyo, Benin, Port Harcourt, and Warri, respectively. These prove positive changing climatic conditions for rainfall in the study area. Erosion and flood control facilities analysis and design in the Niger Delta will require the application of Non-stationary IDF modelling.
文摘The study focused on the detection of indicators of climate change in 24-hourly annual maximum series (AMS) rainfall data collected for 36 years (1982-2017) for Warri Township, using different statistical methods yielded a statistically insignificant positive mild trend. The IMD and MCIMD downscaled model’s time series data respectively produced MK statistics varying from 1.403 to 1.4729, and 1.403 to 1.463 which were less than the critical Z-value of 1.96. Also, the slope magnitude obtained showed a mild increasing trend in variation from 0.0189 to 0.3713, and 0.0175 to 0.5426, with the rate of change in rainfall intensity at 24 hours duration as 0.4536 and 0.42 mm/hr.year (4.536 and 4.2 mm/decade) for the IMD and the MCIMD time series data, respectively. The trend change point date occurred in the year 2000 from the distribution-free CUSUM test with the trend maintaining a significant and steady increase from 2010 to 2015. Thus, this study established the existence of a trend, which is an indication of a changing climate, and satisfied the condition for rainfall Non-stationary intensity-duration-frequency (NS-IDF) modeling required for infrastructural design for combating flooding events.
基金supported by the National Natural Science Foundation of China(No.12271370)。
文摘For the functional partially linear models including flexible nonparametric part and functional linear part,the estimators of the nonlinear function and the slope function have been studied in existing literature.How to test the correlation between response and explanatory variables,however,still seems to be missing.Therefore,a test procedure for testing the linearity in the functional partially linear models will be proposed in this paper.A test statistic is constructed based on the existing estimators of the nonlinear and the slope functions.Further,we prove that the approximately asymptotic distribution of the proposed statistic is a chi-squared distribution under some regularity conditions.Finally,some simulation studies and a real data application are presented to demonstrate the performance of the proposed test statistic.
基金supported by the National Natural Science Foundation of China under Grant Nos.11971064,12371262,and 12171374。
文摘Exclusive hypothesis testing is a new and special class of hypothesis testing.This kind of testing can be applied in survival analysis to understand the association between genomics information and clinical information about the survival time.Besides,it is well known that Cox's proportional hazards model is the most commonly used model for regression analysis of failure time.In this paper,the authors consider doing the exclusive hypothesis testing for Cox's proportional hazards model with right-censored data.The authors propose the comprehensive test statistics to make decision,and show that the corresponding decision rule can control the asymptotic TypeⅠerrors and have good powers in theory.The numerical studies indicate that the proposed approach works well for practical situations and it is applied to a set of real data arising from Rotterdam Breast Cancer Data study that motivated this study.
基金Supported by the Infrastructure Project of China Meteorological Administration(CMA) in 2010~~
文摘In this paper,based on the observation data of air temperature during 1951-2009 in Shenyang,the interannual and interdecadal variation of annual average temperature,maximum and minimum temperature in Shenyang were conducted the statistical analysis by means of linear trend estimation and mutation detection by using Mann-Kendall method.As was demonstrated in the results,the annual average temperature,maximum and minimum temperature in Shenyang showed an upward trend,whose linear tendency rate was 0.231,0.181 and 0.218 respectively.The increment trend of annual average temperature,maximum and minimum temperature was extremely clear.The increase in minimum temperature was more significant than that in mean temperature and maximum temperature.The abrupt change point of annual mean temperature in Shenyang appeared in 1981;the abrupt change point of annual mean maximum temperature appeared in 1994;the annual mean minimum temperature underwent mutation in 1978.
基金This paper is financially supported by the Ministry of Water Conservancy and Electric Power,P.R.China
文摘According to the need of project design for offshore engineering and coastal engineering, this paper statistically analyses the annual extreme data of waves acquired at 19 observation stations along the coast of China. Five kinds of distribution curves are adopted: Pearson III (P-III), Log-Extreme I (LE), Log-Normal(LN), Weibull(W) and Exponential Γ(EΓ) to check the adaptability to the long-term distribution of annual extreme of wave in the China Sea areas. The New Curve Fitting Method (NFIT) method and Probability Weighted Moments (PWM) method are used to estimate the distribution parameters and thereby to derive the design wave parameters with different return periods at 19 observation stations. The test results show that by combining EΓ distribution and NFIT parameter estimation and optimum seeking by computer, the design wave parameters can be estimated with high accuracy, high speed and high efficiency, and the randomness of the estimated results can be avoided.