As shallow resources are increasingly depleted,the mechanics'theory and testing technology of deep insitu rock has become urgent.Traditional coring technologies obtain rock samples without retaining the in-situ en...As shallow resources are increasingly depleted,the mechanics'theory and testing technology of deep insitu rock has become urgent.Traditional coring technologies obtain rock samples without retaining the in-situ environmental conditions,leading to distortion of the measured parameters.Herein,a coring and testing systems retaining in-situ geological conditions is presented:the coring system that obtains in-situ rock samples,and the transfer and testing system that stores and analyzes the rocks under a reconstructed environment.The ICP-Coring system mainly consists of the pressure controller,active insulated core reactor and insulation layer and sealing film.The ultimate bearing strength of 100 MPa for pressurepreservation,temperature control accuracy of 0.97%for temperature-retained are realized.CH_(4)and CO permeability of the optimized sealing film are as low as 3.85 and 0.33 ppm/min.The average tensile elongation of the film is 152.4%and the light transmittance is reduced to 0%.Additionally,the pressure and steady-state temperature accuracy for reconstructing the in-situ environment of transfer and storage system up to 1%and±0.2 is achieved.The error recorded of the noncontact sensor ring made of lowdensity polymer is less than 6%than that of the contact test.The system can provide technical support for the deep in-situ rock mechanics research,improving deep resource acquisition capabilities and further clarifying deep-earth processes.展开更多
Nonparametric time-of-arrival(TOA) estimators for impulse radio ultra-wideband(IR-UWB) signals are proposed. Nonparametric detection is obviously useful in situations where detailed information about the statistic...Nonparametric time-of-arrival(TOA) estimators for impulse radio ultra-wideband(IR-UWB) signals are proposed. Nonparametric detection is obviously useful in situations where detailed information about the statistics of the noise is unavailable or not accurate. Such TOA estimators are obtained based on conditional statistical tests with only a symmetry distribution assumption on the noise probability density function. The nonparametric estimators are attractive choices for low-resolution IR-UWB digital receivers which can be implemented by fast comparators or high sampling rate low resolution analog-to-digital converters(ADCs),in place of high sampling rate high resolution ADCs which may not be available in practice. Simulation results demonstrate that nonparametric TOA estimators provide more effective and robust performance than typical energy detection(ED) based estimators.展开更多
It's a well-known fact that constraint-based algorithms for learning Bayesian network(BN) structure reckon on a large number of conditional independence(C1) tests.Therefore,it is difficult to learn a BN for indica...It's a well-known fact that constraint-based algorithms for learning Bayesian network(BN) structure reckon on a large number of conditional independence(C1) tests.Therefore,it is difficult to learn a BN for indicating the original causal relations in the true graph.In this paper,a two-phase method for learning equivalence class of BN is introduced.The first phase of the method learns a skeleton of the BN by CI tests.In this way,it reduces the number of tests compared with other existing algorithms and decreases the running time drastically.The second phase of the method orients edges that exist in all BN equivalence classes.Our method is tested on the ALARM network and experimental results show that our approach outperforms the other algorithms.展开更多
The main purpose in many randomized trials is to make an inference about the average causal effect of a treatment. Therefore, on a binary outcome, the null hypothesis for the hypothesis test should be that the causal ...The main purpose in many randomized trials is to make an inference about the average causal effect of a treatment. Therefore, on a binary outcome, the null hypothesis for the hypothesis test should be that the causal risks are equal in the two groups. This null hypothesis is referred to as the weak causal null hypothesis. Nevertheless, at present, hypothesis tests applied in actual randomized trials are not for this null hypothesis;Fisher’s exact test is a test for the sharp causal null hypothesis that the causal effect of treatment is the same for all subjects. In general, the rejection of the sharp causal null hypothesis does not mean that the weak causal null hypothesis is rejected. Recently, Chiba developed new exact tests for the weak causal null hypothesis: a conditional exact test, which requires that a marginal total is fixed, and an unconditional exact test, which does not require that a marginal total is fixed and depends rather on the ratio of random assignment. To apply these exact tests in actual randomized trials, it is inevitable that the sample size calculation must be performed during the study design. In this paper, we present a sample size calculation procedure for these exact tests. Given the sample size, the procedure can derive the exact test power, because it examines all the patterns that can be obtained as observed data under the alternative hypothesis without large sample theories and any assumptions.展开更多
Bayesian network is a popular approach to uncertainty knowledge representation and reasoning. Structure learning is the first step to learn a Bayesian network. Score-based methods are one of the most popular ways of l...Bayesian network is a popular approach to uncertainty knowledge representation and reasoning. Structure learning is the first step to learn a Bayesian network. Score-based methods are one of the most popular ways of learning the structure. In most cases, the score of Bayesian network is defined as adding the log-likelihood score and complexity score by using the penalty function. If the penalty function is set unreasonably, it may hurt the performance of structure search. Thus, Bayesian network structure learning is essentially a bi-objective optimization problem. However, the existing bi-objective structure learning algorithms can only be applied to small-scale networks. To this end, this paper proposes a bi-objective evolutionary Bayesian network structure learning algorithm via skeleton constraint (BBS) for the medium-scale networks. To boost the performance of searching, BBS introduces the random order prior (ROP) initial operator. ROP generates a skeleton to constrain the searching space, which is the key to expanding the scale of structure learning problems. Then, the acyclic structures are guaranteed by adding the orders of variables in the initial skeleton. After that, BBS designs the Pareto rank based crossover and skeleton guided mutation operators. The operators operate on the skeleton obtained in ROP to make the search more targeted. Finally, BBS provides a strategy to choose the final solution. The experimental results show that BBS can always find the structure which is closer to the ground truth compared with the single-objective structure learning methods. Furthermore, compared with the existing bi-objective structure learning methods, BBS is scalable and can be applied to medium-scale Bayesian network datasets. On the educational problem of discovering the influencing factors of students’ academic performance, BBS provides higher quality solutions and is featured with the flexibility of solution selection compared with the widely-used Bayesian network structure learning methods.展开更多
Learning Bayesian network structure is one of the most exciting challenges in machine learning. Discovering a correct skeleton of a directed acyclic graph(DAG) is the foundation for dependency analysis algorithms fo...Learning Bayesian network structure is one of the most exciting challenges in machine learning. Discovering a correct skeleton of a directed acyclic graph(DAG) is the foundation for dependency analysis algorithms for this problem. Considering the unreliability of high order condition independence(CI) tests, and to improve the efficiency of a dependency analysis algorithm, the key steps are to use few numbers of CI tests and reduce the sizes of conditioning sets as much as possible. Based on these reasons and inspired by the algorithm PC, we present an algorithm, named fast and efficient PC(FEPC), for learning the adjacent neighbourhood of every variable. FEPC implements the CI tests by three kinds of orders, which reduces the high order CI tests significantly. Compared with current algorithm proposals, the experiment results show that FEPC has better accuracy with fewer numbers of condition independence tests and smaller size of conditioning sets. The highest reduction percentage of CI test is 83.3% by EFPC compared with PC algorithm.展开更多
基金supported by the Program for Guangdong Introducing Innovative and Enterpreneurial Teams(No.2019ZT08G315)National Natural Science Foundation of China(No.51827901,U2013603,and 52004166)。
文摘As shallow resources are increasingly depleted,the mechanics'theory and testing technology of deep insitu rock has become urgent.Traditional coring technologies obtain rock samples without retaining the in-situ environmental conditions,leading to distortion of the measured parameters.Herein,a coring and testing systems retaining in-situ geological conditions is presented:the coring system that obtains in-situ rock samples,and the transfer and testing system that stores and analyzes the rocks under a reconstructed environment.The ICP-Coring system mainly consists of the pressure controller,active insulated core reactor and insulation layer and sealing film.The ultimate bearing strength of 100 MPa for pressurepreservation,temperature control accuracy of 0.97%for temperature-retained are realized.CH_(4)and CO permeability of the optimized sealing film are as low as 3.85 and 0.33 ppm/min.The average tensile elongation of the film is 152.4%and the light transmittance is reduced to 0%.Additionally,the pressure and steady-state temperature accuracy for reconstructing the in-situ environment of transfer and storage system up to 1%and±0.2 is achieved.The error recorded of the noncontact sensor ring made of lowdensity polymer is less than 6%than that of the contact test.The system can provide technical support for the deep in-situ rock mechanics research,improving deep resource acquisition capabilities and further clarifying deep-earth processes.
基金supported by the National High Technology Research and Development Program of China(863 Program)(2009AA011204)
文摘Nonparametric time-of-arrival(TOA) estimators for impulse radio ultra-wideband(IR-UWB) signals are proposed. Nonparametric detection is obviously useful in situations where detailed information about the statistics of the noise is unavailable or not accurate. Such TOA estimators are obtained based on conditional statistical tests with only a symmetry distribution assumption on the noise probability density function. The nonparametric estimators are attractive choices for low-resolution IR-UWB digital receivers which can be implemented by fast comparators or high sampling rate low resolution analog-to-digital converters(ADCs),in place of high sampling rate high resolution ADCs which may not be available in practice. Simulation results demonstrate that nonparametric TOA estimators provide more effective and robust performance than typical energy detection(ED) based estimators.
文摘It's a well-known fact that constraint-based algorithms for learning Bayesian network(BN) structure reckon on a large number of conditional independence(C1) tests.Therefore,it is difficult to learn a BN for indicating the original causal relations in the true graph.In this paper,a two-phase method for learning equivalence class of BN is introduced.The first phase of the method learns a skeleton of the BN by CI tests.In this way,it reduces the number of tests compared with other existing algorithms and decreases the running time drastically.The second phase of the method orients edges that exist in all BN equivalence classes.Our method is tested on the ALARM network and experimental results show that our approach outperforms the other algorithms.
文摘The main purpose in many randomized trials is to make an inference about the average causal effect of a treatment. Therefore, on a binary outcome, the null hypothesis for the hypothesis test should be that the causal risks are equal in the two groups. This null hypothesis is referred to as the weak causal null hypothesis. Nevertheless, at present, hypothesis tests applied in actual randomized trials are not for this null hypothesis;Fisher’s exact test is a test for the sharp causal null hypothesis that the causal effect of treatment is the same for all subjects. In general, the rejection of the sharp causal null hypothesis does not mean that the weak causal null hypothesis is rejected. Recently, Chiba developed new exact tests for the weak causal null hypothesis: a conditional exact test, which requires that a marginal total is fixed, and an unconditional exact test, which does not require that a marginal total is fixed and depends rather on the ratio of random assignment. To apply these exact tests in actual randomized trials, it is inevitable that the sample size calculation must be performed during the study design. In this paper, we present a sample size calculation procedure for these exact tests. Given the sample size, the procedure can derive the exact test power, because it examines all the patterns that can be obtained as observed data under the alternative hypothesis without large sample theories and any assumptions.
基金supported by the Fundamental Research Funds for the Central Universities,the Science and Technology Commission of Shanghai Municipality(No.19511120601)the Scientific and Technological Innovation 2030 Major Projects(No.2018AAA0100902)+1 种基金the CCF-AFSG Research Fund(No.CCF-AFSG RF20220205)the“Chenguang Program”sponsored by Shanghai Education Development Foundation and Shanghai Municipal Education Commission(No.21CGA32).
文摘Bayesian network is a popular approach to uncertainty knowledge representation and reasoning. Structure learning is the first step to learn a Bayesian network. Score-based methods are one of the most popular ways of learning the structure. In most cases, the score of Bayesian network is defined as adding the log-likelihood score and complexity score by using the penalty function. If the penalty function is set unreasonably, it may hurt the performance of structure search. Thus, Bayesian network structure learning is essentially a bi-objective optimization problem. However, the existing bi-objective structure learning algorithms can only be applied to small-scale networks. To this end, this paper proposes a bi-objective evolutionary Bayesian network structure learning algorithm via skeleton constraint (BBS) for the medium-scale networks. To boost the performance of searching, BBS introduces the random order prior (ROP) initial operator. ROP generates a skeleton to constrain the searching space, which is the key to expanding the scale of structure learning problems. Then, the acyclic structures are guaranteed by adding the orders of variables in the initial skeleton. After that, BBS designs the Pareto rank based crossover and skeleton guided mutation operators. The operators operate on the skeleton obtained in ROP to make the search more targeted. Finally, BBS provides a strategy to choose the final solution. The experimental results show that BBS can always find the structure which is closer to the ground truth compared with the single-objective structure learning methods. Furthermore, compared with the existing bi-objective structure learning methods, BBS is scalable and can be applied to medium-scale Bayesian network datasets. On the educational problem of discovering the influencing factors of students’ academic performance, BBS provides higher quality solutions and is featured with the flexibility of solution selection compared with the widely-used Bayesian network structure learning methods.
基金Supported by the National Natural Science Foundation of China(61403290,11301408,11401454)the Foundation for Youths of Shaanxi Province(2014JQ1020)+1 种基金the Foundation of Baoji City(2013R7-3)the Foundation of Baoji University of Arts and Sciences(ZK15081)
文摘Learning Bayesian network structure is one of the most exciting challenges in machine learning. Discovering a correct skeleton of a directed acyclic graph(DAG) is the foundation for dependency analysis algorithms for this problem. Considering the unreliability of high order condition independence(CI) tests, and to improve the efficiency of a dependency analysis algorithm, the key steps are to use few numbers of CI tests and reduce the sizes of conditioning sets as much as possible. Based on these reasons and inspired by the algorithm PC, we present an algorithm, named fast and efficient PC(FEPC), for learning the adjacent neighbourhood of every variable. FEPC implements the CI tests by three kinds of orders, which reduces the high order CI tests significantly. Compared with current algorithm proposals, the experiment results show that FEPC has better accuracy with fewer numbers of condition independence tests and smaller size of conditioning sets. The highest reduction percentage of CI test is 83.3% by EFPC compared with PC algorithm.