期刊文献+
共找到4篇文章
< 1 >
每页显示 20 50 100
CROSSING TIME AND RENEWAL NUMBERS OF TWO PH-RENEWAL PROCESSES
1
作者 SHI DINGHUA AND MIR G. H. TALPUR 《Applied Mathematics(A Journal of Chinese Universities)》 SCIE CSCD 1995年第2期111-122,共12页
This paper considers how to find some joint distributions and their marginal distributions of crossing time and renewal numbers related to two PH-renewal processes by constructing an absorbing Markov process.
关键词 PH-distribution absorbing markov process PH-renewal process.
下载PDF
ANew Theoretical Framework forAnalyzing Stochastic Global Optimization Algorithms 被引量:1
2
作者 SHI Ding hua PENG Jian ping (College of Sciences, Shanghai University) 《Advances in Manufacturing》 SCIE CAS 1999年第3期175-180,共6页
In this paper, we develop a new theoretical framework by means of the absorbing Markov process theory for analyzing some stochastic global optimization algorithms. Applying the framework to the pure random search, we ... In this paper, we develop a new theoretical framework by means of the absorbing Markov process theory for analyzing some stochastic global optimization algorithms. Applying the framework to the pure random search, we prove that the pure random search converges to the global minimum in probability and its time has geometry distribution. We also analyze the pure adaptive search by this framework and turn out that the pure adaptive search converges to the global minimum in probability and its time has Poisson distribution. 展开更多
关键词 Global optimization stochastic global optimization algorithm random search absorbing markov process
下载PDF
Efficient Temporal Difference Learning with Adaptive λ
3
作者 毕金波 吴沧浦 《Journal of Beijing Institute of Technology》 EI CAS 1999年第3期251-257,共7页
Aim To find a more efficient learning method based on temporal difference learning for delayed reinforcement learning tasks. Methods A kind of Q learning algorithm based on truncated TD( λ ) with adaptive scheme... Aim To find a more efficient learning method based on temporal difference learning for delayed reinforcement learning tasks. Methods A kind of Q learning algorithm based on truncated TD( λ ) with adaptive schemes of λ value selection addressed to absorbing Markov decision processes was presented and implemented on computers. Results and Conclusion Simulations on the shortest path searching problems show that using adaptive λ in the Q learning based on TTD( λ ) can speed up its convergence. 展开更多
关键词 dynamic programming delayed reinforcement learning absorbing markov decision processes temporal difference learning Q learning
下载PDF
Estimates on the amplitude of the first Dirichlet eigenvector in discrete frameworks
4
作者 DIACONIS Persi MICLO Laurent 《Science China Mathematics》 SCIE CSCD 2016年第2期205-226,共22页
Consider a finite absorbing Markov generator, irreducible on the non-absorbing states. PerronFrobenius theory ensures the existence of a corresponding positive eigenvector ψ. The goal of the paper is to give bounds o... Consider a finite absorbing Markov generator, irreducible on the non-absorbing states. PerronFrobenius theory ensures the existence of a corresponding positive eigenvector ψ. The goal of the paper is to give bounds on the amplitude max ψ/ min ψ. Two approaches are proposed: One using a path method and the other one, restricted to the reversible situation, based on spectral estimates. The latter approach is extended to denumerable birth and death processes absorbing at 0 for which infinity is an entrance boundary. The interest of estimating the ratio is the reduction of the quantitative study of convergence to quasi-stationarity to the convergence to equilibrium of related ergodic processes, as seen by Diaconis and Miclo(2014). 展开更多
关键词 finite absorbing markov process first Dirichlet eigenvector path method spectral estimates denumerable absorbing birth and death process entrance boundary
原文传递
上一页 1 下一页 到第
使用帮助 返回顶部