The recursion relation of preventive maintenance (PM) cycle is built up concerning the concept of effective age and age setback factor proposed in this paper, which illustrates the dynamic relationship between failure...The recursion relation of preventive maintenance (PM) cycle is built up concerning the concept of effective age and age setback factor proposed in this paper, which illustrates the dynamic relationship between failure rate and preventive maintenance activity. And the nonlinear optimal PM policy model satisfying the reliability constraints in finite time horizon following Weibull distribution is proposed. The model built in this paper avoids the shortcoming of steady analytical PM model in infinite time horizon and can be used to aid scheduling the maintenance plan and providing decision supporting for job shop scheduling.展开更多
In this paper,we study mulit-dimensional oblique reflected backward stochastic differential equations(RBSDEs)in a more general framework over finite or infinite time horizon,corresponding to the pricing problem for a ...In this paper,we study mulit-dimensional oblique reflected backward stochastic differential equations(RBSDEs)in a more general framework over finite or infinite time horizon,corresponding to the pricing problem for a type of real option.We prove that the equation can be solved uniquely in L^(p)(1<p≤2)-space,when the generators are uniformly continuous but each component taking values independently.Furthermore,if the generator of this equation fulfills the infinite time version of Lipschitzian continuity,we can also conclude that the solution to the oblique RBSDE exists and is unique,despite the fact that the values of some generator components may affect one another.展开更多
Under the assumption that the claim size is subexponentially distributed and the insurance surplus is totally invested in risky asset, a simple asymptotic relation of tail probability of discounted aggregate claims fo...Under the assumption that the claim size is subexponentially distributed and the insurance surplus is totally invested in risky asset, a simple asymptotic relation of tail probability of discounted aggregate claims for renewal risk model within finite horizon is obtained. The result extends the corresponding conclusions of related references.展开更多
This paper addresses the simultaneous determination of pricing and inventory replenishment strate- gies under a fluctuating environment. Specifically, we analyze the single item, periodic review model. The demand cons...This paper addresses the simultaneous determination of pricing and inventory replenishment strate- gies under a fluctuating environment. Specifically, we analyze the single item, periodic review model. The demand consists of two parts: the deterministic component, which is influenced by the price, and the stochastic component (perturbation). The distribution of the stochastic component is determined by the current state of an exogenous Markov chain. The price that is charged in any given period can be specified dynamically. A replenishment order may be placed at the beginning of some or all of the periods, and stockouts are fully backlogged. Ordering costs that are lower semicontinuous, and inventory/backlog (or surplus) costs that are continuous with polynomial growth. Finite-horizon and infinite-horizon problems are addressed. Existence of optimal policies is established. Furthermore, optimality of (s,S,p)-type policies is proved when the ordering cost consists of fixed and proportional cost components and the surplus cost (these costs are all state-dependent) is convex.展开更多
In this tutorial paper,the finite-horizon optimal adaptive regulation of linear and nonlinear dynamic systems with unknown system dynamics is presented in a forward-in-time manner using adaptive dynamic programming(AD...In this tutorial paper,the finite-horizon optimal adaptive regulation of linear and nonlinear dynamic systems with unknown system dynamics is presented in a forward-in-time manner using adaptive dynamic programming(ADP).An adaptive estimator(AE)is introduced with the idea of Q-learning to relax the requirement of system dynamics in the case of linear system,while neural network-based identifier is utilised for nonlinear systems.The time-varying nature of the solution to the Bellman/Hamilton–Jacobi–Bellman equation is handled by utilising a time-dependent basis function,while the terminal constraint is incorporated as part of the update law of the AE/Identifier in solving the optimal feedback control.Utilising an initial admissible control,the proposed optimal regulation scheme of the uncertain linear and nonlinear system yields a forward-in-time and online solution without using value and/or policy iterations.An adaptive observer is utilised for linear systems in order to relax the need for state availability so that the optimal adaptive control design depends only on the reconstructed states.Finally,the optimal control is covered for nonlinear-networked control systems where in the feedback loop is closed via a communication network.Effectiveness of the proposed approach is verified by simulation results.The end result is a variant of a roll-out scheme in ADP wherein an initial admissible policy is selected as the base policy and the control policy is enhanced using a one-time policy improvement at each sampling interval.展开更多
基金Natural Science Foundation of China (No. 59889505)
文摘The recursion relation of preventive maintenance (PM) cycle is built up concerning the concept of effective age and age setback factor proposed in this paper, which illustrates the dynamic relationship between failure rate and preventive maintenance activity. And the nonlinear optimal PM policy model satisfying the reliability constraints in finite time horizon following Weibull distribution is proposed. The model built in this paper avoids the shortcoming of steady analytical PM model in infinite time horizon and can be used to aid scheduling the maintenance plan and providing decision supporting for job shop scheduling.
基金supported by the Natural Science Foundation of Shandong Province(Grant Nos.ZR2022MA079 and ZR2021MG049)the National Social Science Funding of China(Grant No.21CJY027)the TianYuan Special Funds of the National Natural Science Foundation of China(Grant No.11626146)。
文摘In this paper,we study mulit-dimensional oblique reflected backward stochastic differential equations(RBSDEs)in a more general framework over finite or infinite time horizon,corresponding to the pricing problem for a type of real option.We prove that the equation can be solved uniquely in L^(p)(1<p≤2)-space,when the generators are uniformly continuous but each component taking values independently.Furthermore,if the generator of this equation fulfills the infinite time version of Lipschitzian continuity,we can also conclude that the solution to the oblique RBSDE exists and is unique,despite the fact that the values of some generator components may affect one another.
基金Supported by the National Natural Science Foundation of China(70871104)the Planning Project of the National Educational Bureau of China(08JA630078)the Project of Key Research Base of Human and Social Sciences(Finance) for Colleges in Zhejiang Province(Grant No. of Academic Education of Zhejiang [2008]255)
文摘Under the assumption that the claim size is subexponentially distributed and the insurance surplus is totally invested in risky asset, a simple asymptotic relation of tail probability of discounted aggregate claims for renewal risk model within finite horizon is obtained. The result extends the corresponding conclusions of related references.
基金Supported by the National Natural Science Foundation of China (No. 70621061,70890082,70671100)Science Foundation of Beijing Jiaotong University (2007RC014)
文摘This paper addresses the simultaneous determination of pricing and inventory replenishment strate- gies under a fluctuating environment. Specifically, we analyze the single item, periodic review model. The demand consists of two parts: the deterministic component, which is influenced by the price, and the stochastic component (perturbation). The distribution of the stochastic component is determined by the current state of an exogenous Markov chain. The price that is charged in any given period can be specified dynamically. A replenishment order may be placed at the beginning of some or all of the periods, and stockouts are fully backlogged. Ordering costs that are lower semicontinuous, and inventory/backlog (or surplus) costs that are continuous with polynomial growth. Finite-horizon and infinite-horizon problems are addressed. Existence of optimal policies is established. Furthermore, optimality of (s,S,p)-type policies is proved when the ordering cost consists of fixed and proportional cost components and the surplus cost (these costs are all state-dependent) is convex.
文摘In this tutorial paper,the finite-horizon optimal adaptive regulation of linear and nonlinear dynamic systems with unknown system dynamics is presented in a forward-in-time manner using adaptive dynamic programming(ADP).An adaptive estimator(AE)is introduced with the idea of Q-learning to relax the requirement of system dynamics in the case of linear system,while neural network-based identifier is utilised for nonlinear systems.The time-varying nature of the solution to the Bellman/Hamilton–Jacobi–Bellman equation is handled by utilising a time-dependent basis function,while the terminal constraint is incorporated as part of the update law of the AE/Identifier in solving the optimal feedback control.Utilising an initial admissible control,the proposed optimal regulation scheme of the uncertain linear and nonlinear system yields a forward-in-time and online solution without using value and/or policy iterations.An adaptive observer is utilised for linear systems in order to relax the need for state availability so that the optimal adaptive control design depends only on the reconstructed states.Finally,the optimal control is covered for nonlinear-networked control systems where in the feedback loop is closed via a communication network.Effectiveness of the proposed approach is verified by simulation results.The end result is a variant of a roll-out scheme in ADP wherein an initial admissible policy is selected as the base policy and the control policy is enhanced using a one-time policy improvement at each sampling interval.