阶梯式碳交易机制以及优化调度模型求解算法是进行园区综合能源系统(community integrated energy system,CIES)优化调度的重要因素,现有文献对这两个因素的考虑不够全面。为此,文中在考虑阶梯式碳交易机制的基础上,提出采用近端策略优...阶梯式碳交易机制以及优化调度模型求解算法是进行园区综合能源系统(community integrated energy system,CIES)优化调度的重要因素,现有文献对这两个因素的考虑不够全面。为此,文中在考虑阶梯式碳交易机制的基础上,提出采用近端策略优化(proximal policy optimization,PPO)算法求解CIES低碳优化调度问题。该方法基于低碳优化调度模型搭建强化学习交互环境,利用设备状态参数及运行参数定义智能体的状态、动作空间及奖励函数,再通过离线训练获取可生成最优策略的智能体。算例分析结果表明,采用PPO算法得到的CIES低碳优化调度方法能够充分发挥阶梯式碳交易机制减少碳排放量和提高能源利用率方面的优势。展开更多
Background: Studies have shown a strong correlation between the growth of E2 in serum and estrone-3-glucuronide (E1-3G) in urine during ovarian stimulation. Thus, we developed theoretical models for using urinary E1-3...Background: Studies have shown a strong correlation between the growth of E2 in serum and estrone-3-glucuronide (E1-3G) in urine during ovarian stimulation. Thus, we developed theoretical models for using urinary E1-3G in ovarian stimulation and focused on their experimental verification and analysis. Methods: A prospective, observational pilot study was conducted involving 54 patients who underwent 54 cycles of ovarian stimulation. The goal was to establish the growth rate of urinary E1-3G during the course of stimulation and to determine the daily upper and lower limits of growth rates at which stimulation is appropriate and safe. Controlled ovarian stimulation was performed using two different stimulation protocols—an antagonist protocol in 25 cases and a progestin-primed ovarian stimulation protocol (PPOS) in 29 cases, with fixed doses of gonadotropins. From the second day of stimulation, patients self-measured their daily urine E1-3G levels at home using a portable analyzer. In parallel, a standard ultrasound follow-up protocol accompanied by a determination of E2, LH, and P levels was applied to optimally control stimulation. Results: The average daily growth rates in both groups were about 50%. The daily increase in E1-3G for the antagonist protocol ranged from 14% to 79%, while they were 28% to 79% for the PPOS protocol. Conclusion: This is the first study to analyze the dynamics of E1-3G in two different protocols and to estimate the limits of its increase during the entire course of the stimulation. The results confirm our theoretical model for the viability of using urinary E1-3G for monitoring ovarian stimulation.展开更多
In Beyond the Fifth Generation(B5G)heterogeneous edge networks,numerous users are multiplexed on a channel or served on the same frequency resource block,in which case the transmitter applies coding and the receiver u...In Beyond the Fifth Generation(B5G)heterogeneous edge networks,numerous users are multiplexed on a channel or served on the same frequency resource block,in which case the transmitter applies coding and the receiver uses interference cancellation.Unfortunately,uncoordinated radio resource allocation can reduce system throughput and lead to user inequity,for this reason,in this paper,channel allocation and power allocation problems are formulated to maximize the system sum rate and minimum user achievable rate.Since the construction model is non-convex and the response variables are high-dimensional,a distributed Deep Reinforcement Learning(DRL)framework called distributed Proximal Policy Optimization(PPO)is proposed to allocate or assign resources.Specifically,several simulated agents are trained in a heterogeneous environment to find robust behaviors that perform well in channel assignment and power allocation.Moreover,agents in the collection stage slow down,which hinders the learning of other agents.Therefore,a preemption strategy is further proposed in this paper to optimize the distributed PPO,form DP-PPO and successfully mitigate the straggler problem.The experimental results show that our mechanism named DP-PPO improves the performance over other DRL methods.展开更多
文摘Background: Studies have shown a strong correlation between the growth of E2 in serum and estrone-3-glucuronide (E1-3G) in urine during ovarian stimulation. Thus, we developed theoretical models for using urinary E1-3G in ovarian stimulation and focused on their experimental verification and analysis. Methods: A prospective, observational pilot study was conducted involving 54 patients who underwent 54 cycles of ovarian stimulation. The goal was to establish the growth rate of urinary E1-3G during the course of stimulation and to determine the daily upper and lower limits of growth rates at which stimulation is appropriate and safe. Controlled ovarian stimulation was performed using two different stimulation protocols—an antagonist protocol in 25 cases and a progestin-primed ovarian stimulation protocol (PPOS) in 29 cases, with fixed doses of gonadotropins. From the second day of stimulation, patients self-measured their daily urine E1-3G levels at home using a portable analyzer. In parallel, a standard ultrasound follow-up protocol accompanied by a determination of E2, LH, and P levels was applied to optimally control stimulation. Results: The average daily growth rates in both groups were about 50%. The daily increase in E1-3G for the antagonist protocol ranged from 14% to 79%, while they were 28% to 79% for the PPOS protocol. Conclusion: This is the first study to analyze the dynamics of E1-3G in two different protocols and to estimate the limits of its increase during the entire course of the stimulation. The results confirm our theoretical model for the viability of using urinary E1-3G for monitoring ovarian stimulation.
基金supported by the Key Research and Development Program of China(No.2022YFC3005401)Key Research and Development Program of China,Yunnan Province(No.202203AA080009,202202AF080003)Postgraduate Research&Practice Innovation Program of Jiangsu Province(No.KYCX21_0482).
文摘In Beyond the Fifth Generation(B5G)heterogeneous edge networks,numerous users are multiplexed on a channel or served on the same frequency resource block,in which case the transmitter applies coding and the receiver uses interference cancellation.Unfortunately,uncoordinated radio resource allocation can reduce system throughput and lead to user inequity,for this reason,in this paper,channel allocation and power allocation problems are formulated to maximize the system sum rate and minimum user achievable rate.Since the construction model is non-convex and the response variables are high-dimensional,a distributed Deep Reinforcement Learning(DRL)framework called distributed Proximal Policy Optimization(PPO)is proposed to allocate or assign resources.Specifically,several simulated agents are trained in a heterogeneous environment to find robust behaviors that perform well in channel assignment and power allocation.Moreover,agents in the collection stage slow down,which hinders the learning of other agents.Therefore,a preemption strategy is further proposed in this paper to optimize the distributed PPO,form DP-PPO and successfully mitigate the straggler problem.The experimental results show that our mechanism named DP-PPO improves the performance over other DRL methods.