期刊文献+
共找到59,406篇文章
< 1 2 250 >
每页显示 20 50 100
改进Q-Learning的路径规划算法研究 被引量:1
1
作者 宋丽君 周紫瑜 +2 位作者 李云龙 侯佳杰 何星 《小型微型计算机系统》 CSCD 北大核心 2024年第4期823-829,共7页
针对Q-Learning算法学习效率低、收敛速度慢且在动态障碍物的环境下路径规划效果不佳的问题,本文提出一种改进Q-Learning的移动机器人路径规划算法.针对该问题,算法根据概率的突变性引入探索因子来平衡探索和利用以加快学习效率;通过在... 针对Q-Learning算法学习效率低、收敛速度慢且在动态障碍物的环境下路径规划效果不佳的问题,本文提出一种改进Q-Learning的移动机器人路径规划算法.针对该问题,算法根据概率的突变性引入探索因子来平衡探索和利用以加快学习效率;通过在更新函数中设计深度学习因子以保证算法探索概率;融合遗传算法,避免陷入局部路径最优同时按阶段探索最优迭代步长次数,以减少动态地图探索重复率;最后提取输出的最优路径关键节点采用贝塞尔曲线进行平滑处理,进一步保证路径平滑度和可行性.实验通过栅格法构建地图,对比实验结果表明,改进后的算法效率相较于传统算法在迭代次数和路径上均有较大优化,且能够较好的实现动态地图下的路径规划,进一步验证所提方法的有效性和实用性. 展开更多
关键词 移动机器人 路径规划 q-learning算法 平滑处理 动态避障
下载PDF
改进的Q-learning蜂群算法求解置换流水车间调度问题
2
作者 杜利珍 宣自风 +1 位作者 唐家琦 王鑫涛 《组合机床与自动化加工技术》 北大核心 2024年第10期175-180,共6页
针对置换流水车间调度问题,提出了一种基于改进的Q-learning算法的人工蜂群算法。该算法设计了一种改进的奖励函数作为人工蜂群算法的环境,根据奖励函数的优劣来判断下一代种群的寻优策略,并通过Q-learning智能选择人工蜂群算法的蜜源... 针对置换流水车间调度问题,提出了一种基于改进的Q-learning算法的人工蜂群算法。该算法设计了一种改进的奖励函数作为人工蜂群算法的环境,根据奖励函数的优劣来判断下一代种群的寻优策略,并通过Q-learning智能选择人工蜂群算法的蜜源的更新维度数大小,根据选择的维度数大小对编码进行更新,提高了收敛速度和精度,最后使用不同规模的置换流水车间调度问题的实例来验证所提算法的性能,通过对标准实例的计算与其它算法对比,证明该算法的准确性。 展开更多
关键词 q-learning算法 人工蜂群算法 置换流水车间调度
下载PDF
基于改进Q-Learning的移动机器人路径规划算法
3
作者 王立勇 王弘轩 +2 位作者 苏清华 王绅同 张鹏博 《电子测量技术》 北大核心 2024年第9期85-92,共8页
随着移动机器人在生产生活中的深入应用,其路径规划能力也需要向快速性和环境适应性兼备发展。为解决现有移动机器人使用强化学习方法进行路径规划时存在的探索前期容易陷入局部最优、反复搜索同一区域,探索后期收敛率低、收敛速度慢的... 随着移动机器人在生产生活中的深入应用,其路径规划能力也需要向快速性和环境适应性兼备发展。为解决现有移动机器人使用强化学习方法进行路径规划时存在的探索前期容易陷入局部最优、反复搜索同一区域,探索后期收敛率低、收敛速度慢的问题,本研究提出一种改进的Q-Learning算法。该算法改进Q矩阵赋值方法,使迭代前期探索过程具有指向性,并降低碰撞的情况;改进Q矩阵迭代方法,使Q矩阵更新具有前瞻性,避免在一个小区域中反复探索;改进随机探索策略,在迭代前期全面利用环境信息,后期向目标点靠近。在不同栅格地图仿真验证结果表明,本文算法在Q-Learning算法的基础上,通过上述改进降低探索过程中的路径长度、减少抖动并提高收敛的速度,具有更高的计算效率。 展开更多
关键词 路径规划 强化学习 移动机器人 q-learning算法 ε-decreasing策略
下载PDF
基于Q-Learning的航空器滑行路径规划研究
4
作者 王兴隆 王睿峰 《中国民航大学学报》 CAS 2024年第3期28-33,共6页
针对传统算法规划航空器滑行路径准确度低、不能根据整体场面运行情况进行路径规划的问题,提出一种基于Q-Learning的路径规划方法。通过对机场飞行区网络结构模型和强化学习的仿真环境分析,设置了状态空间和动作空间,并根据路径的合规... 针对传统算法规划航空器滑行路径准确度低、不能根据整体场面运行情况进行路径规划的问题,提出一种基于Q-Learning的路径规划方法。通过对机场飞行区网络结构模型和强化学习的仿真环境分析,设置了状态空间和动作空间,并根据路径的合规性和合理性设定了奖励函数,将路径合理性评价值设置为滑行路径长度与飞行区平均滑行时间乘积的倒数。最后,分析了动作选择策略参数对路径规划模型的影响。结果表明,与A*算法和Floyd算法相比,基于Q-Learning的路径规划在滑行距离最短的同时,避开了相对繁忙的区域,路径合理性评价值高。 展开更多
关键词 滑行路径规划 机场飞行区 强化学习 q-learning
下载PDF
基于softmax的加权Double Q-Learning算法 被引量:1
5
作者 钟雨昂 袁伟伟 关东海 《计算机科学》 CSCD 北大核心 2024年第S01期46-50,共5页
强化学习作为机器学习的一个分支,用于描述和解决智能体在与环境的交互过程中,通过学习策略以达成回报最大化的问题。Q-Learning作为无模型强化学习的经典方法,存在过估计引起的最大化偏差问题,并且在环境中奖励存在噪声时表现不佳。Dou... 强化学习作为机器学习的一个分支,用于描述和解决智能体在与环境的交互过程中,通过学习策略以达成回报最大化的问题。Q-Learning作为无模型强化学习的经典方法,存在过估计引起的最大化偏差问题,并且在环境中奖励存在噪声时表现不佳。Double Q-Learning(DQL)的出现解决了过估计问题,但同时造成了低估问题。为解决以上算法的高低估问题,提出了基于softmax的加权Q-Learning算法,并将其与DQL相结合,提出了一种新的基于softmax的加权Double Q-Learning算法(WDQL-Softmax)。该算法基于加权双估计器的构造,对样本期望值进行softmax操作得到权重,使用权重估计动作价值,有效平衡对动作价值的高估和低估问题,使估计值更加接近理论值。实验结果表明,在离散动作空间中,相比于Q-Learning算法、DQL算法和WDQL算法,WDQL-Softmax算法的收敛速度更快且估计值与理论值的误差更小。 展开更多
关键词 强化学习 q-learning Double q-learning Softmax
下载PDF
基于Q-Learning的分簇无线传感网信任管理机制
6
作者 赵远亮 王涛 +3 位作者 李平 吴雅婷 孙彦赞 王瑞 《上海大学学报(自然科学版)》 CAS CSCD 北大核心 2024年第2期255-266,共12页
针对无线传感器网络中存在的安全问题,提出了基于Q-Learning的分簇无线传感网信任管理机制(Q-learning based trust management mechanism for clustered wireless sensor networks,QLTMM-CWSN).该机制主要考虑通信信任、数据信任和能... 针对无线传感器网络中存在的安全问题,提出了基于Q-Learning的分簇无线传感网信任管理机制(Q-learning based trust management mechanism for clustered wireless sensor networks,QLTMM-CWSN).该机制主要考虑通信信任、数据信任和能量信任3个方面.在网络运行过程中,基于节点的通信行为、数据分布和能量消耗,使用Q-Learning算法更新节点信任值,并选择簇内信任值最高的节点作为可信簇头节点.当簇中主簇头节点的信任值低于阈值时,可信簇头节点代替主簇头节点管理簇内成员节点,维护正常的数据传输.研究结果表明,QLTMM-CWSN机制能有效抵御通信攻击、伪造本地数据攻击、能量攻击和混合攻击. 展开更多
关键词 无线传感器网络 q-learning 信任管理机制 网络安全
下载PDF
TPACK框架下GeoScene Online与地理教学融合的实践 被引量:2
7
作者 杨可辛 董雯 《地理教育》 2024年第3期10-14,共5页
技术革新影响学科教学方式选择,TPACK模式为解决当下学科教学应用新技术“两张皮”问题提供了新思路。本文从科勒和米什拉的理论出发,尝试将GeoScene Online平台与地理课堂教学融合,提出循序渐进、跨学科、基于真实情境和交互式的融合原... 技术革新影响学科教学方式选择,TPACK模式为解决当下学科教学应用新技术“两张皮”问题提供了新思路。本文从科勒和米什拉的理论出发,尝试将GeoScene Online平台与地理课堂教学融合,提出循序渐进、跨学科、基于真实情境和交互式的融合原则,进而在TPACK模式下将GeoScene Online功能特点与高中地理必修内容进行融合分析,构建以PCK、TCK、TPK三条子路径为导向的GeoScene Online与地理教学融合模式,并以“耕地”为主题进行案例实践探索。 展开更多
关键词 TPACK理论 融合教学 WEBGIS GeoScene online
下载PDF
基于多步信息辅助的Q-learning路径规划算法
8
作者 王越龙 王松艳 晁涛 《系统仿真学报》 CAS CSCD 北大核心 2024年第9期2137-2148,共12页
为提升静态环境下移动机器人路径规划能力,解决传统Q-learning算法在路径规划中收敛速度慢的问题,提出一种基于多步信息辅助机制的Q-learning改进算法。利用ε-greedy策略中贪婪动作的多步信息与历史最优路径长度更新资格迹,使有效的资... 为提升静态环境下移动机器人路径规划能力,解决传统Q-learning算法在路径规划中收敛速度慢的问题,提出一种基于多步信息辅助机制的Q-learning改进算法。利用ε-greedy策略中贪婪动作的多步信息与历史最优路径长度更新资格迹,使有效的资格迹在算法迭代中持续发挥作用,用保存的多步信息解决可能落入的循环陷阱;使用局部多花朵的花授粉算法初始化Q值表,提升机器人前期搜索效率;基于机器人不同探索阶段的目的,结合迭代路径长度的标准差与机器人成功到达目标点的次数设计动作选择策略,以增强算法对环境信息探索与利用的平衡能力。实验结果表明:该算法具有较快的收敛速度,验证了算法的可行性与有效性。 展开更多
关键词 路径规划 q-learning 收敛速度 动作选择策略 栅格地图
下载PDF
Early warning method for thermal runaway of lithium-ion batteries under thermal abuse condition based on online electrochemical impedance monitoring 被引量:1
9
作者 Yuxuan Li Lihua Jiang +5 位作者 Ningjie Zhang Zesen Wei Wenxin Mei Qiangling Duan Jinhua Sun Qingsong Wang 《Journal of Energy Chemistry》 SCIE EI CAS CSCD 2024年第5期74-86,共13页
Early warning of thermal runaway(TR)of lithium-ion batteries(LIBs)is a significant challenge in current application scenarios.Timely and effective TR early warning technology is urgently required considering the curre... Early warning of thermal runaway(TR)of lithium-ion batteries(LIBs)is a significant challenge in current application scenarios.Timely and effective TR early warning technology is urgently required considering the current fire safety situation of LIBs.In this work,we report an early warning method of TR with online electrochemical impedance spectroscopy(EIS)monitoring,which overcomes the shortcomings of warning methods based on traditional signals such as temperature,gas,and pressure with obvious delay and high cost.With in-situ data acquisition through accelerating rate calorimeter(ARC)-EIS test,the crucial features of TR were extracted using the RReliefF algorithm.TR mechanisms corresponding to the features at specific frequencies were analyzed.Finally,a three-level warning strategy for single battery,series module,and parallel module was formulated,which can successfully send out an early warning signal ahead of the self-heating temperature of battery under thermal abuse condition.The technology can provide a reliable basis for the timely intervention of battery thermal management and fire protection systems and is expected to be applied to electric vehicles and energy storage devices to realize early warning and improve battery safety. 展开更多
关键词 online EIS measurement Lithium-ion batterysafety Multistage thermal runaway early warning SENSITIVITYANALYSIS
下载PDF
Efficient unequal error protection for online fountain codes 被引量:1
10
作者 SHI Pengcheng WANG Zhenyong +1 位作者 LI Dezhi LYU Haibo 《Journal of Systems Engineering and Electronics》 SCIE CSCD 2024年第2期286-293,共8页
In this paper,an efficient unequal error protection(UEP)scheme for online fountain codes is proposed.In the buildup phase,the traversing-selection strategy is proposed to select the most important symbols(MIS).Then,in... In this paper,an efficient unequal error protection(UEP)scheme for online fountain codes is proposed.In the buildup phase,the traversing-selection strategy is proposed to select the most important symbols(MIS).Then,in the completion phase,the weighted-selection strategy is applied to provide low overhead.The performance of the proposed scheme is analyzed and compared with the existing UEP online fountain scheme.Simulation results show that in terms of MIS and the least important symbols(LIS),when the bit error ratio is 10-4,the proposed scheme can achieve 85%and 31.58%overhead reduction,respectively. 展开更多
关键词 online fountain code random graph unequal error protection(UEP) rateless code
下载PDF
多无人机辅助边缘计算场景下基于Q-learning的任务卸载优化
11
作者 张露 王康 +2 位作者 燕晶 张博文 王茂励 《曲阜师范大学学报(自然科学版)》 CAS 2024年第4期74-82,共9页
引入多无人机辅助边缘计算系统,由多个无人机和原有边缘服务器共同为移动用户提供通信和计算资源;将优化问题建模为资源竞争和卸载决策约束下的系统总效用最大化问题,系统总效用由用户满意度、任务延迟和系统能耗3个因素组成.由于优化... 引入多无人机辅助边缘计算系统,由多个无人机和原有边缘服务器共同为移动用户提供通信和计算资源;将优化问题建模为资源竞争和卸载决策约束下的系统总效用最大化问题,系统总效用由用户满意度、任务延迟和系统能耗3个因素组成.由于优化模型是一个具有NP难属性的非凸问题,故采用强化学习方法求解得到使系统总效用最大的最优任务卸载决策集.仿真实验结果表明,与贪心顺序调优卸载方案和随机选择卸载方案相比,该文提出的Q-learning方案的系统总效用分别提高了15%和43%以上. 展开更多
关键词 多无人机辅助边缘计算系统 任务卸载 q-learning算法
下载PDF
Modulated-ISRJ rejection using online dictionary learning for synthetic aperture radar imagery 被引量:1
12
作者 WEI Shaopeng ZHANG Lei +1 位作者 LU Jingyue LIU Hongwei 《Journal of Systems Engineering and Electronics》 SCIE CSCD 2024年第2期316-329,共14页
In electromagnetic countermeasures circumstances,synthetic aperture radar(SAR)imagery usually suffers from severe quality degradation from modulated interrupt sampling repeater jamming(MISRJ),which usually owes consid... In electromagnetic countermeasures circumstances,synthetic aperture radar(SAR)imagery usually suffers from severe quality degradation from modulated interrupt sampling repeater jamming(MISRJ),which usually owes considerable coherence with the SAR transmission waveform together with periodical modulation patterns.This paper develops an MISRJ suppression algorithm for SAR imagery with online dictionary learning.In the algorithm,the jamming modulation temporal properties are exploited with extracting and sorting MISRJ slices using fast-time autocorrelation.Online dictionary learning is followed to separate real signals from jamming slices.Under the learned representation,time-varying MISRJs are suppressed effectively.Both simulated and real-measured SAR data are also used to confirm advantages in suppressing time-varying MISRJs over traditional methods. 展开更多
关键词 synthetic aperture radar(SAR) modulated interrupt sampling jamming(MISRJ) online dictionary learning
下载PDF
基于改进Q-learning算法移动机器人局部路径研究
13
作者 方文凯 廖志高 《计算机与数字工程》 2024年第5期1265-1269,1274,共6页
针对局部路径规划时因无法提前获取环境信息导致移动机器人搜索不到合适的路径,以及在采用马尔可夫决策过程中传统强化学习算法应用于局部路径规划时存在着学习效率低下及收敛速度较慢等问题,提出一种改进的Q-learn-ing(QL)算法。首先... 针对局部路径规划时因无法提前获取环境信息导致移动机器人搜索不到合适的路径,以及在采用马尔可夫决策过程中传统强化学习算法应用于局部路径规划时存在着学习效率低下及收敛速度较慢等问题,提出一种改进的Q-learn-ing(QL)算法。首先设计一种动态自适应贪婪策略,用于平衡移动机器人对环境探索和利用之间的问题;其次根据A*算法思想设计启发式学习评估模型,从而动态调整学习因子并为搜索路径提供导向作用;最后引入三阶贝塞尔曲线规划对路径进行平滑处理。通过Pycharm平台仿真结果表明,使得改进后的QL算法所规划的路径长度、搜索效率及路径平滑性等特性上都优于传统Sarsa算法及QL算法,比传统Sarsa算法迭代次数提高32.3%,搜索时间缩短27.08%,比传统QL算法迭代次数提高27.32%,搜索时间缩短17.28%,路径规划的拐点大幅度减少,局部路径优化效果较为明显。 展开更多
关键词 移动机器人 q-learning算法 局部路径 A^(*)算法 贝塞尔曲线
下载PDF
一种基于Q-learning强化学习的导向性处理器安全性模糊测试方案
14
作者 崔云凯 张伟 《北京信息科技大学学报(自然科学版)》 2024年第4期81-87,95,共8页
针对处理器安全性模糊测试在进行细粒度变异时遗传算法存在一定的盲目性,易使生成的测试用例触发相同类型漏洞的问题,提出了一种基于Q-learning强化学习的导向性处理器安全性模糊测试方案。通过测试用例的状态值和所触发的漏洞类型对应... 针对处理器安全性模糊测试在进行细粒度变异时遗传算法存在一定的盲目性,易使生成的测试用例触发相同类型漏洞的问题,提出了一种基于Q-learning强化学习的导向性处理器安全性模糊测试方案。通过测试用例的状态值和所触发的漏洞类型对应的权值构造奖励函数,使用强化学习指导生成具有针对性和导向性的测试用例,快速地触发不同类型的漏洞。在Hikey970平台上的实验验证了基于ARMv8的测试用例生成框架的有效性,并且相较于传统使用遗传算法作为反馈的策略,本文方案在相同时间内生成有效测试用例的的数量多19.15%,发现漏洞类型的数量多80.00%。 展开更多
关键词 处理器漏洞检测 模糊测试 q-learning强化学习 ARMv8 分支预测类漏洞
下载PDF
Online Consensus Control of Nonlinear Affine Systems From Disturbed Data
15
作者 Yifei Li Wenjie Liu +3 位作者 Jian Sun Chen Chen Jia Zhang Gang Wang 《IEEE/CAA Journal of Automatica Sinica》 SCIE EI CSCD 2024年第2期551-553,共3页
Dear Editor,In this letter, we introduce a novel online distributed data-driven robust control approach for learning controllers of unknown nonlinear multi-agent systems(MASs) using state-dependent representations.
关键词 AGENT online online
下载PDF
Q-Learning-Assisted Meta-Heuristics for Scheduling Distributed Hybrid Flow Shop Problems
16
作者 Qianyao Zhu Kaizhou Gao +2 位作者 Wuze Huang Zhenfang Ma Adam Slowik 《Computers, Materials & Continua》 SCIE EI 2024年第9期3573-3589,共17页
The flow shop scheduling problem is important for the manufacturing industry.Effective flow shop scheduling can bring great benefits to the industry.However,there are few types of research on Distributed Hybrid Flow S... The flow shop scheduling problem is important for the manufacturing industry.Effective flow shop scheduling can bring great benefits to the industry.However,there are few types of research on Distributed Hybrid Flow Shop Problems(DHFSP)by learning assisted meta-heuristics.This work addresses a DHFSP with minimizing the maximum completion time(Makespan).First,a mathematical model is developed for the concerned DHFSP.Second,four Q-learning-assisted meta-heuristics,e.g.,genetic algorithm(GA),artificial bee colony algorithm(ABC),particle swarm optimization(PSO),and differential evolution(DE),are proposed.According to the nature of DHFSP,six local search operations are designed for finding high-quality solutions in local space.Instead of randomselection,Q-learning assists meta-heuristics in choosing the appropriate local search operations during iterations.Finally,based on 60 cases,comprehensive numerical experiments are conducted to assess the effectiveness of the proposed algorithms.The experimental results and discussions prove that using Q-learning to select appropriate local search operations is more effective than the random strategy.To verify the competitiveness of the Q-learning assistedmeta-heuristics,they are compared with the improved iterated greedy algorithm(IIG),which is also for solving DHFSP.The Friedman test is executed on the results by five algorithms.It is concluded that the performance of four Q-learning-assisted meta-heuristics are better than IIG,and the Q-learning-assisted PSO shows the best competitiveness. 展开更多
关键词 Distributed scheduling hybrid flow shop META-HEURISTICS local search q-learning
下载PDF
Effective Controller Placement in Software-Defined Internet-of-Things Leveraging Deep Q-Learning (DQL)
17
作者 Jehad Ali Mohammed J.F.Alenazi 《Computers, Materials & Continua》 SCIE EI 2024年第12期4015-4032,共18页
The controller is a main component in the Software-Defined Networking(SDN)framework,which plays a significant role in enabling programmability and orchestration for 5G and next-generation networks.In SDN,frequent comm... The controller is a main component in the Software-Defined Networking(SDN)framework,which plays a significant role in enabling programmability and orchestration for 5G and next-generation networks.In SDN,frequent communication occurs between network switches and the controller,which manages and directs traffic flows.If the controller is not strategically placed within the network,this communication can experience increased delays,negatively affecting network performance.Specifically,an improperly placed controller can lead to higher end-to-end(E2E)delay,as switches must traverse more hops or encounter greater propagation delays when communicating with the controller.This paper introduces a novel approach using Deep Q-Learning(DQL)to dynamically place controllers in Software-Defined Internet of Things(SD-IoT)environments,with the goal of minimizing E2E delay between switches and controllers.E2E delay,a crucial metric for network performance,is influenced by two key factors:hop count,which measures the number of network nodes data must traverse,and propagation delay,which accounts for the physical distance between nodes.Our approach models the controller placement problem as a Markov Decision Process(MDP).In this model,the network configuration at any given time is represented as a“state,”while“actions”correspond to potential decisions regarding the placement of controllers or the reassignment of switches to controllers.Using a Deep Q-Network(DQN)to approximate the Q-function,the system learns the optimal controller placement by maximizing the cumulative reward,which is defined as the negative of the E2E delay.Essentially,the lower the delay,the higher the reward the system receives,enabling it to continuously improve its controller placement strategy.The experimental results show that our DQL-based method significantly reduces E2E delay when compared to traditional benchmark placement strategies.By dynamically learning from the network’s real-time conditions,the proposed method ensures that controller placement remains efficient and responsive,reducing communication delays and enhancing overall network performance. 展开更多
关键词 Software-defined networking deep q-learning controller placement quality of service
下载PDF
Traffic Control Based on Integrated Kalman Filtering and Adaptive Quantized Q-Learning Framework for Internet of Vehicles
18
作者 Othman S.Al-Heety Zahriladha Zakaria +4 位作者 Ahmed Abu-Khadrah Mahamod Ismail Sarmad Nozad Mahmood Mohammed Mudhafar Shakir Hussein Alsariera 《Computer Modeling in Engineering & Sciences》 SCIE EI 2024年第3期2103-2127,共25页
Intelligent traffic control requires accurate estimation of the road states and incorporation of adaptive or dynamically adjusted intelligent algorithms for making the decision.In this article,these issues are handled... Intelligent traffic control requires accurate estimation of the road states and incorporation of adaptive or dynamically adjusted intelligent algorithms for making the decision.In this article,these issues are handled by proposing a novel framework for traffic control using vehicular communications and Internet of Things data.The framework integrates Kalman filtering and Q-learning.Unlike smoothing Kalman filtering,our data fusion Kalman filter incorporates a process-aware model which makes it superior in terms of the prediction error.Unlike traditional Q-learning,our Q-learning algorithm enables adaptive state quantization by changing the threshold of separating low traffic from high traffic on the road according to the maximum number of vehicles in the junction roads.For evaluation,the model has been simulated on a single intersection consisting of four roads:east,west,north,and south.A comparison of the developed adaptive quantized Q-learning(AQQL)framework with state-of-the-art and greedy approaches shows the superiority of AQQL with an improvement percentage in terms of the released number of vehicles of AQQL is 5%over the greedy approach and 340%over the state-of-the-art approach.Hence,AQQL provides an effective traffic control that can be applied in today’s intelligent traffic system. 展开更多
关键词 q-learning intelligent transportation system(ITS) traffic control vehicular communication kalman filtering smart city Internet of Things
下载PDF
Novel Static Security and Stability Control of Power Systems Based on Artificial Emotional Lazy Q-Learning
19
作者 Tao Bao Xiyuan Ma +3 位作者 Zhuohuan Li Duotong Yang Pengyu Wang Changcheng Zhou 《Energy Engineering》 EI 2024年第6期1713-1737,共25页
The stability problem of power grids has become increasingly serious in recent years as the size of novel power systems increases.In order to improve and ensure the stable operation of the novel power system,this stud... The stability problem of power grids has become increasingly serious in recent years as the size of novel power systems increases.In order to improve and ensure the stable operation of the novel power system,this study proposes an artificial emotional lazy Q-learning method,which combines artificial emotion,lazy learning,and reinforcement learning for static security and stability analysis of power systems.Moreover,this study compares the analysis results of the proposed method with those of the small disturbance method for a stand-alone power system and verifies that the proposed lazy Q-learning method is able to effectively screen useful data for learning,and improve the static security stability of the new type of power system more effectively than the traditional proportional-integral-differential control and Q-learning methods. 展开更多
关键词 Artificial sentiment static secure stable analysis q-learning lazy learning data filtering
下载PDF
基于Online-GRU信道预测的星上自适应功率控制方法
20
作者 施文军 朱立东 《太赫兹科学与电子信息学报》 2024年第3期261-268,共8页
针对传统卫星功率控制方法存在资源浪费、时延长的问题,提出一种基于在线-门控循环单元(Online-GRU)信道预测的星上自适应功率控制方法,通过在线训练更新网络参数来解决离线预测算法存在的累积误差的问题。仿真结果表明,提出的在线训练... 针对传统卫星功率控制方法存在资源浪费、时延长的问题,提出一种基于在线-门控循环单元(Online-GRU)信道预测的星上自适应功率控制方法,通过在线训练更新网络参数来解决离线预测算法存在的累积误差的问题。仿真结果表明,提出的在线训练算法比离线算法预测精确度提升了38.30%,相比在线-长短期记忆网络(Online-LSTM)节约了63.21%的训练时间;提出的自适应功率控制方法比固定发射功率的方法节约了55.74%的发射功率;同时,相比基于地面定时反馈信道状态的自适应功率控制方法具备更好的鲁棒性。 展开更多
关键词 星上自适应功率控制 在线训练 在线-门控循环单元 信道预测
下载PDF
上一页 1 2 250 下一页 到第
使用帮助 返回顶部