期刊文献+
共找到79,317篇文章
< 1 2 250 >
每页显示 20 50 100
UAV-Assisted Dynamic Avatar Task Migration for Vehicular Metaverse Services: A Multi-Agent Deep Reinforcement Learning Approach 被引量:1
1
作者 Jiawen Kang Junlong Chen +6 位作者 Minrui Xu Zehui Xiong Yutao Jiao Luchao Han Dusit Niyato Yongju Tong Shengli Xie 《IEEE/CAA Journal of Automatica Sinica》 SCIE EI CSCD 2024年第2期430-445,共16页
Avatars, as promising digital representations and service assistants of users in Metaverses, can enable drivers and passengers to immerse themselves in 3D virtual services and spaces of UAV-assisted vehicular Metavers... Avatars, as promising digital representations and service assistants of users in Metaverses, can enable drivers and passengers to immerse themselves in 3D virtual services and spaces of UAV-assisted vehicular Metaverses. However, avatar tasks include a multitude of human-to-avatar and avatar-to-avatar interactive applications, e.g., augmented reality navigation,which consumes intensive computing resources. It is inefficient and impractical for vehicles to process avatar tasks locally. Fortunately, migrating avatar tasks to the nearest roadside units(RSU)or unmanned aerial vehicles(UAV) for execution is a promising solution to decrease computation overhead and reduce task processing latency, while the high mobility of vehicles brings challenges for vehicles to independently perform avatar migration decisions depending on current and future vehicle status. To address these challenges, in this paper, we propose a novel avatar task migration system based on multi-agent deep reinforcement learning(MADRL) to execute immersive vehicular avatar tasks dynamically. Specifically, we first formulate the problem of avatar task migration from vehicles to RSUs/UAVs as a partially observable Markov decision process that can be solved by MADRL algorithms. We then design the multi-agent proximal policy optimization(MAPPO) approach as the MADRL algorithm for the avatar task migration problem. To overcome slow convergence resulting from the curse of dimensionality and non-stationary issues caused by shared parameters in MAPPO, we further propose a transformer-based MAPPO approach via sequential decision-making models for the efficient representation of relationships among agents. Finally, to motivate terrestrial or non-terrestrial edge servers(e.g., RSUs or UAVs) to share computation resources and ensure traceability of the sharing records, we apply smart contracts and blockchain technologies to achieve secure sharing management. Numerical results demonstrate that the proposed approach outperforms the MAPPO approach by around 2% and effectively reduces approximately 20% of the latency of avatar task execution in UAV-assisted vehicular Metaverses. 展开更多
关键词 AVATAR blockchain metaverses multi-agent deep reinforcement learning transformer UAVS
下载PDF
Cognitive interference decision method for air defense missile fuze based on reinforcement learning 被引量:1
2
作者 Dingkun Huang Xiaopeng Yan +2 位作者 Jian Dai Xinwei Wang Yangtian Liu 《Defence Technology(防务技术)》 SCIE EI CAS CSCD 2024年第2期393-404,共12页
To solve the problem of the low interference success rate of air defense missile radio fuzes due to the unified interference form of the traditional fuze interference system,an interference decision method based Q-lea... To solve the problem of the low interference success rate of air defense missile radio fuzes due to the unified interference form of the traditional fuze interference system,an interference decision method based Q-learning algorithm is proposed.First,dividing the distance between the missile and the target into multiple states to increase the quantity of state spaces.Second,a multidimensional motion space is utilized,and the search range of which changes with the distance of the projectile,to select parameters and minimize the amount of ineffective interference parameters.The interference effect is determined by detecting whether the fuze signal disappears.Finally,a weighted reward function is used to determine the reward value based on the range state,output power,and parameter quantity information of the interference form.The effectiveness of the proposed method in selecting the range of motion space parameters and designing the discrimination degree of the reward function has been verified through offline experiments involving full-range missile rendezvous.The optimal interference form for each distance state has been obtained.Compared with the single-interference decision method,the proposed decision method can effectively improve the success rate of interference. 展开更多
关键词 Cognitive radio Interference decision Radio fuze reinforcement learning Interference strategy optimization
下载PDF
A Comparative Study on the Post-Buckling Behavior of Reinforced Thermoplastic Pipes(RTPs)Under External Pressure Considering Progressive Failure 被引量:1
3
作者 DING Xin-dong WANG Shu-qing +1 位作者 LIU Wen-cheng YE Xiao-han 《China Ocean Engineering》 SCIE EI CSCD 2024年第2期233-246,共14页
The collapse pressure is a key parameter when RTPs are applied in harsh deep-water environments.To investigate the collapse of RTPs,numerical simulations and hydrostatic pressure tests are conducted.For the numerical ... The collapse pressure is a key parameter when RTPs are applied in harsh deep-water environments.To investigate the collapse of RTPs,numerical simulations and hydrostatic pressure tests are conducted.For the numerical simulations,the eigenvalue analysis and Riks analysis are combined,in which the Hashin failure criterion and fracture energy stiffness degradation model are used to simulate the progressive failure of composites,and the“infinite”boundary conditions are applied to eliminate the boundary effects.As for the hydrostatic pressure tests,RTP specimens were placed in a hydrostatic chamber after filled with water.It has been observed that the cross-section of the middle part collapses when it reaches the maximum pressure.The collapse pressure obtained from the numerical simulations agrees well with that in the experiment.Meanwhile,the applicability of NASA SP-8007 formula on the collapse pressure prediction was also discussed.It has a relatively greater difference because of the ignorance of the progressive failure of composites.For the parametric study,it is found that RTPs have much higher first-ply-failure pressure when the winding angles are between 50°and 70°.Besides,the effect of debonding and initial ovality,and the contribution of the liner and coating are also discussed. 展开更多
关键词 reinforced thermoplastic pipes post-buckling behavior progressive failure of composites DEBONDING initial ovality
下载PDF
Recent Progress in Reinforcement Learning and Adaptive Dynamic Programming for Advanced Control Applications 被引量:2
4
作者 Ding Wang Ning Gao +2 位作者 Derong Liu Jinna Li Frank L.Lewis 《IEEE/CAA Journal of Automatica Sinica》 SCIE EI CSCD 2024年第1期18-36,共19页
Reinforcement learning(RL) has roots in dynamic programming and it is called adaptive/approximate dynamic programming(ADP) within the control community. This paper reviews recent developments in ADP along with RL and ... Reinforcement learning(RL) has roots in dynamic programming and it is called adaptive/approximate dynamic programming(ADP) within the control community. This paper reviews recent developments in ADP along with RL and its applications to various advanced control fields. First, the background of the development of ADP is described, emphasizing the significance of regulation and tracking control problems. Some effective offline and online algorithms for ADP/adaptive critic control are displayed, where the main results towards discrete-time systems and continuous-time systems are surveyed, respectively.Then, the research progress on adaptive critic control based on the event-triggered framework and under uncertain environment is discussed, respectively, where event-based design, robust stabilization, and game design are reviewed. Moreover, the extensions of ADP for addressing control problems under complex environment attract enormous attention. The ADP architecture is revisited under the perspective of data-driven and RL frameworks,showing how they promote ADP formulation significantly.Finally, several typical control applications with respect to RL and ADP are summarized, particularly in the fields of wastewater treatment processes and power systems, followed by some general prospects for future research. Overall, the comprehensive survey on ADP and RL for advanced control applications has d emonstrated its remarkable potential within the artificial intelligence era. In addition, it also plays a vital role in promoting environmental protection and industrial intelligence. 展开更多
关键词 Adaptive dynamic programming(ADP) advanced control complex environment data-driven control event-triggered design intelligent control neural networks nonlinear systems optimal control reinforcement learning(RL)
下载PDF
Bridge Bidding via Deep Reinforcement Learning and Belief Monte Carlo Search
5
作者 Zizhang Qiu Shouguang Wang +1 位作者 Dan You MengChu Zhou 《IEEE/CAA Journal of Automatica Sinica》 SCIE EI CSCD 2024年第10期2111-2122,共12页
Contract Bridge,a four-player imperfect information game,comprises two phases:bidding and playing.While computer programs excel at playing,bidding presents a challenging aspect due to the need for information exchange... Contract Bridge,a four-player imperfect information game,comprises two phases:bidding and playing.While computer programs excel at playing,bidding presents a challenging aspect due to the need for information exchange with partners and interference with communication of opponents.In this work,we introduce a Bridge bidding agent that combines supervised learning,deep reinforcement learning via self-play,and a test-time search approach.Our experiments demonstrate that our agent outperforms WBridge5,a highly regarded computer Bridge software that has won multiple world championships,by a performance of 0.98 IMPs(international match points)per deal over 10000 deals,with a much cost-effective approach.The performance significantly surpasses previous state-of-the-art(0.85 IMPs per deal).Note 0.1 IMPs per deal is a significant improvement in Bridge bidding. 展开更多
关键词 Contract Bridge reinforcement learning SEARCH
下载PDF
Distributed Graph Database Load Balancing Method Based on Deep Reinforcement Learning
6
作者 Shuming Sha Naiwang Guo +1 位作者 Wang Luo Yong Zhang 《Computers, Materials & Continua》 SCIE EI 2024年第6期5105-5124,共20页
This paper focuses on the scheduling problem of workflow tasks that exhibit interdependencies.Unlike indepen-dent batch tasks,workflows typically consist of multiple subtasks with intrinsic correlations and dependenci... This paper focuses on the scheduling problem of workflow tasks that exhibit interdependencies.Unlike indepen-dent batch tasks,workflows typically consist of multiple subtasks with intrinsic correlations and dependencies.It necessitates the distribution of various computational tasks to appropriate computing node resources in accor-dance with task dependencies to ensure the smooth completion of the entire workflow.Workflow scheduling must consider an array of factors,including task dependencies,availability of computational resources,and the schedulability of tasks.Therefore,this paper delves into the distributed graph database workflow task scheduling problem and proposes a workflow scheduling methodology based on deep reinforcement learning(DRL).The method optimizes the maximum completion time(makespan)and response time of workflow tasks,aiming to enhance the responsiveness of workflow tasks while ensuring the minimization of the makespan.The experimental results indicate that the Q-learning Deep Reinforcement Learning(Q-DRL)algorithm markedly diminishes the makespan and refines the average response time within distributed graph database environments.In quantifying makespan,Q-DRL achieves mean reductions of 12.4%and 11.9%over established First-fit and Random scheduling strategies,respectively.Additionally,Q-DRL surpasses the performance of both DRL-Cloud and Improved Deep Q-learning Network(IDQN)algorithms,with improvements standing at 4.4%and 2.6%,respectively.With reference to average response time,the Q-DRL approach exhibits a significantly enhanced performance in the scheduling of workflow tasks,decreasing the average by 2.27%and 4.71%when compared to IDQN and DRL-Cloud,respectively.The Q-DRL algorithm also demonstrates a notable increase in the efficiency of system resource utilization,reducing the average idle rate by 5.02%and 9.30%in comparison to IDQN and DRL-Cloud,respectively.These findings support the assertion that Q-DRL not only upholds a lower average idle rate but also effectively curtails the average response time,thereby substantially improving processing efficiency and optimizing resource utilization within distributed graph database systems. 展开更多
关键词 reinforcement learning WORKFLOW task scheduling load balancing
下载PDF
Nano-scale Reinforcements and Properties of Al-Si-Cu Alloy Processed by High-Pressure Torsion
7
作者 DONG Ying WU Siyuan +4 位作者 HE Ziyang LIANG Chen CHENG Feng HE Zuwei QIAN Chenhao 《Journal of Wuhan University of Technology(Materials Science)》 SCIE EI CAS CSCD 2024年第5期1253-1259,共7页
To improve the comprehensive mechanical properties of Al-Si-Cu alloy,it was treated by a high-pressure torsion process,and the effect of the deformation degree on the microstructure and properties of the Al-Si-Cu allo... To improve the comprehensive mechanical properties of Al-Si-Cu alloy,it was treated by a high-pressure torsion process,and the effect of the deformation degree on the microstructure and properties of the Al-Si-Cu alloy was studied.The results show that the reinforcements(β-Si andθ-CuAl_(2)phases)of the Al-Si-Cu alloy are dispersed in theα-Al matrix phase with finer phase size after the treatment.The processed samples exhibit grain sizes in the submicron or even nanometer range,which effectively improves the mechanical properties of the material.The hardness and strength of the deformed alloy are both significantly raised to 268 HV and 390.04 MPa by 10 turns HPT process,and the fracture morphology shows that the material gradually transits from brittle to plastic before and after deformation.The elements interdiffusion at the interface between the phases has also been effectively enhanced.In addition,it is found that the severe plastic deformation at room temperature induces a ternary eutectic reaction,resulting in the formation of ternary Al+Si+CuAl_(2)eutectic. 展开更多
关键词 Al-Si-Cu alloy high-pressure torsion nano-scale reinforcements ternary eutectic
下载PDF
QoS Routing Optimization Based on Deep Reinforcement Learning in SDN
8
作者 Yu Song Xusheng Qian +2 位作者 Nan Zhang Wei Wang Ao Xiong 《Computers, Materials & Continua》 SCIE EI 2024年第5期3007-3021,共15页
To enhance the efficiency and expediency of issuing e-licenses within the power sector, we must confront thechallenge of managing the surging demand for data traffic. Within this realm, the network imposes stringentQu... To enhance the efficiency and expediency of issuing e-licenses within the power sector, we must confront thechallenge of managing the surging demand for data traffic. Within this realm, the network imposes stringentQuality of Service (QoS) requirements, revealing the inadequacies of traditional routing allocation mechanismsin accommodating such extensive data flows. In response to the imperative of handling a substantial influx of datarequests promptly and alleviating the constraints of existing technologies and network congestion, we present anarchitecture forQoS routing optimizationwith in SoftwareDefinedNetwork (SDN), leveraging deep reinforcementlearning. This innovative approach entails the separation of SDN control and transmission functionalities, centralizingcontrol over data forwardingwhile integrating deep reinforcement learning for informed routing decisions. Byfactoring in considerations such as delay, bandwidth, jitter rate, and packet loss rate, we design a reward function toguide theDeepDeterministic PolicyGradient (DDPG) algorithmin learning the optimal routing strategy to furnishsuperior QoS provision. In our empirical investigations, we juxtapose the performance of Deep ReinforcementLearning (DRL) against that of Shortest Path (SP) algorithms in terms of data packet transmission delay. Theexperimental simulation results show that our proposed algorithm has significant efficacy in reducing networkdelay and improving the overall transmission efficiency, which is superior to the traditional methods. 展开更多
关键词 Deep reinforcement learning SDN route optimization QOS
下载PDF
Role Dynamic Allocation of Human-Robot Cooperation Based on Reinforcement Learning in an Installation of Curtain Wall
9
作者 Zhiguang Liu Shilin Wang +2 位作者 Jian Zhao Jianhong Hao Fei Yu 《Computer Modeling in Engineering & Sciences》 SCIE EI 2024年第1期473-487,共15页
A real-time adaptive roles allocation method based on reinforcement learning is proposed to improve humanrobot cooperation performance for a curtain wall installation task.This method breaks the traditional idea that ... A real-time adaptive roles allocation method based on reinforcement learning is proposed to improve humanrobot cooperation performance for a curtain wall installation task.This method breaks the traditional idea that the robot is regarded as the follower or only adjusts the leader and the follower in cooperation.In this paper,a self-learning method is proposed which can dynamically adapt and continuously adjust the initiative weight of the robot according to the change of the task.Firstly,the physical human-robot cooperation model,including the role factor is built.Then,a reinforcement learningmodel that can adjust the role factor in real time is established,and a reward and actionmodel is designed.The role factor can be adjusted continuously according to the comprehensive performance of the human-robot interaction force and the robot’s Jerk during the repeated installation.Finally,the roles adjustment rule established above continuously improves the comprehensive performance.Experiments of the dynamic roles allocation and the effect of the performance weighting coefficient on the result have been verified.The results show that the proposed method can realize the role adaptation and achieve the dual optimization goal of reducing the sum of the cooperator force and the robot’s Jerk. 展开更多
关键词 Human-robot cooperation roles allocation reinforcement learning
下载PDF
Reinforcement learning based edge computing in B5G
10
作者 Jiachen Yang Yiwen Sun +4 位作者 Yutian Lei Zhuo Zhang Yang Li Yongjun Bao Zhihan Lv 《Digital Communications and Networks》 SCIE CSCD 2024年第1期1-6,共6页
The development of communication technology will promote the application of Internet of Things,and Beyond 5G will become a new technology promoter.At the same time,Beyond 5G will become one of the important supports f... The development of communication technology will promote the application of Internet of Things,and Beyond 5G will become a new technology promoter.At the same time,Beyond 5G will become one of the important supports for the development of edge computing technology.This paper proposes a communication task allocation algorithm based on deep reinforcement learning for vehicle-to-pedestrian communication scenarios in edge computing.Through trial and error learning of agent,the optimal spectrum and power can be determined for transmission without global information,so as to balance the communication between vehicle-to-pedestrian and vehicle-to-infrastructure.The results show that the agent can effectively improve vehicle-to-infrastructure communication rate as well as meeting the delay constraints on the vehicle-to-pedestrian link. 展开更多
关键词 reinforcement learning Edge computing Beyond 5G Vehicle-to-pedestrian
下载PDF
Reinforcement Learning in Process Industries:Review and Perspective
11
作者 Oguzhan Dogru Junyao Xie +6 位作者 Om Prakash Ranjith Chiplunkar Jansen Soesanto Hongtian Chen Kirubakaran Velswamy Fadi Ibrahim Biao Huang 《IEEE/CAA Journal of Automatica Sinica》 SCIE EI CSCD 2024年第2期283-300,共18页
This survey paper provides a review and perspective on intermediate and advanced reinforcement learning(RL)techniques in process industries. It offers a holistic approach by covering all levels of the process control ... This survey paper provides a review and perspective on intermediate and advanced reinforcement learning(RL)techniques in process industries. It offers a holistic approach by covering all levels of the process control hierarchy. The survey paper presents a comprehensive overview of RL algorithms,including fundamental concepts like Markov decision processes and different approaches to RL, such as value-based, policy-based, and actor-critic methods, while also discussing the relationship between classical control and RL. It further reviews the wide-ranging applications of RL in process industries, such as soft sensors, low-level control, high-level control, distributed process control, fault detection and fault tolerant control, optimization,planning, scheduling, and supply chain. The survey paper discusses the limitations and advantages, trends and new applications, and opportunities and future prospects for RL in process industries. Moreover, it highlights the need for a holistic approach in complex systems due to the growing importance of digitalization in the process industries. 展开更多
关键词 Process control process systems engineering reinforcement learning
下载PDF
Toward Trustworthy Decision-Making for Autonomous Vehicles:A Robust Reinforcement Learning Approach with Safety Guarantees
12
作者 Xiangkun He Wenhui Huang Chen Lv 《Engineering》 SCIE EI CAS CSCD 2024年第2期77-89,共13页
While autonomous vehicles are vital components of intelligent transportation systems,ensuring the trustworthiness of decision-making remains a substantial challenge in realizing autonomous driving.Therefore,we present... While autonomous vehicles are vital components of intelligent transportation systems,ensuring the trustworthiness of decision-making remains a substantial challenge in realizing autonomous driving.Therefore,we present a novel robust reinforcement learning approach with safety guarantees to attain trustworthy decision-making for autonomous vehicles.The proposed technique ensures decision trustworthiness in terms of policy robustness and collision safety.Specifically,an adversary model is learned online to simulate the worst-case uncertainty by approximating the optimal adversarial perturbations on the observed states and environmental dynamics.In addition,an adversarial robust actor-critic algorithm is developed to enable the agent to learn robust policies against perturbations in observations and dynamics.Moreover,we devise a safety mask to guarantee the collision safety of the autonomous driving agent during both the training and testing processes using an interpretable knowledge model known as the Responsibility-Sensitive Safety Model.Finally,the proposed approach is evaluated through both simulations and experiments.These results indicate that the autonomous driving agent can make trustworthy decisions and drastically reduce the number of collisions through robust safety policies. 展开更多
关键词 Autonomous vehicle DECISION-MAKING reinforcement learning Adversarial attack Safety guarantee
下载PDF
Reinforced tissue matrix to strengthen the abdominal wall following reversal of temporary ostomies or to treat incisional hernias
13
作者 Spencer P Lake Corey R Deeken Amit K Agarwal 《World Journal of Gastrointestinal Surgery》 SCIE 2024年第3期823-832,共10页
BACKGROUND Abdominal wall deficiencies or weakness are a common complication of tem-porary ostomies,and incisional hernias frequently develop after colostomy or ileostomy takedown.The use of synthetic meshes to reinfo... BACKGROUND Abdominal wall deficiencies or weakness are a common complication of tem-porary ostomies,and incisional hernias frequently develop after colostomy or ileostomy takedown.The use of synthetic meshes to reinforce the abdominal wall has reduced hernia occurrence.Biologic meshes have also been used to enhance healing,particularly in contaminated conditions.Reinforced tissue matrices(R-TMs),which include a biologic scaffold of native extracellular matrix and a syn-thetic component for added strength/durability,are designed to take advantage of aspects of both synthetic and biologic materials.To date,RTMs have not been reported to reinforce the abdominal wall following stoma reversal.METHODS Twenty-eight patients were selected with a parastomal and/or incisional hernia who had received a temporary ileostomy or colostomy for fecal diversion after rectal cancer treatment or trauma.Following hernia repair and proximal stoma closure,RTM(OviTex®1S permanent or OviTex®LPR)was placed to reinforce the abdominal wall using a laparoscopic,robotic,or open surgical approach.Post-operative follow-up was performed at 1 month and 1 year.Hernia recurrence was determined by physical examination and,when necessary,via computed tomo-graphy scan.Secondary endpoints included length of hospital stay,time to return to work,and hospital readmissions.Evaluated complications of the wound/repair site included presence of surgical site infection,seroma,hematoma,wound dehiscence,or fistula formation.RESULTS The observational study cohort included 16 male and 12 female patients with average age of 58.5 years±16.3 years and average body mass index of 26.2 kg/m^(2)±4.1 kg/m^(2).Patients presented with a parastomal hernia(75.0%),in-cisional hernia(14.3%),or combined parastomal/incisional hernia(10.7%).Using a laparoscopic(53.6%),robotic(35.7%),or open(10.7%)technique,RTMs(OviTex®LPR:82.1%,OviTex®1S:17.9%)were placed using sublay(82.1%)or intraperitoneal onlay(IPOM;17.9%)mesh positioning.At 1-month and 1-year follow-ups,there were no hernia recurrences(0%).Average hospital stays were 2.1 d±1.2 d and return to work occurred at 8.3 post-operative days±3.0 post-operative days.Three patients(10.7%)were readmitted before the 1-month follow up due to mesh infection and/or gastrointestinal issues.Fistula and mesh infection were observed in two patients each(7.1%),leading to partial mesh removal in one patient(3.6%).There were no complications between 1 month and 1 year(0%).CONCLUSION RTMs were used successfully to treat parastomal and incisional hernias at ileostomy reversal,with no hernia recurrences and favorable outcomes after 1-month and 1-year. 展开更多
关键词 reinforced tissue matrix reinforced forestomach matrix ILEOSTOMY COLOSTOMY Ostomy takedown Incisional hernia Abdominal wall
下载PDF
Combining reinforcement learning with mathematical programming:An approach for optimal design of heat exchanger networks
14
作者 Hui Tan Xiaodong Hong +4 位作者 Zuwei Liao Jingyuan Sun Yao Yang Jingdai Wang Yongrong Yang 《Chinese Journal of Chemical Engineering》 SCIE EI CAS CSCD 2024年第5期63-71,共9页
Heat integration is important for energy-saving in the process industry.It is linked to the persistently challenging task of optimal design of heat exchanger networks(HEN).Due to the inherent highly nonconvex nonlinea... Heat integration is important for energy-saving in the process industry.It is linked to the persistently challenging task of optimal design of heat exchanger networks(HEN).Due to the inherent highly nonconvex nonlinear and combinatorial nature of the HEN problem,it is not easy to find solutions of high quality for large-scale problems.The reinforcement learning(RL)method,which learns strategies through ongoing exploration and exploitation,reveals advantages in such area.However,due to the complexity of the HEN design problem,the RL method for HEN should be dedicated and designed.A hybrid strategy combining RL with mathematical programming is proposed to take better advantage of both methods.An insightful state representation of the HEN structure as well as a customized reward function is introduced.A Q-learning algorithm is applied to update the HEN structure using theε-greedy strategy.Better results are obtained from three literature cases of different scales. 展开更多
关键词 Heat exchanger network reinforcement learning Mathematical programming Process design
下载PDF
Electrochemical Study of the Corrosion Inhibitory Capacity of Calcined Attapulgite in Reinforced Concrete Medium
15
作者 Malang Bodian Kinda Hannawi +3 位作者 Dame Keinde Modou Fall Aveline Darquennes Prince William Agbodjan 《Advances in Materials Physics and Chemistry》 CAS 2024年第5期76-94,共19页
The durability of reinforced concrete structures is greatly influenced by the corrosion of the reinforcement. In addition to air pollution related to the repair of corroded structures, chloride ions are the main facto... The durability of reinforced concrete structures is greatly influenced by the corrosion of the reinforcement. In addition to air pollution related to the repair of corroded structures, chloride ions are the main factors of corrosion of reinforced concrete structures. This study aims to valorize a clay inhibitor against reinforcement corrosion in reinforced concrete. This clay (Attapulgite) was incorporated into reinforced concretes at different percentages of substitution of calcined attapulgite (0%, 5% and 10%) to cement in the formulation. The corrosion inhibitory power of attapulgite is evaluated in reinforced concretes subjected to the action of chloride ions at different intervals in the NaCl solution (1 day, 21 days and 45 days) by electrochemical methods (zero current chronopotentiometry, polarization curves and electrochemical impedance spectroscopy). This study showed that in the presence of chloride ions, the composition based on 10% attapulgite has an appreciable inhibitory effect with an average inhibitory efficiency of 82%. 展开更多
关键词 ATTAPULGITE Electrochemical Methods INHIBITOR reinforced Concrete
下载PDF
Knowledge Reasoning Method Based on Deep Transfer Reinforcement Learning:DTRLpath
16
作者 Shiming Lin Ling Ye +4 位作者 Yijie Zhuang Lingyun Lu Shaoqiu Zheng Chenxi Huang Ng Yin Kwee 《Computers, Materials & Continua》 SCIE EI 2024年第7期299-317,共19页
In recent years,with the continuous development of deep learning and knowledge graph reasoning methods,more and more researchers have shown great interest in improving knowledge graph reasoning methods by inferring mi... In recent years,with the continuous development of deep learning and knowledge graph reasoning methods,more and more researchers have shown great interest in improving knowledge graph reasoning methods by inferring missing facts through reasoning.By searching paths on the knowledge graph and making fact and link predictions based on these paths,deep learning-based Reinforcement Learning(RL)agents can demonstrate good performance and interpretability.Therefore,deep reinforcement learning-based knowledge reasoning methods have rapidly emerged in recent years and have become a hot research topic.However,even in a small and fixed knowledge graph reasoning action space,there are still a large number of invalid actions.It often leads to the interruption of RL agents’wandering due to the selection of invalid actions,resulting in a significant decrease in the success rate of path mining.In order to improve the success rate of RL agents in the early stages of path search,this article proposes a knowledge reasoning method based on Deep Transfer Reinforcement Learning path(DTRLpath).Before supervised pre-training and retraining,a pre-task of searching for effective actions in a single step is added.The RL agent is first trained in the pre-task to improve its ability to search for effective actions.Then,the trained agent is transferred to the target reasoning task for path search training,which improves its success rate in searching for target task paths.Finally,based on the comparative experimental results on the FB15K-237 and NELL-995 datasets,it can be concluded that the proposed method significantly improves the success rate of path search and outperforms similar methods in most reasoning tasks. 展开更多
关键词 Intelligent agent knowledge graph reasoning reinforceMENT transfer learning
下载PDF
Task Offloading and Resource Allocation in NOMA-VEC:A Multi-Agent Deep Graph Reinforcement Learning Algorithm
17
作者 Hu Yonghui Jin Zuodong +1 位作者 Qi Peng Tao Dan 《China Communications》 SCIE CSCD 2024年第8期79-88,共10页
Vehicular edge computing(VEC)is emerging as a promising solution paradigm to meet the requirements of compute-intensive applications in internet of vehicle(IoV).Non-orthogonal multiple access(NOMA)has advantages in im... Vehicular edge computing(VEC)is emerging as a promising solution paradigm to meet the requirements of compute-intensive applications in internet of vehicle(IoV).Non-orthogonal multiple access(NOMA)has advantages in improving spectrum efficiency and dealing with bandwidth scarcity and cost.It is an encouraging progress combining VEC and NOMA.In this paper,we jointly optimize task offloading decision and resource allocation to maximize the service utility of the NOMA-VEC system.To solve the optimization problem,we propose a multiagent deep graph reinforcement learning algorithm.The algorithm extracts the topological features and relationship information between agents from the system state as observations,outputs task offloading decision and resource allocation simultaneously with local policy network,which is updated by a local learner.Simulation results demonstrate that the proposed method achieves a 1.52%∼5.80%improvement compared with the benchmark algorithms in system service utility. 展开更多
关键词 edge computing graph convolutional network reinforcement learning task offloading
下载PDF
Multi-circular formation control with reinforced transient profiles for nonholonomic vehicles:A path-following framework
18
作者 Jintao Zhang Xingling Shao +1 位作者 Wendong Zhang Zongyu Zuo 《Defence Technology(防务技术)》 SCIE EI CAS CSCD 2024年第1期278-287,共10页
This article investigates a multi-circular path-following formation control with reinforced transient profiles for nonholonomic vehicles connected by a digraph.A multi-circular formation controller endowed with the fe... This article investigates a multi-circular path-following formation control with reinforced transient profiles for nonholonomic vehicles connected by a digraph.A multi-circular formation controller endowed with the feature of spatial-temporal decoupling is devised for a group of vehicles guided by a virtual leader evolving along an implicit path,which allows for a circumnavigation on multiple circles with an anticipant angular spacing.In addition,notice that it typically imposes a stringent time constraint on time-sensitive enclosing scenarios,hence an improved prescribed performance control(IPPC)using novel tighter behavior boundaries is presented to enhance transient capabilities with an ensured appointed-time convergence free from any overshoots.The significant merits are that coordinated circumnavigation along different circles can be realized via executing geometric and dynamic assignments independently with modified transient profiles.Furthermore,all variables existing in the entire system are analyzed to be convergent.Simulation and experimental results are provided to validate the utility of suggested solution. 展开更多
关键词 Multi-circular formation reinforced transient profiles Nonholonomic vehicles Path following
下载PDF
Enhancing Image Description Generation through Deep Reinforcement Learning:Fusing Multiple Visual Features and Reward Mechanisms
19
作者 Yan Li Qiyuan Wang Kaidi Jia 《Computers, Materials & Continua》 SCIE EI 2024年第2期2469-2489,共21页
Image description task is the intersection of computer vision and natural language processing,and it has important prospects,including helping computers understand images and obtaining information for the visually imp... Image description task is the intersection of computer vision and natural language processing,and it has important prospects,including helping computers understand images and obtaining information for the visually impaired.This study presents an innovative approach employing deep reinforcement learning to enhance the accuracy of natural language descriptions of images.Our method focuses on refining the reward function in deep reinforcement learning,facilitating the generation of precise descriptions by aligning visual and textual features more closely.Our approach comprises three key architectures.Firstly,it utilizes Residual Network 101(ResNet-101)and Faster Region-based Convolutional Neural Network(Faster R-CNN)to extract average and local image features,respectively,followed by the implementation of a dual attention mechanism for intricate feature fusion.Secondly,the Transformer model is engaged to derive contextual semantic features from textual data.Finally,the generation of descriptive text is executed through a two-layer long short-term memory network(LSTM),directed by the value and reward functions.Compared with the image description method that relies on deep learning,the score of Bilingual Evaluation Understudy(BLEU-1)is 0.762,which is 1.6%higher,and the score of BLEU-4 is 0.299.Consensus-based Image Description Evaluation(CIDEr)scored 0.998,Recall-Oriented Understudy for Gisting Evaluation(ROUGE)scored 0.552,the latter improved by 0.36%.These results not only attest to the viability of our approach but also highlight its superiority in the realm of image description.Future research can explore the integration of our method with other artificial intelligence(AI)domains,such as emotional AI,to create more nuanced and context-aware systems. 展开更多
关键词 Image description deep reinforcement learning attention mechanism
下载PDF
Safety-Constrained Multi-Agent Reinforcement Learning for Power Quality Control in Distributed Renewable Energy Networks
20
作者 Yongjiang Zhao Haoyi Zhong Chang Cyoon Lim 《Computers, Materials & Continua》 SCIE EI 2024年第4期449-471,共23页
This paper examines the difficulties of managing distributed power systems,notably due to the increasing use of renewable energy sources,and focuses on voltage control challenges exacerbated by their variable nature i... This paper examines the difficulties of managing distributed power systems,notably due to the increasing use of renewable energy sources,and focuses on voltage control challenges exacerbated by their variable nature in modern power grids.To tackle the unique challenges of voltage control in distributed renewable energy networks,researchers are increasingly turning towards multi-agent reinforcement learning(MARL).However,MARL raises safety concerns due to the unpredictability in agent actions during their exploration phase.This unpredictability can lead to unsafe control measures.To mitigate these safety concerns in MARL-based voltage control,our study introduces a novel approach:Safety-ConstrainedMulti-Agent Reinforcement Learning(SC-MARL).This approach incorporates a specialized safety constraint module specifically designed for voltage control within the MARL framework.This module ensures that the MARL agents carry out voltage control actions safely.The experiments demonstrate that,in the 33-buses,141-buses,and 322-buses power systems,employing SC-MARL for voltage control resulted in a reduction of the Voltage Out of Control Rate(%V.out)from0.43,0.24,and 2.95 to 0,0.01,and 0.03,respectively.Additionally,the Reactive Power Loss(Q loss)decreased from 0.095,0.547,and 0.017 to 0.062,0.452,and 0.016 in the corresponding systems. 展开更多
关键词 Power quality control multi-agent reinforcement learning safety-constrained MARL
下载PDF
上一页 1 2 250 下一页 到第
使用帮助 返回顶部