The emergence of various new services has posed a huge challenge to the existing network architecture.To improve the network delay and backhaul pressure,caching popular contents at the edge of network has been conside...The emergence of various new services has posed a huge challenge to the existing network architecture.To improve the network delay and backhaul pressure,caching popular contents at the edge of network has been considered as a feasible scheme.However,how to efficiently utilize the limited caching resources to cache diverse contents has been confirmed as a tough problem in the past decade.In this paper,considering the time-varying user requests and the heterogeneous content sizes,a user preference aware hierarchical cooperative caching strategy in edge-user caching architecture is proposed.We divide the caching strategy into three phases,that is,the content placement,the content delivery and the content update.In the content placement phase,a cooperative content placement algorithm for local content popularity is designed to cache contents proactively.In the content delivery phase,a cooperative delivery algorithm is proposed to deliver the cached contents.In the content update phase,a content update algorithm is proposed according to the popularity of the contents.Finally,the proposed caching strategy is validated using the MovieLens dataset,and the results reveal that the proposed strategy improves the delay performance by at least 35.3%compared with the other three benchmark strategies.展开更多
The growing demand for low delay vehicular content has put tremendous strain on the backbone network.As a promising alternative,cooperative content caching among different cache nodes can reduce content access delay.H...The growing demand for low delay vehicular content has put tremendous strain on the backbone network.As a promising alternative,cooperative content caching among different cache nodes can reduce content access delay.However,heterogeneous cache nodes have different communication modes and limited caching capacities.In addition,the high mobility of vehicles renders the more complicated caching environment.Therefore,performing efficient cooperative caching becomes a key issue.In this paper,we propose a cross-tier cooperative caching architecture for all contents,which allows the distributed cache nodes to cooperate.Then,we devise the communication link and content caching model to facilitate timely content delivery.Aiming at minimizing transmission delay and cache cost,an optimization problem is formulated.Furthermore,we use a multi-agent deep reinforcement learning(MADRL)approach to model the decision-making process for caching among heterogeneous cache nodes,where each agent interacts with the environment collectively,receives observations yet a common reward,and learns its own optimal policy.Extensive simulations validate that the MADRL approach can enhance hit ratio while reducing transmission delay and cache cost.展开更多
Mobile Edge Computing(MEC)is a technology designed for the on-demand provisioning of computing and storage services,strategically positioned close to users.In the MEC environment,frequently accessed content can be dep...Mobile Edge Computing(MEC)is a technology designed for the on-demand provisioning of computing and storage services,strategically positioned close to users.In the MEC environment,frequently accessed content can be deployed and cached on edge servers to optimize the efficiency of content delivery,ultimately enhancing the quality of the user experience.However,due to the typical placement of edge devices and nodes at the network’s periphery,these components may face various potential fault tolerance challenges,including network instability,device failures,and resource constraints.Considering the dynamic nature ofMEC,making high-quality content caching decisions for real-time mobile applications,especially those sensitive to latency,by effectively utilizing mobility information,continues to be a significant challenge.In response to this challenge,this paper introduces FT-MAACC,a mobility-aware caching solution grounded in multi-agent deep reinforcement learning and equipped with fault tolerance mechanisms.This approach comprehensively integrates content adaptivity algorithms to evaluate the priority of highly user-adaptive cached content.Furthermore,it relies on collaborative caching strategies based onmulti-agent deep reinforcement learningmodels and establishes a fault-tolerancemodel to ensure the system’s reliability,availability,and persistence.Empirical results unequivocally demonstrate that FTMAACC outperforms its peer methods in cache hit rates and transmission latency.展开更多
Scalable video coding(SVC)has been widely used in video-on-demand(VOD)service,to efficiently satisfy users’different video quality requirements and dynamically adjust video stream to timevariant wireless channels.Und...Scalable video coding(SVC)has been widely used in video-on-demand(VOD)service,to efficiently satisfy users’different video quality requirements and dynamically adjust video stream to timevariant wireless channels.Under the 5G network structure,we consider a cooperative caching scheme inside each cluster with SVC to economically utilize the limited caching storage.A novel multi-agent deep reinforcement learning(MADRL)framework is proposed to jointly optimize the video access delay and users’satisfaction,where an aggregation node is introduced helping individual agents to achieve global observations and overall system rewards.Moreover,to cope with the large action space caused by the large number of videos and users,a dimension decomposition method is embedded into the neural network in each agent,which greatly reduce the computational complexity and memory cost of the reinforcement learning.Experimental results show that:1)the proposed value-decomposed dimensional network(VDDN)algorithm achieves an obvious performance gain versus the traditional MADRL;2)the proposed VDDN algorithm can handle an extremely large action space and quickly converge with a low computational complexity.展开更多
Edge caching is an emerging technology for supporting massive content access in mobile edge networks to address rapidly growing Internet of Things(IoT)services and content applications.However,the edge server is limit...Edge caching is an emerging technology for supporting massive content access in mobile edge networks to address rapidly growing Internet of Things(IoT)services and content applications.However,the edge server is limited with the computation/storage capacity,which causes a low cache hit.Cooperative edge caching jointing neighbor edge servers is regarded as a promising technique to improve cache hit and reduce congestion of the networks.Further,recommender systems can provide personalized content services to meet user’s requirements in the entertainment-oriented mobile networks.Therefore,we investigate the issue of joint cooperative edge caching and recommender systems to achieve additional cache gains by the soft caching framework.To measure the cache profits,the optimization problem is formulated as a 0-1 Integer Linear Programming(ILP),which is NP-hard.Specifically,the method of processing content requests is defined as server actions,we determine the server actions to maximize the quality of experience(QoE).We propose a cachefriendly heuristic algorithm to solve it.Simulation results demonstrate that the proposed framework has superior performance in improving the QoE.展开更多
Considering Wireless Sensor Networks (WSNs) in today’s scenario, sending and receiving uninterrupted sensory data remains a challenge to achieve with minimal latency and energy consumption as low as possible. Energy ...Considering Wireless Sensor Networks (WSNs) in today’s scenario, sending and receiving uninterrupted sensory data remains a challenge to achieve with minimal latency and energy consumption as low as possible. Energy consumption is exponentially growing in computing devices such as computers, embedded systems, portable devices, and wireless sensor networks. Extensive research has been in practice recently to minimize energy consumption without compromising the Quality of Service (QoS) that is to provide data to the requester node with minimum Delay and high Reliability. In this paper, a cooperative caching algorithm is used with the proposed Distributed Energy Aware Routing (DEAR) protocol that attempts to minimize energy consumption by reducing the packet overhead in the network and also providing the data to the requester with minimum delay by retrieving requested datum from the nearby caching node available in the vicinity of the requester or sink node. The simulation results clearly show that the energy consumption is less when the grid-based analytical model is used against the star/cluster based model while keeping the same necessary attributes.展开更多
Integrating the blockchain technology into mobile-edge computing(MEC)networks with multiple cooperative MEC servers(MECS)providing a promising solution to improving resource utilization,and helping establish a secure ...Integrating the blockchain technology into mobile-edge computing(MEC)networks with multiple cooperative MEC servers(MECS)providing a promising solution to improving resource utilization,and helping establish a secure reward mechanism that can facilitate load balancing among MECS.In addition,intelligent management of service caching and load balancing can improve the network utility in MEC blockchain networks with multiple types of workloads.In this paper,we investigate a learningbased joint service caching and load balancing policy for optimizing the communication and computation resources allocation,so as to improve the resource utilization of MEC blockchain networks.We formulate the problem as a challenging long-term network revenue maximization Markov decision process(MDP)problem.To address the highly dynamic and high dimension of system states,we design a joint service caching and load balancing algorithm based on the double-dueling Deep Q network(DQN)approach.The simulation results validate the feasibility and superior performance of our proposed algorithm over several baseline schemes.展开更多
At present, there are many effective ways to achieve high performance in cluster system storage management, including server-end disk, server-end caching, local caching and cooperative caching. The cooperative caching...At present, there are many effective ways to achieve high performance in cluster system storage management, including server-end disk, server-end caching, local caching and cooperative caching. The cooperative caching mechanism shares caches among different clients so as to avoid expensive disk access costs and to improve overall throughput of cluster system. In this paper, a Single Copy Cooperative Cache model is proposed together with block lookup algorithm, block replacement algorithm and the consistency algorithm based on the model. Meanwhile, the prototype system of the model is implemented in PVFS file system. Finally, the performance of this system is tested in InfiniBand Framework, the result of which shows that in contrast to the original PVFS system, read performance of PVFS file system is improved by about two times, while write performance is reduced by nearly ten percent.展开更多
Multimedia streaming served through peer-to-peer (P2P) networks is booming nowadays. However, the end-to-end streaming quality is generally unstable due to the variability of the state of serve-peers. On the other han...Multimedia streaming served through peer-to-peer (P2P) networks is booming nowadays. However, the end-to-end streaming quality is generally unstable due to the variability of the state of serve-peers. On the other hand, proxy caching is a bandwidth-efficient scheme for streaming over the Internet, whereas it is a substantially expensive method needing dedicated powerful proxy servers. In this paper, we present a P2P cooperative streaming architecture combined with the advantages of both P2P networks and multimedia proxy caching techniques to improve the streaming quality of participating clients. In this frame- work, a client will simultaneously retrieve contents from the server and other peers that have viewed and cached the same title before. In the meantime, the client will also selectively cache the aggregated video content so as to serve still future clients. The associate protocol to facilitate the multi-path streaming and a distributed utility-based partial caching scheme are detailedly dis- cussed. We demonstrate the effectiveness of this proposed architecture through extensive simulation experiments on large, Inter- net-like topologies.展开更多
Content-Centric Networking is a novel future network architecture that attracts increasing research interests in recent years. In-network caching has been regarded as a prominent feature of Content-Centric Networking ...Content-Centric Networking is a novel future network architecture that attracts increasing research interests in recent years. In-network caching has been regarded as a prominent feature of Content-Centric Networking since it is able to reduce the network traffic, alleviate the server bottleneck and decrease the user access latency. However, the CCN default caching scheme results in a high caching redundancy, causing an urgent need for an efficient caching scheme. To address this issue, we propose a novel implicit cooperative caching scheme to efficiently reduce the caching redundancy and improve the cache resources utilization. The simulation results show that our design achieves a higher hit ratio and a shorter cache hit distance in comparison with the other typical caching schemes.展开更多
The current Internet is based on host-centric networking, and a user needs to know the host address before reaching a data target in the network. The new architecture of information-centric networking (ICN) facilitate...The current Internet is based on host-centric networking, and a user needs to know the host address before reaching a data target in the network. The new architecture of information-centric networking (ICN) facilitates users to locate data targets by giving their data names without any information about host addresses. In-network caching is one of the prominent features in ICN, which allows network routers to cache data contents. In this paper, we emphasize the management of in-network cache storage, and this includes the mechanisms of cache replacement and cache replication. A new cost function is then proposed to evaluate each cache content and the least valuable content is evicted when cache is full. To increase cache utilization, a cooperative caching policy among neighboring routers is proposed. The proper network locations to cache data contents are also discussed in the paper. Experimental results show the superiority of the proposed caching policy than some traditional caching polices.展开更多
介绍一种有效支持缓存协作的未来网络体系架构:智慧协同网络,然后提出了一种高效的协作缓存机制,称为Co Lo RCache。Co Lo RCache的主要目标是减小缓存冗余和建立缓存共享机制。我们通过仿真结果来验证Co Lo RCache。仿真数据表明,相比...介绍一种有效支持缓存协作的未来网络体系架构:智慧协同网络,然后提出了一种高效的协作缓存机制,称为Co Lo RCache。Co Lo RCache的主要目标是减小缓存冗余和建立缓存共享机制。我们通过仿真结果来验证Co Lo RCache。仿真数据表明,相比较于其他缓存机制,Co Lo RCache能够产生更高的缓存命中率和有着最小的请求命中距离。展开更多
基金supported by Natural Science Foundation of China(Grant 61901070,61801065,62271096,61871062,U20A20157 and 62061007)in part by the Science and Technology Research Program of Chongqing Municipal Education Commission(Grant KJQN202000603 and KJQN201900611)+3 种基金in part by the Natural Science Foundation of Chongqing(Grant CSTB2022NSCQMSX0468,cstc2020jcyjzdxmX0024 and cstc2021jcyjmsxmX0892)in part by University Innovation Research Group of Chongqing(Grant CxQT20017)in part by Youth Innovation Group Support Program of ICE Discipline of CQUPT(SCIE-QN-2022-04)in part by the Chongqing Graduate Student Scientific Research Innovation Project(CYB22246)。
文摘The emergence of various new services has posed a huge challenge to the existing network architecture.To improve the network delay and backhaul pressure,caching popular contents at the edge of network has been considered as a feasible scheme.However,how to efficiently utilize the limited caching resources to cache diverse contents has been confirmed as a tough problem in the past decade.In this paper,considering the time-varying user requests and the heterogeneous content sizes,a user preference aware hierarchical cooperative caching strategy in edge-user caching architecture is proposed.We divide the caching strategy into three phases,that is,the content placement,the content delivery and the content update.In the content placement phase,a cooperative content placement algorithm for local content popularity is designed to cache contents proactively.In the content delivery phase,a cooperative delivery algorithm is proposed to deliver the cached contents.In the content update phase,a content update algorithm is proposed according to the popularity of the contents.Finally,the proposed caching strategy is validated using the MovieLens dataset,and the results reveal that the proposed strategy improves the delay performance by at least 35.3%compared with the other three benchmark strategies.
基金supported by the National Natural Science Foundation of China(62231020,62101401)the Youth Innovation Team of Shaanxi Universities。
文摘The growing demand for low delay vehicular content has put tremendous strain on the backbone network.As a promising alternative,cooperative content caching among different cache nodes can reduce content access delay.However,heterogeneous cache nodes have different communication modes and limited caching capacities.In addition,the high mobility of vehicles renders the more complicated caching environment.Therefore,performing efficient cooperative caching becomes a key issue.In this paper,we propose a cross-tier cooperative caching architecture for all contents,which allows the distributed cache nodes to cooperate.Then,we devise the communication link and content caching model to facilitate timely content delivery.Aiming at minimizing transmission delay and cache cost,an optimization problem is formulated.Furthermore,we use a multi-agent deep reinforcement learning(MADRL)approach to model the decision-making process for caching among heterogeneous cache nodes,where each agent interacts with the environment collectively,receives observations yet a common reward,and learns its own optimal policy.Extensive simulations validate that the MADRL approach can enhance hit ratio while reducing transmission delay and cache cost.
基金supported by the Innovation Fund Project of Jiangxi Normal University(YJS2022065)the Domestic Visiting Program of Jiangxi Normal University.
文摘Mobile Edge Computing(MEC)is a technology designed for the on-demand provisioning of computing and storage services,strategically positioned close to users.In the MEC environment,frequently accessed content can be deployed and cached on edge servers to optimize the efficiency of content delivery,ultimately enhancing the quality of the user experience.However,due to the typical placement of edge devices and nodes at the network’s periphery,these components may face various potential fault tolerance challenges,including network instability,device failures,and resource constraints.Considering the dynamic nature ofMEC,making high-quality content caching decisions for real-time mobile applications,especially those sensitive to latency,by effectively utilizing mobility information,continues to be a significant challenge.In response to this challenge,this paper introduces FT-MAACC,a mobility-aware caching solution grounded in multi-agent deep reinforcement learning and equipped with fault tolerance mechanisms.This approach comprehensively integrates content adaptivity algorithms to evaluate the priority of highly user-adaptive cached content.Furthermore,it relies on collaborative caching strategies based onmulti-agent deep reinforcement learningmodels and establishes a fault-tolerancemodel to ensure the system’s reliability,availability,and persistence.Empirical results unequivocally demonstrate that FTMAACC outperforms its peer methods in cache hit rates and transmission latency.
基金supported by the National Natural Science Foundation of China under Grant No.61801119。
文摘Scalable video coding(SVC)has been widely used in video-on-demand(VOD)service,to efficiently satisfy users’different video quality requirements and dynamically adjust video stream to timevariant wireless channels.Under the 5G network structure,we consider a cooperative caching scheme inside each cluster with SVC to economically utilize the limited caching storage.A novel multi-agent deep reinforcement learning(MADRL)framework is proposed to jointly optimize the video access delay and users’satisfaction,where an aggregation node is introduced helping individual agents to achieve global observations and overall system rewards.Moreover,to cope with the large action space caused by the large number of videos and users,a dimension decomposition method is embedded into the neural network in each agent,which greatly reduce the computational complexity and memory cost of the reinforcement learning.Experimental results show that:1)the proposed value-decomposed dimensional network(VDDN)algorithm achieves an obvious performance gain versus the traditional MADRL;2)the proposed VDDN algorithm can handle an extremely large action space and quickly converge with a low computational complexity.
基金supported in part by National Key R&D Program of China under Grant Nos. 2018YFB2100100 and 2018YFF0214700National NSFC under Grant Nos. 61902044 and 62072060+4 种基金Chongqing Research Program of Basic Research and Frontier Technology under Grant No. CSTC2019-jcyjmsxmX0589Key Research Program of Chongqing Science and Technology Commission under Grant Nos. CSTC2017jcyjBX0025 and CSTC2019jscxzdztzxX0031Fundamental Research Funds for the Central Universities under Grant No.2020CDJQY-A022Chinese National Engineering Laboratory for Big Data System Computing TechnologyCanadian NSERC
文摘Edge caching is an emerging technology for supporting massive content access in mobile edge networks to address rapidly growing Internet of Things(IoT)services and content applications.However,the edge server is limited with the computation/storage capacity,which causes a low cache hit.Cooperative edge caching jointing neighbor edge servers is regarded as a promising technique to improve cache hit and reduce congestion of the networks.Further,recommender systems can provide personalized content services to meet user’s requirements in the entertainment-oriented mobile networks.Therefore,we investigate the issue of joint cooperative edge caching and recommender systems to achieve additional cache gains by the soft caching framework.To measure the cache profits,the optimization problem is formulated as a 0-1 Integer Linear Programming(ILP),which is NP-hard.Specifically,the method of processing content requests is defined as server actions,we determine the server actions to maximize the quality of experience(QoE).We propose a cachefriendly heuristic algorithm to solve it.Simulation results demonstrate that the proposed framework has superior performance in improving the QoE.
文摘Considering Wireless Sensor Networks (WSNs) in today’s scenario, sending and receiving uninterrupted sensory data remains a challenge to achieve with minimal latency and energy consumption as low as possible. Energy consumption is exponentially growing in computing devices such as computers, embedded systems, portable devices, and wireless sensor networks. Extensive research has been in practice recently to minimize energy consumption without compromising the Quality of Service (QoS) that is to provide data to the requester node with minimum Delay and high Reliability. In this paper, a cooperative caching algorithm is used with the proposed Distributed Energy Aware Routing (DEAR) protocol that attempts to minimize energy consumption by reducing the packet overhead in the network and also providing the data to the requester with minimum delay by retrieving requested datum from the nearby caching node available in the vicinity of the requester or sink node. The simulation results clearly show that the energy consumption is less when the grid-based analytical model is used against the star/cluster based model while keeping the same necessary attributes.
基金supported in part by the National Natural Science Foundation of China 62072096the Fundamental Research Funds for the Central Universities under Grant 2232020A-12+4 种基金the International S&T Cooperation Program of Shanghai Science and Technology Commission under Grant 20220713000the Young Top-notch Talent Program in Shanghaithe"Shuguang Program"of Shanghai Education Development Foundation and Shanghai Municipal Education Commissionthe Fundamental Research Funds for the Central Universities and Graduate Student Innovation Fund of Donghua University CUSF-DH-D-2019093supported in part by the NSF under grants CNS-2107190 and ECCS-1923717。
文摘Integrating the blockchain technology into mobile-edge computing(MEC)networks with multiple cooperative MEC servers(MECS)providing a promising solution to improving resource utilization,and helping establish a secure reward mechanism that can facilitate load balancing among MECS.In addition,intelligent management of service caching and load balancing can improve the network utility in MEC blockchain networks with multiple types of workloads.In this paper,we investigate a learningbased joint service caching and load balancing policy for optimizing the communication and computation resources allocation,so as to improve the resource utilization of MEC blockchain networks.We formulate the problem as a challenging long-term network revenue maximization Markov decision process(MDP)problem.To address the highly dynamic and high dimension of system states,we design a joint service caching and load balancing algorithm based on the double-dueling Deep Q network(DQN)approach.The simulation results validate the feasibility and superior performance of our proposed algorithm over several baseline schemes.
基金This work was supported by the National High Technology Development Program of China under Grant(No.2004AA111110,No.2006AA01A109)
文摘At present, there are many effective ways to achieve high performance in cluster system storage management, including server-end disk, server-end caching, local caching and cooperative caching. The cooperative caching mechanism shares caches among different clients so as to avoid expensive disk access costs and to improve overall throughput of cluster system. In this paper, a Single Copy Cooperative Cache model is proposed together with block lookup algorithm, block replacement algorithm and the consistency algorithm based on the model. Meanwhile, the prototype system of the model is implemented in PVFS file system. Finally, the performance of this system is tested in InfiniBand Framework, the result of which shows that in contrast to the original PVFS system, read performance of PVFS file system is improved by about two times, while write performance is reduced by nearly ten percent.
基金Project (Nos. 90412012 and 60673160) supported by the NationalNatural Science Foundation of China
文摘Multimedia streaming served through peer-to-peer (P2P) networks is booming nowadays. However, the end-to-end streaming quality is generally unstable due to the variability of the state of serve-peers. On the other hand, proxy caching is a bandwidth-efficient scheme for streaming over the Internet, whereas it is a substantially expensive method needing dedicated powerful proxy servers. In this paper, we present a P2P cooperative streaming architecture combined with the advantages of both P2P networks and multimedia proxy caching techniques to improve the streaming quality of participating clients. In this frame- work, a client will simultaneously retrieve contents from the server and other peers that have viewed and cached the same title before. In the meantime, the client will also selectively cache the aggregated video content so as to serve still future clients. The associate protocol to facilitate the multi-path streaming and a distributed utility-based partial caching scheme are detailedly dis- cussed. We demonstrate the effectiveness of this proposed architecture through extensive simulation experiments on large, Inter- net-like topologies.
基金supported in part by the 973 Program under Grant No.2013CB329100in part by NSFC under Grant No.61422101,62171200,and 62132017+1 种基金in part by the Ph.D. Programs Foundation of MOE of China under Grant No.20130009110014in part by the Fundamental Research Funds for the Central Universities under Grant No.2016JBZ002
文摘Content-Centric Networking is a novel future network architecture that attracts increasing research interests in recent years. In-network caching has been regarded as a prominent feature of Content-Centric Networking since it is able to reduce the network traffic, alleviate the server bottleneck and decrease the user access latency. However, the CCN default caching scheme results in a high caching redundancy, causing an urgent need for an efficient caching scheme. To address this issue, we propose a novel implicit cooperative caching scheme to efficiently reduce the caching redundancy and improve the cache resources utilization. The simulation results show that our design achieves a higher hit ratio and a shorter cache hit distance in comparison with the other typical caching schemes.
文摘The current Internet is based on host-centric networking, and a user needs to know the host address before reaching a data target in the network. The new architecture of information-centric networking (ICN) facilitates users to locate data targets by giving their data names without any information about host addresses. In-network caching is one of the prominent features in ICN, which allows network routers to cache data contents. In this paper, we emphasize the management of in-network cache storage, and this includes the mechanisms of cache replacement and cache replication. A new cost function is then proposed to evaluate each cache content and the least valuable content is evicted when cache is full. To increase cache utilization, a cooperative caching policy among neighboring routers is proposed. The proper network locations to cache data contents are also discussed in the paper. Experimental results show the superiority of the proposed caching policy than some traditional caching polices.
文摘介绍一种有效支持缓存协作的未来网络体系架构:智慧协同网络,然后提出了一种高效的协作缓存机制,称为Co Lo RCache。Co Lo RCache的主要目标是减小缓存冗余和建立缓存共享机制。我们通过仿真结果来验证Co Lo RCache。仿真数据表明,相比较于其他缓存机制,Co Lo RCache能够产生更高的缓存命中率和有着最小的请求命中距离。