The emergence of various new services has posed a huge challenge to the existing network architecture.To improve the network delay and backhaul pressure,caching popular contents at the edge of network has been conside...The emergence of various new services has posed a huge challenge to the existing network architecture.To improve the network delay and backhaul pressure,caching popular contents at the edge of network has been considered as a feasible scheme.However,how to efficiently utilize the limited caching resources to cache diverse contents has been confirmed as a tough problem in the past decade.In this paper,considering the time-varying user requests and the heterogeneous content sizes,a user preference aware hierarchical cooperative caching strategy in edge-user caching architecture is proposed.We divide the caching strategy into three phases,that is,the content placement,the content delivery and the content update.In the content placement phase,a cooperative content placement algorithm for local content popularity is designed to cache contents proactively.In the content delivery phase,a cooperative delivery algorithm is proposed to deliver the cached contents.In the content update phase,a content update algorithm is proposed according to the popularity of the contents.Finally,the proposed caching strategy is validated using the MovieLens dataset,and the results reveal that the proposed strategy improves the delay performance by at least 35.3%compared with the other three benchmark strategies.展开更多
Scalable video coding(SVC)has been widely used in video-on-demand(VOD)service,to efficiently satisfy users’different video quality requirements and dynamically adjust video stream to timevariant wireless channels.Und...Scalable video coding(SVC)has been widely used in video-on-demand(VOD)service,to efficiently satisfy users’different video quality requirements and dynamically adjust video stream to timevariant wireless channels.Under the 5G network structure,we consider a cooperative caching scheme inside each cluster with SVC to economically utilize the limited caching storage.A novel multi-agent deep reinforcement learning(MADRL)framework is proposed to jointly optimize the video access delay and users’satisfaction,where an aggregation node is introduced helping individual agents to achieve global observations and overall system rewards.Moreover,to cope with the large action space caused by the large number of videos and users,a dimension decomposition method is embedded into the neural network in each agent,which greatly reduce the computational complexity and memory cost of the reinforcement learning.Experimental results show that:1)the proposed value-decomposed dimensional network(VDDN)algorithm achieves an obvious performance gain versus the traditional MADRL;2)the proposed VDDN algorithm can handle an extremely large action space and quickly converge with a low computational complexity.展开更多
Mobile Edge Computing(MEC)is a technology designed for the on-demand provisioning of computing and storage services,strategically positioned close to users.In the MEC environment,frequently accessed content can be dep...Mobile Edge Computing(MEC)is a technology designed for the on-demand provisioning of computing and storage services,strategically positioned close to users.In the MEC environment,frequently accessed content can be deployed and cached on edge servers to optimize the efficiency of content delivery,ultimately enhancing the quality of the user experience.However,due to the typical placement of edge devices and nodes at the network’s periphery,these components may face various potential fault tolerance challenges,including network instability,device failures,and resource constraints.Considering the dynamic nature ofMEC,making high-quality content caching decisions for real-time mobile applications,especially those sensitive to latency,by effectively utilizing mobility information,continues to be a significant challenge.In response to this challenge,this paper introduces FT-MAACC,a mobility-aware caching solution grounded in multi-agent deep reinforcement learning and equipped with fault tolerance mechanisms.This approach comprehensively integrates content adaptivity algorithms to evaluate the priority of highly user-adaptive cached content.Furthermore,it relies on collaborative caching strategies based onmulti-agent deep reinforcement learningmodels and establishes a fault-tolerancemodel to ensure the system’s reliability,availability,and persistence.Empirical results unequivocally demonstrate that FTMAACC outperforms its peer methods in cache hit rates and transmission latency.展开更多
The growing demand for low delay vehicular content has put tremendous strain on the backbone network.As a promising alternative,cooperative content caching among different cache nodes can reduce content access delay.H...The growing demand for low delay vehicular content has put tremendous strain on the backbone network.As a promising alternative,cooperative content caching among different cache nodes can reduce content access delay.However,heterogeneous cache nodes have different communication modes and limited caching capacities.In addition,the high mobility of vehicles renders the more complicated caching environment.Therefore,performing efficient cooperative caching becomes a key issue.In this paper,we propose a cross-tier cooperative caching architecture for all contents,which allows the distributed cache nodes to cooperate.Then,we devise the communication link and content caching model to facilitate timely content delivery.Aiming at minimizing transmission delay and cache cost,an optimization problem is formulated.Furthermore,we use a multi-agent deep reinforcement learning(MADRL)approach to model the decision-making process for caching among heterogeneous cache nodes,where each agent interacts with the environment collectively,receives observations yet a common reward,and learns its own optimal policy.Extensive simulations validate that the MADRL approach can enhance hit ratio while reducing transmission delay and cache cost.展开更多
Edge caching is an emerging technology for supporting massive content access in mobile edge networks to address rapidly growing Internet of Things(IoT)services and content applications.However,the edge server is limit...Edge caching is an emerging technology for supporting massive content access in mobile edge networks to address rapidly growing Internet of Things(IoT)services and content applications.However,the edge server is limited with the computation/storage capacity,which causes a low cache hit.Cooperative edge caching jointing neighbor edge servers is regarded as a promising technique to improve cache hit and reduce congestion of the networks.Further,recommender systems can provide personalized content services to meet user’s requirements in the entertainment-oriented mobile networks.Therefore,we investigate the issue of joint cooperative edge caching and recommender systems to achieve additional cache gains by the soft caching framework.To measure the cache profits,the optimization problem is formulated as a 0-1 Integer Linear Programming(ILP),which is NP-hard.Specifically,the method of processing content requests is defined as server actions,we determine the server actions to maximize the quality of experience(QoE).We propose a cachefriendly heuristic algorithm to solve it.Simulation results demonstrate that the proposed framework has superior performance in improving the QoE.展开更多
At present, there are many effective ways to achieve high performance in cluster system storage management, including server-end disk, server-end caching, local caching and cooperative caching. The cooperative caching...At present, there are many effective ways to achieve high performance in cluster system storage management, including server-end disk, server-end caching, local caching and cooperative caching. The cooperative caching mechanism shares caches among different clients so as to avoid expensive disk access costs and to improve overall throughput of cluster system. In this paper, a Single Copy Cooperative Cache model is proposed together with block lookup algorithm, block replacement algorithm and the consistency algorithm based on the model. Meanwhile, the prototype system of the model is implemented in PVFS file system. Finally, the performance of this system is tested in InfiniBand Framework, the result of which shows that in contrast to the original PVFS system, read performance of PVFS file system is improved by about two times, while write performance is reduced by nearly ten percent.展开更多
Content-Centric Networking is a novel future network architecture that attracts increasing research interests in recent years. In-network caching has been regarded as a prominent feature of Content-Centric Networking ...Content-Centric Networking is a novel future network architecture that attracts increasing research interests in recent years. In-network caching has been regarded as a prominent feature of Content-Centric Networking since it is able to reduce the network traffic, alleviate the server bottleneck and decrease the user access latency. However, the CCN default caching scheme results in a high caching redundancy, causing an urgent need for an efficient caching scheme. To address this issue, we propose a novel implicit cooperative caching scheme to efficiently reduce the caching redundancy and improve the cache resources utilization. The simulation results show that our design achieves a higher hit ratio and a shorter cache hit distance in comparison with the other typical caching schemes.展开更多
In cooperative cache research domain, most of previous work engage in peer-to-peer systems and distributed systems, but do not involve applying cooperative semantic cache in mobile computing environments, which wirele...In cooperative cache research domain, most of previous work engage in peer-to-peer systems and distributed systems, but do not involve applying cooperative semantic cache in mobile computing environments, which wireless communications disconnect at times and clients move frequently. In this paper, we extend the general semantic cache mechanism by enabling mobile clients to share their local semantic caches in a cooperative matter, and the process way and flow chart of the algorithm are described in detail. In addition, we discuss the methods used in cache consistence maintenance, which focus on confirm receiver of the periodic cache invalidation report and the process of validate client's local cache. The experiment results indicate cooperative semantic cache mechanism could reduce query response time and increase cache hit ratio effectively.展开更多
基金supported by Natural Science Foundation of China(Grant 61901070,61801065,62271096,61871062,U20A20157 and 62061007)in part by the Science and Technology Research Program of Chongqing Municipal Education Commission(Grant KJQN202000603 and KJQN201900611)+3 种基金in part by the Natural Science Foundation of Chongqing(Grant CSTB2022NSCQMSX0468,cstc2020jcyjzdxmX0024 and cstc2021jcyjmsxmX0892)in part by University Innovation Research Group of Chongqing(Grant CxQT20017)in part by Youth Innovation Group Support Program of ICE Discipline of CQUPT(SCIE-QN-2022-04)in part by the Chongqing Graduate Student Scientific Research Innovation Project(CYB22246)。
文摘The emergence of various new services has posed a huge challenge to the existing network architecture.To improve the network delay and backhaul pressure,caching popular contents at the edge of network has been considered as a feasible scheme.However,how to efficiently utilize the limited caching resources to cache diverse contents has been confirmed as a tough problem in the past decade.In this paper,considering the time-varying user requests and the heterogeneous content sizes,a user preference aware hierarchical cooperative caching strategy in edge-user caching architecture is proposed.We divide the caching strategy into three phases,that is,the content placement,the content delivery and the content update.In the content placement phase,a cooperative content placement algorithm for local content popularity is designed to cache contents proactively.In the content delivery phase,a cooperative delivery algorithm is proposed to deliver the cached contents.In the content update phase,a content update algorithm is proposed according to the popularity of the contents.Finally,the proposed caching strategy is validated using the MovieLens dataset,and the results reveal that the proposed strategy improves the delay performance by at least 35.3%compared with the other three benchmark strategies.
基金supported by the National Natural Science Foundation of China under Grant No.61801119。
文摘Scalable video coding(SVC)has been widely used in video-on-demand(VOD)service,to efficiently satisfy users’different video quality requirements and dynamically adjust video stream to timevariant wireless channels.Under the 5G network structure,we consider a cooperative caching scheme inside each cluster with SVC to economically utilize the limited caching storage.A novel multi-agent deep reinforcement learning(MADRL)framework is proposed to jointly optimize the video access delay and users’satisfaction,where an aggregation node is introduced helping individual agents to achieve global observations and overall system rewards.Moreover,to cope with the large action space caused by the large number of videos and users,a dimension decomposition method is embedded into the neural network in each agent,which greatly reduce the computational complexity and memory cost of the reinforcement learning.Experimental results show that:1)the proposed value-decomposed dimensional network(VDDN)algorithm achieves an obvious performance gain versus the traditional MADRL;2)the proposed VDDN algorithm can handle an extremely large action space and quickly converge with a low computational complexity.
基金supported by the Innovation Fund Project of Jiangxi Normal University(YJS2022065)the Domestic Visiting Program of Jiangxi Normal University.
文摘Mobile Edge Computing(MEC)is a technology designed for the on-demand provisioning of computing and storage services,strategically positioned close to users.In the MEC environment,frequently accessed content can be deployed and cached on edge servers to optimize the efficiency of content delivery,ultimately enhancing the quality of the user experience.However,due to the typical placement of edge devices and nodes at the network’s periphery,these components may face various potential fault tolerance challenges,including network instability,device failures,and resource constraints.Considering the dynamic nature ofMEC,making high-quality content caching decisions for real-time mobile applications,especially those sensitive to latency,by effectively utilizing mobility information,continues to be a significant challenge.In response to this challenge,this paper introduces FT-MAACC,a mobility-aware caching solution grounded in multi-agent deep reinforcement learning and equipped with fault tolerance mechanisms.This approach comprehensively integrates content adaptivity algorithms to evaluate the priority of highly user-adaptive cached content.Furthermore,it relies on collaborative caching strategies based onmulti-agent deep reinforcement learningmodels and establishes a fault-tolerancemodel to ensure the system’s reliability,availability,and persistence.Empirical results unequivocally demonstrate that FTMAACC outperforms its peer methods in cache hit rates and transmission latency.
基金supported by the National Natural Science Foundation of China(62231020,62101401)the Youth Innovation Team of Shaanxi Universities。
文摘The growing demand for low delay vehicular content has put tremendous strain on the backbone network.As a promising alternative,cooperative content caching among different cache nodes can reduce content access delay.However,heterogeneous cache nodes have different communication modes and limited caching capacities.In addition,the high mobility of vehicles renders the more complicated caching environment.Therefore,performing efficient cooperative caching becomes a key issue.In this paper,we propose a cross-tier cooperative caching architecture for all contents,which allows the distributed cache nodes to cooperate.Then,we devise the communication link and content caching model to facilitate timely content delivery.Aiming at minimizing transmission delay and cache cost,an optimization problem is formulated.Furthermore,we use a multi-agent deep reinforcement learning(MADRL)approach to model the decision-making process for caching among heterogeneous cache nodes,where each agent interacts with the environment collectively,receives observations yet a common reward,and learns its own optimal policy.Extensive simulations validate that the MADRL approach can enhance hit ratio while reducing transmission delay and cache cost.
基金supported in part by National Key R&D Program of China under Grant Nos. 2018YFB2100100 and 2018YFF0214700National NSFC under Grant Nos. 61902044 and 62072060+4 种基金Chongqing Research Program of Basic Research and Frontier Technology under Grant No. CSTC2019-jcyjmsxmX0589Key Research Program of Chongqing Science and Technology Commission under Grant Nos. CSTC2017jcyjBX0025 and CSTC2019jscxzdztzxX0031Fundamental Research Funds for the Central Universities under Grant No.2020CDJQY-A022Chinese National Engineering Laboratory for Big Data System Computing TechnologyCanadian NSERC
文摘Edge caching is an emerging technology for supporting massive content access in mobile edge networks to address rapidly growing Internet of Things(IoT)services and content applications.However,the edge server is limited with the computation/storage capacity,which causes a low cache hit.Cooperative edge caching jointing neighbor edge servers is regarded as a promising technique to improve cache hit and reduce congestion of the networks.Further,recommender systems can provide personalized content services to meet user’s requirements in the entertainment-oriented mobile networks.Therefore,we investigate the issue of joint cooperative edge caching and recommender systems to achieve additional cache gains by the soft caching framework.To measure the cache profits,the optimization problem is formulated as a 0-1 Integer Linear Programming(ILP),which is NP-hard.Specifically,the method of processing content requests is defined as server actions,we determine the server actions to maximize the quality of experience(QoE).We propose a cachefriendly heuristic algorithm to solve it.Simulation results demonstrate that the proposed framework has superior performance in improving the QoE.
基金This work was supported by the National High Technology Development Program of China under Grant(No.2004AA111110,No.2006AA01A109)
文摘At present, there are many effective ways to achieve high performance in cluster system storage management, including server-end disk, server-end caching, local caching and cooperative caching. The cooperative caching mechanism shares caches among different clients so as to avoid expensive disk access costs and to improve overall throughput of cluster system. In this paper, a Single Copy Cooperative Cache model is proposed together with block lookup algorithm, block replacement algorithm and the consistency algorithm based on the model. Meanwhile, the prototype system of the model is implemented in PVFS file system. Finally, the performance of this system is tested in InfiniBand Framework, the result of which shows that in contrast to the original PVFS system, read performance of PVFS file system is improved by about two times, while write performance is reduced by nearly ten percent.
基金supported in part by the 973 Program under Grant No.2013CB329100in part by NSFC under Grant No.61422101,62171200,and 62132017+1 种基金in part by the Ph.D. Programs Foundation of MOE of China under Grant No.20130009110014in part by the Fundamental Research Funds for the Central Universities under Grant No.2016JBZ002
文摘Content-Centric Networking is a novel future network architecture that attracts increasing research interests in recent years. In-network caching has been regarded as a prominent feature of Content-Centric Networking since it is able to reduce the network traffic, alleviate the server bottleneck and decrease the user access latency. However, the CCN default caching scheme results in a high caching redundancy, causing an urgent need for an efficient caching scheme. To address this issue, we propose a novel implicit cooperative caching scheme to efficiently reduce the caching redundancy and improve the cache resources utilization. The simulation results show that our design achieves a higher hit ratio and a shorter cache hit distance in comparison with the other typical caching schemes.
基金supported by the National Basic Research and Development Program of China (2007CB07100, 2007CB07106)the Ministry of Education in China Project of Humanities and Social Sciences (11YJCZH195)
文摘In cooperative cache research domain, most of previous work engage in peer-to-peer systems and distributed systems, but do not involve applying cooperative semantic cache in mobile computing environments, which wireless communications disconnect at times and clients move frequently. In this paper, we extend the general semantic cache mechanism by enabling mobile clients to share their local semantic caches in a cooperative matter, and the process way and flow chart of the algorithm are described in detail. In addition, we discuss the methods used in cache consistence maintenance, which focus on confirm receiver of the periodic cache invalidation report and the process of validate client's local cache. The experiment results indicate cooperative semantic cache mechanism could reduce query response time and increase cache hit ratio effectively.