The demands on conventional communication networks are increasing rapidly because of the exponential expansion of connected multimedia content.In light of the data-centric aspect of contemporary communication,the info...The demands on conventional communication networks are increasing rapidly because of the exponential expansion of connected multimedia content.In light of the data-centric aspect of contemporary communication,the information-centric network(ICN)paradigm offers hope for a solution by emphasizing content retrieval by name instead of location.If 5G networks are to meet the expected data demand surge from expanded connectivity and Internet of Things(IoT)devices,then effective caching solutions will be required tomaximize network throughput andminimize the use of resources.Hence,an ICN-based Cooperative Caching(ICN-CoC)technique has been used to select a cache by considering cache position,content attractiveness,and rate prediction.The findings show that utilizing our suggested approach improves caching regarding the Cache Hit Ratio(CHR)of 84.3%,Average Hop Minimization Ratio(AHMR)of 89.5%,and Mean Access Latency(MAL)of 0.4 s.Within a framework,it suggests improved caching strategies to handle the difficulty of effectively controlling data consumption in 5G networks.These improvements aim to make the network run more smoothly by enhancing content delivery,decreasing latency,and relieving congestion.By improving 5G communication systems’capacity tomanage the demands faced by modern data-centric applications,the research ultimately aids in advancement.展开更多
Computing-intensive and latency-sensitive user requests pose significant challenges to traditional cloud computing.In response to these challenges,mobile edge computing(MEC)has emerged as a new paradigm that extends t...Computing-intensive and latency-sensitive user requests pose significant challenges to traditional cloud computing.In response to these challenges,mobile edge computing(MEC)has emerged as a new paradigm that extends the computational,caching,and communication capabilities of cloud computing.By caching certain services on edge nodes,computational support can be provided for requests that are offloaded to the edges.However,previous studies on task offloading have generally not considered the impact of caching mechanisms and the cache space occupied by services.This oversight can lead to problems,such as high delays in task executions and invalidation of offloading decisions.To optimize task response time and ensure the availability of task offloading decisions,we investigate a task offloading method that considers caching mechanism.First,we incorporate the cache information of MEC into the model of task offloading and reduce the task offloading problem as a mixed integer nonlinear programming(MINLP)problem.Then,we propose an integer particle swarm optimization and improved genetic algorithm(IPSO_IGA)to solve the MINLP.IPSO_IGA exploits the evolutionary framework of particle swarm optimization.And it uses a crossover operator to update the positions of particles and an improved mutation operator to maintain the diversity of particles.Finally,extensive simulation experiments are conducted to evaluate the performance of the proposed algorithm.The experimental results demonstrate that IPSO_IGA can save 20%to 82%of the task completion time,compared with state-of-theart and classical algorithms.Moreover,IPSO_IGA is suitable for scenarios with complex network structures and computing-intensive tasks.展开更多
<div style="text-align:justify;"> <span style="font-family:Verdana;">There has been an explosion in the volume of data that is being accessed from the Internet. As a result, the risk of...<div style="text-align:justify;"> <span style="font-family:Verdana;">There has been an explosion in the volume of data that is being accessed from the Internet. As a result, the risk of a Web server being inundated with requests is ever-present. One approach to reducing the performance degradation that potentially comes from Web server overloading is to employ Web caching where data content is replicated in multiple locations. In this paper, we investigate the use of evolutionary algorithms to dynamically alter partition size in Web caches. We use established modeling techniques to compare the performance of our evolutionary algorithm to that found in statically-partitioned systems. Our results indicate that utilizing an evolutionary algorithm to dynamically alter partition sizes can lead to performance improvements especially in environments where the relative size of large to small pages is high.</span> </div>展开更多
With the increasing maritime activities and the rapidly developing maritime economy, the fifth-generation(5G) mobile communication system is expected to be deployed at the ocean. New technologies need to be explored t...With the increasing maritime activities and the rapidly developing maritime economy, the fifth-generation(5G) mobile communication system is expected to be deployed at the ocean. New technologies need to be explored to meet the requirements of ultra-reliable and low latency communications(URLLC) in the maritime communication network(MCN). Mobile edge computing(MEC) can achieve high energy efficiency in MCN at the cost of suffering from high control plane latency and low reliability. In terms of this issue, the mobile edge communications, computing, and caching(MEC3) technology is proposed to sink mobile computing, network control, and storage to the edge of the network. New methods that enable resource-efficient configurations and reduce redundant data transmissions can enable the reliable implementation of computing-intension and latency-sensitive applications. The key technologies of MEC3 to enable URLLC are analyzed and optimized in MCN. The best response-based offloading algorithm(BROA) is adopted to optimize task offloading. The simulation results show that the task latency can be decreased by 26.5’ ms, and the energy consumption in terminal users can be reduced to 66.6%.展开更多
Heterogeneous cellular networks(HCNs), by introducing caching capability, has been considered as a promising technique in 5 G era, which can bring contents closer to users to reduce the transmission delay, save scarce...Heterogeneous cellular networks(HCNs), by introducing caching capability, has been considered as a promising technique in 5 G era, which can bring contents closer to users to reduce the transmission delay, save scarce bandwidth resource. Although many works have been done for caching in HCNs, from an energy perspective, there still exists much space to develop a more energy-efficient system when considering the fact that the majority of base stations are under-utilized in the most of the time. Therefore, in this paper, by taking the activation mechanism for the base stations into account, we study a joint caching and activation mechanism design to further improve the energy efficiency, then we formulate the optimization problem as an Integer Linear Programming problem(ILP) to maximize the system energy saving. Due to the enormous computation complexity for finding the optimal solution, we introduced a Quantum-inspired Evolutionary Algorithm(QEA) to iteratively provide the global best solution. Numerical results show that our proposed algorithm presents an excellent performance, which is far better than the strategy of only considering caching without deactivation mechanism in the actual, normal situation. We also provide performance comparison amongour QEA, random sleeping algorithm and greedy algorithm, numerical results illustrate our introduced QEA performs best in accuracy and global optimality.展开更多
Streaming media data have strict delay and loss requirements, which cannot be adequately supported by the current Internet. Proxy caching technology is introduced to alleviate the effect of the bottleneck bandwidth on...Streaming media data have strict delay and loss requirements, which cannot be adequately supported by the current Internet. Proxy caching technology is introduced to alleviate the effect of the bottleneck bandwidth on the transporting quality. A new architecture of multimedia proxy caching is given to solve the problem of the traditional Web caching technology for the multimedia streaming. The mixed media management module and the (resource) management module are added to this architecture, the prefetching and replacement method is studied considering the characteristics of streaming media data, the concept of weighed popularity is given to enhance the (exactness) of the replacement algorithm.展开更多
The Web cluster has been a popular solution of network server system because of its scalability and cost effective ness. The cache configured in servers can result in increasing significantly performance, In this pape...The Web cluster has been a popular solution of network server system because of its scalability and cost effective ness. The cache configured in servers can result in increasing significantly performance, In this paper, we discuss the suitable configuration strategies for caching dynamic content by our experimental results. Considering the system itself can provide support for caching static Web page, such as computer memory cache and disk's own cache, we adopt a special pattern that only caches dynamic Web page in some experiments to enlarge cache space. The paper is introduced three different replacement algorithms in our cache proxy module to test the practical effects of caching dynamic pages under different conditions. The paper is chiefly analyzed the influences of generated time and accessed frequency on caching dynamic Web pages. The paper is also provided the detailed experiment results and main conclusions in the paper.展开更多
Modern shared-memory multi-core processors typically have shared Level 2(L2)or Level 3(L3)caches.Cache bottlenecks and replacement strategies are the main problems of such architectures,where multiple cores try to acc...Modern shared-memory multi-core processors typically have shared Level 2(L2)or Level 3(L3)caches.Cache bottlenecks and replacement strategies are the main problems of such architectures,where multiple cores try to access the shared cache simultaneously.The main problem in improving memory performance is the shared cache architecture and cache replacement.This paper documents the implementation of a Dual-Port Content Addressable Memory(DPCAM)and a modified Near-Far Access Replacement Algorithm(NFRA),which was previously proposed as a shared L2 cache layer in a multi-core processor.Standard Performance Evaluation Corporation(SPEC)Central Processing Unit(CPU)2006 benchmark workloads are used to evaluate the benefit of the shared L2 cache layer.Results show improved performance of the multicore processor’s DPCAM and NFRA algorithms,corresponding to a higher number of concurrent accesses to shared memory.The new architecture significantly increases system throughput and records performance improvements of up to 8.7%on various types of SPEC 2006 benchmarks.The miss rate is also improved by about 13%,with some exceptions in the sphinx3 and bzip2 benchmarks.These results could open a new window for solving the long-standing problems with shared cache in multi-core processors.展开更多
多服务移动边缘计算(multiple-services mobile edge computing,MSs-MEC)能根据需求自适应调整服务缓存决策,使得部署在用户侧的边缘服务器能够灵活处理不同服务类型的任务。但在实际应用中,特定类型任务的成功迁移依赖于服务环境的提...多服务移动边缘计算(multiple-services mobile edge computing,MSs-MEC)能根据需求自适应调整服务缓存决策,使得部署在用户侧的边缘服务器能够灵活处理不同服务类型的任务。但在实际应用中,特定类型任务的成功迁移依赖于服务环境的提前安装。此外,同时进行任务迁移和服务缓存可能会因时间冲突而导致计算延时。因此,针对上述相关问题,首先将任务迁移和服务缓存决策进行解耦,针对深度强化学习(deep reinforcement learning,DRL)在具有高维的混合决策空间的性能提升不明显的缺点(例如资源分配时利用率不高),将DRL与Transformer结合,通过在历史数据中学习,输出当前时隙的任务迁移决策和下一时隙的任务决策,保证任务到达边缘服务器时能立即执行。其次,为了提高资源分配问题中的资源利用率,将问题分解为连续资源分配问题和离散的任务迁移与服务缓存问题,利用凸优化技术求解资源分配最优决策。广泛的数值结果表明,与其他基线算法相比,提出的算法能有效地减少任务的平均完成时延,同时在资源利用率和稳定性方面也有优异的表现。展开更多
P2P流媒体cache是一种有效减少带宽开销、提高对象利用率的技术,通常采用FIFO,LRU等算法置换内容.然而,流媒体不同于Web对象,P2P网络也有别于客户/服务器模式.在分布式应用中这些算法可能影响系统的性能,为此,分析了FIFO和LRU置换算法,...P2P流媒体cache是一种有效减少带宽开销、提高对象利用率的技术,通常采用FIFO,LRU等算法置换内容.然而,流媒体不同于Web对象,P2P网络也有别于客户/服务器模式.在分布式应用中这些算法可能影响系统的性能,为此,分析了FIFO和LRU置换算法,提出了基于供求关系的SD算法,以及基于分片副本数量的REP算法,并对其进行评估和比较.针对不同的节点到达间隔,将SD和REP同FIFO,LRU进行比较,发现在启动延迟、媒体副本数量和根节点依赖度方面SD和REP几乎均优于FIFO和LRU.同LSB(least sent bytes)算法相比,某些场景中SD的启动延迟减少了约40%,而REP在副本数量方面远远超过LSB的结果,说明在P2P网络流媒体服务中使用SD和REP缓存置换算法有助于提高系统性能.展开更多
在多核环境下,对共享L2 Cache的优化显得尤为重要,因为当被访问的数据块不在L2 Cache中时(发生L2缺失),CPU需要花费几百个周期访问主存的代价是相当大的.在设计Cache时,替换算法是考虑的一个重要因素,替换算法的好坏直接影响Cache的性...在多核环境下,对共享L2 Cache的优化显得尤为重要,因为当被访问的数据块不在L2 Cache中时(发生L2缺失),CPU需要花费几百个周期访问主存的代价是相当大的.在设计Cache时,替换算法是考虑的一个重要因素,替换算法的好坏直接影响Cache的性能和计算机的整体性能.虽然LRU替换算法已经被广泛应用在片上Cache中,但是也存在着一些不足:当Cache容量小于程序工作集时,容易产生冲突缺失;且LRU替换算法不考虑数据块被访问的频率.文中把冒泡替换算法应用到多核共享Cache中,同时考虑数据块被访问的频率和最近访问的信息.通过分析实验数据,与LRU替换算法相比,采用冒泡替换算法可以使MPKI(Misses per Kilo instructions)和L2 Cache命中率均有所改善.展开更多
基金New Brunswick Innovation Foundation(NBIF)for the financial support of the global project.
文摘The demands on conventional communication networks are increasing rapidly because of the exponential expansion of connected multimedia content.In light of the data-centric aspect of contemporary communication,the information-centric network(ICN)paradigm offers hope for a solution by emphasizing content retrieval by name instead of location.If 5G networks are to meet the expected data demand surge from expanded connectivity and Internet of Things(IoT)devices,then effective caching solutions will be required tomaximize network throughput andminimize the use of resources.Hence,an ICN-based Cooperative Caching(ICN-CoC)technique has been used to select a cache by considering cache position,content attractiveness,and rate prediction.The findings show that utilizing our suggested approach improves caching regarding the Cache Hit Ratio(CHR)of 84.3%,Average Hop Minimization Ratio(AHMR)of 89.5%,and Mean Access Latency(MAL)of 0.4 s.Within a framework,it suggests improved caching strategies to handle the difficulty of effectively controlling data consumption in 5G networks.These improvements aim to make the network run more smoothly by enhancing content delivery,decreasing latency,and relieving congestion.By improving 5G communication systems’capacity tomanage the demands faced by modern data-centric applications,the research ultimately aids in advancement.
基金supported by the Key Scientific and Technological Projects of Henan Province with Grant Nos.232102211084 and 222102210137,both received by B.W.(URL to the sponsor’s website is https://kjt.henan.gov.cn/)the National Natural Science Foundation of China with grant No.61975187,received by Z.Z(the URL to the sponsor’s website is https://www.nsfc.gov.cn/).
文摘Computing-intensive and latency-sensitive user requests pose significant challenges to traditional cloud computing.In response to these challenges,mobile edge computing(MEC)has emerged as a new paradigm that extends the computational,caching,and communication capabilities of cloud computing.By caching certain services on edge nodes,computational support can be provided for requests that are offloaded to the edges.However,previous studies on task offloading have generally not considered the impact of caching mechanisms and the cache space occupied by services.This oversight can lead to problems,such as high delays in task executions and invalidation of offloading decisions.To optimize task response time and ensure the availability of task offloading decisions,we investigate a task offloading method that considers caching mechanism.First,we incorporate the cache information of MEC into the model of task offloading and reduce the task offloading problem as a mixed integer nonlinear programming(MINLP)problem.Then,we propose an integer particle swarm optimization and improved genetic algorithm(IPSO_IGA)to solve the MINLP.IPSO_IGA exploits the evolutionary framework of particle swarm optimization.And it uses a crossover operator to update the positions of particles and an improved mutation operator to maintain the diversity of particles.Finally,extensive simulation experiments are conducted to evaluate the performance of the proposed algorithm.The experimental results demonstrate that IPSO_IGA can save 20%to 82%of the task completion time,compared with state-of-theart and classical algorithms.Moreover,IPSO_IGA is suitable for scenarios with complex network structures and computing-intensive tasks.
文摘<div style="text-align:justify;"> <span style="font-family:Verdana;">There has been an explosion in the volume of data that is being accessed from the Internet. As a result, the risk of a Web server being inundated with requests is ever-present. One approach to reducing the performance degradation that potentially comes from Web server overloading is to employ Web caching where data content is replicated in multiple locations. In this paper, we investigate the use of evolutionary algorithms to dynamically alter partition size in Web caches. We use established modeling techniques to compare the performance of our evolutionary algorithm to that found in statically-partitioned systems. Our results indicate that utilizing an evolutionary algorithm to dynamically alter partition sizes can lead to performance improvements especially in environments where the relative size of large to small pages is high.</span> </div>
基金the National S&T Major Project (No. 2018ZX03001011)the National Key R&D Program(No.2018YFB1801102)+1 种基金the National Natural Science Foundation of China (No. 61671072)the Beijing Natural Science Foundation (No. L192025)
文摘With the increasing maritime activities and the rapidly developing maritime economy, the fifth-generation(5G) mobile communication system is expected to be deployed at the ocean. New technologies need to be explored to meet the requirements of ultra-reliable and low latency communications(URLLC) in the maritime communication network(MCN). Mobile edge computing(MEC) can achieve high energy efficiency in MCN at the cost of suffering from high control plane latency and low reliability. In terms of this issue, the mobile edge communications, computing, and caching(MEC3) technology is proposed to sink mobile computing, network control, and storage to the edge of the network. New methods that enable resource-efficient configurations and reduce redundant data transmissions can enable the reliable implementation of computing-intension and latency-sensitive applications. The key technologies of MEC3 to enable URLLC are analyzed and optimized in MCN. The best response-based offloading algorithm(BROA) is adopted to optimize task offloading. The simulation results show that the task latency can be decreased by 26.5’ ms, and the energy consumption in terminal users can be reduced to 66.6%.
基金jointly supported by the National Natural Science Foundation of China (No.61501042)the National High Technology Research and Development Program(863) of China (2015AA016101)+1 种基金Beijing Nova Program(Z151100000315078)Information Network Open Source Platform and Technology Development Strategy(No.2016-XY-09)
文摘Heterogeneous cellular networks(HCNs), by introducing caching capability, has been considered as a promising technique in 5 G era, which can bring contents closer to users to reduce the transmission delay, save scarce bandwidth resource. Although many works have been done for caching in HCNs, from an energy perspective, there still exists much space to develop a more energy-efficient system when considering the fact that the majority of base stations are under-utilized in the most of the time. Therefore, in this paper, by taking the activation mechanism for the base stations into account, we study a joint caching and activation mechanism design to further improve the energy efficiency, then we formulate the optimization problem as an Integer Linear Programming problem(ILP) to maximize the system energy saving. Due to the enormous computation complexity for finding the optimal solution, we introduced a Quantum-inspired Evolutionary Algorithm(QEA) to iteratively provide the global best solution. Numerical results show that our proposed algorithm presents an excellent performance, which is far better than the strategy of only considering caching without deactivation mechanism in the actual, normal situation. We also provide performance comparison amongour QEA, random sleeping algorithm and greedy algorithm, numerical results illustrate our introduced QEA performs best in accuracy and global optimality.
文摘Streaming media data have strict delay and loss requirements, which cannot be adequately supported by the current Internet. Proxy caching technology is introduced to alleviate the effect of the bottleneck bandwidth on the transporting quality. A new architecture of multimedia proxy caching is given to solve the problem of the traditional Web caching technology for the multimedia streaming. The mixed media management module and the (resource) management module are added to this architecture, the prefetching and replacement method is studied considering the characteristics of streaming media data, the concept of weighed popularity is given to enhance the (exactness) of the replacement algorithm.
基金Supported by the National Natural Science Foun-dation of China (90204008)
文摘The Web cluster has been a popular solution of network server system because of its scalability and cost effective ness. The cache configured in servers can result in increasing significantly performance, In this paper, we discuss the suitable configuration strategies for caching dynamic content by our experimental results. Considering the system itself can provide support for caching static Web page, such as computer memory cache and disk's own cache, we adopt a special pattern that only caches dynamic Web page in some experiments to enlarge cache space. The paper is introduced three different replacement algorithms in our cache proxy module to test the practical effects of caching dynamic pages under different conditions. The paper is chiefly analyzed the influences of generated time and accessed frequency on caching dynamic Web pages. The paper is also provided the detailed experiment results and main conclusions in the paper.
文摘Modern shared-memory multi-core processors typically have shared Level 2(L2)or Level 3(L3)caches.Cache bottlenecks and replacement strategies are the main problems of such architectures,where multiple cores try to access the shared cache simultaneously.The main problem in improving memory performance is the shared cache architecture and cache replacement.This paper documents the implementation of a Dual-Port Content Addressable Memory(DPCAM)and a modified Near-Far Access Replacement Algorithm(NFRA),which was previously proposed as a shared L2 cache layer in a multi-core processor.Standard Performance Evaluation Corporation(SPEC)Central Processing Unit(CPU)2006 benchmark workloads are used to evaluate the benefit of the shared L2 cache layer.Results show improved performance of the multicore processor’s DPCAM and NFRA algorithms,corresponding to a higher number of concurrent accesses to shared memory.The new architecture significantly increases system throughput and records performance improvements of up to 8.7%on various types of SPEC 2006 benchmarks.The miss rate is also improved by about 13%,with some exceptions in the sphinx3 and bzip2 benchmarks.These results could open a new window for solving the long-standing problems with shared cache in multi-core processors.
文摘P2P流媒体cache是一种有效减少带宽开销、提高对象利用率的技术,通常采用FIFO,LRU等算法置换内容.然而,流媒体不同于Web对象,P2P网络也有别于客户/服务器模式.在分布式应用中这些算法可能影响系统的性能,为此,分析了FIFO和LRU置换算法,提出了基于供求关系的SD算法,以及基于分片副本数量的REP算法,并对其进行评估和比较.针对不同的节点到达间隔,将SD和REP同FIFO,LRU进行比较,发现在启动延迟、媒体副本数量和根节点依赖度方面SD和REP几乎均优于FIFO和LRU.同LSB(least sent bytes)算法相比,某些场景中SD的启动延迟减少了约40%,而REP在副本数量方面远远超过LSB的结果,说明在P2P网络流媒体服务中使用SD和REP缓存置换算法有助于提高系统性能.
文摘在多核环境下,对共享L2 Cache的优化显得尤为重要,因为当被访问的数据块不在L2 Cache中时(发生L2缺失),CPU需要花费几百个周期访问主存的代价是相当大的.在设计Cache时,替换算法是考虑的一个重要因素,替换算法的好坏直接影响Cache的性能和计算机的整体性能.虽然LRU替换算法已经被广泛应用在片上Cache中,但是也存在着一些不足:当Cache容量小于程序工作集时,容易产生冲突缺失;且LRU替换算法不考虑数据块被访问的频率.文中把冒泡替换算法应用到多核共享Cache中,同时考虑数据块被访问的频率和最近访问的信息.通过分析实验数据,与LRU替换算法相比,采用冒泡替换算法可以使MPKI(Misses per Kilo instructions)和L2 Cache命中率均有所改善.