期刊文献+
共找到4篇文章
< 1 >
每页显示 20 50 100
A Cache Replacement Policy Based on Multi-Factors for Named Data Networking 被引量:1
1
作者 Meiju Yu Ru Li Yuwen Chen 《Computers, Materials & Continua》 SCIE EI 2020年第10期321-336,共16页
Named Data Networking(NDN)is one of the most excellent future Internet architectures and every router in NDN has the capacity of caching contents passing by.It greatly reduces network traffic and improves the speed of... Named Data Networking(NDN)is one of the most excellent future Internet architectures and every router in NDN has the capacity of caching contents passing by.It greatly reduces network traffic and improves the speed of content distribution and retrieval.In order to make full use of the limited caching space in routers,it is an urgent challenge to make an efficient cache replacement policy.However,the existing cache replacement policies only consider very few factors that affect the cache performance.In this paper,we present a cache replacement policy based on multi-factors for NDN(CRPM),in which the content with the least cache value is evicted from the caching space.CRPM fully analyzes multi-factors that affect the caching performance,puts forward the corresponding calculation methods,and utilize the multi-factors to measure the cache value of contents.Furthermore,a new cache value function is constructed,which makes the content with high value be stored in the router as long as possible,so as to ensure the efficient use of cache resources.The simulation results show that CPRM can effectively improve cache hit ratio,enhance cache resource utilization,reduce energy consumption and decrease hit distance of content acquisition. 展开更多
关键词 Cache replacement policy named data networking content popularity FRESHNESS energy consumption
下载PDF
Optimal chunk caching in network coding-based qualitative communication
2
作者 Lijun Dong Richard Li 《Digital Communications and Networks》 SCIE CSCD 2022年第1期44-50,共7页
Network processing in the current Internet is at the entirety of the data packet,which is problematic when encountering network congestion.The newly proposed Internet service named Qualitative Communication changes th... Network processing in the current Internet is at the entirety of the data packet,which is problematic when encountering network congestion.The newly proposed Internet service named Qualitative Communication changes the network processing paradigm to an even finer granularity,namely chunk level,which obsoletes many existing networking policies and schemes,especially the caching algorithms and cache replacement policies that have been extensively explored in Web Caching,Content Delivery Networks(CDN)or Information-Centric Networks(ICN).This paper outlines all the new factors that are brought by random linear network coding-based Qualitative Communication and proves the importance and necessity of considering them.A novel metric is proposed by taking these new factors into consideration.An optimization problem is formulated to maximize the metric value of all retained chunks in the local storage of network nodes under the constraint of storage limit.A cache replacement scheme that obtains the optimal result in a recursive manner is proposed correspondingly.With the help of the introduced intelligent cache replacement algorithm,the performance evaluations show remarkably reduced end-to-end latency compared to the existing schemes in various network scenarios. 展开更多
关键词 Internet Qualitative communication New IP Chunk caching Random linear network coding End-to-end latency Cache replacement policy Degree of freedom Distance Packet size
下载PDF
A Prefetch-Adaptive Intelligent Cache Replacement Policy Based on Machine Learning 被引量:2
3
作者 杨会静 方娟 +1 位作者 蔡旻 才智 《Journal of Computer Science & Technology》 SCIE EI CSCD 2023年第2期391-404,共14页
Hardware prefetching and replacement policies are two techniques to improve the performance of the memory subsystem.While prefetching hides memory latency and improves performance,interactions take place with the cach... Hardware prefetching and replacement policies are two techniques to improve the performance of the memory subsystem.While prefetching hides memory latency and improves performance,interactions take place with the cache replacement policies,thereby introducing performance variability in the application.To improve the accuracy of reuse of cache blocks in the presence of hardware prefetching,we propose Prefetch-Adaptive Intelligent Cache Replacement Policy(PAIC).PAIC is designed with separate predictors for prefetch and demand requests,and uses machine learning to optimize reuse prediction in the presence of prefetching.By distinguishing reuse predictions for prefetch and demand requests,PAIC can better combine the performance benefits from prefetching and replacement policies.We evaluate PAIC on a set of 27 memory-intensive programs from the SPEC 2006 and SPEC 2017.Under single-core configuration,PAIC improves performance over Least Recently Used(LRU)replacement policy by 37.22%,compared with improvements of 32.93%for Signature-based Hit Predictor(SHiP),34.56%for Hawkeye,and 34.43%for Glider.Under the four-core configuration,PAIC improves performance over LRU by 20.99%,versus 13.23%for SHiP,17.89%for Hawkeye and 15.50%for Glider. 展开更多
关键词 hardware prefetching machine learning Prefetch-Adaptive Intelligent Cache replacement policy(PAIC) replacement policy
原文传递
Fast convergence caching replacement algorithm based on dynamic classification for content-centric networks 被引量:1
4
作者 FANG Chao HUANG Tao +2 位作者 LIU Jiang CHEN Jian-ya LIU Yun-jie 《The Journal of China Universities of Posts and Telecommunications》 EI CSCD 2013年第5期45-50,共6页
One of the key research fields of content-centric networking (CCN) is to develop more efficient cache replacement policies to improve the hit ratio of CCN in-network caching. However, most of existing cache strategi... One of the key research fields of content-centric networking (CCN) is to develop more efficient cache replacement policies to improve the hit ratio of CCN in-network caching. However, most of existing cache strategies designed mainly based on the time or frequency of content access, can not properly deal with the problem of the dynamicity of content popularity in the network. In this paper, we propose a fast convergence caching replacement algorithm based on dynamic classification method for CCN, named as FCDC. It develops a dynamic classification method to reduce the time complexity of cache inquiry, which achieves a higher caching hit rate in comparison to random classification method under dynamic change of content popularity. Meanwhile, in order to relieve the influence brought about by dynamic content popularity, it designs a weighting function to speed up cache hit rate convergence in the CCN router. Experimental results show that the proposed scheme outperforms the replacement policies related to least recently used (LRU) and recent usage frequency (RUF) in cache hit rate and resiliency when content popularity in the network varies. 展开更多
关键词 CCN cache replacement policy dynamic classification fast convergence category popularity
原文传递
上一页 1 下一页 到第
使用帮助 返回顶部