期刊文献+
共找到3篇文章
< 1 >
每页显示 20 50 100
基于保护缓存的x86系统仿真优化 被引量:4
1
作者 董卫宇 王立新 +1 位作者 蒋烈辉 郭玉东 《计算机工程与设计》 CSCD 北大核心 2013年第2期606-610,622,共6页
为降低x86保护机制的仿真开销,提出了一种基于保护缓存的系统仿真优化方法。分析了x86保护机制的仿真开销来源,确定了与保护相关的具有较大仿真开销的指令和操作;给出了保护缓存的概念和工作机制,通过缓存最近访问过的描述符以及对使用... 为降低x86保护机制的仿真开销,提出了一种基于保护缓存的系统仿真优化方法。分析了x86保护机制的仿真开销来源,确定了与保护相关的具有较大仿真开销的指令和操作;给出了保护缓存的概念和工作机制,通过缓存最近访问过的描述符以及对使用该描述符的指令或操作的保护检查结果,来减少虚拟机内存访问和保护检查操作的次数。基于QEMU对保护缓存机制进行了实现,测试结果表明,引入保护缓存后涉及保护检查的x86指令或操作的仿真效率提升了11%~24.5%,且系统调用、页面故障处理和I/O密集型应用的性能均得到了稳定的提升。 展开更多
关键词 系统仿真 保护缓存 系统虚拟机 保护检查 QEMU
下载PDF
暑假期间光驱、刻录机选购指南
2
作者 小鱼 《网迷》 2002年第8期53-54,共2页
关键词 CD-ROM 缓存欠载保护技术 光驱 刻录机
下载PDF
CWLP:coordinated warp scheduling and locality-protected cache allocation on GPUs 被引量:1
3
作者 Yang ZHANG Zuo-cheng XING +1 位作者 Cang LIU Chuan TANG 《Frontiers of Information Technology & Electronic Engineering》 SCIE EI CSCD 2018年第2期206-220,共15页
As we approach the exascale era in supercomputing, designing a balanced computer system with a powerful computing ability and low power requirements has becoming increasingly important. The graphics processing unit(... As we approach the exascale era in supercomputing, designing a balanced computer system with a powerful computing ability and low power requirements has becoming increasingly important. The graphics processing unit(GPU) is an accelerator used widely in most of recent supercomputers. It adopts a large number of threads to hide a long latency with a high energy efficiency. In contrast to their powerful computing ability, GPUs have only a few megabytes of fast on-chip memory storage per streaming multiprocessor(SM). The GPU cache is inefficient due to a mismatch between the throughput-oriented execution model and cache hierarchy design. At the same time, current GPUs fail to handle burst-mode long-access latency due to GPU's poor warp scheduling method.Thus, benefits of GPU's high computing ability are reduced dramatically by the poor cache management and warp scheduling methods, which limit the system performance and energy efficiency. In this paper, we put forward a coordinated warp scheduling and locality-protected(CWLP) cache allocation scheme to make full use of data locality and hide latency. We first present a locality-protected cache allocation method based on the instruction program counter(LPC) to promote cache performance. Specifically, we use a PC-based locality detector to collect the reuse information of each cache line and employ a prioritised cache allocation unit(PCAU) which coordinates the data reuse information with the time-stamp information to evict the lines with the least reuse possibility. Moreover, the locality information is used by the warp scheduler to create an intelligent warp reordering scheme to capture locality and hide latency. Simulation results show that CWLP provides a speedup up to 19.8% and an average improvement of 8.8% over the baseline methods. 展开更多
关键词 LOCALITY Graphics processing unit (GPU) Cache allocation Warp scheduling
原文传递
上一页 1 下一页 到第
使用帮助 返回顶部