期刊文献+

基于机器学习的跨平台缓存划分方法研究 被引量:1

A Multi-platform Cache Partitioning Method Based on Machine Learning
下载PDF
导出
摘要 多核处理器的最后一级缓存(Last Level Cache,LLC)采用共享机制,当多个程序并行随机访问LLC时,可能引起访存冲突,进而导致系统整体性能的大幅下降.已有研究者试图通过引入缓存划分,合理安排不同程序对LLC的访问时机,解决访存冲突问题.然而,现有的缓存划分方法主要采用启发式算法,即通过不断“试错”寻找缓存划分最优方案,寻找过程具有不确定性.这种不确定性可能导致划分开销过大甚至出现无法收敛的问题.为此,我们提出了一种基于机器学习的跨平台缓存划分方法MLPart.该方法借助少量运行参数,利用决策树和序列到序列模型预测剩余各个划分方案的性能,直接找到最优的划分方案,从而保证划分开销相对稳定且有效避免了无法收敛的情况.此外,实验表明如果计算平台之间存在着微小配置差异,即使运行相同程序所产生的运行参数也存在较大差异,因此单一模型通常没有足够的泛化能力适应不同的平台.为解决该问题,我们分别利用迁移学习技术和微调技术优化决策树和序列到序列模型,使其能以较低的计算开销完成不同平台的模型快速迁移部署.我们以当前主流的Intel至强(Xeon)处理器平台展开实验,实验结果表明MLPart能够通过少量划分步骤找到性能最佳的缓存划分方案;并且,与采用启发式缓存划分方法KPart,采用基于贝叶斯优化的缓存划分方法CLITE以及采用机器学习的缓存划分方法C&A相比,MLPart对工作集的性能提升更高且更为稳定. The Last Level Cache(LLC)is shared in a multi-level-cache and multi-core CPU.If multi-programming randomly access LCC,it incurs accessing conflicts,and results in severe performance degradation.Currently,researchers focus on cache partitioning to reduce LLC conflicts.Existing cache partitioning methods adopt heuristic algorithm which needs uncertain steps to optimal performance.Thus it may bring up excessive partitioning overheads or even failure to converge in the partitioning process.To this end,we propose MLPart,a machine learning-based multi-platform adaptive cache partitioning method.With a few of running parameters,MLPart employs the decision tree and sequence-to-sequence models respectively to predict the performance of the different partitioning policies,and finally find out the bestperformance partitioning policy with certain partition steps.In addition,the experiments show that even the same program runs in different computing platforms and result in the totally different runtime parameters.It is hard to ensure that trained models have sufficient generalization ability to be deployed in different platforms.MLPart also adopt transfer learning and tuning techniques to optimize the decision tree and sequence-to-sequence models for reducing the training overheads on other platforms.The experimental results illustrate that MLPart only needs a few of partition steps to find the optimal cache partitioning policy on Intel Xeon processor.The overall performance improvements using MLPart are highest and most stable in the KPart,CLITE and C&A.
作者 邱杰凡 贾逸哲 华宗汉 曹明生 范菁 QIU Jie-Fan;JIA Yi-Zhe;HUA Zong-Han;CAO Ming-Sheng;FAN Jing(Department of Computer Science and Technology,Zhejiang University of Technology,Hangzhou 310023;The People’s Bank of China Tonglu Sub-branch,Hangzhou 311500;University of Electronic Science and Technology of China,Chengdu 610054)
出处 《计算机学报》 EI CAS CSCD 北大核心 2023年第10期2097-2116,共20页 Chinese Journal of Computers
基金 国家重点研发计划“现代服务业共性关键技术研发及应用示范项目”(2018YFB1402800) 浙江省自然科学基金(LY20F020026)资助.
关键词 缓存划分 递归神经网络 决策树 序列到序列模型 迁移学习 cache partition recurrent neural network decision tree sequence to sequence transfer learning
  • 相关文献

同被引文献3

引证文献1

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部