期刊文献+

百亿亿次级系统进程管理接口综述

A survey on the process management interface for exascale computing systems
下载PDF
导出
摘要 高性能计算机不断发展,系统规模日益增加,系统内包含的计算结点数、处理器核数扩展到新的水平。在超大规模系统中,并行应用程序的启动时间成为限制系统运行效率、降低系统易用性的一个重要因素。在并行应用启动阶段,利用进程管理接口为进程部署通信通道,供进程后续通信使用。在百亿亿次级规模系统中,传统进程管理接口无法在启动时快速获得通信信息,导致启动时间过长,系统性能下降。首先介绍进程管理接口在并行程序启动过程中的作用,着重介绍面向百亿亿次级系统的进程管理接口PMIx,而后对比论述PMIx对于改进大规模并行程序启动的作用,分析PMIx在提升系统性能上做出的优化,以及未来发展方向。 With the continuous development of high-performance computing,the scale of the system increases constantly,and the number of nodes and processor cores in the system has expanded to a new level.Under the condition of hyperscale systems,the startup time of parallel applications becomes an important factor,which limits the system’s operating efficiency and reduces the ease of use.The process management interface is used to deploy a communication channel for the process during the parallel application startup phase for subsequent communication of the process.In exascale systems,the traditional process management interface cannot quickly obtain communication information at the startup phase,resulting in long startup time and the reduced system performance.We first introduce the role of the process management interface in the parallel program startup process,focus on the process management interface PMIx for exascale systems,compare and discuss the role of PMIx in improving the startup of large-scale parallel programs,analyze the optimization of PMIx in improving system performance,and discuss future development directions.
作者 张昆 张伟 卢凯 董勇 戴屹钦 ZHANG Kun;ZHANG Wei;LU Kai;DONG Yong;DAI Yi-qin(School of Computer,National University of Defense Technology,Changsha 410073,China)
出处 《计算机工程与科学》 CSCD 北大核心 2020年第10期1774-1783,共10页 Computer Engineering & Science
基金 国家重点研发计划(2018YFB0204301) 国家数值风洞项目(NNW2018-ZT6B13)。
关键词 高性能计算 进程管理接口 进程通信 high-performance computing process management interface process communication
  • 相关文献

参考文献2

二级参考文献27

  • 1Top500, http://www.top500.org, 2013.
  • 2Liao K X, Xiao Q L, Yang Q C, Lu T Y. MilkyWay-2 supercomputer system and application. Submitted to Frontiers of Computer Science, 2013.
  • 3Pritchard H, Gorodetsky I, Buntinas D. A ugni-based mpich2 nemesis network module for the cray xe. In: Proceedings of the 18th European MPI Users' Group Conference on Recent Advances in the Message Passing Interface. 2011, 110--119.
  • 4Xie M, Lu Y, Liu L, Cao H, Yang X. Implementation and evaluation of network interface and message passing services for Tianhe-la supercomputer. In: Proceedings of the 19th IEEE Annual Symposium on High Performance Interconnects. 2011,78-86.
  • 5Chun B N, Mainwaring A, Culler D E. Virtual network transport protocols for myrinet. IEEE Micro, 1998, 18(1): 53--63.
  • 6Araki S, Bilas A, Dubnicki C, Edler J, Konishi K, Philbin J. Userspace communication: a quantitative study. In: Proceedings of the 1998 ACMjlEEE Conference on Supercomputing (CDROM). 1998, 1-16.
  • 7Bhoedjang R A, Ruhl T, Bal H E. User-level network interface protocols. Computer, 1998, 31 (11): 53--60.
  • 8Schoinas I, Hill M D. Address translation mechanisms in network interfaces. In: Proceedings of the 4th International Symposium on HighPerformance Computer Architecture. 1998,219-230.
  • 9InfiniBand Architecture Specification: Release 1.0. InfiniBand Trade Association, 2000.
  • 10Graham R L, Poole S, Shamis P, Bloch G, Bloch N, Chapman H, Kagan M, Shahar A, Rabinovitz I, Shainer G. Overlapping computation and communication: Barrier algorithms and connectx-2 core-direct capabilities. In: Proceedings of the 2010 IEEE International Symposium on Parallel & Distributed Processing, Workshops and Phd Forum. 2010, 1-8.

共引文献26

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部