期刊文献+

基于异构GPU集群的并行分布式编程解决方案 被引量:1

AN PARALLEL AND DISTRIBUTED PROGRAMMING SOLUTION BASED ON HETEROGENEOUS GPU CLUSTER
下载PDF
导出
摘要 由于超强的计算能力、高速访存带宽、支持大规模数据级并行程序设计等特点,GPU已经成为超级计算机和高性能计算(HPC)集群的主流加速器。随着处理单元的发展和集群节点的拓展,GPU集群不仅在节点层面呈现异构化,节点内也趋于异构化,大大提高了在GPU集群中编程的复杂度。主流GPU异构集群系统大多采用针对GPU的异构计算编程模型与面向分布式内存的消息传递模型的简单结合方式,这种方式使得GPU集群程序设计缺乏确定的准则,往往是低效而且易错的。为了提高在GPU集群中编程的效率,降低编程复杂度,以及实现平台无关性,提出一套异构GPU集群的并行分布式编程的解决方案。该方案通过采用扩展语言方法提出了编程框架DISPAR,并实现了预处理器系统StreamCC。实验证明了其可行性。 Due to its characteristics of super powerful computing capability, high-speed memory access bandwidth and supporting large- scale data-level parallel programming, GPUs have become the mainstream accelerators for supercomputers and high performance computing field. GPU-enhanced clusters are showed to be heterogeneous in both node layer and intra-node as the evolvement of the processing elements and the expansion of the cluster nodes, which greatly increases the complexity of GPU cluster programming. Mainstream heterogeneous GPU clusters mostly adopt the way of simple combination of the heterogeneous computing programming model for GPU and the distributed memory- oriented message passing mode], such approach makes the GPU cluster programming lack of determinate criteria and often be inefficient and error-prone. In order to improve the efficiency of programming GPU cluster and to reduce programming complexity, as well as to realise the platform independence, we present a set of parallel and distributed programming solution for heterogeneous GPU cluster. The scheme presents the programming framework DISPAR through the use of the extended language method, and realises the preprocessor system StreamCC. Experiment proves its feasibility.
出处 《计算机应用与软件》 CSCD 北大核心 2014年第9期28-31,176,共5页 Computer Applications and Software
关键词 GPU集群 编程模型 数据级并行 GPU clusters Programming framework Data-level parallel
  • 相关文献

参考文献9

  • 1Kindratenko V,Enos J,Shi G,et al.GPU Clusters for High PerformanceComputing[C]//2009 IEEE International Conference on Cluster Com-puting,Computing and Workshops,2009:1-8.
  • 2Message Passing Interface Forum.MPI:A Message-Passing Interface Stand-ard[EB/ OL].http://www.mcs.anl.gov/research/projects/mpi/.
  • 3Khronos Group.The OpenCL Specification 1.0 revision 43[EB/OL].2011.http://www.khronos.org/opencl/.
  • 4Nvidia Corporation.CUDA Programming Guide Version 4.0[EB/OL].2012.http://www.nvidia.com/object/cuda_home.html.
  • 5OpenACC:Directives for Accelerators[EB/OL].http://www.openacc-standard,org.
  • 6Karunadasa N P,Ranasinghe D N.Accelerating High Performance Ap-plications with CUDA and MPI[C]//2009 International Conference onIndustrial and Information Systems,2009.
  • 7Barak A,BenNum T,Levy E,et al.A Package for OpenCL Based Het-erogeneous Computing on Clusters with Many GPU Devices[C]//Clus-ter Computing Workshops and Posters,2010 IEEE International Confer-ence,2010:1-1.
  • 8Zhang Y,Mueller F.Gstream:A General-Purpose Data Streaming Frame-work on GPU Clusters[C]//2011 International Conference on ParallelProcessing,2011:245-254.
  • 9OpenMP Architecture Review Board,OpenMP Specification:OpenMPapplication program interface[EB/OL].http://www.openmp.org/wp/ openmp-specifications/.

同被引文献4

引证文献1

二级引证文献7

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部