期刊文献+

A High-Performance and Cost-Efficient Interconnection Network for High-Density Servers 被引量:2

A High-Performance and Cost-Efficient Interconnection Network for High-Density Servers
原文传递
导出
摘要 The high-density server is featured as low power, low volume, and high computational density. With the rising use of high-density servers in data-intensive and large-scale web applications, it requires a high-performance and cost-efficient intra-server interconnection network. Most of state-of-the-art high-density servers adopt the fully-connected intra-server network to attain high network performance. Unfortunately, this solution costs too much due to the high degree of nodes. In this paper, we exploit the theoretically optimized Moore graph to interconnect the chips within a server. Accounting for the suitable size of applications, a 50-size Moore graph, called Hoffman-Singleton graph, is adopted. In practice, multiple chips should be integrated onto one processor board, which means that the original graph should be partitioned into homogeneous connected subgraphs. However, the existing partition scheme does not consider above problem and thus generates heterogeneous subgraphs. To address this problem, we propose two equivalent-partition schemes for the Hoffman-Singleton graph. In addition, a logic-based and minimal routing mechanism, which is both time and area efficient, is proposed. Finally, we compare the proposed network architecture with its counterparts, namely the fully-connected, Kautz and Torus networks. The results show that our proposed network can achieve competitive performance as fully-connected network and cost close to Torus. The high-density server is featured as low power, low volume, and high computational density. With the rising use of high-density servers in data-intensive and large-scale web applications, it requires a high-performance and cost-efficient intra-server interconnection network. Most of state-of-the-art high-density servers adopt the fully-connected intra-server network to attain high network performance. Unfortunately, this solution costs too much due to the high degree of nodes. In this paper, we exploit the theoretically optimized Moore graph to interconnect the chips within a server. Accounting for the suitable size of applications, a 50-size Moore graph, called Hoffman-Singleton graph, is adopted. In practice, multiple chips should be integrated onto one processor board, which means that the original graph should be partitioned into homogeneous connected subgraphs. However, the existing partition scheme does not consider above problem and thus generates heterogeneous subgraphs. To address this problem, we propose two equivalent-partition schemes for the Hoffman-Singleton graph. In addition, a logic-based and minimal routing mechanism, which is both time and area efficient, is proposed. Finally, we compare the proposed network architecture with its counterparts, namely the fully-connected, Kautz and Torus networks. The results show that our proposed network can achieve competitive performance as fully-connected network and cost close to Torus.
出处 《Journal of Computer Science & Technology》 SCIE EI CSCD 2014年第2期281-292,共12页 计算机科学技术学报(英文版)
基金 supported by the Strategic Priority Research Program of the Chinese Academy of Sciences under Grant No.XDA06010401 the National Natural Science Foundation of China under Grant Nos.61202056,61331008,61221062 the HuaweiResearch Program of China under Grant No.YBCB2011030
关键词 high-density server interconnection network Moore graph Hoffman-Singleton graph equivalent partition high-density server, interconnection network, Moore graph, Hoffman-Singleton graph, equivalent partition
  • 相关文献

参考文献15

  • 1Montero R S, Huedo E, Llorente I M. Benchmarking of high throughput computing applications on grids. Parallel Com- puting, 2006, 32(4): 267-279.
  • 2Faanes G, Bataineh A, Roweth D, Court T, Froese E, Alver- son B, Johnson T, Kopnick J, Higgins M, Reinhard J. Cray cascade: A scalable HPC system based on a Dragonfly net- work. In Proe. the International Conference for High Performance Computing, Networking, Storage and Analysis (SC2012), November 2012, Article No.103.
  • 3Rao A. SeaMicro technology overview. Technical Report, AMD, January 2012. http://www.seamicro.com/sites/defau- lt/files/SM_TOOL64_v2.5.pdf, December 2013.
  • 4Rajamony R, Stephenson M C, Speight W E. The power 775 architecture at scale. In Proc. the 27th International A CM Conference on International Conference on Supercomputing (ICS2013), June 2013, pp.183-192.
  • 5Rao A. SeaMicro SM10000 system overview. Technical Re- port, AMD, June 2010. http://www.tiger-optics.ru/ down- load/seamicro/SM_TO02_vl.4.pdf, December 2013.
  • 6Hoffman A J, Singleton R R. On Moore graphs with diame- ters 2 and 3. IBM J. Research and Development, 1960, 4(5): 497-504.
  • 7Mattson T G, Van der Wijngaart R, Frumkin M. Program- ruing the Intel 80-core network-on-a-chip terascale processor. In Proc. the International Conference for High Performance Computing, Networking, Storage and Analysis (SC2008), Nov. 2008, Article No.38.
  • 8Bell S, Edwards B, Amann Jet al. TILE64-processor: A 64-core SoC with mesh interconnect. In Proc. Interna- tional Solid-State Circuits Conference (ISSCC2008), Febru- ary 2008, pp.88-89.
  • 9Seo J, Lee H, Jang M. Optimal routing and Hamiltonian cyclein Petersen-Torus networks. In Proc. the 3rd International Conference on Convergence and Hybrid Information Tech- nology (ICCIT2008), November 2008, pp.303-308.
  • 10Barroso L A, Dean J, HStzle U. Web search for a planet: The Coogle cluster architecture. IEEE Micro, 2003, 23(2): 22-28.

同被引文献13

引证文献2

二级引证文献9

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部