期刊文献+
共找到3篇文章
< 1 >
每页显示 20 50 100
Labeled yon Neumann Architecture for Software-Defined Cloud 被引量:8
1
作者 yun-gang bao Sa Wang, Member, CCF 《Journal of Computer Science & Technology》 SCIE EI CSCD 2017年第2期219-223,共5页
As cloud computing is moving forward rapidly, cloud providers have been encountering great challenges: long tail latency, low utilization, and high interference. They intend to co-locate multiple workloads on a singl... As cloud computing is moving forward rapidly, cloud providers have been encountering great challenges: long tail latency, low utilization, and high interference. They intend to co-locate multiple workloads on a single server to improve the resource utilization. But the co-located applications suffer from severe performance interference and long tail latency, which lead to unpredictable user experience. To meet these challenges, software-defined cloud has been proposed to facilitate tighter coordination among application, operating system and hardware. Users' quality of service (QoS) requirements could be propagated all the way down to the hardware with differential management mechanisms. However, there is little hardware support to maintain and guarantee users' QoS requirements. To this end, this paper proposes Labeled von Neumann Architecture (LvNA), which introduces a labelling mechanism to convey more software's semantic information such as QoS and security to the underlying hardware. LvNA is able to correlate labels with various entities, e.g., virtual machine, process and thread, and propagate labels in the whole machine and program differentiated services based on rules. We consider LvNA to be a fundamental hardware support to the software-defined cloud. 展开更多
关键词 software-defined cloud von Neumann architecture tail latency performance interference
原文传递
A Case for Adaptive Resource Management in Alibaba Datacenter Using Neural Networks 被引量:2
2
作者 Sa Wang Yan-Hai Zhu +6 位作者 Shan-Pei Chen Tian-Ze Wu Wen-Jie Li Xu-Sheng Zhan Hai-Yang Ding Wei-Song Shi yun-gang bao 《Journal of Computer Science & Technology》 SCIE EI CSCD 2020年第1期209-220,共12页
Both resource efficiency and application QoS have been big concerns of datacenter operators for a long time,but remain to be irreconcilable.High resource utilization increases the risk of resource contention between c... Both resource efficiency and application QoS have been big concerns of datacenter operators for a long time,but remain to be irreconcilable.High resource utilization increases the risk of resource contention between co-located workload,which makes latency-critical(LC)applications suffer unpredictable,and even unacceptable performance.Plenty of prior work devotes the effort on exploiting effective mechanisms to protect the QoS of LC applications while improving resource efficiency.In this paper,we propose MAGI,a resource management runtime that leverages neural networks to monitor and further pinpoint the root cause of performance interference,and adjusts resource shares of corresponding applications to ensure the QoS of LC applications.MAGI is a practice in Alibaba datacenter to provide on-demand resource adjustment for applications using neural networks.The experimental results show that MAGI could reduce up to 87.3%performance degradation of LC application when co-located with other antagonist applications. 展开更多
关键词 RESOURCE management NEURAL network RESOURCE efficiency TAIL LATENCY
原文传递
高通量计算的兴起(英文) 被引量:1
3
作者 Ning-hui SUN yun-gang bao Dong-rui FAN 《Frontiers of Information Technology & Electronic Engineering》 SCIE EI CSCD 2018年第10期1245-1250,共6页
近年来,云计算、人工智能和物联网等新兴计算应用的出现,对计算机系统设计提出3个共同要求:高利用率、高吞吐量和低延迟。在这里,这些被称为"高通量计算"的要求。我们进一步提出一种新指标,称为"系统熵",用于测量... 近年来,云计算、人工智能和物联网等新兴计算应用的出现,对计算机系统设计提出3个共同要求:高利用率、高吞吐量和低延迟。在这里,这些被称为"高通量计算"的要求。我们进一步提出一种新指标,称为"系统熵",用于测量计算机系统的混乱程度和不确定性。我们认为,与追求高性能和低功耗的传统计算系统的设计不同,高通量计算应致力于实现低并发性。然而,从计算机体系结构角度看,高通量计算面临两大挑战:(1)如何充分利用应用程序数据的并行和并发执行来实现高吞吐量;(2)如何获得低延迟,即便在具有高利用率的数据路径中发生严重争用的情况下。为应对这两个挑战,引入两种技术:片上数据流体系结构和标签化冯诺依曼体系结构。构建了两个可以实现高吞吐量和低延迟的原型,显著降低了系统熵。 展开更多
关键词 HIGH-THROUGHPUT COMPUTING Sysentropy INFORMATION superbahn
原文传递
上一页 1 下一页 到第
使用帮助 返回顶部