期刊文献+
共找到3篇文章
< 1 >
每页显示 20 50 100
supported by the National Natural Science Foundation of China(Nos.62172051,61772085,and 61877005)and Jiangsu Agriculture Science and Technology Innovation Fund(No.CX(18)3054).
1
作者 Jinghui Zhang Yuchen Wang +3 位作者 Tianyu Huang Fang Dong Wei Zhao dian shen 《Tsinghua Science and Technology》 SCIE EI CAS CSCD 2023年第1期82-92,共11页
Numerous neural network(NN)applications are now being deployed to mobile devices.These applications usually have large amounts of calculation and data while requiring low inference latency,which poses challenges to th... Numerous neural network(NN)applications are now being deployed to mobile devices.These applications usually have large amounts of calculation and data while requiring low inference latency,which poses challenges to the computing ability of mobile devices.Moreover,devices’life and performance depend on temperature.Hence,in many scenarios,such as industrial production and automotive systems,where the environmental temperatures are usually high,it is important to control devices’temperatures to maintain steady operations.In this paper,we propose a thermal-aware channel-wise heterogeneous NN inference algorithm.It contains two parts,the thermal-aware dynamic frequency(TADF)algorithm and the heterogeneous-processor single-layer workload distribution(HSWD)algorithm.Depending on a mobile device’s architecture characteristics and environmental temperature,TADF can adjust the appropriate running speed of the central processing unit and graphics processing unit,and then the workload of each layer in the NN model is distributed by HSWD in line with each processor’s running speed and the characteristics of the layers as well as heterogeneous processors.The experimental results,where representative NNs and mobile devices were used,show that the proposed method can considerably improve the speed of the on-device inference by 21%–43%over the traditional inference method. 展开更多
关键词 neural network inference mobile device temperature adjustment channel-wise parallelization
原文传递
VirtCO:Joint Coflow Scheduling and Virtual Machine Placement in Cloud Data Centers 被引量:2
2
作者 dian shen Junzhou Luo +1 位作者 Fang Dong Junxue Zhang 《Tsinghua Science and Technology》 SCIE EI CAS CSCD 2019年第5期630-644,共15页
Cloud data centers, such as Amazon EC2, host myriad big data applications using Virtual Machines(VMs). As these applications are communication-intensive, optimizing network transfer between VMs is critical to the perf... Cloud data centers, such as Amazon EC2, host myriad big data applications using Virtual Machines(VMs). As these applications are communication-intensive, optimizing network transfer between VMs is critical to the performance of these applications and network utilization of data centers. Previous studies have addressed this issue by scheduling network flows with coflow semantics or optimizing VM placement with traffic considerations.However, coflow scheduling and VM placement have been conducted orthogonally. In fact, these two mechanisms are mutually dependent, and optimizing these two complementary degrees of freedom independently turns out to be suboptimal. In this paper, we present VirtCO, a practical framework that jointly schedules coflows and places VMs ahead of VM launch to optimize the overall performance of data center applications. We model the joint coflow scheduling and VM placement optimization problem, and propose effective heuristics for solving it. We further implement VirtCO with OpenStack and deploy it in a testbed environment. Extensive evaluation of real-world traces shows that compared with state-of-the-art solutions, VirtCO greatly reduces the average coflow completion time by up to 36.5%. This new framework is also compatible with and readily deployable within existing data center architectures. 展开更多
关键词 cloud computing data center coflow SCHEDULING Virtual Machine (VM) PLACEMENT
原文传递
Task-Aware Flow Scheduling with Heterogeneous Utility Characteristics for Data Center Networks 被引量:2
3
作者 Fang Dong Xiaolin Guo +1 位作者 Pengcheng Zhou dian shen 《Tsinghua Science and Technology》 SCIE EI CAS CSCD 2019年第4期400-411,共12页
With the continuous enrichment of cloud services, an increasing number of applications are being deployed in data centers. These emerging applications are often communication-intensive and data-parallel, and their per... With the continuous enrichment of cloud services, an increasing number of applications are being deployed in data centers. These emerging applications are often communication-intensive and data-parallel, and their performance is closely related to the underlying network. With their distributed nature, the applications consist of tasks that involve a collection of parallel flows. Traditional techniques to optimize flow-level metrics are agnostic to task-level requirements, leading to poor application-level performance. In this paper, we address the heterogeneous task-level requirements of applications and propose task-aware flow scheduling. First, we model tasks' sensitivity to their completion time by utilities. Second, on the basis of Nash bargaining theory, we establish a flow scheduling model with heterogeneous utility characteristics, and analyze it using Lagrange multiplier method and KKT condition. Third, we propose two utility-aware bandwidth allocation algorithms with different practical constraints. Finally, we present Tasch, a system that enables tasks to maintain high utilities and guarantees the fairness of utilities. To demonstrate the feasibility of our system, we conduct comprehensive evaluations with realworld traffic trace. Communication stages complete up to 1.4 faster on average, task utilities increase up to 2.26,and the fairness of tasks improves up to 8.66 using Tasch in comparison to per-flow mechanisms. 展开更多
关键词 data center networks coflow FLOW SCHEDULING DATA-INTENSIVE applications
原文传递
上一页 1 下一页 到第
使用帮助 返回顶部