This paper studies online scheduling of jobs with kind release times on a single machine. Here "kind release time" means that in online setting, no jobs can be released when the machine is busy. Each job J h...This paper studies online scheduling of jobs with kind release times on a single machine. Here "kind release time" means that in online setting, no jobs can be released when the machine is busy. Each job J has a kind release time r(J) ≥ 0, a processing time p(J) > 0 and a deadline d(J) > 0. The goal is to determine a schedule which maximizes total processing time( p(J)E(J)) or total number( E(J)) of the accepted jobs. For the first objective function p(J)E(J), we first present a lower bound 2(1/2), and then provide an online algorithm LEJ with a competitive ratio of 3. This is the first deterministic algorithm for the problem with a constant competitive ratio. When p(J) ∈ {1, k}, k > 1 is a real number, we first present a lower bound min{(1 + k)/k, 2 k/(1 + k)}, and then we show that LEJ has a competitive ratio of1 + k/k. In particular, when all the k length jobs have tight deadlines, we first present a lower bound max{4/(2 + k), 1}(for p(J)E(J)) and 4/3(for E(J)). Then we prove that LEJ is k/k-competitive for p(J)E(J) and we provide an online algorithm H with a competitive ratio of 2 k/( k + 1) for the second objective function E(J).展开更多
To provide timely results for big data analytics, it is crucial to satisfy deadline requirements for MapReduce jobs in today's production environments. Much effort has been devoted to the problem of meeting deadlines...To provide timely results for big data analytics, it is crucial to satisfy deadline requirements for MapReduce jobs in today's production environments. Much effort has been devoted to the problem of meeting deadlines, and typically there exist two kinds of solutions. The first is to allocate appropriate resources to complete the entire job before the specified time limit, where missed deadlines result because of tight deadline constraints or lack of resources; the second is to run a pre-constructed sample based on deadline constraints, which can satisfy the time requirement but fail to maximize the volumes of processed data. In this paper, we propose a deadline-oriented task scheduling approach, named 'Dart', to address the above problem. Given a specified deadline and restricted resources, Dart uses an iterative estimation method, which is based on both historical data and job running status to precisely estimate the real-time job completion time. Based on the estimated time, Dart uses an approach-revise algorithm to make dynamic scheduling decisions for meeting deadlines while maximizing the amount of processed data and mitigating stragglers. Dart also efficiently handles task failures and data skew, protecting its performance from being harmed. We have validated our approach using workloads from OpenCloud and Facebook on a cluster of 64 virtual machines. The results show that Dart can not only effectively meet the deadline but also process near-maximum volumes of data even with tight deadlines and limited resources.展开更多
Fog computing became a traditional OffLad Destination(OLD)to compute the offloaded tasks of the Internet of Vehicles(IoV).Nevertheless,the limited computing resources of the fog node leads to re-offload these tasks to...Fog computing became a traditional OffLad Destination(OLD)to compute the offloaded tasks of the Internet of Vehicles(IoV).Nevertheless,the limited computing resources of the fog node leads to re-offload these tasks to the neighboring fog nodes or the cloud.Thus,the IoV will incur additional offloading costs.In this paper,we propose a new offloading scheme by utilizing RoadSide Parked Vehicles(RSPV)as an alternative OLD for IoV.The idle computing resources of the RSPVs can compute large tasks with low offloading costs compared with fog nodes and the cloud.Finally,a performance evaluation of the proposed scheme has been presented and discussed with other benchmark offloading schemes.展开更多
针对多DAG(Directed Acyclic Graph)共享一组异构资源在调度吞吐量最大化基础上的费用优化问题,本文提出了一种基于总费用变化量探测的费用优化算法PDTC(based on the Probe of the Total Cost Decrease),目的在于尽可能降低有优化条件...针对多DAG(Directed Acyclic Graph)共享一组异构资源在调度吞吐量最大化基础上的费用优化问题,本文提出了一种基于总费用变化量探测的费用优化算法PDTC(based on the Probe of the Total Cost Decrease),目的在于尽可能降低有优化条件的多个DAG的总费用.实验表明,该算法不仅能使得各DAG充分利用期限内的冗余时间,也能够在一定程度上降低多个DAG调度执行的总费用.展开更多
To fulfill the requirements for hybrid real-time system scheduling, a long-release-interval-first (LRIF) real-time scheduling algorithm is proposed. The algorithm adopts both the fixed priority and the dynamic prior...To fulfill the requirements for hybrid real-time system scheduling, a long-release-interval-first (LRIF) real-time scheduling algorithm is proposed. The algorithm adopts both the fixed priority and the dynamic priority to assign priorities for tasks. By assigning higher priorities to the aperiodic soft real-time jobs with longer release intervals, it guarantees the executions for periodic hard real-time tasks and further probabilistically guarantees the executions for aperiodic soft real-time tasks. The schedulability test approach for the LRIF algorithm is presented. The implementation issues of the LRIF algorithm are also discussed. Simulation result shows that LRIF obtains better schedulable performance than the maximum urgency first (MUF) algorithm, the earliest deadline first (EDF) algorithm and EDF for hybrid tasks. LRIF has great capability to schedule both periodic hard real-time and aperiodic soft real-time tasks.展开更多
The problem of scheduling radar dwells in multifunction phased array radar systems is addressed. A novel dwell scheduling algorithm is proposed. The whole scheduling process is based on an online pulse interleaving te...The problem of scheduling radar dwells in multifunction phased array radar systems is addressed. A novel dwell scheduling algorithm is proposed. The whole scheduling process is based on an online pulse interleaving technique. It takes the system timing and energy constraints into account. In order to adapt the dynamic task load, the algorithm considers both the priorities and deadlines of tasks. The simulation results demonstrate that compared with the conventional adaptive dwell scheduling algorithm, the proposed one can improve the task drop rate and system resource utility effectively.展开更多
In modern datacenters, the most common method to solve the network latency problem is to minimize flow completion time during the transmission process. Following the soft real-time nature, the optimization of transpor...In modern datacenters, the most common method to solve the network latency problem is to minimize flow completion time during the transmission process. Following the soft real-time nature, the optimization of transport latency is relaxed to meet a flow's deadline in deadline-sensitive services. However, none of existing deadline-sensitive protocols consider deadline as a constraint condition of transmission.They can only simplify the objective of meeting a flow's deadline as a deadline-aware mechanism by assigning a higher priority for tight-deadline constrained flows to finish the transmission as soon as possible, which results in an unsatisfactory effect in the condition of high fan-in degree. It drives us to take a step back and rethink whether minimizing flow completion time is the optimal way in meeting flow's deadline. In this paper, we focus on the design of a soft real-time transport protocol with deadline constraint in datacenters and present a flow-based deadline scheduling scheme for datacenter networks(FBDS).FBDS makes the unilateral deadline-aware flow transmission with priority transform into a compound centralized single-machine deadlinebased flow scheduling decision. In addition, FBDS blocks the flow sets and postpones some flows with extra time until their deadlines to make room for the new arriving flows in order to improve the deadline meeting rate. Our simulation resultson flow completion time and deadline meeting rate reveal the potential of FBDS in terms of a considerable deadline-sensitive transport protocol for deadline-sensitive interactive services.展开更多
Network-induced delay and jitter are key factors causing performance degradation and instability of NCSs (networked control systems). The relationships between the sampling periods of the control loops, network-induce...Network-induced delay and jitter are key factors causing performance degradation and instability of NCSs (networked control systems). The relationships between the sampling periods of the control loops, network-induced delay and jitter were studied aimed at token-type networks. A jitter-dependent optimal bandwidth scheduling algorithm for NCSs is proposed, which tries to achieve a tradeoff between bandwidth occupancy and system performance. Simulation tests proved the effectiveness of this optimal scheduling algorithm.展开更多
Cloud-as-the-center computing paradigms face multiple challenges in the 5G and Internet of Things scenarios, where the service requests are usually initiated by the end-user devices located at network edge and have ri...Cloud-as-the-center computing paradigms face multiple challenges in the 5G and Internet of Things scenarios, where the service requests are usually initiated by the end-user devices located at network edge and have rigid time constraints. Therefore, Fog computing, or mobile edge computing, is introduced as a promising solution to the service provision in the tiered IoT infrastructure to compensate the shortage of traditional cloud-only architecture. In this cloud-to-things continuum, several cloudlet or mobile edge server entities are placed at the access network to handle the task offloading and processing problems at the network edge. This raises the resource scheduling problem in this tiered system, which is vital for the promotion of the system efficiency. Therefore, in this paper, a scheduling mechanism for the cloudlets or fog nodes are presented, which takes the mobile tasks’ deadline and resources requirements at the same time while promoting the overall profit of the system. First, the problem at the cloudlet, to which IoT devices offload their tasks, is formulated as a multi-dimensional 0-1 knapsack problem. Second, based on ant colony optimization, a scheduling algorithm is presented which treat this problem as a subset selection problem. Third, to promote the performance of the system in the dynamic environments,a churn-refined algorithm is further put forward. A series of simulation experiments have shown that out proposal outperforms many state-of-the-art algorithms in both profit and guarantee ratio.展开更多
To address the issue of resource co-allocation with constraints to budget and deadline in grid environments, a novel co-allocation model based on virtual resource agent was proposed. The model optimized resources depl...To address the issue of resource co-allocation with constraints to budget and deadline in grid environments, a novel co-allocation model based on virtual resource agent was proposed. The model optimized resources deployment and price scheme through a three-side co-allocation mechanism, and applied queuing system to model the work of grid resources for providing quantitative deadline guarantees for grid applications. The validity and solutions of the model were presented theoretically. Extensive simulations were conducted to examine the effectiveness and the performance of the model by comparing with other co-allocation policies in terms of deadline violation rate, resource benefit and resource utilization. Experimental results show that compared with the three typical co-allocation policies, the proposed model can reduce the deadline violation rate to about 3.5% for the grid applications with constraints to budget and deadline. Also, the system benefits can be increased by about 30% compared with the those widely-used co-allocation policies.展开更多
Renewing warranty can provide customers with better service,and thus help manufacturers to gain market opportunities.In engineering practice,the cost for replacement is usually higher than the cost for maintenance,hen...Renewing warranty can provide customers with better service,and thus help manufacturers to gain market opportunities.In engineering practice,the cost for replacement is usually higher than the cost for maintenance,hence manufacturers often face huge challenge to reduce the warranty service cost.With consideration of the warranty deadline,we propose a two-stage optimization model for renewing warranty.In the first stage,a renewing warranty with deadline(RWD)policy is implemented,where the deadline represents the cumulative uptime threshold.When the cumulative uptime exceeds the deadline,the product will be minimally repaired and kept to the residual warranty period.When RWD is expired,the replacement warranty with limited repairs(RWLR)policy is applied.Under the free replacement and pro-rata warranty policy,the corresponding two-stage cost optimization model is established from the manufacturer’s perspective,the aim is to minimize the cost rate and obtain the optimal warranty period.A numerical example is provided to illustrate the validity of the proposed model,and the sensitivity analysis is also carried out.展开更多
At 02:30 am Baghdad time on January 17, the U. S. and allied forces carried out 1000 sorties air and 100 ship-launcnea cruise missile strikes, blasting Iraqi targets to free Kuwait from Iraqi occupation. The attack be...At 02:30 am Baghdad time on January 17, the U. S. and allied forces carried out 1000 sorties air and 100 ship-launcnea cruise missile strikes, blasting Iraqi targets to free Kuwait from Iraqi occupation. The attack began less than 19 hours after the expiration of the UN-mandated deadline. From the following day to the end of the month, the U.S.-led allied air forces continued massive bombing&against targets in Iraq and occupied Kuwait. From展开更多
In modern data centers, because of the deadline- agnostic congestion control in transmission control protocol (TCP), many deadline-sensitive flows can not finish before their deadlines. Therefore, providing a higher...In modern data centers, because of the deadline- agnostic congestion control in transmission control protocol (TCP), many deadline-sensitive flows can not finish before their deadlines. Therefore, providing a higher deadline meeting ratio becomes a critical challenge in the typical online data intensive (OLDI) ap- plications of data center networks (DCNs). However, a problem named as priority synchronization is found in this paper, which de- creases the deadline meeting ratio badly. To solve this problem, we propose a priority probability deceleration (P2D) deadline-aware TCP. By using the novel probabilistic deceleration, p2D prevents the priority synchronization problem. Simulation results show that P2D increases the deadline meeting ratio by 20% compared with D2TCP.展开更多
Real-time task scheduling is of primary significance in multiprocessor systems.Meeting deadlines and achieving high system utilization are the two main objectives of task scheduling in such systems.In this paper,we re...Real-time task scheduling is of primary significance in multiprocessor systems.Meeting deadlines and achieving high system utilization are the two main objectives of task scheduling in such systems.In this paper,we represent those two goals as the minimization of the average response time and the average task laxity.To achieve this,we propose a genetic-based algorithm with problem-specific and efficient genetic operators.Adaptive control parameters are also employed in our work to improve the genetic algorithms' efficiency.The simulation results show that our proposed algorithm outperforms its counterpart considerably by up to 36% and 35% in terms of the average response time and the average task laxity,respectively.展开更多
基金Supported by the National Natural Science Foundation of China(11501279,11501171,11671188,and11401604)the Young Backbone Teachers of Luoyang Normal University(2018XJGGJS-10)Henan Colleges(2015GGJS-193)
文摘This paper studies online scheduling of jobs with kind release times on a single machine. Here "kind release time" means that in online setting, no jobs can be released when the machine is busy. Each job J has a kind release time r(J) ≥ 0, a processing time p(J) > 0 and a deadline d(J) > 0. The goal is to determine a schedule which maximizes total processing time( p(J)E(J)) or total number( E(J)) of the accepted jobs. For the first objective function p(J)E(J), we first present a lower bound 2(1/2), and then provide an online algorithm LEJ with a competitive ratio of 3. This is the first deterministic algorithm for the problem with a constant competitive ratio. When p(J) ∈ {1, k}, k > 1 is a real number, we first present a lower bound min{(1 + k)/k, 2 k/(1 + k)}, and then we show that LEJ has a competitive ratio of1 + k/k. In particular, when all the k length jobs have tight deadlines, we first present a lower bound max{4/(2 + k), 1}(for p(J)E(J)) and 4/3(for E(J)). Then we prove that LEJ is k/k-competitive for p(J)E(J) and we provide an online algorithm H with a competitive ratio of 2 k/( k + 1) for the second objective function E(J).
基金supported by the National Key Research and Development Program of China(No.2016YFB1000101)
文摘To provide timely results for big data analytics, it is crucial to satisfy deadline requirements for MapReduce jobs in today's production environments. Much effort has been devoted to the problem of meeting deadlines, and typically there exist two kinds of solutions. The first is to allocate appropriate resources to complete the entire job before the specified time limit, where missed deadlines result because of tight deadline constraints or lack of resources; the second is to run a pre-constructed sample based on deadline constraints, which can satisfy the time requirement but fail to maximize the volumes of processed data. In this paper, we propose a deadline-oriented task scheduling approach, named 'Dart', to address the above problem. Given a specified deadline and restricted resources, Dart uses an iterative estimation method, which is based on both historical data and job running status to precisely estimate the real-time job completion time. Based on the estimated time, Dart uses an approach-revise algorithm to make dynamic scheduling decisions for meeting deadlines while maximizing the amount of processed data and mitigating stragglers. Dart also efficiently handles task failures and data skew, protecting its performance from being harmed. We have validated our approach using workloads from OpenCloud and Facebook on a cluster of 64 virtual machines. The results show that Dart can not only effectively meet the deadline but also process near-maximum volumes of data even with tight deadlines and limited resources.
文摘Fog computing became a traditional OffLad Destination(OLD)to compute the offloaded tasks of the Internet of Vehicles(IoV).Nevertheless,the limited computing resources of the fog node leads to re-offload these tasks to the neighboring fog nodes or the cloud.Thus,the IoV will incur additional offloading costs.In this paper,we propose a new offloading scheme by utilizing RoadSide Parked Vehicles(RSPV)as an alternative OLD for IoV.The idle computing resources of the RSPVs can compute large tasks with low offloading costs compared with fog nodes and the cloud.Finally,a performance evaluation of the proposed scheme has been presented and discussed with other benchmark offloading schemes.
文摘针对多DAG(Directed Acyclic Graph)共享一组异构资源在调度吞吐量最大化基础上的费用优化问题,本文提出了一种基于总费用变化量探测的费用优化算法PDTC(based on the Probe of the Total Cost Decrease),目的在于尽可能降低有优化条件的多个DAG的总费用.实验表明,该算法不仅能使得各DAG充分利用期限内的冗余时间,也能够在一定程度上降低多个DAG调度执行的总费用.
基金The Natural Science Foundation of Jiangsu Province(NoBK2005408)
文摘To fulfill the requirements for hybrid real-time system scheduling, a long-release-interval-first (LRIF) real-time scheduling algorithm is proposed. The algorithm adopts both the fixed priority and the dynamic priority to assign priorities for tasks. By assigning higher priorities to the aperiodic soft real-time jobs with longer release intervals, it guarantees the executions for periodic hard real-time tasks and further probabilistically guarantees the executions for aperiodic soft real-time tasks. The schedulability test approach for the LRIF algorithm is presented. The implementation issues of the LRIF algorithm are also discussed. Simulation result shows that LRIF obtains better schedulable performance than the maximum urgency first (MUF) algorithm, the earliest deadline first (EDF) algorithm and EDF for hybrid tasks. LRIF has great capability to schedule both periodic hard real-time and aperiodic soft real-time tasks.
文摘The problem of scheduling radar dwells in multifunction phased array radar systems is addressed. A novel dwell scheduling algorithm is proposed. The whole scheduling process is based on an online pulse interleaving technique. It takes the system timing and energy constraints into account. In order to adapt the dynamic task load, the algorithm considers both the priorities and deadlines of tasks. The simulation results demonstrate that compared with the conventional adaptive dwell scheduling algorithm, the proposed one can improve the task drop rate and system resource utility effectively.
基金supported in part by the Fundamental Research Funds for the Central Universities under Grant No.2014JBM011 and No.2014YJS021in part by NSFC under Grant No.62171200,61422101,and 62132017+2 种基金in part by the Ph.D.Programs Foundation of MOE of China under Grant No.20130009110014in part by "NCET" under Grant No.NCET-12-0767in part by China Postdoctoral Science Foundation under Grant No.2015M570028,2015M580970
文摘In modern datacenters, the most common method to solve the network latency problem is to minimize flow completion time during the transmission process. Following the soft real-time nature, the optimization of transport latency is relaxed to meet a flow's deadline in deadline-sensitive services. However, none of existing deadline-sensitive protocols consider deadline as a constraint condition of transmission.They can only simplify the objective of meeting a flow's deadline as a deadline-aware mechanism by assigning a higher priority for tight-deadline constrained flows to finish the transmission as soon as possible, which results in an unsatisfactory effect in the condition of high fan-in degree. It drives us to take a step back and rethink whether minimizing flow completion time is the optimal way in meeting flow's deadline. In this paper, we focus on the design of a soft real-time transport protocol with deadline constraint in datacenters and present a flow-based deadline scheduling scheme for datacenter networks(FBDS).FBDS makes the unilateral deadline-aware flow transmission with priority transform into a compound centralized single-machine deadlinebased flow scheduling decision. In addition, FBDS blocks the flow sets and postpones some flows with extra time until their deadlines to make room for the new arriving flows in order to improve the deadline meeting rate. Our simulation resultson flow completion time and deadline meeting rate reveal the potential of FBDS in terms of a considerable deadline-sensitive transport protocol for deadline-sensitive interactive services.
基金Project supported by the National Natural Science Foundation ofChina (Nos. 60074011 and 60174009), and Youth Science and Tech-nology Foundation of Shanxi Province (No. 20051020), China
文摘Network-induced delay and jitter are key factors causing performance degradation and instability of NCSs (networked control systems). The relationships between the sampling periods of the control loops, network-induced delay and jitter were studied aimed at token-type networks. A jitter-dependent optimal bandwidth scheduling algorithm for NCSs is proposed, which tries to achieve a tradeoff between bandwidth occupancy and system performance. Simulation tests proved the effectiveness of this optimal scheduling algorithm.
文摘Cloud-as-the-center computing paradigms face multiple challenges in the 5G and Internet of Things scenarios, where the service requests are usually initiated by the end-user devices located at network edge and have rigid time constraints. Therefore, Fog computing, or mobile edge computing, is introduced as a promising solution to the service provision in the tiered IoT infrastructure to compensate the shortage of traditional cloud-only architecture. In this cloud-to-things continuum, several cloudlet or mobile edge server entities are placed at the access network to handle the task offloading and processing problems at the network edge. This raises the resource scheduling problem in this tiered system, which is vital for the promotion of the system efficiency. Therefore, in this paper, a scheduling mechanism for the cloudlets or fog nodes are presented, which takes the mobile tasks’ deadline and resources requirements at the same time while promoting the overall profit of the system. First, the problem at the cloudlet, to which IoT devices offload their tasks, is formulated as a multi-dimensional 0-1 knapsack problem. Second, based on ant colony optimization, a scheduling algorithm is presented which treat this problem as a subset selection problem. Third, to promote the performance of the system in the dynamic environments,a churn-refined algorithm is further put forward. A series of simulation experiments have shown that out proposal outperforms many state-of-the-art algorithms in both profit and guarantee ratio.
基金Project(60673165) supported by the National Natural Science Foundation of China
文摘To address the issue of resource co-allocation with constraints to budget and deadline in grid environments, a novel co-allocation model based on virtual resource agent was proposed. The model optimized resources deployment and price scheme through a three-side co-allocation mechanism, and applied queuing system to model the work of grid resources for providing quantitative deadline guarantees for grid applications. The validity and solutions of the model were presented theoretically. Extensive simulations were conducted to examine the effectiveness and the performance of the model by comparing with other co-allocation policies in terms of deadline violation rate, resource benefit and resource utilization. Experimental results show that compared with the three typical co-allocation policies, the proposed model can reduce the deadline violation rate to about 3.5% for the grid applications with constraints to budget and deadline. Also, the system benefits can be increased by about 30% compared with the those widely-used co-allocation policies.
基金Project(71671035) supported by the National Natural Science Foundation of China
文摘Renewing warranty can provide customers with better service,and thus help manufacturers to gain market opportunities.In engineering practice,the cost for replacement is usually higher than the cost for maintenance,hence manufacturers often face huge challenge to reduce the warranty service cost.With consideration of the warranty deadline,we propose a two-stage optimization model for renewing warranty.In the first stage,a renewing warranty with deadline(RWD)policy is implemented,where the deadline represents the cumulative uptime threshold.When the cumulative uptime exceeds the deadline,the product will be minimally repaired and kept to the residual warranty period.When RWD is expired,the replacement warranty with limited repairs(RWLR)policy is applied.Under the free replacement and pro-rata warranty policy,the corresponding two-stage cost optimization model is established from the manufacturer’s perspective,the aim is to minimize the cost rate and obtain the optimal warranty period.A numerical example is provided to illustrate the validity of the proposed model,and the sensitivity analysis is also carried out.
文摘At 02:30 am Baghdad time on January 17, the U. S. and allied forces carried out 1000 sorties air and 100 ship-launcnea cruise missile strikes, blasting Iraqi targets to free Kuwait from Iraqi occupation. The attack began less than 19 hours after the expiration of the UN-mandated deadline. From the following day to the end of the month, the U.S.-led allied air forces continued massive bombing&against targets in Iraq and occupied Kuwait. From
基金supported by the National Natural Science Foundation of China(611630606110320461462007)
文摘In modern data centers, because of the deadline- agnostic congestion control in transmission control protocol (TCP), many deadline-sensitive flows can not finish before their deadlines. Therefore, providing a higher deadline meeting ratio becomes a critical challenge in the typical online data intensive (OLDI) ap- plications of data center networks (DCNs). However, a problem named as priority synchronization is found in this paper, which de- creases the deadline meeting ratio badly. To solve this problem, we propose a priority probability deceleration (P2D) deadline-aware TCP. By using the novel probabilistic deceleration, p2D prevents the priority synchronization problem. Simulation results show that P2D increases the deadline meeting ratio by 20% compared with D2TCP.
文摘Real-time task scheduling is of primary significance in multiprocessor systems.Meeting deadlines and achieving high system utilization are the two main objectives of task scheduling in such systems.In this paper,we represent those two goals as the minimization of the average response time and the average task laxity.To achieve this,we propose a genetic-based algorithm with problem-specific and efficient genetic operators.Adaptive control parameters are also employed in our work to improve the genetic algorithms' efficiency.The simulation results show that our proposed algorithm outperforms its counterpart considerably by up to 36% and 35% in terms of the average response time and the average task laxity,respectively.