Energy-efficient Mapping of Big Data Workflows under Deadline Constraints

Similar documents
Energy-Efficient Real-Time Task Scheduling in Multiprocessor DVS Systems

STATISTICAL PERFORMANCE

Che-Wei Chang Department of Computer Science and Information Engineering, Chang Gung University

Real-Time Scheduling and Resource Management

Embedded Systems 14. Overview of embedded systems design

Scheduling Lecture 1: Scheduling on One Machine

Embedded Systems 15. REVIEW: Aperiodic scheduling. C i J i 0 a i s i f i d i

CPU SCHEDULING RONG ZHENG

Paper Presentation. Amo Guangmo Tong. University of Taxes at Dallas February 11, 2014

How to deal with uncertainties and dynamicity?

Scheduling Lecture 1: Scheduling on One Machine

Multiprocessor Scheduling I: Partitioned Scheduling. LS 12, TU Dortmund

An Energy-efficient Task Scheduler for Multi-core Platforms with per-core DVFS Based on Task Characteristics

CHAPTER 5 - PROCESS SCHEDULING

EDF Scheduling. Giuseppe Lipari May 11, Scuola Superiore Sant Anna Pisa

Algorithms. Outline! Approximation Algorithms. The class APX. The intelligence behind the hardware. ! Based on

Non-Preemptive and Limited Preemptive Scheduling. LS 12, TU Dortmund

TDDB68 Concurrent programming and operating systems. Lecture: CPU Scheduling II

Real-time operating systems course. 6 Definitions Non real-time scheduling algorithms Real-time scheduling algorithm

Task Models and Scheduling

Segment-Fixed Priority Scheduling for Self-Suspending Real-Time Tasks

0-1 Knapsack Problem in parallel Progetto del corso di Calcolo Parallelo AA

Aperiodic Task Scheduling

Non-Work-Conserving Non-Preemptive Scheduling: Motivations, Challenges, and Potential Solutions

Resilient and energy-aware algorithms

An Efficient Knapsack-Based Approach for Calculating the Worst-Case Demand of AVR Tasks

There are three priority driven approaches that we will look at

CPU Consolidation versus Dynamic Voltage and Frequency Scaling in a Virtualized Multi-Core Server: Which is More Effective and When

This means that we can assume each list ) is

ONLINE SCHEDULING OF MALLEABLE PARALLEL JOBS

Embedded Systems Design: Optimization Challenges. Paul Pop Embedded Systems Lab (ESLAB) Linköping University, Sweden

Real-Time Systems. Event-Driven Scheduling

Scheduling I. Today. Next Time. ! Introduction to scheduling! Classical algorithms. ! Advanced topics on scheduling

APTAS for Bin Packing

Tuan V. Dinh, Lachlan Andrew and Philip Branch

Lecture 13. Real-Time Scheduling. Daniel Kästner AbsInt GmbH 2013

Andrew Morton University of Waterloo Canada

CS 6901 (Applied Algorithms) Lecture 3

Online Scheduling Switch for Maintaining Data Freshness in Flexible Real-Time Systems

Lecture 2: Metrics to Evaluate Systems

Revenue Maximization in a Cloud Federation

Scheduling Periodic Real-Time Tasks on Uniprocessor Systems. LS 12, TU Dortmund

Online Energy-Aware I/O Device Scheduling for Hard Real-Time Systems with Shared Resources

Energy-efficient scheduling

MICROPROCESSOR REPORT. THE INSIDER S GUIDE TO MICROPROCESSOR HARDWARE

Scheduling Parallel Jobs with Linear Speedup

Scheduling I. Today Introduction to scheduling Classical algorithms. Next Time Advanced topics on scheduling

Paper Presentation. Amo Guangmo Tong. University of Taxes at Dallas January 24, 2014

Optimal Voltage Allocation Techniques for Dynamically Variable Voltage Processors

On Minimizing Total Energy Consumption in the Scheduling of Virtual Machine Reservations

Clock-driven scheduling

Lecture 2: Scheduling on Parallel Machines

Minimizing Mean Flowtime and Makespan on Master-Slave Systems

Performance Metrics for Computer Systems. CASS 2018 Lavanya Ramapantulu

Metode şi Algoritmi de Planificare (MAP) Curs 2 Introducere în problematica planificării

Single Machine Problems Polynomial Cases

RCPSP Single Machine Problems

Module 5: CPU Scheduling

CSE 380 Computer Operating Systems

Chapter 6: CPU Scheduling

Spatio-Temporal Thermal-Aware Scheduling for Homogeneous High-Performance Computing Datacenters

A Dynamic Real-time Scheduling Algorithm for Reduced Energy Consumption

P C max. NP-complete from partition. Example j p j What is the makespan on 2 machines? 3 machines? 4 machines?

Real-Time Systems. Event-Driven Scheduling

Proportional Share Resource Allocation Outline. Proportional Share Resource Allocation Concept

TDDI04, K. Arvidsson, IDA, Linköpings universitet CPU Scheduling. Overview: CPU Scheduling. [SGG7] Chapter 5. Basic Concepts.

UC Santa Barbara. Operating Systems. Christopher Kruegel Department of Computer Science UC Santa Barbara

CycleTandem: Energy-Saving Scheduling for Real-Time Systems with Hardware Accelerators

Parallel Performance Theory - 1

Maximizing Rewards for Real-Time Applications with Energy Constraints

Energy-Constrained Scheduling for Weakly-Hard Real-Time Systems

Performance Metrics & Architectural Adaptivity. ELEC8106/ELEC6102 Spring 2010 Hayden Kwok-Hay So

Saving Energy in Sparse and Dense Linear Algebra Computations

Evaluation and Validation

EDF Scheduling. Giuseppe Lipari CRIStAL - Université de Lille 1. October 4, 2015

COSC 341: Lecture 25 Coping with NP-hardness (2)

2/5/07 CSE 30341: Operating Systems Principles

Algorithm Design. Scheduling Algorithms. Part 2. Parallel machines. Open-shop Scheduling. Job-shop Scheduling.

Energy-efficient Scheduling Algorithms for Real-time Parallel Applications on Heterogeneous Distributed Embedded Systems

Schedulability of Periodic and Sporadic Task Sets on Uniprocessor Systems

Power-Aware Linear Programming Based Scheduling for Heterogeneous Computer Clusters

EDF Feasibility and Hardware Accelerators

On Two Class-Constrained Versions of the Multiple Knapsack Problem

RUN-TIME EFFICIENT FEASIBILITY ANALYSIS OF UNI-PROCESSOR SYSTEMS WITH STATIC PRIORITIES

ILP Formulations for the Lazy Bureaucrat Problem

5 Integer Linear Programming (ILP) E. Amaldi Foundations of Operations Research Politecnico di Milano 1

Temperature-Aware Analysis and Scheduling

On the Complexity of Mapping Pipelined Filtering Services on Heterogeneous Platforms

Power-Saving Scheduling for Weakly Dynamic Voltage Scaling Devices

Optimizing Energy Consumption under Flow and Stretch Constraints

Networked Embedded Systems WS 2016/17

Enhancing Multicore Reliability Through Wear Compensation in Online Assignment and Scheduling. Tam Chantem Electrical & Computer Engineering

Special Nodes for Interface

State-dependent and Energy-aware Control of Server Farm

Computers and Intractability. The Bandersnatch problem. The Bandersnatch problem. The Bandersnatch problem. A Guide to the Theory of NP-Completeness

Reclaiming the energy of a schedule: models and algorithms

Computers and Intractability

A New Task Model and Utilization Bound for Uniform Multiprocessors

Batch delivery scheduling with simple linear deterioration on a single machine 1

Scheduling divisible loads with return messages on heterogeneous master-worker platforms

Transcription:

Energy-efficient Mapping of Big Data Workflows under Deadline Constraints Presenter: Tong Shu Authors: Tong Shu and Prof. Chase Q. Wu Big Data Center Department of Computer Science New Jersey Institute of Technology, Newark, NJ, USA WORKS on Nov. 14 th, 2016 in Salt Lake City, UT, USA

Outline Introduction Related Work Motivation Problem Formulation Computational Complexity Analysis Algorithm Design A Pipeline-structured Workflow on a Homogenous Cluster An Arbitrary Workflow on a Heterogeneous Cluster Performance Evaluation Summary of Contributions 2

Introduction Problem Large-scale workflows for big data analytics in Hadoop systems consume a significant amount of energy Existing main green computing techniques Insufficient to address the energy efficiency issue of big data workflows Existing techniques v Task consolidation to reduce static energy consumption (SEC) by turning off idle servers v Load balancing to reduce dynamic energy consumption (DEC) through dynamic voltage and frequency scaling (DVFS) Weakness v Frequently switching on and off a server may reduce its lifespan or cause unnecessary peaks of energy consumption v DVFS may not be always available on all servers in a cluster Our solution Reduce DEC by adaptively determining the degree of parallelism in each MapReduce job to mitigate the workload overhead while meeting a given performance requirement 3

Related Work in Energy-efficient Scheduling Workflow Scheduling Ø Virtualized environments o Migrate active VMs onto energy-efficient PMs [TCC 16] o Consolidate jobs with complementary resource demands [SC 10] Ø Physical environments [TPDS 11, TC 12, IS 15, IS 16] o Extend virtual deadlines of jobs with less dependencies [TPDS 16] J Existing work only considers workflows comprised of serial or rigid jobs, while our work is focused on workflows comprised of moldable jobs MapReduce Job Scheduling Ø Heterogeneous machines [TC 12, TPDS 15, ICDCS 15] Ø Renewable energy [EuroSys 12, IPDPS 15] Ø Overhead reduction [ICDCS 13, Middleware 15] J Existing work only considers independent MapReduce jobs, while our work deals with precedence constraints among MapReduce jobs Moldable/Malleable Job Scheduling Ø DVFS: P solution (p = f α ) and FPTAS (p = f α + δ) [Euro-Par 12] J Existing work schedules malleable jobs based on theoretical power consumption models, while our work schedules moldable jobs based on measurement-based power consumption models 4

Motivation Rigid Jobs ØRunning on a fixed number of processors ØMulti-threaded programs Parallel Jobs Moldable Jobs ØRunning on any number of processors decided prior to execution ØMapReduce jobs in Hadoop Malleable Jobs ØRunning on a variable number of processors at runtime Performance model The workload of each component task decreases and the total workload, proportional to DEC, increases as the number of allotted processors increases deadline 6 6 time = 6 workload = 6 (DEC = 6) 4 deadline < 6 3 deadline < 4 3 3 3 4 4 time = 4 workload = 8 (DEC = 8) time = 3 workload = 9 (DEC = 9) time DEC 5

Performance Model Validation Experimental Results in Hadoop/YARN A single real-life MapReduce job runs on a computer server equipped with 2 processors of Intel(R) Xeon(R) CPU E5-2630, each with 6 cores of 2.30GHz and 64GB memory The execution time of the job increases as the number of tasks increases when the server is fully utilized by the job, which means the total workload increases with the number of tasks 6

Cost Models Cluster Model Cluster M = {machine m i (N i, p i, o i, P i )} Available resource-time (ART) table Workflow Model DAG with a deadline: f(g(v, A), d), where MapReduce Model Each task s j,k in job v j requires a memory of size o j and spends a percentage µ i,j of time executing CPU-bound instructions on a CPU core of machine m i As the number K j of parallel tasks increases, the workload w j,k (K j ) of each task s j,k decreases and the total workload w j (K j ) = K j w j,k (K j ) of all tasks increases Execution time of each task: Used resources:, Energy Model Dynamic power consumption proportional to utilization Mapping Function 7

Problem Definition: EEWM Input Given a cluster {m i (N i, p i, o i, P i )} of machines with an available resource-time table, and a workflow request f(g(v,a),d), where each job v j has a set {w j (K j ) K j = 1, 2,, K' j } of workloads for different task partitions, and each task in job v j has a memory demand o j and a percentage µ i,j of execution time for CPU-bound instructions on machine m i, Question Find a mapping function to minimize dynamic energy consumption: Constraints Precedence: Deadline: Resource: Computational Complexity: Strongly NP-complete => No FPTAS 8

Special Case: A Pipeline-structured Workflow on a Homogenous Cluster (PHO) Computational Complexity Analysis The two-choice knapsack problem (TCKP) is NP-hard The knapsack problem is NP-hard Reduce the knapsack problem to TCKP $4 $5 $3 3 kg 4 kg $7 2 kg $0 $0 5 kg $0 0 kg 0 kg $0 0 kg? 0 kg $1 $9 1 kg 6 kg $0 0 kg max total value $0 0 kg 9

Special Case: A Pipeline-structured Workflow on a Homogenous Cluster (2) Computational Complexity Analysis: NP-hardness Reduce TCKP to EEWM-PHO max total value V total weight limit W min energy E = 1 j J V j V deadline d = W Item 1 Item 2 Class J Class 1 Class j value v j,1 weight w j,1 value v j,2 weight w j,2 v j,1 > v j,2 w j,1 > w j,2 > 0 V j = (2w j,2 v j,1 w j,1 v j,2 )/(2w j,2 w j,1 ) Job j Option1 (1 task): energy E j (1) = V j v j,1 time t j (1) = w j,1 Option 2 (2 tasks): energy E j (2) = V j v j,2 time t j (2) = w j,2 Job 1... Job j... Job J deadline d 10

Special Case: A Pipeline-structured Workflow on a Homogenous Cluster (3) Fully Polynomial-Time Approximation Scheme (FPTAS) Reducing EEWM-PHO to RSP Task partition options Task partition options edge[time,energy] u 1,1 u J,1 e 1,1 [t 1 (1),E 1 (1)] e 2,1 [0,0] u 1,2 e 2J-1,1 [t J (1), E J (1)] u J,2 e 2J,1 [0,0] e 1,2 [t 1 (2),E 1 (2)] e 1,k [t 1 (k),e 1 (k)] u 0 u 1,k e 2,2 [0,0] e 2,k [0,0] u 1 e 3,k e 2J-2,k u J-1 e 2J-1,2 [t J (2), E J (2)] e 2J,2 [0,0] e 2J-1,k [t J (k),e J (k)] e 2J,k [0,0] u J,k u J e 1,K [t 1 (K ), E 1 (K )] e 2,K [0,0] 1 1 1 1 J J J e 2J-1,K [t J (K ), E J (K )] e 2J,K [0,0] J u 1,K 1 u J,K J 1 st Job J th Job deadline d The restricted shortest path (RSP) problem is weakly NP-complete and solvable with FPTAS, so is EEWM-PHO EEWM-PHO-FPTAS Time Complexity: O(JK'(log K' + 1/ε)), where K' = max 1 j J K' j Near optimality and low time complexity (linear w.r.t. 1/ε) 11

Generalized Problem: An Arbitrary Workflow on a Heterogeneous Cluster Heuristic -- BAWMEE Iterative critical path selection Iteratively computes a critical path with the earliest last finish time (LFT) from the remaining unmapped workflow branches according to the average execution time of each job running in serial on all the machines. Pipeline mapping (considering machine heterogeneity) Resource/time sufficiency: In energy-efficient pipeline mapping (EEPM), we map all the homogeneous tasks in the same job onto a homogeneous sub-cluster, hence using EEWM-PHO-FPTAS to balance the trade-off between execution time and dynamic energy consumption for each job on a pipeline. Resource/time insufficiency: In minimum delay pipeline mapping (MDPM) with energy awareness, we search for a good task partitioning to minimize the end time of each job by exponentially relaxing the limit on the maximum number of tasks in a job to make a tradeoff between the optimality and the time complexity. A critical path is the longest execution path in a workflow. 12

Example for Resource/time sufficiency Pipeline 1: Pipeline 2: (1,C 1 ):0-3 (1,C 1 ):3-6 (1,C 2 ):6-11 (1,C 1 ):11-14 (1,C 2 ):14-19 (1,C 1 ):0-3 (1,C 1 ):3-6 (1,C 2 ):6-11 (1,C 1 ):11-14 (1,C 2 ):14-19 v 1 v 2 v 3 v 4 v 5 v 6 v 7 v 8 v 1 v 2 v 3? v 4 v 5 v 6 v 7 v 8 0 19 (1,C 2 ):3-8 (1,C 1 ):8-11 (1,C 1 ):11-14 3 14 (1,C 1 ):3-6 3 6 13

Example for Resource/time sufficiency (2) Pipeline 3: (1,C 1 ):0-3 (1,C 1 ):3-6 (1,C 2 ):6-11 (1,C 1 ):11-14 (1,C 2 ):14-19 v 1 v 2 v 4 v 3 v 5? v 6 (1,C 1 ):3-6 (1,C 1 ):6-9 (1,C 2 ):9-14 6 14 v 7 v 8 Mapping Results BAWMEE (DEC = 45) 0 3 6 11 14 19 m 1 v 1 v 2 v 6 m 2 v 3 v 7 m 3 v 4 v 8 m 4 v 5 (1,C 2 ):6-11 Pipeline 4: (1,C 1 ):0-3 (1,C 1 ):3-6 (1,C 2 ):6-11 (1,C 1 ):11-14 (1,C 2 ):14-19 v 1 v 2 6 11 v 3 v 5 v 4 v 6 v 7 v 8 Optimal (DEC = 44) 0 3 6 11 16 19 m 1 v 1 v 2 v 8 m 2 v 3 m 3 v 4 v 6 m 4 v 5 v 7 (1,C 1 ):3-6 (1,C 2 ):6-11 (1,C 1 ):11-14 11 14 14

Simulation Settings Cluster Homogeneous sub-clusters Each sub-cluster has the same number of machines Workflows # of edges = # of jobs x 1.5 Arrivals: Poisson distribution with the average arrival interval of 30 min Simulation Duration per run: 3 days The number of runs: 20 MapReduce jobs The maximum number of tasks per job: 12 to 48 The workload w(k) of a job 1 task: 0.6 to 21.6 Tera CPU cycles k>1 tasks: w(k 1)[1+0.2/ (k 1)] to w(k 1)[1+0.6/(k 1)] The percentage of execution time for CPU-bound instructions: 0.1-1.0 Memory demand per task: 0.5GB to 4GB at a step of 0.5GB Approximation ε = 0.2 15

Simulation Results Problem Size Deadline factor: 0.1 Workflow Structure: random 16

Simulation Results (2) Deadline Constraint Workflow size: 40-60 jobs Cluster size: 128 machines Workflow Structure: random 17

Simulation Results (3) Workflow Size Deadline factor: 0.1 Workflow size variation: ±2 jobs Cluster size: 128 machines Workflow Structure: random 18

Simulation Results (4) Cluster Size Deadline factor: 0.1 Workflow size: 40-60 jobs Workflow Structure: random 19

Simulation Results (5) Workflow Structure Deadline factor: 0.1 Workflow size: 40-60 jobs Cluster size: 128 machines 20

Summary of Contributions The first to study energy-efficient mapping of big data workflows comprised of moldable jobs in Hadoop systems Proved a deadline-constrained pipeline-structured workflow mapping problem for minimum total (energy) cost to be weakly NP-complete and designed an FPTAS for it The performance superiority of the proposed heuristic for the general workflow mapping problem was illustrated by simulation results in Hadoop/YARN Dynamic energy saving Deadline missing rate 21

22