Reliability at Scale

Size: px
Start display at page:

Download "Reliability at Scale"

Transcription

1 Reliability at Scale Intelligent Storage Workshop 5 James Nunez Los Alamos National lab LA-UR & LA-UR May 15, 2007

2 A Word about scale Petaflop class machines LLNL Blue Gene 350 Tflops 128k processors 2005 Storage Demands: 896 TB LANL Road Runner 1.4 Pflops Hybrid System 2008 Storage Demands: 3 PBytes (projected)

3 Why do you need so much storage?

4 What Drives Us? ASCI Balanced System Approach Disk Parallel I/O GigaBytes/sec Memory TeraBytes TeraBytes Computing Speed TFLOP/s Year 2012 Application Performance Network Speed Gigabits/sec Archival Storage PetaBytes Computational Computational Resource Resource Scaling Scaling for for ASCI ASCI Physics Physics Applications Applications 1 1 FLOP/s FLOP/s Peak Peak Compute Compute Byte/FLOP/s Byte/FLOP/s Memory Memory Byte/FLOP/s Byte/FLOP/s Disk Disk Byte/s/FLOP/s Byte/s/FLOP/s Peak Peak Parallel Parallel I/O I/O bit/s/flop/s bit/s/flop/s Peak Peak Network Network Byte/FLOP/s Byte/FLOP/s Archive Archive

5 Machines are Getting Faster Number of Processing Units in different machines Number of Processing Units Blue Mountain 3TF (1996) White 12TF (2001) Q 20TF (2003) Purple 100TF (2006) BG/L 300TF (2006) Machine - Year Cpu Speeds vs. Chip Type (year) 3500 Speed (MHz) Intel 8088 (1987) HP PA (1989) MIPS R4000 (1991) DEC Alpha 4 (1994) DEC Alpha 5 (1996) DEC Alpha 21264A (1999) U N C L A Proccessor S S I F I E type D (year) Intel Itanium (2001) Sun Ultra SPARC IIIcu (2001) AMD Opteron (2002) Intel Xeon (2006)

6 Disk Drives are Getting Denser Storage Capacity vs. Disk Type MegaBytes Ramac Fuji Eagle 690M-36k Seagate Elite 1.4G-54k Disk Type Seagate Barracuda-4 4.3G-72k Seagate Barracuda 18G- 72k Seagate Cheetah 73G- 10k Data Transfer Rate vs. Disk Type MegaBytes/Sec Ramac Fuji Eagle 690M-36k Seagate Elite 1.4G-54k Seagate Barracuda-4 4.3G-72k Seagate Barracuda 18G- 72k Seagate Cheetah 73G- 10k Disk Type

7 Density versus Bandwidth Disks are getting faster, but not at same rate as density read/write bandwidth is not keeping up with machines Data Rate / MB (density) 0.12 MB/Sec per MB Ramac Fuji Eagle 690M-36k Seagate Elite 1.4G-54k Seagate Barracuda-4 4.3G-72k Seagate Barracuda 18G- 72k Seagate Cheetah 73G- 10k Disk Type

8 The ASC I/O Ratio and past over engineering of the BW ASC ratios (1 GByte/sec per Tflop and 20 Bytes/flop disk) In 1996 on a 3 Tflop system, 20 bytes/flop is 60 TBytes of disk, which yielded about 48 GigaBytes/sec which was over engineered by a factor of 16X for BW In 2002 on a 20 Tflop system, 20 bytes/flop is 400 Tbyte of disk, which yielded about 40 Gigabytes/sec which was over engineered by a factor of 2X for BW Today for a 100 Tflop machine, 20 bytes/flop is 2000 Tbytes of disk yields a little over 100 Gigabytes/sec, which is not over engineered at all. We do not enjoy having far more BW than we really needed to get the space anymore!

9 Classical RAID Plus 1 Rebuild Read the remaining disks, XOR, and write the result. Speed ultimately governed by write speed of a (single) target disk or read of N disks This is true for N+1 and N+2 with Classical RAID A0+B0+D0+P0 Storage Blade Storage Blade Storage Blade Storage Blade Storage Blade

10 Classical RAID Rebuild Time Due to increase in density and little increase in data rate, rebuild times get worse and worse, from minutes, to hours, to days raising chances of 2-3 disk failure more and more Classical Raid Rebuild Time hours minimum busy tf tf tf tf tf years

11 More bad news, rebuild is under pressure, unrecoverable bit error rates (UBER) on read High end 73GB FC 2.5 inch 10k drives UBER=10 to the 15 th bits SATA 400GB 3.5 inch drives UBER=10 to the 14 th bits Rebuild High end 4+N have to read 292GBytes or 2.3*10 to the 12 bits (error on rebuild every few thousand) SATA 4+N have to read 1600GBytes or 1.2*10 to the 13 bits (error on rebuild every few tens) High end 8+N have to read 584GBytes or 4.6*10 to the 12 bits (error on rebuild every few thousand) SATA 8+N have to read 3200GBytes or 2.6*10 to the 13 bits (error on rebuild every few tens)

12 Plus 1 RAID 5 Independent data disks with distributed parity blocks. Data is striped across a number of storage devices and a parity stripe is written for fault tolerance. Parity load is shared. With disk blocksizes getting bigger you must aggregate more and more data to do an efficient write (no read, update, write)

13 RAID N Plus 2 Examples RAID 6 Row Diagonal parity Lets look at one scheme Normal XOR parity is calculated straight across the disk blocks Diagonal parity is calculated on diagonals, there are other methods based on polynomials You need to have way more data around to do efficient parity calculation This means you have to aggregate more data to get efficient writes (no read,update,write)

14 What do our applications do?

15 The Applications versus the Industry CPU s are not getting faster, so we are getting more CPU s. Memory per processor is not going up appreciably, in some cases it is going down Therefore, apps are not going to write larger writes (and writes are already too small for current storage systems) But RAID/Disks are requiring larger and larger write ops for efficiency

16 N-to-N example T P0 P P0 H P0 Process 0 T P1 P P1 H P1 Process 1 T P2 P P2 H P2 Process 2 T P0 P P0 H P0 file0 T P2 P P2 H P2 file2 T P1 P P1 H P1 file1

17 N-to-1 non-strided example Process 0 T P0 P P0 H P0 Process 1 T P1 P P1 H P1 Process 2 T P2 P P2 H P2 T P0 P P0 H P0 T P1 P P1 H P1 T P2 P P2 H P2

18 N-to-1 strided example Process 0 T P0 P P0 H P0 Process 1 T P1 P P1 H P1 Process 2 T P2 P P2 H P2 T P0 T P1 T P2 P P0 P P1 P P2 H P0 H P1 H P2

19 N-to-1 strided evaluation Advantages Simplest book-keeping for N-to-M restart Read each element contiguously, resplit for M Simplest formatting for visualization Visualization typically only interested in small number of variables, can read each contiguously Smaller number of files to manage Can help with archiving and with managing your data Disadvantages Small, unaligned writes False sharing Read-modify-writes Note: All advantages relative to user, All disadvantages relative to storage system.

20 Applications seem to want to use N to 1 strided for convenience, for big writes this is not an issue but small writes are problematic

21 Microprocessor trends are changing Moore s law still holds but is now being realized differently Clock frequency, chip power, & instructionlevel-parallelism (ILP) have all plateaued Multi-core is here today and manycore ( 32 ) looks to be the future Scalability of full complex & cache-based core designs to manycore designs is likely problemmatic Memory bandwidth and memory capacity per core are headed downward (predominantly caused by increased core counts) Key findings of Jan IDC Study: Next Phase in HPC new ways of dealing with parallelism will be required must focus more heavily on bandwidth (flow of data) and less on processor References: IDC report #205025, January 2007 UC Berkeley UCB/EECS LASCI-06 Burton Smith keynote Reinventing Computing U N C L A S S I F I E From D Burton Smith, LASCI-06 keynote, with permission; data from Herb Sutter, Microsoft From LA-UR Salishan RR Talk Intel 80-core

22 A Disturbing Summary ASC ratio driven BW over engineering is no more You have to involve more disks to do the job Number of disks to get the BW is going through the roof Rebuild times get worse and worse Plus 2 technologies don t really solve the problem reliability/rebuild problem It takes larger and larger write operations to be efficient Applications are having each process write less data as cores go up

23 So, do we go home and call it quits?

24 Some Solutions Data Aggregation in Middleware ROMIO s (existing) aggregation Northwestern s work on DAChe and Persistent File Domains in MPI-IO (ROMIO) Per object Parity

25 Per Object Parity: Enables scalable rebuild and helps address UBER issue! Parity by object (not by disk) Different files are striped across different disks Scalable object rebuild Gets list of objects that are partially on the bad disk Starts many parallel processes to read the other related disks for each object and recalculate and write missing piece on any disk in the system and updates map Rebuild is scalable, not gated by single disk write anymore The more rebuilding processes the faster rebuild Each rebuilder is gated by reading N drives and writing to any one drive Rebuild ratio N reads to 1 write UBER based read error only yields one bad object not bad entire disk Object or Disk Server Object or Disk Server File System Clients Object or Disk Server Object or Disk Server Obj Obj Object or Disk Server P Object or Disk Server 4 P Obj P

26 Some Solutions Per object Parity - Where At the disk In the network In the client RAID10 Forget about parity, make two or three copies of the data

27 Per Object Parity + RAID10 : Scalable Rebuild All same as before but Rebuild ratio is 1 to 1 (very nice and very simple) File System Clients Object or Disk Server Object or Disk Server Object or Disk Server Object or Disk Server Object or Disk Server Object or Disk Server Obj Obj Obj

28 RAID10 versus RAID5 Results

29 RAID10 versus RAID5 Results

30 RAID10, are there any drawbacks Sure, sending two copies of the data to disk Doubles the space needed Total BW is cut in half But Scales well for all problems, N to N, N to 1 non strided and N to N strided (even small) Simple, and that is important in a world with100k disks Helps with scalable rebuild which is needed for faster rebuild Helps with per object RAID which helps the UBER problem And besides that, we can use that density that we are getting more or less for free

31 Bottom line on reliability at scale! Had to get away from RAID5 due to rebuilds and number of disks RAID6 could work with lots of cache in front of it Because we will be implementing a scalable global parallel file system that is shared between all future clusters It will have a life well beyond any one cluster It will become extremely large and need reliability and availability requirements well beyond any single cluster file system Per object RAID is fundamental approach to solving our problem Hurts small unaligned writes which is becoming more prevalent We need scalable rebuild - implies per object parity, done on the client RAID10 is simpler, and scales better But we need RAID10 to fix the N to 1 small strided application writing problem RAID10 seems to have addresses this and is the direction we are following

416 Distributed Systems

416 Distributed Systems 416 Distributed Systems RAID, Feb 26 2018 Thanks to Greg Ganger and Remzi Arapaci-Dusseau for slides Outline Using multiple disks Why have multiple disks? problem and approaches RAID levels and performance

More information

A Tale of Two Erasure Codes in HDFS

A Tale of Two Erasure Codes in HDFS A Tale of Two Erasure Codes in HDFS Dynamo Mingyuan Xia *, Mohit Saxena +, Mario Blaum +, and David A. Pease + * McGill University, + IBM Research Almaden FAST 15 何军权 2015-04-30 1 Outline Introduction

More information

Some thoughts about energy efficient application execution on NEC LX Series compute clusters

Some thoughts about energy efficient application execution on NEC LX Series compute clusters Some thoughts about energy efficient application execution on NEC LX Series compute clusters G. Wellein, G. Hager, J. Treibig, M. Wittmann Erlangen Regional Computing Center & Department of Computer Science

More information

Redundant Array of Independent Disks

Redundant Array of Independent Disks Redundant Array of Independent Disks Yashwant K. Malaiya 1 Redundant Array of Independent Disks (RAID) Enables greater levels of performance and/or reliability How? By concurrent use of two or more hard

More information

High-Performance Scientific Computing

High-Performance Scientific Computing High-Performance Scientific Computing Instructor: Randy LeVeque TA: Grady Lemoine Applied Mathematics 483/583, Spring 2011 http://www.amath.washington.edu/~rjl/am583 World s fastest computers http://top500.org

More information

CS 700: Quantitative Methods & Experimental Design in Computer Science

CS 700: Quantitative Methods & Experimental Design in Computer Science CS 700: Quantitative Methods & Experimental Design in Computer Science Sanjeev Setia Dept of Computer Science George Mason University Logistics Grade: 35% project, 25% Homework assignments 20% midterm,

More information

CSE 4201, Ch. 6. Storage Systems. Hennessy and Patterson

CSE 4201, Ch. 6. Storage Systems. Hennessy and Patterson CSE 4201, Ch. 6 Storage Systems Hennessy and Patterson Challenge to the Disk The graveyard is full of suitors Ever heard of Bubble Memory? There are some technologies that refuse to die (silicon, copper...).

More information

Essentials of Large Volume Data Management - from Practical Experience. George Purvis MASS Data Manager Met Office

Essentials of Large Volume Data Management - from Practical Experience. George Purvis MASS Data Manager Met Office Essentials of Large Volume Data Management - from Practical Experience George Purvis MASS Data Manager Met Office There lies trouble ahead Once upon a time a Project Manager was tasked to go forth and

More information

PERFORMANCE METRICS. Mahdi Nazm Bojnordi. CS/ECE 6810: Computer Architecture. Assistant Professor School of Computing University of Utah

PERFORMANCE METRICS. Mahdi Nazm Bojnordi. CS/ECE 6810: Computer Architecture. Assistant Professor School of Computing University of Utah PERFORMANCE METRICS Mahdi Nazm Bojnordi Assistant Professor School of Computing University of Utah CS/ECE 6810: Computer Architecture Overview Announcement Jan. 17 th : Homework 1 release (due on Jan.

More information

NEC PerforCache. Influence on M-Series Disk Array Behavior and Performance. Version 1.0

NEC PerforCache. Influence on M-Series Disk Array Behavior and Performance. Version 1.0 NEC PerforCache Influence on M-Series Disk Array Behavior and Performance. Version 1.0 Preface This document describes L2 (Level 2) Cache Technology which is a feature of NEC M-Series Disk Array implemented

More information

One Optimized I/O Configuration per HPC Application

One Optimized I/O Configuration per HPC Application One Optimized I/O Configuration per HPC Application Leveraging I/O Configurability of Amazon EC2 Cloud Mingliang Liu, Jidong Zhai, Yan Zhai Tsinghua University Xiaosong Ma North Carolina State University

More information

Coding for loss tolerant systems

Coding for loss tolerant systems Coding for loss tolerant systems Workshop APRETAF, 22 janvier 2009 Mathieu Cunche, Vincent Roca INRIA, équipe Planète INRIA Rhône-Alpes Mathieu Cunche, Vincent Roca The erasure channel Erasure codes Reed-Solomon

More information

Redundant Array of Independent Disks

Redundant Array of Independent Disks Redundant Array of Independent Disks Yashwant K. Malaiya 1 Redundant Array of Independent Disks (RAID) Enables greater levels of performance and/or reliability How? By concurrent use of two or more hard

More information

HYCOM and Navy ESPC Future High Performance Computing Needs. Alan J. Wallcraft. COAPS Short Seminar November 6, 2017

HYCOM and Navy ESPC Future High Performance Computing Needs. Alan J. Wallcraft. COAPS Short Seminar November 6, 2017 HYCOM and Navy ESPC Future High Performance Computing Needs Alan J. Wallcraft COAPS Short Seminar November 6, 2017 Forecasting Architectural Trends 3 NAVY OPERATIONAL GLOBAL OCEAN PREDICTION Trend is higher

More information

CMP 338: Third Class

CMP 338: Third Class CMP 338: Third Class HW 2 solution Conversion between bases The TINY processor Abstraction and separation of concerns Circuit design big picture Moore s law and chip fabrication cost Performance What does

More information

GPU Computing Activities in KISTI

GPU Computing Activities in KISTI International Advanced Research Workshop on High Performance Computing, Grids and Clouds 2010 June 21~June 25 2010, Cetraro, Italy HPC Infrastructure and GPU Computing Activities in KISTI Hongsuk Yi hsyi@kisti.re.kr

More information

Outline. policies for the first part. with some potential answers... MCS 260 Lecture 10.0 Introduction to Computer Science Jan Verschelde, 9 July 2014

Outline. policies for the first part. with some potential answers... MCS 260 Lecture 10.0 Introduction to Computer Science Jan Verschelde, 9 July 2014 Outline 1 midterm exam on Friday 11 July 2014 policies for the first part 2 questions with some potential answers... MCS 260 Lecture 10.0 Introduction to Computer Science Jan Verschelde, 9 July 2014 Intro

More information

RAID+: Deterministic and Balanced Data Distribution for Large Disk Enclosures

RAID+: Deterministic and Balanced Data Distribution for Large Disk Enclosures RAID+: Deterministic and Balanced Data Distribution for Large Disk Enclosures Guangyan Zhang, Zican Huang, Xiaosong Ma SonglinYang, Zhufan Wang, Weimin Zheng Tsinghua University Qatar Computing Research

More information

Weather Research and Forecasting (WRF) Performance Benchmark and Profiling. July 2012

Weather Research and Forecasting (WRF) Performance Benchmark and Profiling. July 2012 Weather Research and Forecasting (WRF) Performance Benchmark and Profiling July 2012 Note The following research was performed under the HPC Advisory Council activities Participating vendors: Intel, Dell,

More information

Quantum Effect or HPC without FLOPS. Lugano March 23, 2016

Quantum Effect or HPC without FLOPS. Lugano March 23, 2016 Quantum Effect or HPC without FLOPS Lugano March 23, 2016 Electronics April 19, 1965 2016 D-Wave Systems Inc. All Rights Reserved 2 Moore s Law 2016 D-Wave Systems Inc. All Rights Reserved 3 www.economist.com/technology-quarterly/2016-03-12/aftermoores-law

More information

Claude Tadonki. MINES ParisTech PSL Research University Centre de Recherche Informatique

Claude Tadonki. MINES ParisTech PSL Research University Centre de Recherche Informatique Claude Tadonki MINES ParisTech PSL Research University Centre de Recherche Informatique claude.tadonki@mines-paristech.fr Monthly CRI Seminar MINES ParisTech - CRI June 06, 2016, Fontainebleau (France)

More information

Marla Meehl Manager of NCAR/UCAR Networking and Front Range GigaPoP (FRGP)

Marla Meehl Manager of NCAR/UCAR Networking and Front Range GigaPoP (FRGP) Big Data at the National Center for Atmospheric Research (NCAR) & expanding network bandwidth to NCAR over Pacific Wave and Western Regional Network (WRN) Marla Meehl Manager of NCAR/UCAR Networking and

More information

CS425: Algorithms for Web Scale Data

CS425: Algorithms for Web Scale Data CS425: Algorithms for Web Scale Data Most of the slides are from the Mining of Massive Datasets book. These slides have been modified for CS425. The original slides can be accessed at: www.mmds.org Challenges

More information

CONTEMPORARY ANALYTICAL ECOSYSTEM PATRICK HALL, SAS INSTITUTE

CONTEMPORARY ANALYTICAL ECOSYSTEM PATRICK HALL, SAS INSTITUTE CONTEMPORARY ANALYTICAL ECOSYSTEM PATRICK HALL, SAS INSTITUTE Copyright 2013, SAS Institute Inc. All rights reserved. Agenda (Optional) History Lesson 2015 Buzzwords Machine Learning for X Citizen Data

More information

Lecture 9: Clocking, Clock Skew, Clock Jitter, Clock Distribution and some FM

Lecture 9: Clocking, Clock Skew, Clock Jitter, Clock Distribution and some FM Lecture 9: Clocking, Clock Skew, Clock Jitter, Clock Distribution and some FM Mark McDermott Electrical and Computer Engineering The University of Texas at Austin 9/27/18 VLSI-1 Class Notes Why Clocking?

More information

Power in Digital CMOS Circuits. Fruits of Scaling SpecInt 2000

Power in Digital CMOS Circuits. Fruits of Scaling SpecInt 2000 Power in Digital CMOS Circuits Mark Horowitz Computer Systems Laboratory Stanford University horowitz@stanford.edu Copyright 2004 by Mark Horowitz MAH 1 Fruits of Scaling SpecInt 2000 1000.00 100.00 10.00

More information

PI SERVER 2012 Do. More. Faster. Now! Copyr i g h t 2012 O S Is o f t, L L C. 1

PI SERVER 2012 Do. More. Faster. Now! Copyr i g h t 2012 O S Is o f t, L L C. 1 PI SERVER 2012 Do. More. Faster. Now! Copyr i g h t 2012 O S Is o f t, L L C. 1 AUGUST 7, 2007 APRIL 14, 2010 APRIL 24, 2012 Copyr i g h t 2012 O S Is o f t, L L C. 2 PI Data Archive Security PI Asset

More information

Lecture 19. Architectural Directions

Lecture 19. Architectural Directions Lecture 19 Architectural Directions Today s lecture Advanced Architectures NUMA Blue Gene 2010 Scott B. Baden / CSE 160 / Winter 2010 2 Final examination Announcements Thursday, March 17, in this room:

More information

Chapter 7. Sequential Circuits Registers, Counters, RAM

Chapter 7. Sequential Circuits Registers, Counters, RAM Chapter 7. Sequential Circuits Registers, Counters, RAM Register - a group of binary storage elements suitable for holding binary info A group of FFs constitutes a register Commonly used as temporary storage

More information

Lecture 12: Energy and Power. James C. Hoe Department of ECE Carnegie Mellon University

Lecture 12: Energy and Power. James C. Hoe Department of ECE Carnegie Mellon University 18 447 Lecture 12: Energy and Power James C. Hoe Department of ECE Carnegie Mellon University 18 447 S18 L12 S1, James C. Hoe, CMU/ECE/CALCM, 2018 Housekeeping Your goal today a working understanding of

More information

P214 Efficient Computation of Passive Seismic Interferometry

P214 Efficient Computation of Passive Seismic Interferometry P214 Efficient Computation of Passive Seismic Interferometry J.W. Thorbecke* (Delft University of Technology) & G.G. Drijkoningen (Delft University of Technology) SUMMARY Seismic interferometry is from

More information

2.6 Complexity Theory for Map-Reduce. Star Joins 2.6. COMPLEXITY THEORY FOR MAP-REDUCE 51

2.6 Complexity Theory for Map-Reduce. Star Joins 2.6. COMPLEXITY THEORY FOR MAP-REDUCE 51 2.6. COMPLEXITY THEORY FOR MAP-REDUCE 51 Star Joins A common structure for data mining of commercial data is the star join. For example, a chain store like Walmart keeps a fact table whose tuples each

More information

Coping with disk crashes

Coping with disk crashes Lecture 04.03 Coping with disk crashes By Marina Barsky Winter 2016, University of Toronto Disk failure types Intermittent failure Disk crash the entire disk becomes unreadable, suddenly and permanently

More information

1 Overview. 2 Adapting to computing system evolution. 11 th European LS-DYNA Conference 2017, Salzburg, Austria

1 Overview. 2 Adapting to computing system evolution. 11 th European LS-DYNA Conference 2017, Salzburg, Austria 1 Overview Improving LSTC s Multifrontal Linear Solver Roger Grimes 3, Robert Lucas 3, Nick Meng 2, Francois-Henry Rouet 3, Clement Weisbecker 3, and Ting-Ting Zhu 1 1 Cray Incorporated 2 Intel Corporation

More information

The Memory Intensive System

The Memory Intensive System DiRAC@Durham The Memory Intensive System The DiRAC-2.5x Memory Intensive system at Durham in partnership with Dell Dr Lydia Heck, Technical Director ICC HPC and DiRAC Technical Manager 1 DiRAC Who we are:

More information

Scaling of MOS Circuits. 4. International Technology Roadmap for Semiconductors (ITRS) 6. Scaling factors for device parameters

Scaling of MOS Circuits. 4. International Technology Roadmap for Semiconductors (ITRS) 6. Scaling factors for device parameters 1 Scaling of MOS Circuits CONTENTS 1. What is scaling?. Why scaling? 3. Figure(s) of Merit (FoM) for scaling 4. International Technology Roadmap for Semiconductors (ITRS) 5. Scaling models 6. Scaling factors

More information

Qualitative vs Quantitative metrics

Qualitative vs Quantitative metrics Qualitative vs Quantitative metrics Quantitative: hard numbers, measurable Time, Energy, Space Signal-to-Noise, Frames-per-second, Memory Usage Money (?) Qualitative: feelings, opinions Complexity: Simple,

More information

Cactus Tools for Petascale Computing

Cactus Tools for Petascale Computing Cactus Tools for Petascale Computing Erik Schnetter Reno, November 2007 Gamma Ray Bursts ~10 7 km He Protoneutron Star Accretion Collapse to a Black Hole Jet Formation and Sustainment Fe-group nuclei Si

More information

Parallel Asynchronous Hybrid Krylov Methods for Minimization of Energy Consumption. Langshi CHEN 1,2,3 Supervised by Serge PETITON 2

Parallel Asynchronous Hybrid Krylov Methods for Minimization of Energy Consumption. Langshi CHEN 1,2,3 Supervised by Serge PETITON 2 1 / 23 Parallel Asynchronous Hybrid Krylov Methods for Minimization of Energy Consumption Langshi CHEN 1,2,3 Supervised by Serge PETITON 2 Maison de la Simulation Lille 1 University CNRS March 18, 2013

More information

Introduction to magnetic recording + recording materials

Introduction to magnetic recording + recording materials Introduction to magnetic recording + recording materials Laurent Ranno Institut Néel, Nanoscience Dept, CNRS-UJF, Grenoble, France I will give two lectures about magnetic recording. In the first one, I

More information

The Green Index (TGI): A Metric for Evalua:ng Energy Efficiency in HPC Systems

The Green Index (TGI): A Metric for Evalua:ng Energy Efficiency in HPC Systems The Green Index (TGI): A Metric for Evalua:ng Energy Efficiency in HPC Systems Wu Feng and Balaji Subramaniam Metrics for Energy Efficiency Energy- Delay Product (EDP) Used primarily in circuit design

More information

FAS and Solver Performance

FAS and Solver Performance FAS and Solver Performance Matthew Knepley Mathematics and Computer Science Division Argonne National Laboratory Fall AMS Central Section Meeting Chicago, IL Oct 05 06, 2007 M. Knepley (ANL) FAS AMS 07

More information

Lecture 2: Metrics to Evaluate Systems

Lecture 2: Metrics to Evaluate Systems Lecture 2: Metrics to Evaluate Systems Topics: Metrics: power, reliability, cost, benchmark suites, performance equation, summarizing performance with AM, GM, HM Sign up for the class mailing list! Video

More information

IBM Research Report. Performance Metrics for Erasure Codes in Storage Systems

IBM Research Report. Performance Metrics for Erasure Codes in Storage Systems RJ 10321 (A0408-003) August 2, 2004 Computer Science IBM Research Report Performance Metrics for Erasure Codes in Storage Systems James Lee Hafner, Veera Deenadhayalan, Tapas Kanungo, KK Rao IBM Research

More information

ab initio Electronic Structure Calculations

ab initio Electronic Structure Calculations ab initio Electronic Structure Calculations New scalability frontiers using the BG/L Supercomputer C. Bekas, A. Curioni and W. Andreoni IBM, Zurich Research Laboratory Rueschlikon 8803, Switzerland ab

More information

From Physics to Logic

From Physics to Logic From Physics to Logic This course aims to introduce you to the layers of abstraction of modern computer systems. We won t spend much time below the level of bits, bytes, words, and functional units, but

More information

ww.padasalai.net

ww.padasalai.net t w w ADHITHYA TRB- TET COACHING CENTRE KANCHIPURAM SUNDER MATRIC SCHOOL - 9786851468 TEST - 2 COMPUTER SCIENC PG - TRB DATE : 17. 03. 2019 t et t et t t t t UNIT 1 COMPUTER SYSTEM ARCHITECTURE t t t t

More information

Lecture 15: Scaling & Economics

Lecture 15: Scaling & Economics Lecture 15: Scaling & Economics Outline Scaling Transistors Interconnect Future Challenges Economics 2 Moore s Law Recall that Moore s Law has been driving CMOS [Moore65] Corollary: clock speeds have improved

More information

Periodic I/O Scheduling for Supercomputers

Periodic I/O Scheduling for Supercomputers Periodic I/O Scheduling for Supercomputers Guillaume Aupy 1, Ana Gainaru 2, Valentin Le Fèvre 3 1 Inria & U. of Bordeaux 2 Vanderbilt University 3 ENS Lyon & Inria PMBS Workshop, November 217 Slides available

More information

Stochastic Modelling of Electron Transport on different HPC architectures

Stochastic Modelling of Electron Transport on different HPC architectures Stochastic Modelling of Electron Transport on different HPC architectures www.hp-see.eu E. Atanassov, T. Gurov, A. Karaivan ova Institute of Information and Communication Technologies Bulgarian Academy

More information

Knowledge Discovery and Data Mining 1 (VO) ( )

Knowledge Discovery and Data Mining 1 (VO) ( ) Knowledge Discovery and Data Mining 1 (VO) (707.003) Map-Reduce Denis Helic KTI, TU Graz Oct 24, 2013 Denis Helic (KTI, TU Graz) KDDM1 Oct 24, 2013 1 / 82 Big picture: KDDM Probability Theory Linear Algebra

More information

SPECIAL PROJECT PROGRESS REPORT

SPECIAL PROJECT PROGRESS REPORT SPECIAL PROJECT PROGRESS REPORT Progress Reports should be 2 to 10 pages in length, depending on importance of the project. All the following mandatory information needs to be provided. Reporting year

More information

Experience in Factoring Large Integers Using Quadratic Sieve

Experience in Factoring Large Integers Using Quadratic Sieve Experience in Factoring Large Integers Using Quadratic Sieve D. J. Guan Department of Computer Science, National Sun Yat-Sen University, Kaohsiung, Taiwan 80424 guan@cse.nsysu.edu.tw April 19, 2005 Abstract

More information

Parallelization of the Molecular Orbital Program MOS-F

Parallelization of the Molecular Orbital Program MOS-F Parallelization of the Molecular Orbital Program MOS-F Akira Asato, Satoshi Onodera, Yoshie Inada, Elena Akhmatskaya, Ross Nobes, Azuma Matsuura, Atsuya Takahashi November 2003 Fujitsu Laboratories of

More information

Lecture 13: Sequential Circuits, FSM

Lecture 13: Sequential Circuits, FSM Lecture 13: Sequential Circuits, FSM Today s topics: Sequential circuits Finite state machines 1 Clocks A microprocessor is composed of many different circuits that are operating simultaneously if each

More information

GRAPE and Project Milkyway. Jun Makino. University of Tokyo

GRAPE and Project Milkyway. Jun Makino. University of Tokyo GRAPE and Project Milkyway Jun Makino University of Tokyo Talk overview GRAPE Project Science with GRAPEs Next Generation GRAPE the GRAPE-DR Project Milkyway GRAPE project GOAL: Design and build specialized

More information

Measurement & Performance

Measurement & Performance Measurement & Performance Timers Performance measures Time-based metrics Rate-based metrics Benchmarking Amdahl s law Topics 2 Page The Nature of Time real (i.e. wall clock) time = User Time: time spent

More information

Measurement & Performance

Measurement & Performance Measurement & Performance Topics Timers Performance measures Time-based metrics Rate-based metrics Benchmarking Amdahl s law 2 The Nature of Time real (i.e. wall clock) time = User Time: time spent executing

More information

ECMWF Computing & Forecasting System

ECMWF Computing & Forecasting System ECMWF Computing & Forecasting System icas 2015, Annecy, Sept 2015 Isabella Weger, Deputy Director of Computing ECMWF September 17, 2015 October 29, 2014 ATMOSPHERE MONITORING SERVICE CLIMATE CHANGE SERVICE

More information

ECE 571 Advanced Microprocessor-Based Design Lecture 9

ECE 571 Advanced Microprocessor-Based Design Lecture 9 ECE 571 Advanced Microprocessor-Based Design Lecture 9 Vince Weaver http://web.eece.maine.edu/~vweaver vincent.weaver@maine.edu 20 February 2018 Announcements HW#4 was posted. About branch predictors Don

More information

Large-Scale Behavioral Targeting

Large-Scale Behavioral Targeting Large-Scale Behavioral Targeting Ye Chen, Dmitry Pavlov, John Canny ebay, Yandex, UC Berkeley (This work was conducted at Yahoo! Labs.) June 30, 2009 Chen et al. (KDD 09) Large-Scale Behavioral Targeting

More information

TR A Comparison of the Performance of SaP::GPU and Intel s Math Kernel Library (MKL) for Solving Dense Banded Linear Systems

TR A Comparison of the Performance of SaP::GPU and Intel s Math Kernel Library (MKL) for Solving Dense Banded Linear Systems TR-0-07 A Comparison of the Performance of ::GPU and Intel s Math Kernel Library (MKL) for Solving Dense Banded Linear Systems Ang Li, Omkar Deshmukh, Radu Serban, Dan Negrut May, 0 Abstract ::GPU is a

More information

Climate Variability Experiments on Cray XT4 and NERSC. John Dennis and Richard Loft National Center for Atmospheric Research Boulder, Colorado

Climate Variability Experiments on Cray XT4 and NERSC. John Dennis and Richard Loft National Center for Atmospheric Research Boulder, Colorado Optimizing High-Resolution Climate Variability Experiments on Cray XT4 and Cray XT5 Systems at NICS and NERSC John Dennis and Richard Loft National Center for Atmospheric Research Boulder, Colorado August

More information

The QMC Petascale Project

The QMC Petascale Project The QMC Petascale Project Richard G. Hennig What will a petascale computer look like? What are the limitations of current QMC algorithms for petascale computers? How can Quantum Monte Carlo algorithms

More information

Symmetric Pivoting in ScaLAPACK Craig Lucas University of Manchester Cray User Group 8 May 2006, Lugano

Symmetric Pivoting in ScaLAPACK Craig Lucas University of Manchester Cray User Group 8 May 2006, Lugano Symmetric Pivoting in ScaLAPACK Craig Lucas University of Manchester Cray User Group 8 May 2006, Lugano Introduction Introduction We wanted to parallelize a serial algorithm for the pivoted Cholesky factorization

More information

An Overview of HPC at the Met Office

An Overview of HPC at the Met Office An Overview of HPC at the Met Office Paul Selwood Crown copyright 2006 Page 1 Introduction The Met Office National Weather Service for the UK Climate Prediction (Hadley Centre) Operational and Research

More information

Error Detection, Correction and Erasure Codes for Implementation in a Cluster File-system

Error Detection, Correction and Erasure Codes for Implementation in a Cluster File-system Error Detection, Correction and Erasure Codes for Implementation in a Cluster File-system Steve Baker December 6, 2011 Abstract. The evaluation of various error detection and correction algorithms and

More information

A Data Communication Reliability and Trustability Study for Cluster Computing

A Data Communication Reliability and Trustability Study for Cluster Computing A Data Communication Reliability and Trustability Study for Cluster Computing Speaker: Eduardo Colmenares Midwestern State University Wichita Falls, TX HPC Introduction Relevant to a variety of sciences,

More information

Sparse BLAS-3 Reduction

Sparse BLAS-3 Reduction Sparse BLAS-3 Reduction to Banded Upper Triangular (Spar3Bnd) Gary Howell, HPC/OIT NC State University gary howell@ncsu.edu Sparse BLAS-3 Reduction p.1/27 Acknowledgements James Demmel, Gene Golub, Franc

More information

Analysis of the Weather Research and Forecasting (WRF) Model on Large-Scale Systems

Analysis of the Weather Research and Forecasting (WRF) Model on Large-Scale Systems John von Neumann Institute for Computing Analysis of the Weather Research and Forecasting (WRF) Model on Large-Scale Systems Darren J. Kerbyson, Kevin J. Barker, Kei Davis published in Parallel Computing:

More information

80386DX. 32-Bit Microprocessor FEATURES: DESCRIPTION: Logic Diagram

80386DX. 32-Bit Microprocessor FEATURES: DESCRIPTION: Logic Diagram 32-Bit Microprocessor 21 1 22 1 2 10 3 103 FEATURES: 32-Bit microprocessor RAD-PAK radiation-hardened agait natural space radiation Total dose hardness: - >100 Krad (Si), dependent upon space mission Single

More information

Large-scale Electronic Structure Simulations with MVAPICH2 on Intel Knights Landing Manycore Processors

Large-scale Electronic Structure Simulations with MVAPICH2 on Intel Knights Landing Manycore Processors Large-scale Electronic Structure Simulations with MVAPICH2 on Intel Knights Landing Manycore Processors Hoon Ryu, Ph.D. (E: elec1020@kisti.re.kr) Principal Researcher / Korea Institute of Science and Technology

More information

Organisasi dan Arsitektur Komputer L#1: Fundamental Concepts Amil A. Ilham

Organisasi dan Arsitektur Komputer   L#1: Fundamental Concepts Amil A. Ilham Organisasi dan Arsitektur Komputer http://www.unhas.ac.id/amil/stmik2016/arsikom/ L#1: Fundamental Concepts Amil A. Ilham http://www.unhas.ac.id/amil Administrasi Kuliah ADMINISTRASI KULIAH 2 Penilaian

More information

An Algorithm for a Two-Disk Fault-Tolerant Array with (Prime 1) Disks

An Algorithm for a Two-Disk Fault-Tolerant Array with (Prime 1) Disks An Algorithm for a Two-Disk Fault-Tolerant Array with (Prime 1) Disks Sanjeeb Nanda and Narsingh Deo School of Computer Science University of Central Florida Orlando, Florida 32816-2362 sanjeeb@earthlink.net,

More information

Using R for Iterative and Incremental Processing

Using R for Iterative and Incremental Processing Using R for Iterative and Incremental Processing Shivaram Venkataraman, Indrajit Roy, Alvin AuYoung, Robert Schreiber UC Berkeley and HP Labs UC BERKELEY Big Data, Complex Algorithms PageRank (Dominant

More information

GPU-accelerated Computing at Scale. Dirk Pleiter I GTC Europe 10 October 2018

GPU-accelerated Computing at Scale. Dirk Pleiter I GTC Europe 10 October 2018 GPU-accelerated Computing at Scale irk Pleiter I GTC Europe 10 October 2018 Outline Supercomputers at JSC Future science challenges Outlook and conclusions 2 3 Supercomputers at JSC JUQUEEN (until 2018)

More information

AstroPortal: A Science Gateway for Large-scale Astronomy Data Analysis

AstroPortal: A Science Gateway for Large-scale Astronomy Data Analysis AstroPortal: A Science Gateway for Large-scale Astronomy Data Analysis Ioan Raicu Distributed Systems Laboratory Computer Science Department University of Chicago Joint work with: Ian Foster: Univ. of

More information

Lecture 3, Performance

Lecture 3, Performance Repeating some definitions: Lecture 3, Performance CPI MHz MIPS MOPS Clocks Per Instruction megahertz, millions of cycles per second Millions of Instructions Per Second = MHz / CPI Millions of Operations

More information

EECS150 - Digital Design Lecture 27 - misc2

EECS150 - Digital Design Lecture 27 - misc2 EECS150 - Digital Design Lecture 27 - misc2 May 1, 2002 John Wawrzynek Spring 2002 EECS150 - Lec27-misc2 Page 1 Outline Linear Feedback Shift Registers Theory and practice Simple hardware division algorithms

More information

Lecture 3, Performance

Lecture 3, Performance Lecture 3, Performance Repeating some definitions: CPI Clocks Per Instruction MHz megahertz, millions of cycles per second MIPS Millions of Instructions Per Second = MHz / CPI MOPS Millions of Operations

More information

AstroPortal: A Science Gateway for Large-scale Astronomy Data Analysis

AstroPortal: A Science Gateway for Large-scale Astronomy Data Analysis AstroPortal: A Science Gateway for Large-scale Astronomy Data Analysis Ioan Raicu Distributed Systems Laboratory Computer Science Department University of Chicago Joint work with: Ian Foster: Univ. of

More information

Performance Metrics for Computer Systems. CASS 2018 Lavanya Ramapantulu

Performance Metrics for Computer Systems. CASS 2018 Lavanya Ramapantulu Performance Metrics for Computer Systems CASS 2018 Lavanya Ramapantulu Eight Great Ideas in Computer Architecture Design for Moore s Law Use abstraction to simplify design Make the common case fast Performance

More information

2. Accelerated Computations

2. Accelerated Computations 2. Accelerated Computations 2.1. Bent Function Enumeration by a Circular Pipeline Implemented on an FPGA Stuart W. Schneider Jon T. Butler 2.1.1. Background A naive approach to encoding a plaintext message

More information

Quantum Computation. Dr Austin Fowler Centre for Quantum Computer Technology. New Scientist, 10/11/07

Quantum Computation. Dr Austin Fowler Centre for Quantum Computer Technology. New Scientist, 10/11/07 Quantum Computation Dr Austin Fowler Centre for Quantum Computer Technology New Scientist, 10/11/07 Overview what is a quantum computer? bits vs qubits superpositions and measurement implementations why

More information

METHOD OF WCS CLIENT BASED ON PYRAMID MODEL

METHOD OF WCS CLIENT BASED ON PYRAMID MODEL METHOD OF WCS CLIENT BASED ON PYRAMID MODEL Shen Shengyu, Wu Huayi* State Key Laboratory of Information Engineering in Surveying, Mapping and Remote Sensing, Wuhan University, 430079 Wuhan, China - shshy.whu@gmail.com,

More information

Fundamentals of Computational Science

Fundamentals of Computational Science Fundamentals of Computational Science Dr. Hyrum D. Carroll August 23, 2016 Introductions Each student: Name Undergraduate school & major Masters & major Previous research (if any) Why Computational Science

More information

XXL-BIOMD. Large Scale Biomolecular Dynamics Simulations. onsdag, 2009 maj 13

XXL-BIOMD. Large Scale Biomolecular Dynamics Simulations. onsdag, 2009 maj 13 XXL-BIOMD Large Scale Biomolecular Dynamics Simulations David van der Spoel, PI Aatto Laaksonen Peter Coveney Siewert-Jan Marrink Mikael Peräkylä Uppsala, Sweden Stockholm, Sweden London, UK Groningen,

More information

Leveraging Task-Parallelism in Energy-Efficient ILU Preconditioners

Leveraging Task-Parallelism in Energy-Efficient ILU Preconditioners Leveraging Task-Parallelism in Energy-Efficient ILU Preconditioners José I. Aliaga Leveraging task-parallelism in energy-efficient ILU preconditioners Universidad Jaime I (Castellón, Spain) José I. Aliaga

More information

EECS150 - Digital Design Lecture 23 - FFs revisited, FIFOs, ECCs, LSFRs. Cross-coupled NOR gates

EECS150 - Digital Design Lecture 23 - FFs revisited, FIFOs, ECCs, LSFRs. Cross-coupled NOR gates EECS150 - Digital Design Lecture 23 - FFs revisited, FIFOs, ECCs, LSFRs April 16, 2009 John Wawrzynek Spring 2009 EECS150 - Lec24-blocks Page 1 Cross-coupled NOR gates remember, If both R=0 & S=0, then

More information

Scheduling I. Today. Next Time. ! Introduction to scheduling! Classical algorithms. ! Advanced topics on scheduling

Scheduling I. Today. Next Time. ! Introduction to scheduling! Classical algorithms. ! Advanced topics on scheduling Scheduling I Today! Introduction to scheduling! Classical algorithms Next Time! Advanced topics on scheduling Scheduling out there! You are the manager of a supermarket (ok, things don t always turn out

More information

Scalable and Power-Efficient Data Mining Kernels

Scalable and Power-Efficient Data Mining Kernels Scalable and Power-Efficient Data Mining Kernels Alok Choudhary, John G. Searle Professor Dept. of Electrical Engineering and Computer Science and Professor, Kellogg School of Management Director of the

More information

IBM Research Report. Notes on Reliability Models for Non-MDS Erasure Codes

IBM Research Report. Notes on Reliability Models for Non-MDS Erasure Codes RJ10391 (A0610-035) October 24, 2006 Computer Science IBM Research Report Notes on Reliability Models for Non-MDS Erasure Codes James Lee Hafner, KK Rao IBM Research Division Almaden Research Center 650

More information

Reflecting on the Goal and Baseline of Exascale Computing

Reflecting on the Goal and Baseline of Exascale Computing Reflecting on the Goal and Baseline of Exascale Computing Thomas C. Schulthess!1 Tracking supercomputer performance over time? Linpack benchmark solves: Ax = b!2 Tracking supercomputer performance over

More information

Performance and Application of Observation Sensitivity to Global Forecasts on the KMA Cray XE6

Performance and Application of Observation Sensitivity to Global Forecasts on the KMA Cray XE6 Performance and Application of Observation Sensitivity to Global Forecasts on the KMA Cray XE6 Sangwon Joo, Yoonjae Kim, Hyuncheol Shin, Eunhee Lee, Eunjung Kim (Korea Meteorological Administration) Tae-Hun

More information

15-451/651: Design & Analysis of Algorithms September 13, 2018 Lecture #6: Streaming Algorithms last changed: August 30, 2018

15-451/651: Design & Analysis of Algorithms September 13, 2018 Lecture #6: Streaming Algorithms last changed: August 30, 2018 15-451/651: Design & Analysis of Algorithms September 13, 2018 Lecture #6: Streaming Algorithms last changed: August 30, 2018 Today we ll talk about a topic that is both very old (as far as computer science

More information

Improvements for Implicit Linear Equation Solvers

Improvements for Implicit Linear Equation Solvers Improvements for Implicit Linear Equation Solvers Roger Grimes, Bob Lucas, Clement Weisbecker Livermore Software Technology Corporation Abstract Solving large sparse linear systems of equations is often

More information

ECEN 248: INTRODUCTION TO DIGITAL SYSTEMS DESIGN. Week 9 Dr. Srinivas Shakkottai Dept. of Electrical and Computer Engineering

ECEN 248: INTRODUCTION TO DIGITAL SYSTEMS DESIGN. Week 9 Dr. Srinivas Shakkottai Dept. of Electrical and Computer Engineering ECEN 248: INTRODUCTION TO DIGITAL SYSTEMS DESIGN Week 9 Dr. Srinivas Shakkottai Dept. of Electrical and Computer Engineering TIMING ANALYSIS Overview Circuits do not respond instantaneously to input changes

More information

ALU A functional unit

ALU A functional unit ALU A functional unit that performs arithmetic operations such as ADD, SUB, MPY logical operations such as AND, OR, XOR, NOT on given data types: 8-,16-,32-, or 64-bit values A n-1 A n-2... A 1 A 0 B n-1

More information

I/O Devices. Device. Lecture Notes Week 8

I/O Devices. Device. Lecture Notes Week 8 I/O Devices CPU PC ALU System bus Memory bus Bus interface I/O bridge Main memory USB Graphics adapter I/O bus Disk other devices such as network adapters Mouse Keyboard Disk hello executable stored on

More information

Performance of the fusion code GYRO on three four generations of Crays. Mark Fahey University of Tennessee, Knoxville

Performance of the fusion code GYRO on three four generations of Crays. Mark Fahey University of Tennessee, Knoxville Performance of the fusion code GYRO on three four generations of Crays Mark Fahey mfahey@utk.edu University of Tennessee, Knoxville Contents Introduction GYRO Overview Benchmark Problem Test Platforms

More information