An Overview of HPC at the Met Office

Similar documents
ECMWF Scalability Programme

Deutscher Wetterdienst

Scalability Programme at ECMWF

MSC HPC Infrastructure Update. Alain St-Denis Canadian Meteorological Centre Meteorological Service of Canada

History of the partnership between SMHI and NSC. Per Undén

Medium-range Ensemble Forecasts at the Met Office

The next-generation supercomputer and NWP system of the JMA

Progress in NWP on Intel HPC architecture at Australian Bureau of Meteorology

Nowcasting for the London Olympics 2012 Brian Golding, Susan Ballard, Nigel Roberts & Ken Mylne Met Office, UK. Crown copyright Met Office

Exploiting ensemble members: forecaster-driven EPS applications at the met office

Performance of WRF using UPC

Performance and Application of Observation Sensitivity to Global Forecasts on the KMA Cray XE6

HYCOM and Navy ESPC Future High Performance Computing Needs. Alan J. Wallcraft. COAPS Short Seminar November 6, 2017

Operational and research activities at ECMWF now and in the future

ACCESS AGREPS Ensemble Prediction System

ECMWF Computing & Forecasting System

Performance of Met Office Weather and Climate Codes on Cavium ThunderX2 Processors. Adam Voysey, Maff Glover HPC Optimisation Team

Model error and parameter estimation

NCAS Unified Model Introduction. Part I: Overview of the UM system

The Performance Evolution of the Parallel Ocean Program on the Cray X1

Porting and optimisation of the Met Office Unified Model on Petascale architectures

Performance Analysis of Lattice QCD Application with APGAS Programming Model

Scaling the Software and Advancing the Science of Global Modeling and Assimilation Systems at NASA. Bill Putman

Exascale I/O challenges for Numerical Weather Prediction

Center Report from KMA

The Met Office climate model HadSM3 and climateprediction.net

Introduction of a Stabilized Bi-Conjugate Gradient iterative solver for Helmholtz s Equation on the CMA GRAPES Global and Regional models.

Predicting rainfall using ensemble forecasts

4DEnVar. Four-Dimensional Ensemble-Variational Data Assimilation. Colloque National sur l'assimilation de données

Performance Evaluation of Scientific Applications on POWER8

JOINT WMO TECHNICAL PROGRESS REPORT ON THE GLOBAL DATA PROCESSING AND FORECASTING SYSTEM AND NUMERICAL WEATHER PREDICTION RESEARCH ACTIVITIES FOR 2007

Swedish Meteorological and Hydrological Institute

Convective-scale NWP for Singapore

Using Weather Pattern Analysis to Identify Periods of Heightened Coastal Flood Risk in the Medium to Long Range

Convective Scale Ensemble for NWP

Exascale challenges for Numerical Weather Prediction : the ESCAPE project

Essentials of Large Volume Data Management - from Practical Experience. George Purvis MASS Data Manager Met Office

SPECIAL PROJECT PROGRESS REPORT

A diagnostic interface for ICON Coping with the challenges of high-resolution climate simulations

A. Doerenbecher 1, M. Leutbecher 2, D. S. Richardson 3 & collaboration with G. N. Petersen 4

TIGGE-LAM archive development in the frame of GEOWOW. Richard Mladek (ECMWF)

Progress in Numerical Methods at ECMWF

International Civil Aviation Organization. Seventh Meeting of CNS/MET Sub-Group of APANPIRG and Tenth Meeting of CNS/ATM IC Sub-Group of APANPIRG

SLOVAKIA. Slovak Hydrometeorological Institute

The Panel: What does the future look like for NPW application development? 17 th ECMWF Workshop on High Performance Computing in Meteorology

Quarterly numerical weather prediction model performance summaries April to June 2010 and July to September 2010

Crossing the Chasm. On the Paths to Exascale: Presented by Mike Rezny, Monash University, Australia

Deutscher Wetterdienst

Application and verification of ECMWF products 2012

State of the art of wind forecasting and planned improvements for NWP Helmut Frank (DWD), Malte Mülller (met.no), Clive Wilson (UKMO)

ECMWF global reanalyses: Resources for the wind energy community

Skilful seasonal predictions for the European Energy Industry

MOGREPS short-range ensemble forecasting and the PREVIEW Windstorms Project

Improvements to the NCEP Global and Regional Data Assimilation Systems

Current and future configurations of MOGREPS-UK. Susanna Hagelin EWGLAM/SRNWP, Rome, 4 Oct 2016

Recent advances in Tropical Cyclone prediction using ensembles

LATE REQUEST FOR A SPECIAL PROJECT

Piz Daint & Piz Kesch : from general purpose supercomputing to an appliance for weather forecasting. Thomas C. Schulthess

JOINT WMO TECHNICAL PROGRESS REPORT ON THE GLOBAL DATA PROCESSING AND FORECASTING SYSTEM AND NUMERICAL WEATHER PREDICTION RESEARCH ACTIVITIES FOR 2006

Model enhancement & delivery plans, RAI

ECMWF 10 th workshop on Meteorological Operational Systems

WRF Modeling System Overview

NCEP Global Ensemble Forecast System (GEFS) Yuejian Zhu Ensemble Team Leader Environmental Modeling Center NCEP/NWS/NOAA February

Ensemble of Data Assimilations methods for the initialization of EPS

Current Limited Area Applications

Recent activities related to EPS (operational aspects)

Does the ATOVS RARS Network Matter for Global NWP? Brett Candy, Nigel Atkinson & Stephen English

ESM development at the Met Office Hadley Centre

The ENSEMBLES Project

Probabilistic Weather Prediction

Regional Atmosphere. Developing a unified science configuration for Convection- Permitting Climate and NWP simulations

ECMWF Forecasting System Research and Development

Thirteenth Workshop on Use of High Performance Computing in Meteorology 3 7 November Programme

Weather and Climate Modeling on GPU and Xeon Phi Accelerated Systems

Relative Merits of 4D-Var and Ensemble Kalman Filter

Hellenic National Meteorological Service (HNMS) GREECE

Optimization strategy for MASNUM surface wave model

Computational Challenges in Big Data Assimilation with Extreme-scale Simulations

Workstations at Met Éireann. Kieran Commins Head Applications Development

NCEP Applications -- HPC Performance and Strategies. Mark Iredell software team lead USDOC/NOAA/NWS/NCEP/EMC

JOINT WMO TECHNICAL PROGRESS REPORT ON THE GLOBAL DATA PROCESSING AND FORECASTING SYSTEM AND NUMERICAL WEATHER PREDICTION RESEARCH ACTIVITIES FOR 2007

The benefits and developments in ensemble wind forecasting

Probabilistic Weather Forecasting and the EPS at ECMWF

Application and verification of ECMWF products 2009

SPECIAL PROJECT PROGRESS REPORT

WRF Modeling System Overview

Performance Evaluation of MPI on Weather and Hydrological Models

Building Ensemble-Based Data Assimilation Systems. for High-Dimensional Models

Validation of operational NWP forecasts for global, diffuse and direct solar exposure over Australia

11A.3 The Impact on Tropical Cyclone Predictions of a Major Upgrade to the Met Office Global Model

Current Issues and Challenges in Ensemble Forecasting

Visualising and communicating probabilistic flow forecasts in The Netherlands

JOINT WMO TECHNICAL PROGRESS REPORT ON THE GLOBAL DATA PROCESSING AND FORECASTING SYSTEM AND NUMERICAL WEATHER PREDICTION RESEARCH ACTIVITIES FOR 2016

Five years of limited-area ensemble activities at ARPA-SIM: the COSMO-LEPS system

MACSSIMIZE. Measurements of Arctic Clouds, Snow, and Sea Ice nearby the Marginal Ice ZonE. Principal investigator. Chawn Harlow

Model Error Diagnosis across Timescales

Recent Developments of JMA Operational NWP Systems and WGNE Intercomparison of Tropical Cyclone Track Forecast

M.Sc. in Meteorology. Numerical Weather Prediction

ASCAT NRT Data Processing and Distribution at NOAA/NESDIS

Regional Production, Quarterly report on the daily analyses and forecasts activities, and verification of the CHIMERE performances

Transcription:

An Overview of HPC at the Met Office Paul Selwood Crown copyright 2006 Page 1

Introduction

The Met Office National Weather Service for the UK Climate Prediction (Hadley Centre) Operational and Research activities Headquarters relocated to Exeter from Bracknell 2003/4 150 th Anniversary in 2004 Crown copyright 2006 Page 3

Met Office Supercomputers NWP 19 node SX-6 Climate 15 node SX-6 January 2005 16 node SX-8 installed. 8 CPU/node 64GB FCRAM/node September 2006 4 additional SX-8 nodes for Climate November 2006 1 additional SX-8 node due for NWP Crown copyright 2006 Page 4

SX-8; differences from SX-6 Passed acceptance and reliability tests at first attempt. Easiest supercomputer install we ve had! Code compiled for SX-6 runs well on SX-8. To allow operational backup on SX-6. SX-8 specific code (sqrt vector pipe) gives little improvement SX-8 : SX-6 ratio ~2.1 for computation. Bank caching on SX-8 gives much better lookup table performance. Need to double size of cache for local I/O, else saturation gives variable performance. Crown copyright 2006 Page 5

The Unified Model

The Unified Model Climate and Forecast model Atmosphere, Ocean and Coupled (also seaice, atmospheric chemistry, aerosols, river transport, ) Atmosphere Non-hydrostatic, semi-lagrangian, semi-implicit, Arakawa C grid, Charney-Phillips vertical coordinate ~ 700K LOC MPI parallelisation Crown copyright 2006 Page 7

Deterministic Forecasts 12km 38 levels Up to 48hr f/c 6-hourly update 3 SX-8 nodes 4D-Var 4km Up to 36hr f/c 1 SX-8 node 3D-Var 40km - 50 levels Up to 7 day f/c 6-hourly update 3 SX-8 nodes 4D-Var Crown copyright 2006 Page 8

MOGREPS 24 members Run on 3 nodes of SX-6 Global Run to T+72 N144 (~ 90 km) Uses Ensemble Transform Kalman Filter (ETKF) for generating initial perturbations Stochastic physics random perturbation of parameterisation schemes LAM Run to T+36 24 km North-Atlantic Europe Takes initial and boundary conditions from global model Stochastic physics Crown copyright 2006 Page 9

THORPEX - TIGGE The Met Office medium-range ensemble forecast system is running on ECMWF hpcd (soon hpce). Based on short-range MOGREPS-Global system, extended to 15 days. Resolution N144 (0.833 x 1.25 ), 38 levels 24 members (control + 23 perturbed), run twice a day (0 and 12 UTC). Initial data, with ETKF perturbations, created at Met Office and copied to ECMWF. Crown copyright 2006 Page 10

Unified Model Performance

I/O the problem Unified Model I/O initially very slow Route to GFS disk depends on packet size < 64KB nfs (slow) >= 64KB GFS (fast) Frequency 14000 12000 10000 8000 6000 4000 Application buffering improved I/O rates from 40 MB/s to 140MB/s 2000 0 0 1 2 4 8 16 32 64 KBytes 128 256 512 1024 More Crown copyright 2006 Page 12

I/O Local disk Local disks have cache, GFS doesn t Application can see > 1 GB/s transfer rate Only used for operational work Only enough disk space for certain output streams Needs careful data management Crown copyright 2006 Page 13

Local I/O Unified Model File 2 File 1 Post Processing File 3 Format Conversion Transfer To IBM Local Disk Multiple Files Unified Model File 1 Post Processing Format Conversion Transfer To IBM Crown copyright 2006 Page 14

I/O Server processes Unified Model typically only does output on certain timesteps I/O Server process can process the output asynchronously Initial work on NEC had little benefit 1% improvement for 15% cost Small numbers of CPUs Bob Carruthers (IBM) extended code for multiple server processes and improved scheduling 8% improvement for 3% cost Need to re-evaluate on NEC Crown copyright 2006 Page 15

I/O in Scripting Analysis of full operational suite showed inefficiencies in scripting I/O cp on SX node has poor blocksize dd allows user control of blocksize Typically save 60-70% of cost (30s per file) Multiple appends to a file taking 2-4 minutes Buffer via perl array with a single write takes cost to 2-4 seconds. Crown copyright 2006 Page 16

Communications improvements made T3E coding practices Unnecessary barriers removed Naïve SHMEM GCOM conversion improved (1326s improved to 25s overloaded MPI buffers?) Gathering/Scattering 3D fields level-by-level Optimised by copying into temporary buffers and doing one communication per CPU pair Halves cost of these communications >6000 halo exchanges in 6 hour forecast 1500 & many other communications removed from a single diagnostic calculation! Amalgamating communications only minor benefit Crown copyright 2006 Page 17

40km Global Model 40km Operational Global Model 640 x 481 x 50 7 day forecast in ~45 minutes on 3 SX-8 nodes 4.5 4 N320L50 Scalability Scaling relative to 8 CPUs 3.5 3 2.5 2 1.5 1 0.5 N320L50 Perfect (wrt CPUs) Perfect (wrt nodes) 0 8 12 16 20 24 28 CPUs Crown copyright 2006 Page 18

North Atlantic/Europe Model Scalability 1 day forecast 7 cpu/node Limited output Scaling 18 16 14 12 10 8 6 4 2 0 0 5 10 15 20 Nodes Crown copyright 2006 Page 19

UM TIGGE on IBM - optimisation Port to IBM from NEC straightforward Tuning of physics segmentation (like NPROMA) Paul Burton & Deborah Salmond made improvements to communications. John Hague (IBM) implemented OpenMP for 90% of runtime. Crown copyright 2006 Page 20

UM TIGGE on IBM - timings 32 64 128 hpcd 265 125 67 hpce 129 70 49 hpce (2 thread) 133 70 39 Crown copyright 2006 Page 21

Procurement and RAPS

HPC Procurement timeline Business case for new supercomputer and mass-storage accepted Possible partnership with UK academia Expected tender July 2007 Award contract late Summer 2008 Delivery of first hardware Winter 2008/9 Acceptance Spring 2009 Crown copyright 2006 Page 23

RAPS First Met Office RAPS release Initially a Unified Model benchmark N512L76 main resolution 1 day forecast, with or without I/O N320L50 and N48L38 supporting resolutions Available now standard Met Office benchmarking licence Plans for a Data Assimilation benchmark in Spring 2007 Observation processing 4D-Var Crown copyright 2006 Page 24

RAPS performance on NEC SX-8 2, 4, 5 nodes run N512L76 noio case CPUs Time 7 cpu/node 14 7140 ~70 GB memory required 28 4200 35 3420 Crown copyright 2006 Page 25

Questions?