MSC HPC Infrastructure Update. Alain St-Denis Canadian Meteorological Centre Meteorological Service of Canada

Similar documents
Ninth Workshop on Meteorological Operational Systems. Timeliness and Impact of Observations in the CMC Global NWP system

An Overview of HPC at the Met Office

JOINT WMO TECHNICAL PROGRESS REPORT ON THE GLOBAL DATA PROCESSING AND FORECASTING SYSTEM AND NUMERICAL WEATHER PREDICTION RESEARCH ACTIVITIES FOR 2007

JOINT WMO TECHNICAL PROGRESS REPORT ON THE GLOBAL DATA PROCESSING AND FORECASTING SYSTEM AND NUMERICAL WEATHER PREDICTION RESEARCH ACTIVITIES FOR 2006

Deutscher Wetterdienst

The next-generation supercomputer and NWP system of the JMA

ECMWF Computing & Forecasting System

JOINT WMO TECHNICAL PROGRESS REPORT ON THE GLOBAL DATA PROCESSING AND FORECASTING SYSTEM AND NUMERICAL WEATHER PREDICTION RESEARCH ACTIVITIES FOR 2006

Data Impact Studies in the CMC Global NWP system

Computational Challenges in Big Data Assimilation with Extreme-scale Simulations

Impact Evaluation of New Radiance data, Reduced Thinning and Higher Analysis Resolution in the GEM Global Deterministic Prediction System

Performance and Application of Observation Sensitivity to Global Forecasts on the KMA Cray XE6

Numerical Weather prediction at the European Centre for Medium-Range Weather Forecasts

Center Report from KMA

Hellenic National Meteorological Service (HNMS) GREECE

The Canadian Land Data Assimilation System (CaLDAS)

JOINT WMO TECHNICAL PROGRESS REPORT ON THE GLOBAL DATA PROCESSING AND FORECASTING SYSTEM AND NUMERICAL WEATHER PREDICTION RESEARCH ACTIVITIES FOR 2007

Scalability Programme at ECMWF

Recent experience at Météo-France on the assimilation of observations at high temporal frequency Cliquez pour modifier le style du titre

Weather Forecasting. March 26, 2009

Pierre Gauthier Department of Earth and Atmospheric Sciences Université du Québec à Montréal (UQAM)

Progress in NWP on Intel HPC architecture at Australian Bureau of Meteorology

HYCOM and Navy ESPC Future High Performance Computing Needs. Alan J. Wallcraft. COAPS Short Seminar November 6, 2017

ASSIMILATION OF ATOVS RETRIEVALS AND AMSU-A RADIANCES AT THE ITALIAN WEATHER SERVICE: CURRENT STATUS AND PERSPECTIVES

Status and Plans of Next Generation Japanese Geostationary Meteorological Satellites Himawari 8/9

LETKF Data Assimilation System for KIAPS AGCM: Progress and Plan

Recent Data Assimilation Activities at Environment Canada

Direct Readout at the Meteorological Service of Canada

Himawari 8/9 data distribution/dissemination plan

Swedish Meteorological and Hydrological Institute

SLOVAKIA. Slovak Hydrometeorological Institute

OSSE to infer the impact of Arctic AMVs extracted from highly elliptical orbit imagery

Chile / Dirección Meteorológica de Chile (Chilean Weather Service)

WRF Modeling System Overview

Development of the Canadian Precipitation Analysis (CaPA) and the Canadian Land Data Assimilation System (CaLDAS)

WRF Modeling System Overview

The satellite winds in the operational NWP system at Météo-France

PI SERVER 2012 Do. More. Faster. Now! Copyr i g h t 2012 O S Is o f t, L L C. 1

ECMWF Scalability Programme

CERA-SAT: A coupled reanalysis at higher resolution (WP1)

Exascale I/O challenges for Numerical Weather Prediction

Near-real-time satellite data processing at NIWA with Cylc

Introduction to Data Assimilation. Saroja Polavarapu Meteorological Service of Canada University of Toronto

WRF Model Simulated Proxy Datasets Used for GOES-R Research Activities

Weather Research and Forecasting (WRF) Performance Benchmark and Profiling. July 2012

1. Current atmospheric DA systems 2. Coupling surface/atmospheric DA 3. Trends & ideas

Current Limited Area Applications

HY-2A Satellite User s Guide

Environment Canada s Regional Ensemble Kalman Filter

The Changing Landscape of Land Administration

JOINT WMO TECHNICAL PROGRESS REPORT ON THE GLOBAL DATA PROCESSING AND FORECASTING SYSTEM AND NUMERICAL WEATHER PREDICTION RESEARCH ACTIVITIES FOR 2016

Meteorological Service of Canada Perspectives. WMO Coordination Group on Satellite Data Requirements for RAIII/IV

DEVELOPMENT OF THE ECMWF FORECASTING SYSTEM

Synergie PC Updates and Plans. A. Lasserre-Bigorry, B. Benech M-F Voidrot, & R. Giraud

NCEP Applications -- HPC Performance and Strategies. Mark Iredell software team lead USDOC/NOAA/NWS/NCEP/EMC

Convective-scale NWP for Singapore

O.M Smedstad 1, E.J. Metzger 2, R.A. Allard 2, R. Broome 1, D.S. Franklin 1 and A.J. Wallcraft 2. QinetiQ North America 2. Naval Research Laboratory

Scaling the Software and Advancing the Science of Global Modeling and Assimilation Systems at NASA. Bill Putman

AMVs in the operational ECMWF system

History of the partnership between SMHI and NSC. Per Undén

Operational Use of Scatterometer Winds in the JMA Data Assimilation System

Deutscher Wetterdienst

Performance of WRF using UPC

Global and Regional OSEs at JMA

Enhancing information transfer from observations to unobserved state variables for mesoscale radar data assimilation

NWC-SAF Satellite Application Facility in Support to Nowcasting and Very Short Range Forecasting

ERA-CLIM: Developing reanalyses of the coupled climate system

LATE REQUEST FOR A SPECIAL PROJECT

THE WEATHER RESEARCH AND FORECAST MODEL VERSION 2.0

ERA5 and the use of ERA data

AMVs in the ECMWF system:

Radar data assimilation using a modular programming approach with the Ensemble Kalman Filter: preliminary results

Inter-comparison of 4D-Var and EnKF systems for operational deterministic NWP

The National Operational Hydrologic Remote Sensing Center Operational Snow Analysis

EUMETSAT SAF NETWORK. Lothar Schüller, EUMETSAT SAF Network Manager

Application and verification of ECMWF products 2012

Operational and research activities at ECMWF now and in the future

FUTURE PLAN AND RECENT ACTIVITIES FOR THE JAPANESE FOLLOW-ON GEOSTATIONARY METEOROLOGICAL SATELLITE HIMAWARI-8/9

REQUIREMENTS FOR WEATHER RADAR DATA. Review of the current and likely future hydrological requirements for Weather Radar data

North Atlantic Coast Comprehensive Study Storm Simulation and Statistical Analysis Part II Production System

Bias correction of satellite data at Météo-France

Data Short description Parameters to be used for analysis SYNOP. Surface observations by ships, oil rigs and moored buoys

Environment Canada s Regional Ensemble Kalman Filter

Exploiting ensemble members: forecaster-driven EPS applications at the met office

Nowcasting and Urban Interactive Modeling Using Robotic and Remotely Sensed Data James Cogan, Robert Dumais, and Yansen Wang

Strategic Radar Enhancement Project (SREP) Forecast Demonstration Project (FDP) The future is here and now

onboard of Metop-A COSMIC Workshop 2009 Boulder, USA

Compounding insights Thermo Scientific Compound Discoverer Software

THE USE OF SATELLITE OBSERVATIONS IN NWP

Advanced Computing Systems for Scientific Research

The National Weather Service of the Future: Building a Weather-Ready Nation

Orchestrating a Climate Modeling Pipeline. using Python and RAM Disk

WMO Coordination Group on Satellite Data Requirements for Region III and IV Sept 5-8, 2016 Willemstad, Curaçao

Using Aziz Supercomputer

EUMETSAT SAF NETWORK. Lothar Schüller, EUMETSAT SAF Network Manager

The NOAA/NESDIS/STAR IASI Near Real-Time Product Processing and Distribution System

Satellite data assimilation for Numerical Weather Prediction II

Introduction to Data Assimilation

Observation usage in Météo-France data assimilation systems

TCC News 1 No. 29 Summer 2012

Transcription:

MSC HPC Infrastructure Update Alain St-Denis Canadian Meteorological Centre Meteorological Service of Canada

Outline HPC Infrastructure Overview Supercomputer Configuration Scientific Direction 2

IT Infrastructure Overview

4 NEC SX-532M2 NEC SX-4/80M3 NEC SX-4/16 NEC SX-3/44R Cray 1S CDC 176 Cray XMP 416 CDC 7600 NEC SX-3/44 NEC SX-6/80M10 1 10 100 1000 10000 100000 1000000 10000000 1974 1976 1978 1980 1982 1984 1986 1988 1990 1992 1994 1996 1998 2000 2002 2004 MFLOPFs Cray XMP 28 IBM P690 MSC s Supercomputing History

5

Supercomputer: IBM IBM 128 nodes 944 PE 2.19 TB Memory 15 TB Disks 4.825 Tflops (peak) 6

IBM Committed Sustained Performance in MSC TFlops 6 5.89251 Sustained TFlops 5 4 3 2 1 1.042 2.47996 IBM 0 0 2.5 5 Initial, Upgrade & Optional Upgrade 7

Front-End Servers Twin Servers: Production & R&D systems Production Server & R&D: SGI 3000: 20 R14000 600MHz CPUs, 20GB RAM Together both systems have ~13TB disk Production Server: 80% Batch 8

Front-Ends: Linux Cluster Configuration: 17 nodes with a 3TB (raw) SAN Each node: 2 processors, P3-1.26GHz / P4 2.4GHz 2 GB memory 1 FC HBA 1 GE Compute nodes batch-only Linux Clusters are growing like weed 9

Automated Archiving System SGI Origin 300 with 8 processors & 8 GB memory 1 TB of FC disk drives 144 TB Grau AML/E robot with 4 AMPEX DST-312 70 TB Scalar 10K with 10 LTO tape drives HSM is FileServ (ADIC) + home grown backup management software 10

Automated Archiving System Main NWP archive (operational run data, R&D) Main climate research archive Main backup for all systems including the climate archive. Limited satellite and radar imagery archive 11

Automated Archiving System 19,000 tape mounts/month 20 TB (Tera Bytes or 10 12 bytes) growth/month 370 TB in use. 12

Automated Archiving System 13

Supercomputer Configuration

Software Levels AIX 5.1 ML6 PSSP 3.5 LoadLeveler 3.1 GPFS 2.1 15

Software Maintenance Two hours window for maintenance: we use alternate disk installs. Try to stay up to date, but not obvious. Still not completely familiar with IBM s software distribution methods 16

GPFS Configuration 18 VSD Servers, dual FC, 17 FASTt700 controllers. 4 VSD reserved for production i/o (1 file system). 14 remaining servers for development (15 file systems). 17

Batch Sub-system Use of gang scheduler only for preemption. Two main classes: production, development. Negotiator and Scheduler running on the service nodes. One schedd. 18

Batch Sub-system 19

Batch Sub-system Issues We need enforced memory limits!!! Various issues related to preemption. 20

System Usage 21

Monitoring Home grown tools Passive: log gathering and massaging Active: scripts pro-actively testing various items All syslogs redirected to the CWS How to cope with non-syslog logs? Still thinking Heterogeneous solution. 22

Home grown stuff LL SGE gateway Refurbished HPM toolkit Double buffered rcp. 23

On-going Support The systems must achieve a minimum of 99 % availability. Remedial maintenance 24/7: 30 minute response time between 8 A.M. and 5 P.M. on weekdays. One hour response outside above periods. Preventive maintenance or engineering changes: maximum of eight (8) hours a month, in blocks of time not exceeding two (2) or four (4) hours per day (subject to certains conditions). Software support: Provision of emergency and nonemergency assistance. 24

SCIENTIFIC DIRECTIONS

Global System Now: Uniform resolution of 100 km (400 X 200 X 28) 3D-Var at T108 on model levels, 6-hr cycle Use of raw radiances from AMSUA, AMSUB and GOES Use of MODIS satellite winds and profiler data. 2005: Resolution to 35 km (800 X 600 X 58) 4D-Var assimilation, 6-hr time window with 3 outer loops at full model resolution and inner loops at T108 (cpu equivalent of a 5- day forecast of full resolution model) new datasets: QuikScat 2006+: Additional datasets (AIRS, MSG, MTSAT, IASI, GIFTS, COSMIC) Improved data assimilation 26

Regional System Now: Variable resolution, uniform region at 15 km (575 X 641 X 58) 3D-Var assimilation on model levels at T108, 12-hr spin-up 2004: Resolution to 15 km in uniform region (576 X 641 X 58) Inclusion of AMSU-B and GOES data in assimilation cycle New datasets: profilers, MODIS winds 2005: Four model runs a day (instead of two) 2006+: LAM 4D-Var data assimilation Limited area model at 10 km resolution (800 X 800 X 60) Assimilation of Radar data 27

Ensemble Prediction System Now: 16 members global system (300 X 150 X 28) Forecasts up to 10 days once a day at 00Z Optimal Interpolation assimilation system, 6-hr cycle,use of derived radiance data (Satems) End 2004 (currently running in parallel): Ensemble Kalman Filter assimilation system, 6-hr cycle, use of raw radiances from AMSUA, AMSUB and GOES Forecasts extended to 15 days (instead of 10) 28

...Ensemble Prediction System 2005: Increased resolution to 100 km (400 X 200 X 58) Increased members to 32 Additional datasets such as in global deterministic system Two forecast runs per day (12Z run added) 2007: Prototype regional ensemble 10 members LAM (500 X 500 X 58) No distinct data assimilation; initial and boundary conditions 29 from global EPS

Mesoscale System Now: Variable resolution, uniform region at 10 km (290 X 371 X 35) Two windows; no data assimilation Prototype Limited Area Model at 2.5 km (500 X 500 X 58) over one area 2006: Five Limited Area Model windows at 2.5 km (500 X 500 X 58) 2007+: 4D data assimilation 30

Coupled Models Today In R&D: coupled atmosphere, ocean, ice, wave, hydrology, biology & chemistry In Production: storm surge, wave 2005 Regional system coupled with ocean/ice model over Gulf of St Lawrence in operations. Future Global coupled model for both prediction & data assimilation 31