J-Zephyr Sebastian D. Eastham

Slides:



Advertisements
Similar presentations
GEMS Kick- off MPI -Hamburg CTM - IFS interfaces GEMS- GRG Review of meeting in January and more recent thoughts Johannes Flemming.
Advertisements

Weather Research & Forecasting: A General Overview
Ch. 2 Protocol Architecture. 2.1 The Need for a Protocol Architecture Same set of layered functions need to exist in the two communicating systems. Key.
Computational Challenges in Air Pollution Modelling Z. Zlatev National Environmental Research Institute 1. Why air pollution modelling? 2. Major physical.
TUPEC057 Advances With Merlin – A Beam Tracking Code J. Molson, R.J. Barlow, H.L. Owen, A. Toader MERLIN is a.
1 Coven a Framework for High Performance Problem Solving Environments Nathan A. DeBardeleben Walter B. Ligon III Sourabh Pandit Dan C. Stanzione Jr. Parallel.
Quantitative retrievals of NO 2 from GOME Lara Gunn 1, Martyn Chipperfield 1, Richard Siddans 2 and Brian Kerridge 2 School of Earth and Environment Institute.
Meteorological Data Issues for Class II Increment Analysis.
Using beryllium-7 to assess stratosphere-to- troposphere transport in global models 4 th GEOS-Chem Users’ Meeting Harvard University, April 7-10, 2009.
Atmospheric modelling activities inside the Danish AMAP program Jesper H. Christensen NERI-ATMI, Frederiksborgvej Roskilde.
FPGA BASED IMAGE PROCESSING Texas A&M University / Prairie View A&M University Over the past few decades, the improvements from machine language to objected.
PhD/Master course, Uppsala  Understanding the interaction between your program and computer  Structuring the code  Optimizing the code  Debugging.
Effects of Global Change on Air Quality in the United States Shiliang Wu, Harvard 2 nd GEOS-CHEM users’ meeting Apr 5, 2005 Look into the Future with GEOS-CHEM.
Advance the understanding and the prediction of mesoscale precipitation systems and to promote closer ties between the research and operational forecasting.
S. Maksyutov, P.K. Patra and M. Ishizawa Jena; 13 May 2003 TDI experiment with NIES model and interannually varying NCEP winds.
A TWO-FLUID NUMERICAL MODEL OF THE LIMPET OWC CG Mingham, L Qian, DM Causon and DM Ingram Centre for Mathematical Modelling and Flow Analysis Manchester.
Introduction to Symmetric Multiprocessors Süha TUNA Bilişim Enstitüsü UHeM Yaz Çalıştayı
Template Development of a Plume-in-Grid Version of Global-through-Urban WRF/Chem Prakash Karamchandani, Krish Vijayaraghavan, Shu-Yun Chen ENVIRON International.
Status of Dynamical Core C++ Rewrite (Task 5) Oliver Fuhrer (MeteoSwiss), Tobias Gysi (SCS), Men Muhheim (SCS), Katharina Riedinger (SCS), David Müller.
1 CCOS Seasonal Modeling: The Computing Environment S.Tonse, N.J.Brown & R. Harley Lawrence Berkeley National Laboratory University Of California at Berkeley.
University of Veszprém Department of Image Processing and Neurocomputing Emulated Digital CNN-UM Implementation of a 3-dimensional Ocean Model on FPGAs.
Institute for Mathematical Modeling RAS 1 Dynamic load balancing. Overview. Simulation of combustion problems using multiprocessor computer systems For.
Development of WRF-CMAQ Interface Processor (WCIP)
Protocols and the TCP/IP Suite
Understanding the USEPA’s AERMOD Modeling System for Environmental Managers Ashok Kumar Abhilash Vijayan Kanwar Siddharth Bhardwaj University of Toledo.
Sept COMP60611 Fundamentals of Concurrency Lab Exercise 2 Notes Notes on the finite difference performance model example – for the lab… Graham Riley,
HIRLAM 1D model tool B.H. Sass December Overview 1) Introduction 1) Introduction 2) History of 1D-model(s) 2) History of 1D-model(s) 3) Basics of.
2 March 2007 Running MATCH. 2 March 2007 MATCH is running on: Linux, Solaris PCs (not Windows PC!) Unix workstations Possible to compile on most hardware.
Georgia Institute of Technology Initial Application of the Adaptive Grid Air Quality Model Dr. M. Talat Odman, Maudood N. Khan Georgia Institute of Technology.
Overview of the tracer code in RegCM Tracers Aerosols ( droplets, smoke particles, dust, pollens, flying cats …) Gazeous phase, chemical species Evolution.
August 2001 Parallelizing ROMS for Distributed Memory Machines using the Scalable Modeling System (SMS) Dan Schaffer NOAA Forecast Systems Laboratory (FSL)
TEMPLATE DESIGN © A high-order accurate and monotonic advection scheme is used as a local interpolator to redistribute.
MESQUITE: Mesh Optimization Toolkit Brian Miller, LLNL
Advancing Galaxy Formation Modeling Andrew Benson California Institute of Technology.
Lagrangian particle models are three-dimensional models for the simulation of airborne pollutant dispersion, able to account for flow and turbulence space-time.
CS 732: Advance Machine Learning
CSCI 465 D ata Communications and Networks Lecture 2 Martin van Bommel CSCI 465 Data Communications & Networks 1.
___________________________________________________________________________CMAQ Basics ___________________________________________________Community Modeling.
Module 6 MM5: Overview William J. Gutowski, Jr. Iowa State University.
STDP and Network Architectures Parallel ODE Solver and Event Detector Eugene Lubenov.
Kriging for Estimation of Mineral Resources GISELA/EPIKH School Exequiel Sepúlveda Department of Mining Engineering, University of Chile, Chile ALGES Laboratory,
Applied Operating System Concepts
Kai Li, Allen D. Malony, Sameer Shende, Robert Bell
Test 2 Review Outline.
Introduction to Parallel Computing: MPI, OpenMP and Hybrid Programming
Transport Working Group
RAID, Programmed I/O, Interrupt Driven I/O, DMA, Operating System
Status of the COSMO-Software and Documentation
CMAQ Programs and Options
Advances in BUB3D-OM Development
WindNinja Model Domain/Objective
Implementing Simplified Molecular Dynamics Simulation in Different Parallel Paradigms Chao Mei April 27th, 2006 CS498LVK.
Parallel NetCDF + MASS Development
CMAQ PARALLEL PERFORMANCE WITH MPI AND OpenMP George Delic, Ph
NASA ROSES 2007: Decision Support through Earth Science Research Results Improving an Air Quality Decision Support System through the Integration of Satellite.
Assimilating Tropospheric Emission Spectrometer profiles in GEOS-Chem
Models of atmospheric chemistry
GENERAL VIEW OF KRATOS MULTIPHYSICS
FUJIN: a parallel framework for meteorological models
Operating System Concepts
WRF-GC: on-line two-way coupling of WRF and GEOS-Chem
FIRST GEOS-CHEM USERS’ MEETING June
Shiliang Wu1 Loretta J. Mickley1, Daniel J
Hybrid Programming with OpenMP and MPI
Some thoughts on future air quality models from a WRF-Chem modeler
COMP60621 Designing for Parallelism
Operating System Concepts
A brief introduction to NEMS
Development and Evaluation of a Hybrid Eulerian-Lagrangian Modeling Approach Beata Czader, Peter Percell, Daewon Byun, Yunsoo Choi University of Houston.
Inputs, Outputs and Assignment
Presentation transcript:

J-Zephyr Sebastian D. Eastham Website: LAE.MIT.EDU Twitter: @MIT_LAE

MATLAB: Slow, accessible Well suited to analysis Project aim Modeling INCREASING SPEED FORTRAN: Fast Somewhat hostile Analysis Julia…? MATLAB: Slow, accessible Well suited to analysis

Base model: GEOS-Chem Chemistry Transport Model (CTM) Simulates atmosphere using combination of CFD with chemistry solver

New model: J-Zephyr Feature GEOS-Chem J-Zephyr Language FORTRAN 77/90 Julia v0.3.0 Chemistry? Yes No Transport? 3-D 2-D* PBL mixing? WIP* Convection? Parallelization OpenMP DArrays Maximum CPUs Physical cores Maximum mahcines 1 Network size* Dev time? 10 years 3 months

Parallelization: OpenMP v MPI HEAD GEOS-Chem WORKER HEAD Time Loops independently parallel Arrays redistributed each time WORKER J-Zephyr Both GC and JZ, head proc coordinates [click] GC uses OpenMP Each loop is broken out [click] requiring data to be redistributed In JZ, use Darrays [click] Data is stored in subgrids between loops [click], reducing communication Data stored in DArrays Low worker communication

Conflict of interest Chemistry and transport processes favor different domain decompositions SPECIES Transport operates independently on each species, mostly horizontal Requires halo data Ideal decomposition: parallel by species EXPLAIN GRID! SPATIAL

Conflict of interest Chemistry and transport processes favor different domain decompositions SPECIES Chemistry concerned only with small spatial domain Often concerned with columnar processes and many species at once Ideal decomposition: parallel by longitude/latitude J-Zephyr uses this SPATIAL

Read input file Set up DArrays Read in GEOS-5 data Begin simulation Model structure Read input file Set up DArrays Read in GEOS-5 data Begin simulation Customized input file establishes simulation parameters Basic grid parameters are set up, and various constants and multipliers are calculated in advance on the head node GEOS-5 meteorological data including surface pressures, wind velocities, moisture etc are read in from FORTRAN “binary punch card” files and stored in Darrays. These will need to be updated every 3-6 simulation hours A set of nested loops, optimized with respect to handling the met data transfers

Input Input file allows specification of: Date range (2004-2011) Machine file Output directory Time step length Tracers (unlimited number)

DArrays Grid distribution automatically determined on a 2-D (surface pressure) array in initialization All subsequent DArrays use this grid as their basis All cores have access to: Restricted latitudes Restricted longitudes All pressure levels All tracers

Modules Initially intended to use “common-block” approach, holding DArrays as module variables This caused a huge number of problems Now pass large data structures (e.g. meteorological state variable holds 74 72x46x72 DArrays) between functions

Project status Implemented so far: Model initialization Single machine parallelization Read-in of meteorological data from FORTRAN formatted records (BPCH) Derivation of 3-D pressure fluxes from meteorological data Simple emissions scheme Horizontal advection and mixing

Lessons Excellent learning experience in parallel Inter-worker communication prohibitively slow Parallel debugging much slower than serial! 5000+ lines of code do not guarantee results

Conclusions Project intended to give insight into difficulties that crop up when solving large problems in parallel fashion Highlighted disproportionate penalties associated with communication Demonstrated importance of intelligent domain decomposition Showcased Julia capabilities – extremely rough first pass still capable of emulating CTM on simple 6-core machine

Next steps Current domain decomposition is far from ideal Priority 1: Optimize “halo exchange” Cell [I,J] here requires data from 3 other workers Problem worsens at poles Should assign polar caps to single workers J+2 J+1 I-2 I-1 I,J I+1 I+2 J-1 J-2

Future work Add vertical transport Routines in place but not yet tested Verify horizontal and vertical transport against GEOS-Chem Perform timing analysis Excessive inter-worker communication Add deep convection Add PBL mixing

Laboratory for Aviation and the Environment Massachusetts Institute of Technology Presenter Name seastham@mit.edu Website: LAE.MIT.EDU Twitter: @MIT_LAE