I/O and the SciDAC Software API Robert Edwards U.S. SciDAC Software Coordinating Committee May 2, 2003.

Slides:



Advertisements
Similar presentations
UKQCD GridPP NeSCAC Irving, 4/2/041 9 th GridPP Collaboration Meeting QCDgrid: Status and Future Alan Irving University of Liverpool.
Advertisements

Introduction to the BinX Library eDIKT project team Ted Wen Robert Carroll
JLab Lattice Portal – Data Grid Web Service Ying Chen, Chip Watson Thomas Jefferson National Accelerator Facility.
SciDAC Software Infrastructure for Lattice Gauge Theory
Nuclear Physics in the SciDAC Era Robert Edwards Jefferson Lab SciDAC 2009 TexPoint fonts used in EMF. Read the TexPoint manual before you delete this.
Lattice QCD Comes of Age y Richard C. Brower XLIst Rencontres de Moriond March QCD and Hadronic interactions at high energy.
Data-Parallel Programming Model Basic uniform operations across lattice: C(x) = A(x)*B(x) Distribute problem grid across a machine grid Want API to hide.
QDP++ and Chroma Robert Edwards Jefferson Lab
ILDG File Format Chip Watson, for Middleware & MetaData Working Groups.
SciDAC Software Infrastructure for Lattice Gauge Theory DOE meeting on Strategic Plan --- April 15, 2002 Software Co-ordinating Committee Rich Brower ---
HackLatt MILC with SciDAC C Carleton DeTar HackLatt 2008.
MILC Code Basics Carleton DeTar KITPC MILC Code Capabilities Molecular dynamics evolution –Staggered fermion actions (Asqtad, Fat7, HISQ,
HackLatt MILC Code Basics Carleton DeTar HackLatt 2008.
SciDAC Software Infrastructure for Lattice Gauge Theory DOE Grant ’01 -- ’03 (-- ’05?) All Hands Meeting: FNAL Feb. 21, 2003 Richard C.Brower Quick Overview.
I/O and the SciDAC Software API Robert Edwards U.S. SciDAC Software Coordinating Committee May 2, 2003.
The science of simulation falsification algorithms phenomenology machines better theories computer architectures non-perturbative QFT experimental tests.
Status of netCDF-3, netCDF-4, and CF Conventions Russ Rew Community Standards for Unstructured Grids Workshop, Boulder
QCDgrid Technology James Perry, George Beckett, Lorna Smith EPCC, The University Of Edinburgh.
The QCDOC Project Overview and Status Norman H. Christ DOE LGT Review May 24-25, 2005.
LQCD Project Overview Don Holmgren LQCD Project Progress Review May 25-26, 2006 Fermilab.
SciDAC Software Infrastructure for Lattice Gauge Theory Richard C. Brower Annual Progress Review JLab, May 14, 2007 Code distribution see
ICOM 5995: Performance Instrumentation and Visualization for High Performance Computer Systems Lecture 7 October 16, 2002 Nayda G. Santiago.
1 High level view of HDF5 Data structures and library HDF Summit Boeing Seattle September 19, 2006.
QCD Project Overview Ying Zhang September 26, 2005.
A Metadata Based Approach For Supporting Subsetting Queries Over Parallel HDF5 Datasets Vignesh Santhanagopalan Graduate Student Department Of CSE.
Chroma I: A High Level View Bálint Joó Jefferson Lab, Newport News, VA given at HackLatt'06 NeSC, Edinburgh March 29, 2006.
SciDAC Software Infrastructure for Lattice Gauge Theory Richard C. Brower & Robert Edwards June 24, 2003.
CHEP 2000, Giuseppe Andronico Grid portal based data management for Lattice QCD data ACAT03, Tsukuba, work in collaboration with A.
IPlant Collaborative Tools and Services Workshop iPlant Collaborative Tools and Services Workshop Collaborating with iPlant.
QCDGrid Progress James Perry, Andrew Jackson, Stephen Booth, Lorna Smith EPCC, The University Of Edinburgh.
Scalable Systems Software Center Resource Management and Accounting Working Group Face-to-Face Meeting October 10-11, 2002.
JLab SciDAC Activities QCD-API design and other activities at JLab include: –Messaging passing design and code (Level 1) [Watson, Edwards] First implementation.
ALMA Integrated Computing Team Coordination & Planning Meeting #1 Santiago, April 2013 Relational APDM & Relational ASDM models effort done in online.
IPlant Collaborative Tools and Services Workshop iPlant Collaborative Tools and Services Workshop Collaborating with iPlant.
The netCDF-4 data model and format Russ Rew, UCAR Unidata NetCDF Workshop 25 October 2012.
An Overview of Berkeley Lab’s Linux Checkpoint/Restart (BLCR) Paul Hargrove with Jason Duell and Eric.
Lattice QCD and the SciDAC-2 LQCD Computing Project Lattice QCD Workflow Workshop Fermilab, December 18, 2006 Don Holmgren,
SciDAC All Hands Meeting, March 2-3, 2005 Northwestern University PIs:Alok Choudhary, Wei-keng Liao Graduate Students:Avery Ching, Kenin Coloma, Jianwei.
The Future of the iPlant Cyberinfrastructure: Coming Attractions.
HackLatt MILC Code Basics Carleton DeTar First presented at Edinburgh EPCC HackLatt 2008 Updated 2013.
Parallel and Grid I/O Infrastructure W. Gropp, R. Ross, R. Thakur Argonne National Lab A. Choudhary, W. Liao Northwestern University G. Abdulla, T. Eliassi-Rad.
ILDG Middleware Status Bálint Joó UKQCD University of Edinburgh, School of Physics on behalf of ILDG Middleware Working Group alternative title: Report.
Lattice QCD Data Grid Middleware: status report M. Sato, CCS, University of Tsukuba ILDG6, May, 12, 2005.
_______________________________________________________________CMAQ Libraries and Utilities ___________________________________________________Community.
Numerical Libraries Project Microsoft Incubation Group Mary Beth Hribar Microsoft Corporation CSCAPES Workshop June 10, 2008 Copyright Microsoft Corporation,
Chroma: An Application of the SciDAC QCD API(s) Bálint Joó School of Physics University of Edinburgh UKQCD Collaboration Soon to be moving to the JLAB.
1 1 What does Performance Across the Software Stack mean?  High level view: Providing performance for physics simulations meaningful to applications 
SciDAC Software Infrastructure for Lattice Gauge Theory Richard C. Brower QCD Project Review May 24-25, 2005 Code distribution see
Earth System Modeling Framework Python Interface (ESMP) October 2011 Ryan O’Kuinghttons Robert Oehmke Cecelia DeLuca.
CEOS Working Group on Information Systems and Services - 1 Data Services Task Team Discussions on GRID and GRIDftp Stuart Doescher, USGS WGISS-15 May 2003.
An Introduction to Device Drivers Ted Baker  Andy Wang COP 5641 / CIS 4930.
1 Cluster Development at Fermilab Don Holmgren All-Hands Meeting Jefferson Lab June 1-2, 2005.
Connections to Other Packages The Cactus Team Albert Einstein Institute
WRF Software Development and Performance John Michalakes, NCAR NCAR: W. Skamarock, J. Dudhia, D. Gill, A. Bourgeois, W. Wang, C. Deluca, R. Loft NOAA/NCEP:
HEP and NP SciDAC projects: Key ideas presented in the SciDAC II white papers Robert D. Ryne.
ESMF,WRF and ROMS. Purposes Not a tutorial Not a tutorial Educational and conceptual Educational and conceptual Relation to our work Relation to our work.
May 30-31, 2012 HDF5 Workshop at PSI May Metadata Journaling Dana Robinson The HDF Group Efficient Use of HDF5 With High Data Rate X-Ray Detectors.
QDP++ and Chroma Robert Edwards Jefferson Lab Collaborators: Balint Joo.
Unidata Infrastructure for Data Services Russ Rew GO-ESSP Workshop, LLNL
Fermi National Accelerator Laboratory & Thomas Jefferson National Accelerator Facility SciDAC LQCD Software The Department of Energy (DOE) Office of Science.
Parallel Virtual File System (PVFS) a.k.a. OrangeFS
LQCD Computing Project Overview
Jim Fawcett CSE687 – Object Oriented Design Spring 2016
Computational Requirements
LQCD Computing Operations
ILDG Implementation Status
Chroma: An Application of the SciDAC QCD API(s)
MPJ: A Java-based Parallel Computing System
International Lattice Data Grid
Jim Fawcett CSE687 – Object Oriented Design Spring 2015
Presentation transcript:

I/O and the SciDAC Software API Robert Edwards U.S. SciDAC Software Coordinating Committee May 2, 2003

Cast of Characters Committee: R.Brower, C.DeTar, R.Edwards, D.Holmgren, R.Mawhinney, C.Mendes, C.Watson Additional: J.Chen, E.Gregory, J.Hetrick, C.Jung, J.Osborn, K.Petrov, A.Pochinsky, J.Simone IO: C.DeTar, R.Edwards, J.Osborn, J.Simone, B.Joó

Sci D A C Through entific iscovery dvanced omputing

SciDAC Project Goals Portable, scalable software High performance optimization on two target architectures Exploitation and Optimization of existing application base Infrastructure for (US) national community Sharing of valuable lattice data, and data management GRID (ILDG) ClustersQCDOC

Optimised Dirac Operators, Inverters Level 3 QDP (QCD Data Parallel) Lattice Wide Operations, Data shifts Level 2 QMP (QCD Message Passing) QLA (QCD Linear Algebra) Level 1 QIO XML I/O DIME SciDAC Software Structure Exists in C/C++, implemented over MPI, GM, QCDOC, gigE Optimised for P4 and QCDOC Focus of talk Exists in C/C++

Data Parallel QDP/C,C++ API Hides architecture and layout Operates on lattice fields across sites Linear algebra tailored for QCD Shifts and permutation maps across sites Reductions Subsets

Data-parallel Operations Unary and binary: -a; a-b; … Unary functions: adj(a), cos(a), sin(a), … Random numbers: // platform independent random(a), gaussian(a) Comparisons (booleans) a <= b, … Broadcasts: a = 0, … Reductions: sum(a), …

QDP Expressions Can create expressions QDP/C++ code multi1d u(Nd); LatticeDiracFermion b, c, d; int mu; c = u[mu] * shift(b,mu) + 2 * d; PETE: Portable Expression Template Engine Temporaries eliminated, expressions optimised

Generic QDP Binary File Formats Composed of 1 or more application records Single application record has 1 QDP field or an array of fields Binary data in lexicographic site major order Physics metadata for file and for each record Using DIME to package

Metadata Use XML for file and record metadata File and record metadata managed at user convenience No agreed minimum standard Use binX to describe binary binX not in record metadata – provides serialization info

Gauge Fields For published data – still converging on metadata Considering adopting ILDG-like schema Need more cases, like Asqtad, domain wall Likely that non-published/private configs will use reduced set of metadata Write arrays of fields as one record – all 3 rows Site major order – slowest varying Will adopt single format and byte ordering

File Format File physics metadata Application record 1 Physics metadata binX description Binary data [may have array indices within sites] Checksum Record 2 - possible additional records Physics metadata binX description Binary data Checksum Record 3….

Data Hierarchy Project built from datasets (e.g. gauge fields and propagators) Dataset built from files (e.g. gauge fields) File built from records (e.g. eigenvectors) Record = QDP field and metadata

Direct Internet Message Encapsulation (DIME) Data written to (read from) a list of records Each record has DIME Type (required) URL or like MIME type DIME Id (optional URL) Maximum record size is 2Gb Data larger than 2Gb can be split into successive record chunks Chunking easy, file size > 2Gb a problem

QIO: Grid Friendly I/O Metadata & Physics data Reader / Writer API Read/Write simple XML documents Not using data binding Metadata used like a buffer, physics data like a stream QDP IO (QIO) Serial – all nodes stream through one node Parallel – if available, many nodes to parallel filesystem MetaWriter file_xml, record_xml; SerialFileWriter out(file_xml,foo.dat); LatticeDiracFermion psi; out.write(record_xml, psi);

MetaReader struct foo_t foo; struct bar_t bar; double kappa; MetaReader in; char *key=/foo/bar/kappa; File xml: XML Reader/Writer supports recursive serialization To/From buffers (strings - MetaData) To/From files (PhysicsData) Intended to drive codes rather than DataGrid C,C++ versions in.get (foo,/foo) in.get (kappa,key);

Current Status Releases and documentation QMP, QDP/C,C++ in first release Performance improvements/testing underway Porting & development efforts of physics codes over QDP on-going QIO near completion DIME completed (Bálint Joó) XML Reader/Writer in development