Big Data Remote Access Interfaces for Experimental Physics

Slides:



Advertisements
Similar presentations
A Workflow Engine with Multi-Level Parallelism Supports Qifeng Huang and Yan Huang School of Computer Science Cardiff University
Advertisements

Experiment Workflow Pipelines at APS: Message Queuing and HDF5 Claude Saunders, Nicholas Schwarz, John Hammonds Software Services Group Advanced Photon.
Issues of HPC software From the experience of TH-1A Lu Yutong NUDT.
ASCR Data Science Centers Infrastructure Demonstration S. Canon, N. Desai, M. Ernst, K. Kleese-Van Dam, G. Shipman, B. Tierney.
SALSA HPC Group School of Informatics and Computing Indiana University.
RCAC Research Computing Presents: DiaGird Overview Tuesday, September 24, 2013.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
Presented by Scalable Systems Software Project Al Geist Computer Science Research Group Computer Science and Mathematics Division Research supported by.
Astrophysics, Biology, Climate, Combustion, Fusion, Nanoscience Working Group on Simulation-Driven Applications 10 CS, 10 Sim, 1 VR.
The new The new MONARC Simulation Framework Iosif Legrand  California Institute of Technology.
Swift: A Scientist’s Gateway to Campus Clusters, Grids and Supercomputers Swift project: Presenter contact:
Web-based Portal for Discovery, Retrieval and Visualization of Earth Science Datasets in Grid Environment Zhenping (Jane) Liu.
A Workflow-Aware Storage System Emalayan Vairavanathan 1 Samer Al-Kiswany, Lauro Beltrão Costa, Zhao Zhang, Daniel S. Katz, Michael Wilde, Matei Ripeanu.
Lecture 8 – Platform as a Service. Introduction We have discussed the SPI model of Cloud Computing – IaaS – PaaS – SaaS.
 Cloud computing  Workflow  Workflow lifecycle  Workflow design  Workflow tools : xcp, eucalyptus, open nebula.
Descriptive Data Analysis of File Transfer Data Sudarshan Srinivasan Victor Hazlewood Gregory D. Peterson.
DISTRIBUTED DATA FLOW WEB-SERVICES FOR ACCESSING AND PROCESSING OF BIG DATA SETS IN EARTH SCIENCES A.A. Poyda 1, M.N. Zhizhin 1, D.P. Medvedev 2, D.Y.
Emalayan Vairavanathan
DISTRIBUTED COMPUTING
CCA Common Component Architecture Manoj Krishnan Pacific Northwest National Laboratory MCMD Programming and Implementation Issues.
IPlant Collaborative Tools and Services Workshop iPlant Collaborative Tools and Services Workshop Collaborating with iPlant.
Integrating HPC into the ATLAS Distributed Computing environment Doug Benjamin Duke University.
SC 2012 © LLNL / JSC 1 HPCToolkit / Rice University Performance Analysis through callpath sampling  Designed for low overhead  Hot path analysis  Recovery.
IPlant Collaborative Tools and Services Workshop iPlant Collaborative Tools and Services Workshop Collaborating with iPlant.
The Future of the iPlant Cyberinfrastructure: Coming Attractions.
MATRIX MULTIPLY WITH DRYAD B649 Course Project Introduction.
March 27, 2007HPC 07 - Norfolk, VA1 C++ Reflection for High Performance Problem Solving Environments Tharaka Devadithya 1, Kenneth Chiu 2, Wei Lu 1 1.
Pascucci-1 Valerio Pascucci Director, CEDMAV Professor, SCI Institute & School of Computing Laboratory Fellow, PNNL Massive Data Management, Analysis,
Scalable Systems Software for Terascale Computer Centers Coordinator: Al Geist Participating Organizations ORNL ANL LBNL.
Interactive Workflows Branislav Šimo, Ondrej Habala, Ladislav Hluchý Institute of Informatics, Slovak Academy of Sciences.
CCGrid, 2012 Supporting User Defined Subsetting and Aggregation over Parallel NetCDF Datasets Yu Su and Gagan Agrawal Department of Computer Science and.
May 6, 2002Earth System Grid - Williams The Earth System Grid Presented by Dean N. Williams PI’s: Ian Foster (ANL); Don Middleton (NCAR); and Dean Williams.
Swift/T: Dataflow Composition of Tcl Scripts for Petascale Computing Justin M Wozniak Argonne National Laboratory and University of Chicago
High Performance Computational Experiments with Swift/T Justin M. Wozniak, Argonne National Laboratory & The University of Chicago
DIRAC Project A.Tsaregorodtsev (CPPM) on behalf of the LHCb DIRAC team A Community Grid Solution The DIRAC (Distributed Infrastructure with Remote Agent.
Compilation of XSLT into Dataflow Graphs for Web Service Composition Peter Kelly Paul Coddington Andrew Wendelborn.
Globus online Delivering a scalable service Steve Tuecke Computation Institute University of Chicago and Argonne National Laboratory.
TeraGrid Capability Discovery John-Paul “JP” Navarro TeraGrid Area Co-Director for Software Integration University of Chicago/Argonne National Laboratory.
Towards a High Performance Extensible Grid Architecture Klaus Krauter Muthucumaru Maheswaran {krauter,
Scientific Data Processing Portal and Heterogeneous Computing Resources at NRC “Kurchatov Institute” V. Aulov, D. Drizhuk, A. Klimentov, R. Mashinistov,
Petr Škoda, Jakub Koza Astronomical Institute Academy of Sciences
HPC In The Cloud Case Study: Proteomics Workflow
VisIt Project Overview
Workload Management Workpackage
Organizations Are Embracing New Opportunities
Deployment of Flows Loretta Auvil
CI Updates and Planning Discussion
HPC In The Cloud Case Study: Proteomics Workflow
Introduction to Distributed Platforms
Tools and Services Workshop
University of Chicago and ANL
Seismic Hazard Analysis Using Distributed Workflows
Pathology Spatial Analysis February 2017
Spark Presentation.
Distributed Data Access and Resource Management in the D0 SAM System
Joseph JaJa, Mike Smorul, and Sangchul Song
Parallel Programming By J. H. Wang May 2, 2017.
Platform as a Service.
The Client/Server Database Environment
Abstract Machine Layer Research in VGrADS
Introduction to Spark.
Weaving Abstractions into Workflows
Hadoop Technopoints.
Overview of big data tools
Spark and Scala.
Software for Neutron Imaging Analysis
Overview of Workflows: Why Use Them?
High Throughput Computing for Astronomers
gLite The EGEE Middleware Distribution
Question 1 How are you going to provide language and/or library (or other?) support in Fortran, C/C++, or another language for massively parallel programming.
Presentation transcript:

Big Data Remote Access Interfaces for Experimental Physics Justin M Wozniak <wozniak@mcs.anl.gov> 2nd S2I2 CS/HEP Workshop Princeton – May 2, 2017

Advanced Photon Source (APS) Chicago Supercomputers Advanced Photon Source (APS)

Advanced Photon Source (APS) Moves electrons at electrons at >99.999999% of the speed of light. Magnets bend electron trajectories, producing x-rays, highly focused onto a small area X-rays strike targets in 35 different laboratories – each a lead-lined, radiation-proof experiment station

Proximity means we can closely couple computing in novel ways Terabits/s in the near future Petabits/s are possible ALCF MCS APS

Swift/T: Enabling high-performance workflows Write site-independent scripts Automatic parallelization and data movement Run native code, script fragments as applications Rapidly subdivide large partitions for MPI jobs Move work to data locations Swift worker process C C++ Fortran Swift/T worker Swift control process Swift/T control process MPI 64K cores of Blue Waters 2 billion Python tasks 14 million Pythons/s www.ci.uchicago.edu/swift www.mcs.anl.gov/exm

Swift programming model: all progress driven by concurrent dataflow (int r) myproc (int i, int j) { int x = F(i); int y = G(j); r = x + y; } F() and G() implemented in native code or external programs F() and G()run in concurrently in different processes r is computed when they are both done This parallelism is automatic Works recursively throughout the program’s call graph

Swift/T: Fully parallel evaluation of complex scripts int X = 100, Y = 100; int A[][]; int B[]; foreach x in [0:X-1] { foreach y in [0:Y-1] { if (check(x, y)) { A[x][y] = g(f(x), f(y)); } else { A[x][y] = 0; } B[x] = sum(A[x]); Swift/T: Scalable data flow programming for distributed-memory task-parallel applications . Proc. CCGrid, 2013. Compiler techniques for massively scalable implicit task parallelism. Proc. SC 2014.

Features for Big Data analysis Location-aware scheduling User and runtime coordinate data/task locations Collective I/O User and runtime coordinate data/task locations Application I/O hook Application Dataflow, annotations Runtime MPI-IO transfers Runtime Hard/soft locations Distributed data Distributed data Parallel FS F. Duro et al. Flexible data-aware scheduling for workflows over an in-memory object store. Proc. CCGrid 2016. Wozniak et al. Big data staging with MPI-IO for interactive X-ray science. Proc. Big Data Computing, 2014.

NeXpy: A Python Toolbox for Big Data A toolbox for manipulating and visualizing arbitrary NeXus (HDF5) data of any size A scripting engine for GUI applications Uses Python bindings for HDF5 A portal to Globus Catalog A demonstration of the value of combining: a flexible data model a powerful scripting language http://nexpy.github.io/nexpy $ pip install nexpy + =

Mullite

NeXpy in the Pipeline Use of NeXpy throughout the analysis pipeline

The NeXus File Service (NXFS)

NXFS Performance Faster than application-agnostic remote filesystem technologies Compared Pyro to Chirp and SSHFS from inside ANL (L) and AWS EC2 (W) Plus ability to invoke remote methods! File open (10-1s) Metadata read (10-2s) Pixel read (1s) Operation and Time Scale

Ad Hoc Cornell CHESS Pipeline

It worked! 21 TB in a few days

Possible interactions with the Institute Swift: Deploying loosely-coupled workloads on HPC systems Decomposing/recomposing workloads in different ways Data access: What are the data standards for HEP How are these datasets accessed by workflows Reconfiguration: What are the intersections of ad hoc and permanent computing infrastructure How do technologies scale, including how projects scale