Data Management and Software Centre Mark Hagen Head of DMSC IKON6, February 27 th 2014.

Slides:



Advertisements
Similar presentations
ASCR Data Science Centers Infrastructure Demonstration S. Canon, N. Desai, M. Ernst, K. Kleese-Van Dam, G. Shipman, B. Tierney.
Advertisements

Status of the European Spallation Source Timo Korhonen Chief Engineer, Integrated Control System Division October 21, 2014.
Data Management and Software Centre Mark Hagen Head of DMSC Institutional Export Conference, Tallinn,
Report from DANSE Workshop Sept. 3-8, 2003 Goals: 1) To explain DANSE to selected scientists and engineers who develop software for neutron scattering.
Summary Role of Software (1 slide) ARCS Software Architecture (4 slides) SNS -- Caltech Interactions (3 slides)
The Data Management Requirements at SNS Shelly Ren & Steve Miller Scientific Computing Group, SNS-ORNL December 11, 2006.
Chronopolis: Preserving Our Digital Heritage David Minor UC San Diego San Diego Supercomputer Center.
Experimental Facilities DivisionORNL - SNS June 22, 2004 SNS Update – Team Building Steve Miller June 22, 2004 DANSE Meeting at Caltech.
Experimental Facilities Division ANL-ORNL SNS Experimental Data Standards (Status) Richard Riedel SNS Data Acquisition Group Leader.
The Role of DANSE at SNS Steve Miller Scientific Computing Group Leader January 22, 2007.
Developing PANDORA Mark Corbould Director, IT Business Systems.
CERN IT Department CH-1211 Genève 23 Switzerland t Integrating Lemon Monitoring and Alarming System with the new CERN Agile Infrastructure.
TeraGrid Gateway User Concept – Supporting Users V. E. Lynch, M. L. Chen, J. W. Cobb, J. A. Kohl, S. D. Miller, S. S. Vazhkudai Oak Ridge National Laboratory.
Data Acquisition at the NSLS II Leo Dalesio, (NSLS II control group) Oct 22, 2014 (not 2010)
EPICS Archiving Appliance Test at ESS
Nick Draper Teswww.mantidproject.orgwww.mantidproject.org Instrument Independent Reduction and Analysis at ISIS and SNS.
Artdaq Introduction artdaq is a toolkit for creating the event building and filtering portions of a DAQ. A set of ready-to-use components along with hooks.
Managed by UT-Battelle for the Department of Energy 1 Integrated Catalogue (ICAT) Auto Update System Presented by Jessica Feng Research Alliance in Math.
Chapter © 2006 The McGraw-Hill Companies, Inc. All rights reserved.McGraw-Hill/ Irwin Chapter 7 IT INFRASTRUCTURES Business-Driven Technologies 7.
The Red Storm High Performance Computer March 19, 2008 Sue Kelly Sandia National Laboratories Abstract: Sandia National.
Mantid Scientific Steering Committee Nick Draper 10/11/2010.
Grid Architecture William E. Johnston Lawrence Berkeley National Lab and NASA Ames Research Center (These slides are available at grid.lbl.gov/~wej/Grids)
Jamie Hall (ILL). SciencePAD Persistent Identifiers Workshop PANData Software Catalogue January 30th 2013 Jamie Hall Developer IT Services, Institut Laue-Langevin.
Chapter 5 McGraw-Hill/Irwin Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved.
Management of the LHCb DAQ Network Guoming Liu * †, Niko Neufeld * * CERN, Switzerland † University of Ferrara, Italy.
GRID Overview Internet2 Member Meeting Spring 2003 Sandra Redman Information Technology and Systems Center and Information Technology Research Center National.
Online Software 8-July-98 Commissioning Working Group DØ Workshop S. Fuess Objective: Define for you, the customers of the Online system, the products.
ESFRI & e-Infrastructure Collaborations, EGEE’09 Krzysztof Wrona September 21 st, 2009 European XFEL.
GCRC Meeting 2004 BIRN Coordinating Center Software Development Vicky Rowley.
Mantid Stakeholder Review Nick Draper 01/11/2007.
TeraGrid Gateway User Concept – Supporting Users V. E. Lynch, M. L. Chen, J. W. Cobb, J. A. Kohl, S. D. Miller, S. S. Vazhkudai Oak Ridge National Laboratory.
Experimental Facilities Division ORNL DAQ System Interfaces Neutron Science Software Workshop Oct 13, 2003 Rick Riedel.
Comprehensive Scientific Support Of Large Scale Parallel Computation David Skinner, NERSC.
CD FY09 Tactical Plan Status FY09 Tactical Plan Status Report for Neutrino Program (MINOS, MINERvA, General) Margaret Votava April 21, 2009 Tactical plan.
Computing at SSRL: Experimental User Support Timothy M. McPhillips Stanford Synchrotron Radiation Laboratory.
Mantid: A new approach to data analysis at large scale facilities Jon Taylor Project WIKI (Binary downloads)
ICAT Status Alistair Mills Project Manager Scientific Computing Department.
Data Management and Software Centre Mark Hagen Head of DMSC IKON7, September 15th 2014.
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
Electrical Engineering for NSS - Status and Planning- Thomas Gahl Electrical Engineering Group Leader February 26, 2014.
PEER 2003 Meeting 03/08/031 Interdisciplinary Framework Major focus areas Structural Representation Fault Systems Earthquake Source Physics Ground Motions.
Northwest Indiana Computational Grid Preston Smith Rosen Center for Advanced Computing Purdue University - West Lafayette West Lafayette Calumet.
ORNL is managed by UT-Battelle for the US Department of Energy Status Report: Data Acquisition and Instrument Controls for the Spallation Neutron Source.
Nigel Lockyer Fermilab Operations Review 16 th -18 th May 2016 Fermilab in the Context of the DOE Mission.
BrightnESS is funded by the European Union's Framework Programme for Research and Innovation Horizon 2020, under grant agreement Building a Research.
Instrument Construction Phase 1 and beyond Rob Connatser Chief Instrument Project Engineer September, 2014.
Fermilab Scientific Computing Division Fermi National Accelerator Laboratory, Batavia, Illinois, USA. Off-the-Shelf Hardware and Software DAQ Performance.
EGI-InSPIRE RI EGI Compute and Data Services for Open Access in H2020 Tiziana Ferrari Technical Director, EGI.eu
Data Acquisition, Diagnostics & Controls (DAQ)
Accessing the VI-SEEM infrastructure
WP18, High-speed data recording Krzysztof Wrona, European XFEL
Clouds , Grids and Clusters
MANAGEMENT OF STATISTICAL PRODUCTION PROCESS METADATA IN ISIS
Joseph JaJa, Mike Smorul, and Sangchul Song
Grid Portal Services IeSE (the Integrated e-Science Environment)
Recap: introduction to e-science
Pandata Service Verification
Computing Infrastructure for DAQ, DM and SC
Mirjam van Daalen, (Stephan Egli, Derek Feichtinger) :: Paul Scherrer Institut Status Report PSI PaNDaaS2 meeting Grenoble 6 – 7 July 2016.
Future Data Architecture Cloud Hosting at USGS
Scientific Data Management with MXCuBE
DMSC services Stig Skelboe
Mirjam van Daalen, (Stephan Egli, Derek Feichtinger) :: Paul Scherrer Institut Status Report PSI PaNDaaS2 meeting Grenoble 12 – 13 December 2016.
ICS update 7th Experiment Control Workshop
Defining the Grid Fabrizio Gagliardi EMEA Director Technical Computing
Afonso Mukai Scientific Software Engineer
Imaging & Engineering STAP Meeting 12th-13th of April 2018
Screaming Udder: data streaming at ISIS
Readout Systems Update
NOAA OneStop and the Cloud
Presentation transcript:

Data Management and Software Centre Mark Hagen Head of DMSC IKON6, February 27 th 2014

What is DMSC ? 2 o Data Management and Software Centre (DMSC) o A Division of ESS Science Directorate… Just like Instrument Technologies, Neutron Instruments etc. Two campuses: ESS Lund & ESS Copenhagen (Universitetsparken, Københavns Universitet) DMSC building to be constructed in Copenhagen o Responsibility: design, develop & implement for the ESS instruments: Software (user control interfaces, data acquisition, reduction & analysis) Hardware (servers, networks, workstations, clusters, disks, pfs etc.)

Organization 3 DMSC Data Systems & Technologies Inst. Control, Data Acq. & Reduction Data Management Data Analysis & Modeling User Office Software Copenhagen Data Centre DMSC servers in Lund Clusters, Workstations Disks, Parallel File System Networks (inc. Lund – CPH) Data transfer & Back-Up External Servers Instrument Control User Interfaces EPICS read/write Streaming data (ADARA) Data reduction (MANTID) File writers (ADARA) Data Catalogues Workflow Management Post-Processing………… Reduction ---- Analysis Messaging Services Web Interfaces Initially one work package (2 work units) MCSTAS support + dev. Instrument Integrators Analysis codes (e.g. SANSview, Rietveld,…) MD + DFT Framework User Database Proposal System Training Database Publications Database

4 ESS is an experimental facility – users will come here to do neutron scattering experiments: Collect the most (appropriate) data during their beam time Understand that data as much as possible/practical DMSC’s mission is to create a computational infrastructure that gives ESS users: The ability to acquire (capture) neutron & associated data & control the instrument The ability to reduce, and analyze, the data (as it is being acquired) Data files created instantly after acquisition (no matter how big) The ability to reduce a data set post-acquisition (could be 1 min, 1 month, 1 year later) in ~1 minute Archival, and cataloguing, of the data Access to the resources (software/hardware) for users to do post-acquisition reduction, analysis, visualization, modeling How do we make it happen? Mission

5 General Framework + Customization Generic Data Framework o Event mode data for later reprocessing/filtering o Stream the data  on the fly data processing (live view)  on the fly file creation  to a location where appropriate post-acq. resources are available o Create standard HDF5 data files (NeXus or other) o (Where possible) Automate data reduction & analysis o Catalogue the data & meta-data for fast processing Don’t re-invent the wheel  Existing projects: ICAT, MANTID, ADARA – Join these Customize for ESS Instruments o User Control Interface, detector type, IOC’s for sample environment etc.

6 Control Box Fast Sample Environment Detectors & Monitors Timing INTEGRATED CONTROL SYSTEMS NEUTRON TECHNOLOGIES Data Acquisition, Reduction & Control Data Aggregator & Streamer ( ) DATA MANAGEMENT & SOFTWARE CENTRE Bulk Data Readout (Detector Group) Lund Server Room Copenhagen Server Room User Control Interface Instrument Control Room Data Analysis Interfaces Automated Data Reduction Automated Data Reduction Live, Local & Remote Data Reduction Generic Framework o Instrument Control o Data Acquisition o Data Reduction & Visualization Choppers Sample Environment Motion Control

7 Data Acquisition, Streaming & Reduction Potential Collaborations Data Management Used by ESS accelerator/target, SLS, Diamond, US light sources, to be used by ISIS & SNS Publish/subscribe software & protocol for streaming data (neutron + meta) Data reduction framework in Python & C++ developed by ISIS & SNS ICAT data cataloguing software developed under NMI3 by PanData collaboration of 19 European facilities (+ SNS in US)

Data Systems & Technologies 8 o DMSC will not be running a “supercomputer” centre! o Data (disk) storage: Back of the envelope  2 – 3 PBytes/yr Spectrum of file sizes ~100MByte - ~10’s GByte – ~1TByte Fast disk (200MByte/s) & Parallel File System (10GByte/s) o Cluster(s) for data reduction & (modest) data analysis - ~2048 cores Architecture – CPU, GPU… visualization cluster/server o Interim Data Centre (funded by Denmark in pre-construction) 888 core cluster + ~100TByte Parallel File System Used by neutronics (target), instrument concepts, optics group

Data Systems & Technologies 9 o Archive – 2yrs on fast disk – 5yrs on disk - > 5yrs slow or ?? Back-up – somewhere else – supercomputer centre – commercial?? o Data download servers – sftp & gridftp o Remote login capability for ESS users: Re-reduce data using cluster Data analysis software available for users o Software development servers – repositories, bug trackers, build servers o DMSC could obtain a grant of supercomputer time for use by ESS users who need it for modeling

o Data analysis – surely all facilities need data analysis software ? Talking with ILL, ISIS, SINQ/PSI, JCNS Interoperability of software Modularization, file formats/converters… Economies of scale o Could enable new developments  Automated analysis  Streaming analysis  Feedback from analysis to data acquisition… o Surely every facility needs “User Office Software”  Are there opportunities for development of existing software (don’t re-invent the wheel) Potential Collaborations

11 Structure of Nanomaterials o Data on disk is useless! It is published results from the data that makes progress o Need to ensure that ESS users have access to appropriate software packages for data analysis the necessary computational resources to exploit the software to obtain those results analysis software during experiment to influence the data taking strategies o Roll out in-sync with instruments Data Analysis Polarized SANS demonstrated that these nanoparticles have uniform nuclear structure but core-shell magnetic structure. Required development of both data reduction and data analysis methods and tools. K. L. Krycka et al. Core Shell Magnetic Morphology of Structurally Uniform Magnetite Nanoparticles PRL 104, (2010)

Data Analysis Development Original force-field Optimized force-field Integration of Analysis with Advanced Modeling Techniques Molecular Dynamics & Density Functional Theory (DFT) Techniques Jose Borreguero (SNS/NDAV) Mike Crawford (Dupont) & Niina Jalarvo (Julich): BASIS experiment / MD simulation studies of Methyl rotations in methyl-Polyhedral oligomeric silsesquioxanes (POSS) Optimization of dihedral potential governing the rotational jumps in methyl hydrogens. Preliminary simulations indicate that the dihedral potential barrier must be decreased ~37% in order to match experimental data

13 o Primary goal: To be ready for hot commissioning of first ESS instruments in o Includes significant amount of NRE (Non- Recurrent Engineering) for subsequent ESS instruments. o Establish the basic facilities at both campuses. o Front loaded with instrument-centric work + scope to grow analysis in an integrated way During the construction years

14 Conclusion o DMSC’s mission is the computational (software/hardware) infrastructure for the ESS data chain - instrument control, data acquisition, reduction & analysis o Generic framework customized for each of the instruments – utilizes data streaming to account for large data files/live processing o Don’t re-invent the wheel – work with collaborators ISIS, SNS, PanData (+ others?) to develop/customize existing software – EPICS, ADARA, MANTID, ICAT… o Customize for each of the ESS instruments in-sync as they roll out o In-sync with instrument roll out work with instrument teams to ensure analysis software is available (doesn’t necessarily mean develop) o Longer term goal to develop new data analysis methods for ESS instruments

Questions QUESTIONS