Snapshot of DAQ challenges for Diamond Martin Walsh.

Slides:



Advertisements
Similar presentations
Founded in 2010: UCL, Southampton, Oxford and Bristol Key Objectives of the Consortium: Prove the concept of shared, regional e-infrastructure services.
Advertisements

ICAT + Information Model Brian Matthews Scientific Information Group E-Science Centre STFC Rutherford Appleton Laboratory
STFC and the UK e-Infrastructure Initiative The Hartree Centre Prof. John Bancroft Project Director, the Hartree Centre Member, e-Infrastructure Leadership.
Slide: 1 Welcome to the workshop ESRFUP-WP7 User Single Entry Point.
1 SESAME as International Research Lab by: Salman Matalgah Synchrotron-light for Experimental Science and Applications in the Middle.
NSLS-II Life Science Breakout Session. Agenda  Introduction (Miller)  Keynote Speaker: Carolyn Larabell (ALS, UCSF) (30 min)  Technique talks (4 min.
Challenges choosing storage systems for experimental data Nick Rees.
1 ECE 453 – CS 447 – SE 465 Software Testing & Quality Assurance Case Studies Instructor Paulo Alencar.
Data Analysis I19 Upgrade Workshop 11 Feb Overview Short history of automated processing for Diamond MX beamlines Effects of adding Pilatus detectors.
State supporting science and technology challenge. Measures for modern scientific landscape. By Lyudmila Ogorodova Deputy Minister, Ministry of Education.
Effective User Services for High Performance Computing A White Paper by the TeraGrid Science Advisory Board May 2009.
Shaping Future Science Introducing the National Science and Innovation Campuses.
Integrated e-Infrastructure for Scientific Facilities Kerstin Kleese van Dam STFC- e-Science Centre Daresbury Laboratory
Chipir A new neutron single event test facility at the UK’s ISIS Neutron and Muon Source.
Results of the HPC in Europe Taskforce (HET) e-IRG Workshop Kimmo Koski CSC – The Finnish IT Center for Science April 19 th, 2007.
Chapter 4 Realtime Widely Distributed Instrumention System.
Context and Linking in the Research Lifecycle CERIF and other standards Catherine Jones Scientific Information Group Scientific Computing Department STFC.
Systems and Communications1 CERN Openlab – 10/11 December 2013 ESRF: IT Challenges.
Dept of Materials Engineering l Use of a Virtual Laboratory to plan, execute and analyse Neutron Strain Scanning experiments. John A James 1, Javier R.
WP5 – Virtual Laboratories. WP5 Deliverables  D5.1: Specific requirements for the virtual laboratories M6  D5.2: Deployment of Specification of the.
Introduction to Synchrotron Radiation
Next Generation Operating Systems Zeljko Susnjar, Cisco CTG June 2015.
“Live” Tomographic Reconstructions Alun Ashton Mark Basham.
E-Science Research Councils awarded e-Science funds ” science increasingly done through distributed global collaborations enabled by the Internet, using.
ESFRI & e-Infrastructure Collaborations, EGEE’09 Krzysztof Wrona September 21 st, 2009 European XFEL.
Project Database Handler The Project Database Handler is a brokering application that mediates interactions between the project database and the external.
1 Recommendations Now that 40 GbE has been adopted as part of the 802.3ba Task Force, there is a need to consider inter-switch links applications at 40.
High Speed Detectors at Diamond Nick Rees. A few words about HDF5 PSI and Dectris held a workshop in May 2012 which identified issues with HDF5: –HDF5.
Workshop Overview, Charge and Deliverables Joachim Stöhr.
Towards a Structural Biology Work Bench Chris Morris, STFC.
Steve Worm – Gateway CentresApril 25, The Gateway Centres and their context in the STFC Strategy Steve Worm Rutherford Appleton Laboratory.
Using Technology to Study Cellular and Molecular Biology.
ATTRACT is a proposal for an EU-funded R&D programme for sensor, imaging and related computing devlopment Its purpose is to demonstrate the value of European.
3D-ASICS for X-ray Science
Sensors and Instrumentation Computational and Data Challenges in Environmental Modelling Dr Peter M Allan Director, Hartree Centre, STFC.
RC ICT Conference 17 May 2004 Research Councils ICT Conference The UK e-Science Programme David Wallace, Chair, e-Science Steering Committee.
UK X-FEL National Laboratory Perspective Susan Smith STFC ASTeC IoP PAB/STFC Workshop Towards a UK XFEL 16 th February 2016.
WHAT SURF DOES FOR RESEARCH SURF’s Science Engagement TNC15 June 18, 2015 Sylvia Kuijpers (SURFnet)
Supporting the “Solving Business Problems with Environmental Data” Competition 24 th October 2013 Vlad Stoiljkovic.
ESS TDR Internal Review Chapter and 12. October 2012 Oliver Kirstein.
Develop a Robust User Community Sarah Fisher STFC.
STFC’s National Laboratories Round Table on Synergies and Complementarity among Laboratories John Womersley Chief Executive, STFC 13 th Pisa meeting on.
Nigel Lockyer Fermilab Operations Review 16 th -18 th May 2016 Fermilab in the Context of the DOE Mission.
Usecases: 1.ISIS Neutron Source 2.DP for HEP Matthew Viljoen STFC, UK APARSEN-EGI workshop: preserving big data for research Amsterdam Science Park 4-6.
The Helmholtz Association Project „Large Scale Data Management and Analysis“ (LSDMA) Kilian Schwarz, GSI; Christopher Jung, KIT.
Dr Andrew Peter Hammersley ESRF ESRF MX COMPUTIONAL AND NETWORK CAPABILITIES (Note: This is not my field, so detailed questions will have to be relayed.
E-infrastructure requirements from the ESFRI Physics, Astronomy and Analytical Facilities cluster Provisional material based on outcome of workshop held.
ORNL is managed by UT-Battelle for the US Department of Energy OLCF HPSS Performance Then and Now Jason Hill HPC Operations Storage Team Lead
A U.S. Department of Energy laboratory managed by UChicago Argonne, LLC. Introduction APS Engineering Support Division –Beamline Controls and Data Acquisition.
Project Database Handler The Project Database Handler is a brokering application which will mediate interactions between the project database and other.
An short overview of INSTRUCT and its computational requirements Alexandre M.J.J. Bonvin Project coordinator Bijvoet Center for Biomolecular Research Faculty.
Capabilities and Programmes of STFC’s Accelerator Science & Technology Centre (ASTeC)
National SYNCHROTRON Radiation Centre SOLARIS Jagiellonian University, Krakow, Poland Piotr Goryl Motion Workshop, Soleil,
Budget JRA2 Beneficiaries Description TOT Costs incl travel
MSc-Student Activities at the European XFEL
Joslynn Lee – Data Science Educator
Data interpretation Creative Biostructure now can provide a collection of tools and analysis package for biomolecular structure determination, refinement.
Data Management & Analysis in MATTER
Graeme Winter STFC Computational Science & Engineering
Centre d’Excellence en Technologies de l’Information et de la Communication EC Workshop Delivery of industrial-strength Grid middleware: establishing an.
Data processing Creative Biostructure now can provide a collection of tools and analysis package for biomolecular structure determination, refinement and.
Data Processing Data Processing and Interpretation in Creative Biostructure.
DIGITAL LIBRARY.
EOSCpilot All Hands Meeting 8 March 2018 Pisa
MAX IV Laboratory National Synchrotron Light Source – two storage rings (1.5 & 3.0 GeV) and a Short Pulse Facility Characteristics of User Program: 15.
Physics Dept. Cornell University
Brian Matthews STFC EOSCpilot Brian Matthews STFC
DMSC services Stig Skelboe
Imaging & Engineering STAP Meeting 12th-13th of April 2018
Pitch Deck.
Presentation transcript:

Snapshot of DAQ challenges for Diamond Martin Walsh

Your role in all this... -SAC is the advisory body for Diamond: how do we make the most of your collective knowledge, experience and wisdom ? -Providing the information you need -Organisation and content of meetings -The most effective forum for discussion -Efficient transmission of your advice to us -Informing you how Diamond has acted on your advice – and what the result has been Harwell Campus ISIS CLF RAL Space Mary Lyon centre mouse functional genomics International Space Innovation Centre (ISIC) MRC Harwell Rutherford Appleton Laboratory Public Health England The European Centre for Space Applications and Telecommunications (ECSAT) Research Complex

Beamlines by Village Macromolecular Crystallography Soft Condensed Matter Spectroscopy Materials Engineering and Environment Surfaces and Interfaces eBIC

< 100GB/day < 1TB/day > 1TB/day Per Beamline Data Rates Tomography Beamlines have collected nearly 2PB of data, more than the rest of Diamond beamlines put together. New Arrival EM 2  5TB/day  XFEL from 2017

Some Numbers for Total number of user proposals: 642, delivered shifts 7964 (1 shift =8 hours) Total number of users 7,696 –(4,988 on site remote) –MX remote use now exceeds 50% of use Total number of Unique PhD’s 857 Total Journal papers published 3,883 (677 published in 2014)

Increasing resolution CryoET Single particle cryo EM X-ray crystallography B21 X-ray Solution Scattering Cryo-electron tomography One major player – integrated Structural biology Increasing biological complexity and integrity Fluorescence microscopy (B24/CLF) B24 X-ray microscopy Cellular cryo-electron tomography B22 - Infrared microspectroscopy

Cell/tissueSolutionCrystalline Electron Microscopy XFEL Life Science & DLS B22 Infrared B24 Cryo X-ray microscopy I18, I20, B18, I14 X-ray spectroscopy I22/B21 SAXS B23 CD Spectroscopy MX village ( I02, I03, I04) (I24, I04-1) (I23, VMX) National facility for EM in life & Physical Sciences UK Hub for XFEL sample and software developments I08 X-ray STXM I13 X-ray tomography & coherent diffraction

Cell Biology OPPF-UK MPL Cell Biology OPPF-UK MPL Diamond Beamlines: Macromolecular Crystallography, Scattering, X-ray spectroscopy ISIS beamlines: SANS Neutron Reflection (NR) Diamond Beamlines: Macromolecular Crystallography, Scattering, X-ray spectroscopy ISIS beamlines: SANS Neutron Reflection (NR) Computational environment / CCP4, CCP-EM HPC Synchrotron Imaging UK XFEL Cryo-EM/ET Electron Bio- Imaging Centre (eBIC) Cryo-EM/ET Electron Bio- Imaging Centre (eBIC) An integrated Approach to Structural Biology Fluorescence microscopy (CLF(STFC & DLS) Fluorescence microscopy (CLF(STFC & DLS)

Diamond Data Rates/Volumes History Early 2007: –Diamond first user. –No detector faster than ~10 MB/sec. Early 2009: –first Lustre system (DDN S2A9900) –first Pilatus 6M 60 MB/s. Early 2011: –second Lustre system (DDN SFA10K) –first 25Hz Pilatus 6M MB/s. Early 2013: –first GPFS system (DDN SFA12K) –First 100 Hz Pilatus 6M 600 MB/sec –~10 beamlines with 10 GbE detectors (mainly Pilatus and PCO Edge). Early 2016: –delivery of Eiger 16M for MX (initially at 6.75 GB/s, potential 13.5 GB/s) Doubling time = 7.5 months

Challenges Hardware life cycles are fast, and hardware problems can be solved with sufficient money. –So detector data rates are not the problem. Software life cycles are slow – our analysis routines have a clear lineage often dating back 40 years. –Software is a problem Synchrotron have to support a diverse range of techniques. –Systems and skills developed for one beamline are not appropriate for all beamlines. –Need to be able to attract talented software scientists AND software engineers Remote access to large scale facilities such as synchrotrons, XFELs, national facilities (e.g. Electron microscopy, HPC) –Need for dedicated light paths between these facilities to deal with data volumes generated

Use CASE: Structural Biology Numbers: –Raw Data Macromolecular crystallography Currently DIAMOND ( 3-6TB/day) 2016 – detector technology will enable easily X10 increase in data. Upgrades to beamlines will enable better exploitation (hardware can currently produce >100TB/day if samples available. Near future expect to produce 5PB MX data/year –this is at DIAMOND ALONE! Including SR MX beamlines over Europe expect to reach/exceed (25PB/year) European XFEL – SFX beamline at full operations potential to generate 300 TB/day –Raw Data Electron microscopy Electron Microscopy currently at 1TB/day/microscope – high resolution experiments to start in November ‘16 which will produce >5TB/day/microscope High resolution EM work from Jan 2016 will generate >10TB/day of data –Data reduction and analysis Requirement for Light paths to be established between large scale and national facilities – SR, CryoEM, Data Centers etc UK will have dedicated light path from European XFEL SFX beamline to DIAMOND – plans needed to extend... Currently large investment in software for data analysis is required to exploit developments in parallelized systems /new HPC storage

The future A lot of software will need to be redeveloped: –Incorporate modern paradigms like map-reduce. –May include middle layers processing that runs close to distributed data chunks. –Intermediate data will be cached between processing steps. Synchrotrons/Structural biology infrastructure will become turnkey sites. –Users may not come to site/facility –Results will be in the form of processed, not raw data. –There must be trust between the site and the user, backed up by data provenance and full metadata. –High speed light links between centers required.

Overview -First Impressions - Science highlights - Technical developments - Industrial engagement - Plans for the future - Finance Thanks for your attention

Example data access rates Example of 12GB/s …typically at 3-4GB/s

Tomography rates

MX/EM data storage 2015