High Speed Detectors at Diamond Nick Rees. A few words about HDF5 PSI and Dectris held a workshop in May 2012 which identified issues with HDF5: –HDF5.

Slides:



Advertisements
Similar presentations
Buffers & Spoolers J L Martin Think about it… All I/O is relatively slow. For most of us, input by typing is painfully slow. From the CPUs point.
Advertisements

Operating System.
areaDetector Developments
Fast Communication Firefly RPC Lightweight RPC  CS 614  Tuesday March 13, 2001  Jeff Hoy.
AreaDetector Data Processing Pipeline In EPICS V4 Dave Hickin Diamond Light Source EPICS Collaboration Meeting Diamond Light Source 01/05/2013.
ACAT 2002, Moscow June 24-28thJ. Hernández. DESY-Zeuthen1 Offline Mass Data Processing using Online Computing Resources at HERA-B José Hernández DESY-Zeuthen.
COM S 614 Advanced Systems Novel Communications U-Net and Active Messages.
Operating Systems.
Challenges choosing storage systems for experimental data Nick Rees.
Sven Ubik, Petr Žejdl CESNET TNC2008, Brugges, 19 May 2008 Passive monitoring of 10 Gb/s lines with PC hardware.
Large scale data flow in local and GRID environment V.Kolosov, I.Korolko, S.Makarychev ITEP Moscow.
Data Acquisition at the NSLS II Leo Dalesio, (NSLS II control group) Oct 22, 2014 (not 2010)
Higher Computing Computer Systems S. McCrossan 1 Higher Grade Computing Studies 4. Peripherals Input Devices Keyboard Mouse Scanners Microphone Digital.
ECE 526 – Network Processing Systems Design Network Processor Architecture and Scalability Chapter 13,14: D. E. Comer.
What is a Protocol A set of definitions and rules defining the method by which data is transferred between two or more entities or systems. The key elements.
Operating System. Architecture of Computer System Hardware Operating System (OS) Programming Language (e.g. PASCAL) Application Programs (e.g. WORD, EXCEL)
CDF data production models 1 Data production models for the CDF experiment S. Hou for the CDF data production team.
Chapter Three Network Protocols By JD McGuire ARP Address Resolution Protocol Address Resolution Protocol The core protocol in the TCP/IP suite that.
Uncovering the Multicore Processor Bottlenecks Server Design Summit Shay Gal-On Director of Technology, EEMBC.
Cluster Computers. Introduction Cluster computing –Standard PCs or workstations connected by a fast network –Good price/performance ratio –Exploit existing.
The HDF Group Multi-threading in HDF5: Paths Forward Current implementation - Future directions May 30-31, 2012HDF5 Workshop at PSI 1.
Processes and OS basics. RHS – SOC 2 OS Basics An Operating System (OS) is essentially an abstraction of a computer As a user or programmer, I do not.
Data and Computer Communications Circuit Switching and Packet Switching.
Selling the Storage Edition for Oracle November 2000.
ATCA based LLRF system design review DESY Control servers for ATCA based LLRF system Piotr Pucyk - DESY, Warsaw University of Technology Jaroslaw.
PFPC: A Parallel Compressor for Floating-Point Data Martin Burtscher 1 and Paruj Ratanaworabhan 2 1 The University of Texas at Austin 2 Cornell University.
Frontiers in Massive Data Analysis Chapter 3.  Difficult to include data from multiple sources  Each organization develops a unique way of representing.
Detectors for Light Sources Contribution to the eXtreme Data Workshop of Nicola Tartoni Diamond Light Source.
DOE PI Meeting at BNL 1 Lightweight High-performance I/O for Data-intensive Computing Jun Wang Computer Architecture and Storage System Laboratory (CASS)
Computer Security Workshops Networking 101. Reasons To Know Networking In Regard to Computer Security To understand the flow of information on the Internet.
DYNES Storage Infrastructure Artur Barczyk California Institute of Technology LHCOPN Meeting Geneva, October 07, 2010.
Systems and Communications1 CERN Openlab – 10/11 December 2013 ESRF: IT Challenges.
EGEE is a project funded by the European Union under contract IST HEP Use Cases for Grid Computing J. A. Templon Undecided (NIKHEF) Grid Tutorial,
Clara Gaspar, March 2005 LHCb Online & the Conditions DB.
Next Generation Operating Systems Zeljko Susnjar, Cisco CTG June 2015.
“Live” Tomographic Reconstructions Alun Ashton Mark Basham.
Petra III Status Teresa Núñez Hasylab-DESY Tango Meeting DESY,
AreaDetector workshop Summary and notes…. Overview iocBuilder File Writer Changes areaDetector driver pattern shared vector Asyn client without an IOC.
Data Recording Model at XFEL CRISP 2 nd Annual meeting March 18-19, 2013 Djelloul Boukhelef 1Djelloul Boukhelef - XFEL.
Modeling Big Data Execution speed limited by: –Model complexity –Software Efficiency –Spatial and temporal extent and resolution –Data size & access speed.
IPHC - DRS Gilles CLAUS 04/04/20061/20 EUDET JRA1 Meeting, April 2006 MAPS Test & DAQ Strasbourg OUTLINE Summary of MimoStar 2 Workshop CCMOS DAQ Status.
Sep. 17, 2002BESIII Review Meeting BESIII DAQ System BESIII Review Meeting IHEP · Beijing · China Sep , 2002.
CIS250 OPERATING SYSTEMS Chapter One Introduction.
Snapshot of DAQ challenges for Diamond Martin Walsh.
3/12/2013Computer Engg, IIT(BHU)1 PARALLEL COMPUTERS- 2.
Cluster Computers. Introduction Cluster computing –Standard PCs or workstations connected by a fast network –Good price/performance ratio –Exploit existing.
CODA Graham Heyes Computer Center Director Data Acquisition Support group leader.
NI Big Physics Summit CERN
Markus Frank (CERN) & Albert Puig (UB).  An opportunity (Motivation)  Adopted approach  Implementation specifics  Status  Conclusions 2.
Report on PaNdata Work Package 8 Scalability Acknowledgements: Mark Basham, Nick Rees, Tobias Richter,Ulrik Pedersen, Heiner Billich et al (PSI/SLS), Frank.
CRISP WP18, High-speed data recording Krzysztof Wrona, European XFEL PSI, 18 March 2013.
Using Deduplicating Storage for Efficient Disk Image Deployment Xing Lin, Mike Hibler, Eric Eide, Robert Ricci University of Utah.
Dr Andrew Peter Hammersley ESRF ESRF MX COMPUTIONAL AND NETWORK CAPABILITIES (Note: This is not my field, so detailed questions will have to be relayed.
Scalable Readout System Data Acquisition using LabVIEW Riccardo de Asmundis INFN Napoli [Certified LabVIEW Developer]
ORNL is managed by UT-Battelle for the US Department of Energy ADnED Handling V4 Neutron Event Data Matt Pearson SNS ORNL 18 th -22 nd May 2015.
Virtual Machine Monitors
Modeling Big Data Execution speed limited by: Model complexity
WP18, High-speed data recording Krzysztof Wrona, European XFEL
HDR Handling at EMBL-HH Status Report
Eiger-Related Development at
Operating System.
Introduction to HDF5 Session Five Reading & Writing Raw Data Values
Enrico Gamberini, Giovanna Lehmann Miotto, Roland Sipos
Spark Presentation.
Status of the Merlin Readout System
Virtualization, Cloud Computing and Big Data
Algorithms for Big Data Delivery over the Internet of Things
(Dectris Eiger) HDF5 Stream Writer
Chapter 2: The Linux System Part 1
Lecture 4: Instruction Set Design/Pipelining
Presentation transcript:

High Speed Detectors at Diamond Nick Rees

A few words about HDF5 PSI and Dectris held a workshop in May 2012 which identified issues with HDF5: –HDF5 was (is) single threaded Makes it difficult to benefit from parallel systems. –pHDF5 could not store compressed data Compression is essential to reduce the data sizes. –HDF5 files could not be open for read and write simultaneously. Analysis software could not run while data taking

What did we do? Direct chunk write –Funded by Dectris and PSI –Released in HDF5 version in May 2013 Dynamically loaded filters –Funded by DESY –Also released in HDF5 version Single Writer/Multiple Reader –Funded by Diamond, ESRF and Dectris –Available in HDF scheduled for March 2016 Virtual data set. –Funded by Diamond, DESY and the Percival project. –Also available in HDF

Diamond History Early 2007: –Diamond first user. –No detector faster than ~10 MB/sec. Early 2009: –first Lustre system (DDN S2A9900) –first Pilatus 6M 60 MB/s. Early 2011: –second Lustre system (DDN SFA10K) –first 25Hz Pilatus 6M MB/s. Early 2013: –first GPFS system (DDN SFA12K) –First 100 Hz Pilatus 6M 600 MB/sec –~10 beamlines with 10 GbE detectors (mainly Pilatus and PCO Edge). 2016: –Percival and parallel Eiger 16M (5000 MB/sec). Doubling time = 7.5 months

Diamond Parallel Detectors Excalibur –Medipix sensor array developed with STFC 2048x1532, 3M, 100 Hz, 5 Gb/s Uses 6x1 Gbps servers. –Commissioned in Percival –Low energy MAPS array with 3 proposed configurations: 1484x1408 2M, 300 Hz, 20 Gb/s 3710x M, 120 Hz, 50 Gb/s –First system in 2016 Eiger (proposal) –Talking to Dectris about combining 4x4M sensors ~200 Gb/s uncompressed data rate. Compression works better at high data rates because there are fewer photons/frame. Scientific benefit is to be able to monitor sample degradation. –Delivery 2016

Diamond Data Flow Model

General Parallel Detector Design Transfer between FPGA and processing nodes is UDP so need to be careful about handling packets Switch has deep buffers so links can have momentary (factor 10) overloads – tested HP 5920 Deep-Buffer Switch. Processing nodes all write in parallel to parallel file system. Use VDS to map parallel data streams into one dataset. Typically LVDS SensorFPGA Readout boards Front End 1 Front End 2 10Gbit/s links Switch 10Gbit/s links Processing nodesFile storage 10Gbit/s links PC servers Parallel writing of HDF5 files PC

Temporal and spatial processing distribution Computer 1 Frame 1 Frame 2 Frame 3 Frame 4 Computer 2Computer 3Computer 4

Data Receiver Readout Hardware (n copies) Data Processing: 2 bit gain handling DCS subtraction Pixel re-arrangement Rate correction(?) Flat field Dark subtraction Efficiency correction Data Processing: 2 bit gain handling DCS subtraction Pixel re-arrangement Rate correction(?) Flat field Dark subtraction Efficiency correction Data Compression HDF5 File Writer Detector API Control and Data Protocol Control Hardware Configuration Control Server Control Driver Cmd Status HDF5 file Python/ Matlab Tango/ Lima EPICS/ Area Detector Detector Control Software Detector Data Stream (n copies) Controlled Interfaces Beamline Control Software Detector Engineer Software Key Actual/potential network or CPU socket boundaries

Use of existing features - chunking areaDetector is well suited to high speed detectors –multi-threading to take advantage of modern CPU’s –support for high performance libraries such as hdf5. hdf5 file support provides –complex data structures –ability to store multiple frames in one file, reducing file system overhead. –parallel writing –chunking, to optimise read back speeds in, for example, tomography

Why the new features are important Direct chunk write –In library compression is limited to ~500 MB/sec/process of raw data. –Direct chunk write can write at 2 GB/sec/process of compressed data. Dynamically loaded filters –Dectris has been developing LZ4 based filters that are very well suited to zero noise diffraction detectors. –Compression ratios typically a factor of 10. Single Writer/Multiple Reader –Particularly useful for slow data taking Spectroscopic mapping experiments can take several hours and needs immediate feedback. –Now can use off-line data analysis software on-line Virtual data set –Main use is for high speed parallel data writing Datasets can be compressed as well as written in parallel. –Supports different write patterns for different detectors. –Can be used for combining sparse scans in larger datasets.

What next? Most Exascale developments are looking at object storage systems rather than traditional file systems –Intel is working with LLNS and The HDF Group on Extreme-Scale Computing Research and Development (Fast Forward) Storage and I/O. –Seagate is working on ProjectSAGE in a consortium of 10 companies. Both aim match processing and storage so processing can happen closer to storage in data dominated applications. Both are looking at special interfaces to HDF5

Summary Detectors are challenging our ways of taking and storing data. We continue to work with The HDF Group to enhance HDF5 for our applications. Open Source Software is not free. –We need to sort out the HDF5 support model