Eiger at the Australian Synchrotron

Slides:



Advertisements
Similar presentations
Buffers & Spoolers J L Martin Think about it… All I/O is relatively slow. For most of us, input by typing is painfully slow. From the CPUs point.
Advertisements

Alastair Dewhurst, Dimitrios Zilaskos RAL Tier1 Acknowledgements: RAL Tier1 team, especially John Kelly and James Adams Maximising job throughput using.
ONYX RIP Version Technical Training General. Overview General Messaging and What’s New in X10 High Level Print and Cut & Profiling Overviews In Depth.
Report of Liverpool HEP Computing during 2007 Executive Summary. Substantial and significant improvements in the local computing facilities during the.
Click Here to Begin. Objectives Purchasing a PC can be a difficult process full of complex questions. This Computer Based Training Module will walk you.
Cacti Workshop Tony Roman Agenda What is Cacti? The Origins of Cacti Large Installation Considerations Automation The Current.
1 CS 501 Spring 2005 CS 501: Software Engineering Lecture 22 Performance of Computer Systems.
Cluster Components Compute Server Disk Storage Image Server.
THE CPU Cpu brands AMD cpu Intel cpu By Nathan Ferguson.
Using Virtualization in the Classroom. Using Virtualization in the Classroom Session Objectives Define virtualization Compare major virtualization programs.
COMPUTER CONCEPTS.
Digital Graphics and Computers. Hardware and Software Working with graphic images requires suitable hardware and software to produce the best results.
Fundamentals of Networking Discovery 1, Chapter 2 Operating Systems.
Computer communication
TPB Models Development Status Report Presentation to the Travel Forecasting Subcommittee Ron Milone National Capital Region Transportation Planning Board.
About the Presentations The presentations cover the objectives found in the opening of each chapter. All chapter objectives are listed in the beginning.
Google MapReduce Simplified Data Processing on Large Clusters Jeff Dean, Sanjay Ghemawat Google, Inc. Presented by Conroy Whitney 4 th year CS – Web Development.
Servers Last Update Copyright Kenneth M. Chipps Ph.D. 1.
How to buy a PC Brad Leach David Howarth James Sawruk Andrew U.
INFO1 – Practical problem solving in the digital world
Using Virtualization in the Classroom. Using Virtualization in the Classroom Session Objectives Define virtualization Compare major virtualization programs.
Planning and Designing Server Virtualisation.
Ch 6. Performance Rating Windows 7 adjusts itself to match the ability of the hardware –Aero Theme v. Windows Basic –Gaming features –TV recording –Video.
David Hutchcroft on behalf of John Bland Rob Fay Steve Jones And Mike Houlden [ret.] * /.\ /..‘\ /'.‘\ /.''.'\ /.'.'.\ /'.''.'.\ ^^^[_]^^^ * /.\ /..‘\
Comp 335 – File Structures Why File Structures?. Goal of the Class To develop an understanding of the file I/O process. Software must be able to interact.
EPICS Developments at the Australian Synchrotron DSP EPICS driver for the General Standards 16AIO analog card EPICS driver for the Galil range of motor.
Block1 Wrapping Your Nugget Around Distributed Processing.
Hardware. Make sure you have paper and pen to hand as you will need to take notes and write down answers and thoughts that you can refer to later on.
Virtualization for the LHCb Online system CHEP Taipei Dedicato a Zio Renato Enrico Bonaccorsi, (CERN)
“Live” Tomographic Reconstructions Alun Ashton Mark Basham.
Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Implementation of a reliable and expandable on-line storage for compute clusters Jos van Wezel.
High Speed Detectors at Diamond Nick Rees. A few words about HDF5 PSI and Dectris held a workshop in May 2012 which identified issues with HDF5: –HDF5.
COMMON INTERFACE FOR EMBEDDED SOFTWARE CONFIGURATION by Yatiraj Bhumkar Advisor Dr. Chung-E Wang Department of Computer Science CALIFORNIA STATE UNIVERSITY,
 The End to the Means › (According to IBM ) › 03.ibm.com/innovation/us/thesmartercity/in dex_flash.html?cmp=blank&cm=v&csr=chap ter_edu&cr=youtube&ct=usbrv111&cn=agus.
Randy MelenApril 14, Stanford Linear Accelerator Center Site Report April 1999 Randy Melen SLAC Computing Services/Systems HPC Team Leader.
Next Generation of Apache Hadoop MapReduce Owen
Extending Auto-Tiering to the Cloud For additional, on-demand, offsite storage resources 1.
Dr Andrew Peter Hammersley ESRF ESRF MX COMPUTIONAL AND NETWORK CAPABILITIES (Note: This is not my field, so detailed questions will have to be relayed.
Canadian Bioinformatics Workshops
ESS Timing System Plans Timo Korhonen Chief Engineer, Integrated Control System Division Nov.27, 2014.
3rd July 2007Open Repository, Open Source 1 Where we were, What we've learnt, And what it means to you.
Using Virtualization in the Classroom
Compute and Storage For the Farm at Jlab
Unit 2 Technology Systems
Integrating Disk into Backup for Faster Restores
Title of the Poster Supervised By: Prof.*********
Big Data is a Big Deal!.
Experience of Lustre at QMUL
WP18, High-speed data recording Krzysztof Wrona, European XFEL
HDR Handling at EMBL-HH Status Report
Eiger-Related Development at
Tango Integration of Modern 2D Detectors
Enrico Bonaccorsi, (CERN) Loic Brarda, (CERN) Gary Moine, (CERN)
Experience of Lustre at a Tier-2 site
How can a detector saturate a 10Gb link through a remote file system
Статус ГРИД-кластера ИЯФ СО РАН.
(Dectris Eiger) HDF5 Stream Writer
Graeme Winter STFC Computational Science & Engineering
Hadoop Clusters Tess Fulkerson.
Computing Infrastructure for DAQ, DM and SC
TYPES OFF OPERATING SYSTEM
Autoprocessing updates at the MX beamlines
Migration Strategies – Business Desktop Deployment (BDD) Overview
INFO 344 Web Tools And Development
Hyper-Converged Technology a User's Perspective
Specifications Clean & Match
CS246: Search-Engine Scale
LO2 – Understand Computer Software
CS 295: Modern Systems Organizing Storage Devices
Introduction to research computing using Condor
Presentation transcript:

Eiger at the Australian Synchrotron Tom Caradoc-Davies, Jun Aishima, David Aragao Australian Synchrotron / Monash University Melbourne, Australia

Eiger - First beam, Friday, 3rd February General user operations from 18/2/17

Eiger Cover – ADS Engineering Roller cover Interlocked to hutch search state Working very well

Eiger for direct beam measurement. Used Eiger in 4M ROI at 750Hz for direct beam tests. Adjusting cryocooler settings fixed split beam issue. Also used to measure incident flux but issues with count rate correction gave 1.9% stdev in Io. Counts from 1500-2500 per pixel (333e6 cts/sec/mm2). Used oil in loop and scatter instead of direct beam to get more accurate data. Counts per pixel 25-45 (6e6 cts/sec/mm2). Now stdev 0.243% on Io for channel-cut crystal and 0.305% for double-crystal set. Matches other diagnostics. Eiger a very powerful tool for measuring beam at sample (shape, position and intensity).

MX2 collection and processing system outline Collect Server Python Wrapper Oscillation AD eiger DCU & Eiger QEGui Epics Running on a galil 4080 HW triggering on omega encoder SIMPLON API PyEPICS ZMQ, via python BL library

Screening A new screening method is being tested Small wedges of data are collected with wider delta (angle per image) Processed on same nodes as regular datasets with same code, with higher priority Aim for processing stats in less than a minute to help user make judgements about full data collection Aim to provide angle/image (ideally 0.1 * mosaicity/image), dose, resolution Pros: Reliable indexing In general matches stats in dataset (I/sigI, Rmerge etc) Cons: Slow. Between 30-60 seconds for results Dose.

Software used to process data at the MX beamlines Custom pipeline software – Nathan Mudie - https://github.com/AustralianSynchrotron/mx-auto-dataset Custom web application software – Nathan Mudie - https://github.com/AustralianSynchrotron/mx-sample-manager xdsme – my upgrades to add handling inverted rotation axis, improve integration speed, etc.- (BSD license following from original) https://github.com/JunAishima/xdsme Pierre Legrand (Soleil) (BSD license) https://github.com/legrandp/xdsme XDS http://xds.mpimf-heidelberg.mpg.de/ Pointless, Aimless (CCP4) Sadabs and Xprep for CX data

Dectris’s solution - Eiger Processing Unit A computer able to retrieve data from the detector via DCU, write it to disk (/data), able to write onto other external disk drives, and ability to do some processing 4x E7-8870 v4 – 2.1 GHz, 20 cores each (40 hyperthreaded) 755 GB RAM – 512 GB can be used for a Ramdisk to store data pulled from DCU. 16 TB RAID spinning disk storage – meant to hold 1-5 days’ worth of data We have # 002!

Benchmarking the EPU 9k image lysozyme dataset collected at the SLS – kindly provided for us by Andreas Forster (Dectris) More representative of fine sliced experiments than the publicly available 16M 900 image insulin dataset or 9M 1800 image transthryretin dataset Goal was to be able to process a dataset (collected at the maximum rate) in less than 4x the data collection time when using 3 of the 4 CPUs This was verified by Dectris and us using the 9k image dataset and 120 threads. Dectris have recommended that 16-40 threads be reserved for Furka/Grimsel Comparison with two Dell 730 servers (node1 and node2) 2x E5-2650 v3 at 2.3 GHz, total 20 cores (40 hyperthreaded) 64 GB RAM

EPU alone is not sufficient to replace a cluster Dataset/computer Eiger 16M 9k images (Dectris, using xdsme) Screening (MX2, typical) EPU (1 CPU) 13 mins 30 s – 1 min EPU (3 CPUs) 5:30 mins (not tested) Node1/2 (2x E5-2650 v3 40threads) 64 GB RAM 12 mins EPU (3 CPUs) approaches Dectris design target of processing < 4x collection time Phase 1: Many jobs with multiple slow jobs Phase 2: Fewer jobs on EPU, multiple slow jobs Phase 3: One fast EPU and some screening jobs Slow sample changer (~4.5 minutes per change) prevents multiple datasets from queueing – saves us for now Create a processing system where they are equivalent Easily expandable by adding more computing nodes – simply subscribe the new node to the processing Redis queue

Processing results - setup Initial EPU: 6 jobs (3x datasets+3x screening) Node 1: 1 jobs (dataset) Node 2: Current 2 jobs (dataset+screening) Node1: Node2: Future 1 job (140 threads) 1-2 jobs (1-2x screening) Node... Initial Issues: Ram limited Full swap, unstable machines User changer needs to kill zombies Current -Investigating : XDS fork integrate Cluster design Ramdisk use on EPU Spotfinder hardware needs

Impact of no cluster Processing is far slower than collection. Users have 2 choices: Wait for processing. Slows data collection down drastically. Collect without processing information. Can collect a lot of poor data especially when learning how to use Eiger e.g. incorrect attenuation (too low = no signal, too high = fast radiation damage). No option for rastering (no spare computing power). Cluster being scoped.

MX2 processing system outline newdataset.py IP:*.2.44 http://processing/retrigger/submit redis queues Screening autodataset Screening-cpu 2 host:EPU Screening-host:Node1 Screening- host:Node2 autodataset-cpu1 host:EPU autodataset host:EPU02 autodataset host:EPU03 EPU Node1/2

Network/Filesystem DCU->EPU link on 10Gbps network. EPU has bonded dual 40Gbps (Mellanox card) to switch. EPU->730 link on dual 10Gbps. Filesystem is vanilla linux EXT4. Processing from EPU file system (not RAM-disk).

EPU Summary Works well as a data storage device Integrates well with DCU – makes master, then data files available seamlessly via NFS The CPUs are very capable processing nodes Dectris designed the EPU for all three CPUs to be used for rapid processing of 1 job. The 3 CPUs combined processed the data in about 4x the collection time We tested using the three CPUs independently allowing us to have more processing jobs instead of a single fast job. Running 6 jobs however, drastically slowed the system as it became memory limited and 100% swapped out. Some issues with Furka/Grimsel (a bug with large files, now fixed by Dectris). Ramdisk would fill and needed cleaning. Using custom Python with lsyncd to synchronise EPU storage to SAN. We have EPU serial #002 so problems aren’t unexpected…

Data Storage Long term data storage (6 months to 3 years: TBD retention period) Facility SAN is a commercial IBM product (EOL: Jun 2017) ~80 Tb free for 9 beamlines (3 high demading) Temporary storage as part of new custom build Facility Cluster ~100 Tb available New facility new storage in build process (~Sep 2017)  CEPH 0.5 Pb but might increase if we can make our case Current user data rate generation Commercial companies: ~3 Tb/20h Academic users: 0.5 to 2 Tb/20h  Still learning After experiment we compress all data in a single file using Squashfs (~70% of original size: Master file compresses a lot) How would this look like with a cluster, rastering, spotfinder, etc?

MX team – Tom Caradoc-Davies, David Aragao, Daniel Ericsson, Stephen Harrop*, Sofia Macedo, Santosh Panjikar, Jason Price, Alan Riboldi-Tunnicliffe, Rachel Williamson Nathan Mudie IT Group – Uli Felzmann*, Robbie Clarken, Andreas Moll Dectris – Andreas Forster, Diego Gaemperle SLS – Simon Ebner, Zac Panepucci, Meitian Wang HDRMX – http://hdrmx.medsbio.org NSLS2 – Bruno Martins for AD Eiger Centre for Advanced Molecular Imaging – for jointly funding my position!