Break out group coordinator:

Slides:



Advertisements
Similar presentations
Cloud Computing: Theirs, Mine and Ours Belinda G. Watkins, VP EIS - Network Computing FedEx Services March 11, 2011.
Advertisements

Summary Role of Software (1 slide) ARCS Software Architecture (4 slides) SNS -- Caltech Interactions (3 slides)
FI-WARE – Future Internet Core Platform FI-WARE Cloud Hosting July 2011 High-level description.
Marilyn T. Smith, Head, MIT Information Services & Technology DataSpace IS&T Data CenterMIT Optical Network 1.
DATA PRESERVATION IN ALICE FEDERICO CARMINATI. MOTIVATION ALICE is a 150 M CHF investment by a large scientific community The ALICE data is unique and.
Ian M. Fisk Fermilab February 23, Global Schedule External Items ➨ gLite 3.0 is released for pre-production in mid-April ➨ gLite 3.0 is rolled onto.
GridPP Steve Lloyd, Chair of the GridPP Collaboration Board.
Ian Bird LHCC Referees’ meeting; CERN, 11 th June 2013 March 6, 2013
High Energy Physics At OSCER A User Perspective OU Supercomputing Symposium 2003 Joel Snow, Langston U.
PolarGrid Geoffrey Fox (PI) Indiana University Associate Dean for Graduate Studies and Research, School of Informatics and Computing, Indiana University.
PCGRID ‘08 Workshop, Miami, FL April 18, 2008 Preston Smith Implementing an Industrial-Strength Academic Cyberinfrastructure at Purdue University.
BINP/GCF Status Report BINP LCG Site Registration Oct 2009
A. Aimar - EP/SFT LCG - Software Process & Infrastructure1 Software Process panel SPI GRIDPP 7 th Collaboration Meeting 30 June – 2 July 2003 A.Aimar -
GENIUS. WP600 Support Activities Participants: University of Barcelona: 6PM CESCA (Center for Scientific and Academic Services of Catalonia): 24PM University.
CRISP & SKA WP19 Status. Overview Staffing SKA Preconstruction phase Tiered Data Delivery Infrastructure Prototype deployment.
14 Aug 08DOE Review John Huth ATLAS Computing at Harvard John Huth.
SouthGrid SouthGrid SouthGrid is a distributed Tier 2 centre, one of four setup in the UK as part of the GridPP project. SouthGrid.
Next Steps: becoming users of the NGS Mike Mineter
ESFRI & e-Infrastructure Collaborations, EGEE’09 Krzysztof Wrona September 21 st, 2009 European XFEL.
Grid User Interface for ATLAS & LHCb A more recent UK mini production used input data stored on RAL’s tape server, the requirements in JDL and the IC Resource.
Near Real-Time Verification At The Forecast Systems Laboratory: An Operational Perspective Michael P. Kay (CIRES/FSL/NOAA) Jennifer L. Mahoney (FSL/NOAA)
ATLAS WAN Requirements at BNL Slides Extracted From Presentation Given By Bruce G. Gibbard 13 December 2004.
Slide David Britton, University of Glasgow IET, Oct 09 1 Prof. David Britton GridPP Project leader University of Glasgow UK-T0 Meeting 21 st Oct 2015 GridPP.
PI in a Modern Power Plant – American National Power, Inc. PI User Conference ‘03 Presented by: Brian M. Wood, American National Power, Inc More Uses Than.
Data Transfer Service Challenge Infrastructure Ian Bird GDB 12 th January 2005.
1 st EGI CTA VT meeting 18 January 2013 C. Vuerli (INAF, Italy), N. Neyroud (CNRS/IN2P3/LAPP, France)
12 NOVEMBER 2015 LoS Engagement in the Netherlands The Support4research project Jan Bot, SURFsara Support4research team SURF.
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
GridPP2 Data Management work area J Jensen / RAL GridPP2 Data Management Work Area – Part 2 Mass storage & local storage mgmt J Jensen
1 Open Science Grid: Project Statement & Vision Transform compute and data intensive science through a cross- domain self-managed national distributed.
STFC in INDIGO DataCloud WP3 INDIGO DataCloud Kickoff Meeting Bologna April 2015 Ian Collier
IBERGRID as RC Total Capacity: > 10k-20K cores, > 3 Petabytes Evolving to cloud (conditioned by WLCG in some cases) Capacity may substantially increase.
CMB & LSS Virtual Research Community Marcos López-Caniego Enrique Martínez Isabel Campos Jesús Marco Instituto de Física de Cantabria (CSIC-UC) EGI Community.
A worldwide e-Infrastructure and Virtual Research Community for NMR and structural biology Alexandre M.J.J. Bonvin Project coordinator Bijvoet Center for.
Building on virtualization capabilities for ExTENCI Carol Song and Preston Smith Rosen Center for Advanced Computing Purdue University ExTENCI Kickoff.
Computing infrastructures for the LHC: current status and challenges of the High Luminosity LHC future Worldwide LHC Computing Grid (WLCG): Distributed.
EGI-InSPIRE RI EGI Compute and Data Services for Open Access in H2020 Tiziana Ferrari Technical Director, EGI.eu
EGI-InSPIRE RI An Introduction to European Grid Infrastructure (EGI) March An Introduction to the European Grid Infrastructure.
EGI… …is a Federation of over 300 computing and data centres spread across 56 countries in Europe and worldwide …delivers advanced computing.
September 27th, 2016 Challenges Posed by Processing Scientific Data at Extreme Light Infrastructures Tamás Gaizer.
Status of WLCG FCPPL project
BEST CLOUD COMPUTING PLATFORM Skype : mukesh.k.bansal.
Ian Bird WLCG Workshop San Francisco, 8th October 2016
H2020, COEs and PRACE.
WP18, High-speed data recording Krzysztof Wrona, European XFEL
Benefits of harmonising R&E network and NGI Research Engagement activities Jan Bot, SURFsara 31/5/2017 Research Engagement and the Sociology of Global.
Computing models, facilities, distributed computing
Presenter: T. Gaizer (ELI-ALPS), co-authors: G. Beckett, J. Chudoba, M
INFN Computing Outlook The Bologna Initiative
Break out group coordinator:
Pasquale Migliozzi INFN Napoli
Design your e-infrastructure. egi
Design your e-infrastructure!
EGI use case description and development planning template Use case(s): Provider: …
Christos Markou Institute of Nuclear Physics NCSR ‘Demokritos’
Dagmar Adamova (NPI AS CR Prague/Rez) and Maarten Litmaath (CERN)
Thoughts on Computing Upgrade Activities
Статус ГРИД-кластера ИЯФ СО РАН.
Exploitation of ISS Scientific data - sustainability
ELITRANS project and DIRAC
The INFN Tier-1 Storage Implementation
Computing Infrastructure for DAQ, DM and SC
Design your e-infrastructure. egi
EISCAT-3D: a data centric design for extreme scale computing
Experience with an IT Asset Management System
"Cloud services" - what it is.
Design your e-infrastructure. egi
Design your e-infrastructure. egi
Break out group coordinator:
Break out group coordinator:
Presentation transcript:

Break out group coordinator: Design your e-infrastructure! https://indico.egi.eu/indico/event/3025/ Use case: Extreme Light Infrastructure Break out group coordinator: Jan Bot Krakow, 27. September, 2016.

Group members Tamas Gaizer (Software Architect at ELI-ALPS) Teodor Ivanoaica (Network Administrator ELI-NP) Catalin Condurache (STFC RAL) LHC Tier-1 group Abraham Nieva (Cardiff Universisty/ENVRI+) Claudio Cacciari (Cineca & EUDAT) Vincenzo Capone (GÉANT) Jan Bot (SURFsara)

First break-out Background and Users

Who will be the user? Can the users be characterised? How many are they? People who run the experiment: 600 people per year more people could be processing the data Large variety in computational capabilities, however, many of the researchers should be able to fend for themselves Misc Data made public after two years Online simulations For now no other partners

Something about the data High repetition laser raw data: <10 TB per experiment (2 weeks) Forecast for cleaned data not possible to make. Experimentation will take place in the coming two years. Derived data …? Other two experiments pretty much the same numbers. Metadata generation needs to be thought about (laser & gamma control systems, calibration). Some should be linked directly to the data, some should just be kept for maintenance purposes.

What value will the envisaged system deliver for them (the whole setup)? What will the system exactly deliver to them? T0 (local system) Online storage & processing Buffer space for measurements Keep data at the facility for 5 years 7 PB usable storage as buffer 5000 – 10 000 cores for all processing in house T1 (external system) Running simulations Offline processing Long term archive of experiment data (both ‘raw’ unpublished data and data linked to from scientific papers) Hidden need for extra compute & storage from people now taking their data home

How should they use the system? Training can be made mandatory (also for the IT part) T0 (local system) Data taken on facility User measuring the data also responsible for running the cleaning pipeline (helped by local staff) User can use part of the system for further processing, gets x amount of time to do so (cpu time needs to be pre-specified) T1 (external system) More diverse group (not just people who requested beam time) of researchers (small group in total) People quite tech savvy, can be relied upon to arrange computations themselves  this could be a challenge Web interface would be useful, should support some major applications. If people want something else they should arrange this themselves

What's the timeline for development, testing and large-scale operation What's the timeline for development, testing and large-scale operation? (Consecutive releases can/should be considered.) Installation of end-points 2017 First ’friendly user test’ 2017 Beam line operation: >2018 q3 First simulations probably done on PRACE equipment (can be done now) or locally (in own data center) Pool of storage online early 2017

Design and implementation plan Second break-out Design and implementation plan

What should the first version include What should the first version include? - The most basic product prototype imaginable already bringing value to the users (the so-called Minimal Viable Product - MVP) T0 Storage facilities (buffer space for experiment) Small computing resource to validate results from experiments T1 Basic working pipeline to do simulations

Which components/services already exist in this architecture? Storage: next year (tender is ongoing) Software component of the simulation system (validate experiment) T1 Experience with grid & DIRAC

Which components/services are under development (and by who)? Entire system under development T1 None

Which components/services should be still brought into the system Which components/services should be still brought into the system? Which EGI/EUDAT/GEANT/OpenAire partner can do it? T0 None T1 Storage resources Archive (near-line) Online storage Computational resources (HPC, HTC) Any (PRACE, EGI, Commercial) Portal to resources (EGI, EUDAT) Connectivity (NREN, GÉANT) Ring infrastructure?

Are there gaps in the EGI/EUDAT/GEANT/OpenAIRE service catalogues that should be filled to implement the use case? Which service provider could fill the gap? Capacity might pose a problem (what is the lead time?)

Next steps Gather requirements Study & test compute and storage resources (end in February / March) – for T1 Upgrade network when necessary Setup VO and provision resources Setup & test portal