Fabric for Frontier Experiments at Fermilab Gabriele Garzoglio Grid and Cloud Services Department, Scientific Computing Division, Fermilab ISGC – Thu,

Slides:



Advertisements
Similar presentations
 Contributing >30% of throughput to ATLAS and CMS in Worldwide LHC Computing Grid  Reliant on production and advanced networking from ESNET, LHCNET and.
Advertisements

1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
S. Gadomski, "ATLAS computing in Geneva", journee de reflexion, 14 Sept ATLAS computing in Geneva Szymon Gadomski description of the hardware the.
O. Stézowski IPN Lyon AGATA Week September 2003 Legnaro Data Analysis – Team #3 ROOT as a framework for AGATA.
Office of Science U.S. Department of Energy Grids and Portals at NERSC Presented by Steve Chan.
JIM Deployment for the CDF Experiment M. Burgon-Lyon 1, A. Baranowski 2, V. Bartsch 3,S. Belforte 4, G. Garzoglio 2, R. Herber 2, R. Illingworth 2, R.
Minerva Infrastructure Meeting – October 04, 2011.
Ian M. Fisk Fermilab February 23, Global Schedule External Items ➨ gLite 3.0 is released for pre-production in mid-April ➨ gLite 3.0 is rolled onto.
The SAM-Grid Fabric Services Gabriele Garzoglio (for the SAM-Grid team) Computing Division Fermilab.
CERN - IT Department CH-1211 Genève 23 Switzerland t Monitoring the ATLAS Distributed Data Management System Ricardo Rocha (CERN) on behalf.
Open Science Grid Software Stack, Virtual Data Toolkit and Interoperability Activities D. Olson, LBNL for the OSG International.
Alexandre A. P. Suaide VI DOSAR workshop, São Paulo, 2005 STAR grid activities and São Paulo experience.
Grid Job and Information Management (JIM) for D0 and CDF Gabriele Garzoglio for the JIM Team.
CDF data production models 1 Data production models for the CDF experiment S. Hou for the CDF data production team.
Building a distributed software environment for CDF within the ESLEA framework V. Bartsch, M. Lancaster University College London.
Miguel Branco CERN/University of Southampton Enabling provenance on large-scale e-Science applications.
D0 SAM – status and needs Plagarized from: D0 Experiment SAM Project Fermilab Computing Division.
3rd June 2004 CDF Grid SAM:Metadata and Middleware Components Mòrag Burgon-Lyon University of Glasgow.
Jean-Yves Nief CC-IN2P3, Lyon HEPiX-HEPNT, Fermilab October 22nd – 25th, 2002.
CHEP'07 September D0 data reprocessing on OSG Authors Andrew Baranovski (Fermilab) for B. Abbot, M. Diesburg, G. Garzoglio, T. Kurca, P. Mhashilkar.
CERN IT Department CH-1211 Genève 23 Switzerland t Internet Services Job Monitoring for the LHC experiments Irina Sidorova (CERN, JINR) on.
Sep 21, 20101/14 LSST Simulations on OSG Sep 21, 2010 Gabriele Garzoglio for the OSG Task Force on LSST Computing Division, Fermilab Overview OSG Engagement.
SAMGrid as a Stakeholder of FermiGrid Valeria Bartsch Computing Division Fermilab.
Instrumentation of the SAM-Grid Gabriele Garzoglio CSC 426 Research Proposal.
Data Grid projects in HENP R. Pordes, Fermilab Many HENP projects are working on the infrastructure for global distributed simulated data production, data.
1. Maria Girone, CERN  Q WLCG Resource Utilization  Commissioning the HLT for data reprocessing and MC production  Preparing for Run II  Data.
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
Tier-2  Data Analysis  MC simulation  Import data from Tier-1 and export MC data CMS GRID COMPUTING AT THE SPANISH TIER-1 AND TIER-2 SITES P. Garcia-Abia.
Grid Computing at Yahoo! Sameer Paranjpye Mahadev Konar Yahoo!
Workshop on Computing for Neutrino Experiments - Summary April 24, 2009 Lee Lueking, Heidi Schellman NOvA Collaboration Meeting.
NA-MIC National Alliance for Medical Image Computing UCSD: Engineering Core 2 Portal and Grid Infrastructure.
1/12 Scientific Data Processing Solutions – Department Meeting – Dec 2014 Scientific Data Processing Solutions Department Meeting Dec 2014 Dec 04, 2014.
What is SAM-Grid? Job Handling Data Handling Monitoring and Information.
Data reprocessing for DZero on the SAM-Grid Gabriele Garzoglio for the SAM-Grid Team Fermilab, Computing Division.
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
Grid User Interface for ATLAS & LHCb A more recent UK mini production used input data stored on RAL’s tape server, the requirements in JDL and the IC Resource.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
High Energy FermiLab Two physics detectors (5 stories tall each) to understand smallest scale of matter Each experiment has ~500 people doing.
Scientific Storage at FNAL Gerard Bernabeu Altayo Dmitry Litvintsev Gene Oleynik 14/10/2015.
CD FY09 Tactical Plan Status FY09 Tactical Plan Status Report for Neutrino Program (MINOS, MINERvA, General) Margaret Votava April 21, 2009 Tactical plan.
Adapting SAM for CDF Gabriele Garzoglio Fermilab/CD/CCF/MAP CHEP 2003.
Parag Mhashilkar Computing Division, Fermi National Accelerator Laboratory.
April 25, 2006Parag Mhashilkar, Fermilab1 Resource Selection in OSG & SAM-On-The-Fly Parag Mhashilkar Fermi National Accelerator Laboratory Condor Week.
Ian Bird Overview Board; CERN, 8 th March 2013 March 6, 2013
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
CMS: T1 Disk/Tape separation Nicolò Magini, CERN IT/SDC Oliver Gutsche, FNAL November 11 th 2013.
Victoria A. White Head, Computing Division, Fermilab Fermilab Grid Computing – CDF, D0 and more..
Nigel Lockyer Fermilab Operations Review 16 th -18 th May 2016 Fermilab in the Context of the DOE Mission.
A Data Handling System for Modern and Future Fermilab Experiments Robert Illingworth Fermilab Scientific Computing Division.
1 Open Science Grid: Project Statement & Vision Transform compute and data intensive science through a cross- domain self-managed national distributed.
Experiment Support CERN IT Department CH-1211 Geneva 23 Switzerland t DBES The Common Solutions Strategy of the Experiment Support group.
Nigel Lockyer Fermilab Operations Review 16 th -18 th May 2016 Fermilab in the Context of the DOE Mission.
FIFE Architecture Figures for V1.2 of document. Servers Desktops and Laptops Desktops and Laptops Off-Site Computing Off-Site Computing Interactive ComputingSoftware.
On-demand Services for the Scientific Program at Fermilab Gabriele Garzoglio Grid and Cloud Services Department, Scientific Computing Division, Fermilab.
Recent Evolution of the Offline Computing Model of the NOA Experiment Talk #200 Craig Group & Alec Habig CHEP 2015 Okinawa, Japan.
Recommendations from Previous Review Greg Bock Fermilab Science & Technology Review November
Building on virtualization capabilities for ExTENCI Carol Song and Preston Smith Rosen Center for Advanced Computing Purdue University ExTENCI Kickoff.
Gavin S. Davies Iowa State University On behalf of the NOvA Collaboration FIFE Workshop, June th 2014 FIFE workshop.
Hao Wu, Shangping Ren, Gabriele Garzoglio, Steven Timm, Gerard Bernabeu, Hyun Woo Kim, Keith Chadwick, Seo-Young Noh A Reference Model for Virtual Machine.
Mu2e FY15 and FY16 Computing Needs Rob Kutschke Scientific Computing Portfolio Management Team (SC-PMT) Review 4 March 2015.
Scientific Computing at Fermilab Lothar Bauerdick, Deputy Head Scientific Computing Division 1 of 7 10k slot tape robots.
VO Experiences with Open Science Grid Storage OSG Storage Forum | Wednesday September 22, 2010 (10:30am)
Fermilab Scientific Computing Division Fermi National Accelerator Laboratory, Batavia, Illinois, USA. Off-the-Shelf Hardware and Software DAQ Performance.
Fermilab Budget Briefing FY 2014 Intensity Frontier Proton Research KA Breakout February 28, 2013 Office of High Energy Physics Germantown, MD.
1 P. Murat, Mini-review of the CDF Computing Plan 2006, 2005/10/18 An Update to the CDF Offline Plan and FY2006 Budget ● Outline: – CDF computing model.
Particle Physics Sector Young-Kee Kim / Greg Bock Leadership Team Strategic Planning Winter Workshop January 29, 2013.
Thoughts on Computing Upgrade Activities
Simulation use cases for T2 in ALICE
Leigh Grundhoefer Indiana University
LHC Data Analysis using a worldwide computing grid
Presentation transcript:

Fabric for Frontier Experiments at Fermilab Gabriele Garzoglio Grid and Cloud Services Department, Scientific Computing Division, Fermilab ISGC – Thu, March 27, 2014

3/27/2014Fabric for Frontier Experiments at Fermilab Overview Frontier Physics at Fermilab: Computing Challenges Strategy: common infrastructure on distributed resources Early successes with multiple communities 3/27/20142

Fabric for Frontier Experiments at Fermilab Diverse program of experiments 3 LBNE 35T Lariat

3/27/2014Fabric for Frontier Experiments at Fermilab Fermilab Experiment Schedule Measurements at all frontiers – Electroweak physics, neutrino oscillations, muon g-2, dark energy, dark matter 8 major experiments in 3 frontiers running simultaneously in 2016 Sharing both beam and computing resources Impressive breadth of experiments at FNAL FY16 4 IF Experiments

3/27/2014Fabric for Frontier Experiments at Fermilab Computing requirements for experiments higher intensity and higher precision measurements are driving request for more computing resources than previous “small” experiments beam simulations to optimize experiments - make every particle count detector design studies - cost effectiveness and sensitivity projections higher bandwidth DAQ and greater detector granularity event generation and detector response simulation reconstruction and analysis algorithms Fermilab Scientific Computing Review* 5 * Number of slots have been updated periodically since

3/27/2014Fabric for Frontier Experiments at Fermilab Available onsite resources allowed experiments to develop architectures based on local resources now reintegrating them into a distributed computing model 30,000 slots on-site available through Open Science Grid interfaces –~7,000 slots for smaller experiments, but aggregate peak need is twice as much –off-site resources are needed to meet experiments needs –OSG provides a framework for universities and laboratories to contribute resources to experiments remotely 30k jobs FermiGrid 500k hours / day 6

3/27/2014Fabric for Frontier Experiments at Fermilab Transition to OSG important for smaller experiments Most experiments are relatively small (~100 researchers) –Unlike Tevatron or LHC experiments, no resources to develop large scale computing infrastructure Implementing a unique computing infrastructure is too hard Fermilab SCD goal: integrate tools to meet the needs of smaller experiments, focusing on local resources, OSG, and Clouds Need to inform experiments and gather requirements from them –what resources are available? –how do you get a job onto the OSG? And onto Amazon? –what storage elements are available? –how does it all fit together? FIFE represents an integrated solution from a new architecture 7

3/27/2014Fabric for Frontier Experiments at Fermilab FabrIc for Frontier Experiments (FIFE) ✤ provide resources while experiments focus on science ✤ integrate existing solutions into a consistent model Adam Lyon 8

3/27/2014Fabric for Frontier Experiments at Fermilab FIFE Strategy support the experiments in their computing needs modular: experiments can take only what they need designed so that while underlying solution may change, interface will be consistent provide mechanism for experiments to incorporate their tools and solutions help experiments utilize computing beyond the Fermilab campus integrate new tools and resources from outside Fermilab and other communities as they develop 9 generic offline workflow

3/27/2014Fabric for Frontier Experiments at Fermilab Current FIFE implementations Software Framework – ART supports integrated services (Data Handling, Geant4, ROOT) build environment and distribution –git, svn, gmake, cmake and distribution with CERN Virtual Machine File System (CVMFS) –build process and machine in development data handling - access to file catalog, tape storage, cache disk, and file transfer job submission infrastructure – based on GlideinWMS pilot system database infrastructure and access shared software (LArSoft for Liquid Argon TPC reconstruction) additional infrastructure - authentication, electronic control room log book, new user accounts 10 generic offline workflow

3/27/2014Fabric for Frontier Experiments at Fermilab Common Software Layers across the experiments MinosMinervaNoVAuBooneLBNEGm2Mu2eLariat 35T prototype Software Framework root/SRTgaudiart software processing larsoft GeneratorsGenie Genie, Flugg, Cry and Coriska Genie Genie, Cry Data Management SAM NoneSAM Job Submission Jobsub/gli deinwms glideinwms Jobsub/glidei nwms Onsite Grid Enabled Yes CVMFSYes In Progress YesPlannedIn Design OSG Enabled In Progress PartiallyIn ProgressPlanned In Design e-control room logbook Yes PlannedYes Beam Condition DB Yes Scientific Document Management (docdb) Yes tape storage enstore data cache dcache

3/27/2014Fabric for Frontier Experiments at Fermilab §1 – Integration success: early NOvA MonteCarlo on OSG Initial contained computing campaign with a single well- understood application 1,000,000 ev generated with ~10k jobs for 88,535 CPU h in 2 weeks of ops and 2 TB of data Run on OSG at SMU (dedicated), UNL, Uwisc, UC, UCSD and O(100) jobs at FermiCloud Operations consisted of 1 submission + 2 recoveries Spent about 10% more resources than expected due to preemption 12 using same submission command from user perspective, launched jobs “on-demand” at FermiCloud to provide proof-of-principle for Cloud bursting grid jobs ~90k CPU h

3/27/2014Fabric for Frontier Experiments at Fermilab §2 – Integration success: Ongoing NOvA computation on OSG … 13 …and Amazon Web Services “On-demand Services for the Scientific Program at Fermilab” Thu 3/27 16:25 – Rm1

3/27/2014Fabric for Frontier Experiments at Fermilab §3 – Integration success: early experience of MicroBoone on OSG LArSoft: package common to multiple Intensity Frontier experiments – Liquid Argon Simulation of Time-Projection Chambers Code with dependencies were made fully portable and deployed through CVMFS. Demonstrated the ability to run interactively at OSG sites (Fermilab and SLAC) Preparing for batch job submission to OSG sites – Nebraska, UCSD, SLAC 14

3/27/2014Fabric for Frontier Experiments at Fermilab Plans for improvements FIFE architecture has recently undergone a re-evaluation and re-architecture process job submission infrastructure modify to client-server model local storage element making transition to shared dCache pools data handling project continues to integrate new resources without any change in user interface starting new integration push for several experiments in both the Intensity Frontier and Cosmic Frontier 15

3/27/2014Fabric for Frontier Experiments at Fermilab Conclusions Fermilab physics program is extremely active at all three frontiers large scale computing needs for new “small” experiments encourages shared services designed to ease integration to distributed resource – OSG, Commercial, and Community Clouds FIFE project has developed a strategy and architecture to achieve this currently integrated OSG processing into NOvA and MicroBoone. Next, targeting DarkSide50, mu2e, and Muon g-2 looking forward to bringing more resources and techniques to enable increased scientific output from frontier experiments 16