The Grid & Cloud Computing Department at Fermilab and the KISTI Collaboration Meeting with KISTI Nov 1, 2011 Gabriele Garzoglio Grid & Cloud Computing.

Slides:



Advertisements
Similar presentations
Fermilab, the Grid & Cloud Computing Department and the KISTI Collaboration GSDC - KISTI Workshop Jangsan-ri, Nov 4, 2011 Gabriele Garzoglio Grid & Cloud.
Advertisements

Dec 14, 20061/10 VO Services Project – Status Report Gabriele Garzoglio VO Services Project WBS Dec 14, 2006 OSG Executive Board Meeting Gabriele Garzoglio.
Computing Strategy Victoria White, Associate Lab Director for Computing and CIO Fermilab PAC June 24, 2011.
Office of Science U.S. Department of Energy Grids and Portals at NERSC Presented by Steve Chan.
Idle virtual machine detection in FermiCloud Giovanni Franzini September 21, 2012 Scientific Computing Division Grid and Cloud Computing Department.
Minerva Infrastructure Meeting – October 04, 2011.
Jun 29, 20101/25 Storage Evaluation on FG, FC, and GPCF Jun 29, 2010 Gabriele Garzoglio Computing Division, Fermilab Overview Introduction Lustre Evaluation:
The SAM-Grid Fabric Services Gabriele Garzoglio (for the SAM-Grid team) Computing Division Fermilab.
F Run II Experiments and the Grid Amber Boehnlein Fermilab September 16, 2005.
SCD FIFE Workshop - GlideinWMS Overview GlideinWMS Overview FIFE Workshop (June 04, 2013) - Parag Mhashilkar Why GlideinWMS? GlideinWMS Architecture Summary.
Open Science Grid For CI-Days Internet2: Fall Member Meeting, 2007 John McGee – OSG Engagement Manager Renaissance Computing Institute.
The Fermilab Campus Grid (FermiGrid) Keith Chadwick Fermilab Work supported by the U.S. Department of Energy under contract No. DE-AC02-07CH11359.
HEP Experiment Integration within GriPhyN/PPDG/iVDGL Rick Cavanaugh University of Florida DataTAG/WP4 Meeting 23 May, 2002.
Building a distributed software environment for CDF within the ESLEA framework V. Bartsch, M. Lancaster University College London.
May 8, 20071/15 VO Services Project – Status Report Gabriele Garzoglio VO Services Project – Status Report Overview and Plans May 8, 2007 Computing Division,
Impacts of Tevatron Extension Pier Oddone P5 Meeting, October 15 th, P. Oddone, P5 Meeting, October 15, 2010.
Apr 30, 20081/11 VO Services Project – Stakeholders’ Meeting Gabriele Garzoglio VO Services Project Stakeholders’ Meeting Apr 30, 2008 Gabriele Garzoglio.
GlobusWorld 2012: Experience with EXPERIENCE WITH GLOBUS ONLINE AT FERMILAB Gabriele Garzoglio Computing Sector Fermi National Accelerator.
Virtualization within FermiGrid Keith Chadwick Work supported by the U.S. Department of Energy under contract No. DE-AC02-07CH11359.
November 16, 2012 Seo-Young Noh Haengjin Jang {rsyoung, Status Updates on STAR Computing at KISTI.
SAMGrid as a Stakeholder of FermiGrid Valeria Bartsch Computing Division Fermilab.
Virtualization within FermiGrid Keith Chadwick Fermilab Work supported by the U.S. Department of Energy under contract No. DE-AC02-07CH11359.
Instrumentation of the SAM-Grid Gabriele Garzoglio CSC 426 Research Proposal.
Mar 28, 20071/9 VO Services Project Gabriele Garzoglio The VO Services Project Don Petravick for Gabriele Garzoglio Computing Division, Fermilab ISGC 2007.
10/24/2015OSG at CANS1 Open Science Grid Ruth Pordes Fermilab
Workshop on Computing for Neutrino Experiments - Summary April 24, 2009 Lee Lueking, Heidi Schellman NOvA Collaboration Meeting.
Fermilab Site Report Spring 2012 HEPiX Keith Chadwick Fermilab Work supported by the U.S. Department of Energy under contract No. DE-AC02-07CH11359.
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
US LHC OSG Technology Roadmap May 4-5th, 2005 Welcome. Thank you to Deirdre for the arrangements.
4/25/2006Condor Week 1 FermiGrid Steven Timm Fermilab Computing Division Fermilab Grid Support Center.
GLIDEINWMS - PARAG MHASHILKAR Department Meeting, August 07, 2013.
FermiCloud: Enabling Scientific Workflows with Federation and Interoperability Steven C. Timm FermiCloud Project Lead Grid & Cloud Computing Department.
Farms User Meeting April Steven Timm 1 Farms Users meeting 4/27/2005
CD FY09 Tactical Plan Status FY09 Tactical Plan Status Report for Neutrino Program (MINOS, MINERvA, General) Margaret Votava April 21, 2009 Tactical plan.
Fermilab Site Report HEPiX Fall 2011 Keith Chadwick Work supported by the U.S. Department of Energy under contract No. DE-AC02-07CH11359.
Final Implementation of a High Performance Computing Cluster at Florida Tech P. FORD, X. FAVE, K. GNANVO, R. HOCH, M. HOHLMANN, D. MITRA Physics and Space.
Eileen Berman. Condor in the Fermilab Grid FacilitiesApril 30, 2008  Fermi National Accelerator Laboratory is a high energy physics laboratory outside.
Sep 25, 20071/5 Grid Services Activities on Security Gabriele Garzoglio Grid Services Activities on Security Gabriele Garzoglio Computing Division, Fermilab.
Fermilab: Present and Future Young-Kee Kim Data Preservation Workshop May 16, 2011.
An Introduction to Campus Grids 19-Apr-2010 Keith Chadwick & Steve Timm.
Parag Mhashilkar Computing Division, Fermi National Accelerator Laboratory.
FermiGrid Keith Chadwick. Overall Deployment Summary 5 Racks in FCC:  3 Dell Racks on FCC1 –Can be relocated to FCC2 in FY2009. –Would prefer a location.
1 Open Science Grid.. An introduction Ruth Pordes Fermilab.
Victoria A. White Head, Computing Division, Fermilab Fermilab Grid Computing – CDF, D0 and more..
FermiCloud Project Overview and Progress Keith Chadwick Grid & Cloud Computing Department Head Fermilab Work supported by the U.S. Department of Energy.
Fermilab Site Report Keith Chadwick Grid & Cloud Computing Department Head Fermilab Work supported by the U.S. Department of Energy under contract No.
Nigel Lockyer Fermilab Operations Review 16 th -18 th May 2016 Fermilab in the Context of the DOE Mission.
1 Open Science Grid: Project Statement & Vision Transform compute and data intensive science through a cross- domain self-managed national distributed.
Auxiliary services Web page Secrets repository RSV Nagios Monitoring Ganglia NIS server Syslog Forward FermiCloud: A private cloud to support Fermilab.
Fermilab / FermiGrid / FermiCloud Security Update Work supported by the U.S. Department of Energy under contract No. DE-AC02-07CH11359 Keith Chadwick Grid.
Nigel Lockyer Fermilab Operations Review 16 th -18 th May 2016 Fermilab in the Context of the DOE Mission.
FIFE Architecture Figures for V1.2 of document. Servers Desktops and Laptops Desktops and Laptops Off-Site Computing Off-Site Computing Interactive ComputingSoftware.
Claudio Grandi INFN Bologna Virtual Pools for Interactive Analysis and Software Development through an Integrated Cloud Environment Claudio Grandi (INFN.
GlobusWorld 2012: Experience with EXPERIENCE WITH GLOBUS ONLINE AT FERMILAB Gabriele Garzoglio Computing Sector Fermi National Accelerator.
Fabric for Frontier Experiments at Fermilab Gabriele Garzoglio Grid and Cloud Services Department, Scientific Computing Division, Fermilab ISGC – Thu,
Fermilab: Introduction Young-Kee Kim DOE KA12 (Electron Research)Review June 22-23, 2010.
Building on virtualization capabilities for ExTENCI Carol Song and Preston Smith Rosen Center for Advanced Computing Purdue University ExTENCI Kickoff.
April 18, 2006FermiGrid Project1 FermiGrid Project Status April 18, 2006 Keith Chadwick.
Hao Wu, Shangping Ren, Gabriele Garzoglio, Steven Timm, Gerard Bernabeu, Hyun Woo Kim, Keith Chadwick, Seo-Young Noh A Reference Model for Virtual Machine.
CPM 2012, Fermilab D. MacFarlane & N. Holtkamp The Snowmass process and SLAC plans for HEP.
GPCF* Update Present status as a series of questions / answers related to decisions made / yet to be made * General Physics Computing Facility (GPCF) is.
Parag Mhashilkar (Fermi National Accelerator Laboratory)
FermiCloud Review Response to Questions Keith Chadwick Steve Timm Gabriele Garzoglio Work supported by the U.S. Department of Energy under contract No.
Scientific Computing at Fermilab Lothar Bauerdick, Deputy Head Scientific Computing Division 1 of 7 10k slot tape robots.
FermiGrid The Fermilab Campus Grid 28-Oct-2010 Keith Chadwick Work supported by the U.S. Department of Energy under contract No. DE-AC02-07CH11359.
Virtualization within FermiGrid Keith Chadwick Work supported by the U.S. Department of Energy under contract No. DE-AC02-07CH11359.
Particle Physics Sector Young-Kee Kim / Greg Bock Leadership Team Strategic Planning Winter Workshop January 29, 2013.
Margaret Votava / Scientific Computing Division FIFE Workshop 20 June 2016 State of the Facilities.
July 18, 2011S. Timm FermiCloud Enabling Scientific Computing with Integrated Private Cloud Infrastructures Steven Timm.
f f FermiGrid – Site AuthoriZation (SAZ) Service
Presentation transcript:

The Grid & Cloud Computing Department at Fermilab and the KISTI Collaboration Meeting with KISTI Nov 1, 2011 Gabriele Garzoglio Grid & Cloud Computing Department, Associate Head Computing Sector, Fermilab Overview Fermilab, the Computing Sector, and the Grid & Cloud Computing department Collaboration with KISTI

Gabriele Garzoglio, The GCC Dept. at FNAL and the KISTI Collaboration, Nov 1, 2011 Fermi National Accelerator Laboratory 2

Gabriele Garzoglio, The GCC Dept. at FNAL and the KISTI Collaboration, Nov 1, 2011

5

Physics! CDF & D0 - Top Quark Asymmetry Results. CDF - Discovery of Ξ b 0. CDF - Λ c (2595) baryon. Combined CDF & D0 Limits on standard model higgs mass. 6

Gabriele Garzoglio, The GCC Dept. at FNAL and the KISTI Collaboration, Nov 1, 2011 CDF & D0 Publications CDFD0 7

Gabriele Garzoglio, The GCC Dept. at FNAL and the KISTI Collaboration, Nov 1, 2011 TeVatron Shutdown On Friday 30-Sep-2011, the Fermilab TeVatron was shut down following 28 years of operation, The collider reached peak luminosities of 4 x per centimeter squared per second, The CDF and Dzero detectors recorded 8.63 PB and 7.54 PB of data respectively, corresponding to nearly 12 inverse femtobarns of data, CDF and Dzero data analysis continues, Fermilab has committed to 5+ years of support for analysis and 10+ years for access to data. 8

Gabriele Garzoglio, The GCC Dept. at FNAL and the KISTI Collaboration, Nov 1, 2011

11

Gabriele Garzoglio, The GCC Dept. at FNAL and the KISTI Collaboration, Nov 1, 2011 Present Plan: Energy Frontier Tevatron LHC ILC, CLIC or Muon Collider Now 2016 LHC Upgrades ILC??

Gabriele Garzoglio, The GCC Dept. at FNAL and the KISTI Collaboration, Nov 1, 2011

14 Present Plan: Intensity Frontier MINOS MiniBooNE MINERvA SeaQuest NOvA MicroBooNE g-2? SeaQuest Now 2016 LBNE Mu2e Project X+LBNE , K, nuclear, … Factory ??

Gabriele Garzoglio, The GCC Dept. at FNAL and the KISTI Collaboration, Nov 1,

Gabriele Garzoglio, The GCC Dept. at FNAL and the KISTI Collaboration, Nov 1, 2011 Present Plan: Cosmic Frontier Now DM: ~10 kg DE: SDSS P. Auger DM: ~100 kg DE: DES P. Auger Holometer? DM: ~1 ton DE: LSST WFIRST?? BigBOSS?? DE: LSST WFIRST??

Gabriele Garzoglio, The GCC Dept. at FNAL and the KISTI Collaboration, Nov 1, 2011 The Feynman Computing Center 17

Gabriele Garzoglio, The GCC Dept. at FNAL and the KISTI Collaboration, Nov 1, 2011 Fermilab Computing Facilities Feynman Computing Center:  FCC2  FCC3  High availability services – e.g. core network, , etc.  Tape Robotic Storage ( slot libraries)  UPS & Standby Power Generation  ARRA project: upgrade cooling and add HA computing room - completed Grid Computing Center:  3 Computer Rooms – GCC-[A,B,C]  Tape Robot Room – GCC-TRR  High Density Computational Computing  CMS, RUNII, Grid Farm batch worker nodes  Lattice HPC nodes  Tape Robotic Storage ( slot libraries)  UPS & taps for portable generators Lattice Computing Center:  High Performance Computing (HPC)  Accelerator Simulation, Cosmology nodes  No UPS EPA Energy Star award for

Gabriele Garzoglio, The GCC Dept. at FNAL and the KISTI Collaboration, Nov 1, 2011 Fermilab CPU Core Count 19

Gabriele Garzoglio, The GCC Dept. at FNAL and the KISTI Collaboration, Nov 1, 2011 Data Storage at Fermilab 20

Gabriele Garzoglio, The GCC Dept. at FNAL and the KISTI Collaboration, Nov 1, 2011 High Speed Networking 21

Gabriele Garzoglio, The GCC Dept. at FNAL and the KISTI Collaboration, Nov 1, 2011 To establish and maintain Fermilab as a high performance, robust, highly available, expertly supported and documented Premier Grid Facility which supports the scientific program based on Computing Division and Laboratory priorities. The department provides production infrastructure and software, together with leading edge and innovative computing solutions, to meet the needs of the Fermilab Grid community and takes a strong leadership role in the development and operation of the global computing infrastructure of the Open Science Grid.

Gabriele Garzoglio, The GCC Dept. at FNAL and the KISTI Collaboration, Nov 1, 2011 FermiGrid Occupancy & Utilization Cluster(s) Current Size (Slots) Average Size (Slots) Average Occupancy Average Utilization CDF %67% CMS %87% D %53% GP %68% Total %70% 23

Gabriele Garzoglio, The GCC Dept. at FNAL and the KISTI Collaboration, Nov 1, 2011 Overall Occupancy & Utilization 24

Gabriele Garzoglio, The GCC Dept. at FNAL and the KISTI Collaboration, Nov 1, 2011 FermiGrid-HA2 Service Availability Service Raw Availability HA Configuration Measured HA Availability Minutes of Downtime VOMS – VO Management Service %Active-Active %0 GUMS – Grid User Mapping Service %Active-Active %0 SAZ – Site AuthoriZation Service %Active-Active %0 Squid – Web Cache99.640%Active-Active %0 MyProxy – Grid Proxy Server %Active-Standby99.954%240 ReSS – Resource Selection Service %Active-Active %0 Gratia – Fermilab and OSH Accounting %Active-Standby99.997%120 Databases99.765%Active-Active99.988%60 25

Gabriele Garzoglio, The GCC Dept. at FNAL and the KISTI Collaboration, Nov 1, 2011 FY2011 Worker Node Acquisition Specifications: Quad processor, 8 core, AMD 6128/6128HE, 2.0 GHz, 64 Gbytes DDR3 memory, 3x2 Tbytes disk, 4 year warranty, $3,654 each. Retire- mentsBaseOptionPurchaseAssign CDF CMS D IF GP Total

Gabriele Garzoglio, The GCC Dept. at FNAL and the KISTI Collaboration, Nov 1, 2011 Other Significant FY2011 Purchases Storage:  5.52 Petabytes of Nexsan E60 SATA drives for raw cache disk (will mostly be used for dCache with some Lustre).  180 Terabytes of BlueArc SATA disk.  60 Terabytes of BlueArc 15K SAS disk. Servers:  119 servers,  16 different configurations,  All done in a single order. 27

Gabriele Garzoglio, The GCC Dept. at FNAL and the KISTI Collaboration, Nov 1, 2011 FermiCloud Utilization

Gabriele Garzoglio, The GCC Dept. at FNAL and the KISTI Collaboration, Nov 1, 2011 KISTI and GCC: working together… Maintaining frequent in-person visits at KISTI and FNAL Working side-by-side in the Grid and Cloud Computing department at FNAL  Seo-Young: 3 mo in the Summer 2011  Hyunwoo Kim: 6 months since Spring 2011 Sharing information about Grid & Cloud computing and FNAL ITIL service management Consulting help on operations for CDF data processing and OSG 29

Gabriele Garzoglio, The GCC Dept. at FNAL and the KISTI Collaboration, Nov 1, 2011 Super Computing Fermilab Posters at SC12 mention KISTI and our collaboration

Gabriele Garzoglio, The GCC Dept. at FNAL and the KISTI Collaboration, Nov 1, 2011 FermiCloud Technology Investigations Testing storage services with real neutrino experiment codes. Evaluate ceph as a FS for the image repository. * Using Infiniband interface to create sandbox for MPI applications. * Batch queue look-ahead to create worker node VM's on demand. * Submission of multiple worker node VM's, grid cluster in the cloud. * Bulk launching of VMs and interaction with private nets Idle VM detection and suspension, backfill with worker node VM's. Leverage site “network jail” for new virtual machines. IPv6 support. Testing dCache NFS4.1 support with multiple clients in the cloud. Interest in OpenID/SAML assertion-based authentication. Design a high-availability solution across buildings Interoperability: CERNVM, HEPiX, Glidein WMS, ExTENCI, Desktop * In Collaboration with KISTI – Seo-Young and Hyunwoo: Summer 2011 – Seo-Young / KISTI: Summer 2011; Proposed Fall / Winter 2011 – Hyunwoo: Proposed Fall / Winter

Gabriele Garzoglio, The GCC Dept. at FNAL and the KISTI Collaboration, Nov 1, 2011 vcluster and FermiCloud A System to Dynamically Extend Grid Clusters Through Virtual Worker Nodes Expand Grid resources with worker nodes deployed dynamically as VM at clouds (e.g.EC2) vcluster is a virtual cluster system that uses computing resources from heterogeneous cloud systems provides a uniform view for the jobs managed by the system distributes batch jobs to VM over clouds depending on the status of queue and system pool. vcluster is cloud and batch system agnostic via a plug-in architecture. Seo-Young Noh has been leading the vcluster effort. 32

Gabriele Garzoglio, The GCC Dept. at FNAL and the KISTI Collaboration, Nov 1, 2011 InfiniBand and FermiCloud InfiniBand Support on FermiCloud Virtual Machines The project enabled FermiCloud support for InfiniBand cards within VM. Goal: deploy high performance computing- like environments and prototype MPI-based applications. The project evaluated techniques to expose IB cards of the host to the FermiCloud VM Approach: Now: software-based resource sharing Future: hardware-based resource sharing Hyunwoo Kim has been leading the effort to provide InfiniBand support on virtual machines. 33

Gabriele Garzoglio, The GCC Dept. at FNAL and the KISTI Collaboration, Nov 1, 2011 KISTI and OSG KISTI has participated in an ambitious data reprocessing program for CDF Data moved via SAM and job submitted through OSG interfaces We look forward to further collaborations  Resource federation  Opportunistic resource usage

Gabriele Garzoglio, The GCC Dept. at FNAL and the KISTI Collaboration, Nov 1, 2011 Conclusions Fermilab future is well planned and full of activities on all Frontiers The Grid and Cloud Computing department contributes with  FermiGrid operations  FermiCloud commissioning & operations  Distributed Offline Computing development and integration We are thrilled about continuing our collaboration with KISTI  Successful visiting program: FermiCloud enhancement & Grid knowledge exchange  Remote Support for CDF and OSG interfaces  SuperComputing 2011 Posters on Collaboration 35