Minerva Infrastructure Meeting – October 04, 2011.

Slides:



Advertisements
Similar presentations
Wei Lu 1, Kate Keahey 2, Tim Freeman 2, Frank Siebenlist 2 1 Indiana University, 2 Argonne National Lab
Advertisements

FP7-INFRA Enabling Grids for E-sciencE EGEE Induction Grid training for users, Institute of Physics Belgrade, Serbia Sep. 19, 2008.
Esma Yildirim Department of Computer Engineering Fatih University Istanbul, Turkey DATACLOUD 2013.
GridFTP: File Transfer Protocol in Grid Computing Networks
Data Grids: Globus vs SRB. Maturity SRB  Older code base  Widely accepted across multiple communities  Core components are tightly integrated Globus.
70-290: MCSE Guide to Managing a Microsoft Windows Server 2003 Environment Chapter 8: Implementing and Managing Printers.
70-290: MCSE Guide to Managing a Microsoft Windows Server 2003 Environment, Enhanced Chapter 8: Implementing and Managing Printers.
70-290: MCSE Guide to Managing a Microsoft Windows Server 2003 Environment Chapter 8: Implementing and Managing Printers.
Zach Miller Condor Project Computer Sciences Department University of Wisconsin-Madison Flexible Data Placement Mechanisms in Condor.
Hands-On Microsoft Windows Server 2008 Chapter 1 Introduction to Windows Server 2008.
CONDOR DAGMan and Pegasus Selim Kalayci Florida International University 07/28/2009 Note: Slides are compiled from various TeraGrid Documentations.
High Throughput Parallel Computing (HTPC) Dan Fraser, UChicago Greg Thain, Uwisc.
MCTS Guide to Microsoft Windows Server 2008 Applications Infrastructure Configuration (Exam # ) Chapter Two Deploying Windows Servers.
OSG End User Tools Overview OSG Grid school – March 19, 2009 Marco Mambelli - University of Chicago A brief summary about the system.
National Alliance for Medical Image Computing Grid Computing with BatchMake Julien Jomier Kitware Inc.
OSG Public Storage and iRODS
Microsoft and Community Tour 2011 – Infrastrutture in evoluzione Community Tour 2011 Infrastrutture in evoluzione.
03/27/2003CHEP20031 Remote Operation of a Monte Carlo Production Farm Using Globus Dirk Hufnagel, Teela Pulliam, Thomas Allmendinger, Klaus Honscheid (Ohio.
VOX Project Status T. Levshina. Talk Overview VOX Status –Registration –Globus callouts/Plug-ins –LRAS –SAZ Collaboration with VOMS EDG team Preparation.
1 Evolution of OSG to support virtualization and multi-core applications (Perspective of a Condor Guy) Dan Bradley University of Wisconsin Workshop on.
Active Directory Windows2003 Server. Agenda What is Active Directory What is Active Directory Building an Active Directory Building an Active Directory.
Integrating HPC into the ATLAS Distributed Computing environment Doug Benjamin Duke University.
Scalable Systems Software Center Resource Management and Accounting Working Group Face-to-Face Meeting October 10-11, 2002.
A. Sim, CRD, L B N L 1 OSG Applications Workshop 6/1/2005 OSG SRM/DRM Readiness and Plan Alex Sim / Jorge Rodriguez Scientific Data Management Group Computational.
SAMGrid as a Stakeholder of FermiGrid Valeria Bartsch Computing Division Fermilab.
1 Introduction to Microsoft Windows 2000 Windows 2000 Overview Windows 2000 Architecture Overview Windows 2000 Directory Services Overview Logging On to.
Grid Execution Management for Legacy Code Applications Grid Enabling Legacy Code Applications Tamas Kiss Centre for Parallel.
Module 3 Configuring File Access and Printers on Windows ® 7 Clients.
A Brief Documentation.  Provides basic information about connection, server, and client.
Communicating Security Assertions over the GridFTP Control Channel Rajkumar Kettimuthu 1,2, Liu Wantao 3,4, Frank Siebenlist 1,2 and Ian Foster 1,2,3 1.
Module 3 Configuring File Access and Printers on Windows 7 Clients.
Module 3: Configuring File Access and Printers on Windows 7 Clients
Grid Execution Management for Legacy Code Applications Grid Enabling Legacy Applications.
CEOS Working Group on Information Systems and Services - 1 Data Services Task Team Discussions on GRID and GRIDftp Stuart Doescher, USGS WGISS-15 May 2003.
US LHC OSG Technology Roadmap May 4-5th, 2005 Welcome. Thank you to Deirdre for the arrangements.
Campus grids: e-Infrastructure within a University Mike Mineter National e-Science Centre 14 February 2006.
CEDPS Data Services Ann Chervenak USC Information Sciences Institute.
GLIDEINWMS - PARAG MHASHILKAR Department Meeting, August 07, 2013.
System Center & SharePoint On- Prem Matija Blagus, Acceleratio
AliEn AliEn at OSC The ALICE distributed computing environment by Bjørn S. Nilsen The Ohio State University.
1 e-Science AHM st Aug – 3 rd Sept 2004 Nottingham Distributed Storage management using SRB on UK National Grid Service Manandhar A, Haines K,
December 26, 2015 RHIC/USATLAS Grid Computing Facility Overview Dantong Yu Brookhaven National Lab.
Tier 3 Status at Panjab V. Bhatnagar, S. Gautam India-CMS Meeting, July 20-21, 2007 BARC, Mumbai Centre of Advanced Study in Physics, Panjab University,
FermiGrid School Steven Timm FermiGrid School FermiGrid 201 Scripting and running Grid Jobs.
ALCF Argonne Leadership Computing Facility GridFTP Roadmap Bill Allcock (on behalf of the GridFTP team) Argonne National Laboratory.
Dynamic Creation and Management of Runtime Environments in the Grid Kate Keahey Matei Ripeanu Karl Doering.
1 Overview of Microsoft Windows 2000 Multipurpose OS Reduces total cost of ownership (TCO)
Bulk Data Transfer Activities We regard data transfers as “first class citizens,” just like computational jobs. We have transferred ~3 TB of DPOSS data.
The Gateway Computational Web Portal Marlon Pierce Indiana University March 15, 2002.
Parag Mhashilkar Computing Division, Fermi National Accelerator Laboratory.
Open Science Grid Build a Grid Session Siddhartha E.S University of Florida.
R. Krempaska, October, 2013 Wir schaffen Wissen – heute für morgen Controls Security at PSI Current Status R. Krempaska, A. Bertrand, C. Higgs, R. Kapeller,
VOX Project Status T. Levshina. 5/7/2003LCG SEC meetings2 Goals, team and collaborators Purpose: To facilitate the remote participation of US based physicists.
LHCC Referees Meeting – 28 June LCG-2 Data Management Planning Ian Bird LHCC Referees Meeting 28 th June 2004.
Grid Execution Management for Legacy Code Architecture Exposing legacy applications as Grid services: the GEMLCA approach Centre.
TANYA LEVSHINA Monitoring, Diagnostics and Accounting.
Breaking the frontiers of the Grid R. Graciani EGI TF 2012.
StoRM + Lustre Proposal YAN Tian On behalf of Distributed Computing Group
Open Science Grid Consortium Storage on Open Science Grid Placing, Using and Retrieving Data on OSG Resources Abhishek Singh Rana OSG Users Meeting July.
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI EGI Services for Distributed e-Infrastructure Access Tiziana Ferrari on behalf.
9 Copyright © 2004, Oracle. All rights reserved. Getting Started with Oracle Migration Workbench.
Job submission overview Marco Mambelli – August OSG Summer Workshop TTU - Lubbock, TX THE UNIVERSITY OF CHICAGO.
Parag Mhashilkar (Fermi National Accelerator Laboratory)
Fermilab Scientific Computing Division Fermi National Accelerator Laboratory, Batavia, Illinois, USA. Off-the-Shelf Hardware and Software DAQ Performance.
Condor Week Apr 30, 2008Pseudo Interactive monitoring - I. Sfiligoi1 Condor Week 2008 Pseudo-interactive monitoring in Condor by Igor Sfiligoi.
Minos Computing Infrastructure Arthur Kreymer 1 ● Young Minos issues: – /local/stage1/minosgli temporary reprieve – need grid shared area – Retiring.
and Alexandre Duarte OurGrid/EELA Interoperability Meeting
Concept of VLAN (Virtual LAN) and Benefits
Getting Started.
Getting Started.
Presentation transcript:

Minerva Infrastructure Meeting – October 04, 2011

Overview  Problem Statement  Requirements  Proto-type Running jobs on FermiGrid Running jobs outside Fermilab 10/04/2011End-To-End Solution: GlideinWMS & GlobusOnline2

Problem Statement  Current Model Community uses glideinWMS to run workflows Transfer output files - ○ to Bluearc experiment mount points ○ using custom scripts (CPN)  Problem Ownership of the output files after transfer is not preserved correctly  Proposed Solution Provide an end-to-end solution that would integrate well with glideinWMS and the custom data transfer solutions Two sub-projects ○ Setup gridftp server to facilitate data transfer Required component to preserve the file ownership ○ Integrate with the Globus Online Service Optional component that leverages on the modern techniques and solutions like Globus Online as provided by the Grid community Futuristic – CPN script can not work outside Fermigrid setup Work was done as part of CEDPS activity in the Fermilab (Ended Sep 2011) Project Home: Project Document: 10/04/2011End-To-End Solution: GlideinWMS & GlobusOnline3

Requirements  Key Requirements: Solution must – Use user membership & account information from VOMS & NIS servers Preserve the file ownership properties of a transferred file. i.e. uid.gid Be compatible with existing grid processing infrastructure Must support existing strong authentications i.e. Kerberos & X509 Must not adversely affect the performance of the central disk services. i.e. support for appropriate access locks and bandwidth throttling to the central disk Support a minimum transfer bandwidth of 50Mb/s per client, and an aggregated bandwidth sufficient to saturate the network connection to the storage destination Be able to scale to support a minimum of 500 simultaneous client jobs slots per experiment  Other Requirements: Solution should – Be convenient enough to replicate for different IF experiments Not impact other experiments that are using the provided grid infrastructure Fermigrid members must have administrative rights on the machine Include diagnostic tools and support for system monitoring 10/04/20114End-To-End Solution: GlideinWMS & GlobusOnline

Proto-type  Services hosted on FermiCloud VMs  Core setup Setup Gridftp service on FermiCloud VM (if-gridftp.fnal.gov) ○ Each experiment has their own VM (minerva-gridftp.fnal.gov, minos-gridftp.fnal.gov, nova-gridftp.fnal.gov …) ○ VMs are identical Easy to clone for different experiments Base image to clone from available Differ in options used to invoke the tools ○ Setup multiple VMs behind LVS to increase scalability Tools to generate - ○ Gridftp access list (gridmap file) populated from experiment subgroup in Fermilab VOMS server ○ Local user accounts from experiment NIS servers Scale testing by Dennis Box ○ Ran ~500 jobs & ~600 files transferred using globus-url-copy with locking mechanism ○ More scale testing in progress Enhance CPN to use globus-url-copy ○ Work in progress by Dennis Box 10/04/2011End-To-End Solution: GlideinWMS & GlobusOnline5

Proto-type Scenario: Computing on FermiGrid resources FermiCloud 10/04/20116End-To-End Solution: GlideinWMS & GlobusOnline FermiGrid Site Transfers through if-gridftp.fnal.gov using custom scripts (CPN)

End-To-End Solution using GlideinWMS & Globus Online FermiCloud 10/04/20117End-To-End Solution: GlideinWMS & GlobusOnline Grid Site (Not FermiGrid) Transfers through if-gridftp.fnal.gov using custom scripts (CPN)

Proto-type […contd.]  Setup using third-party file transfer services Running workflows on non Fermilab sites ○ Experiment disks not mounted on worker nodes ○ CPN script cannot be used for locking without major changes – is it even possible to use CPN? Other solutions ○ Let infrastructure transfer the files rather than application ○ Use condor file transfer plugin Globus Online plugin through glideinWMS Reuse the infrastructure from core setup Register if-gridftp.fnal.gov as a Globus Online end point ○ Scheme can be extended to other file transfer protocols 10/04/2011End-To-End Solution: GlideinWMS & GlobusOnline8

Deployment: Steps taken  Setup if-gridftp.fnal.gov Core setup - ○ Create a FermiCloud VM ○ Install VDT & configure gridftp service ○ Tool to create gridmap file from experiment specific VOMS group ○ Tool to create local user accounts from experiment specific NIS server Setup related to Globus Online ○ Create a Globus Online account (few simple steps) ○ Register the gridftp service with Globus Online service  glideinWMS deployment (fermicloud049.fnal.gov) Configure Frontend to use globusonline_plugin.py Run test jobs 10/04/2011End-To-End Solution: GlideinWMS & GlobusOnline9

Sample Condor JDF # CONDOR JDF for test job universe = vanilla executable = /local/home/testuser/testjobs/info-from-go.sh initialdir = /local/home/testuser/testjobs/joboutput output = out.$(cluster).$(process) error = err.$(cluster).$(process) log = log.$(cluster).$(process) should_transfer_files = YES when_to_transfer_output = ON_EXIT_OR_EVICT # Input files transfer_input_files = globusonline://parag:paragtest:/grid/data/parag/data # Output sandbox transfer output_destination = globusonline://parag:if-go-ep:/tmp/uploaded_data-fnalgrid transfer_output_files = data_to_upload x509userproxy=/local/home/testuser/testjobs/x509up_u11017.fermilab +DESIRED_Sites = "GR7x1“ Requirements = (stringListMember(GLIDEIN_Site,DESIRED_Sites)) && (Disk > 0 && (Arch == "INTEL" || Arch == "X86_64")) Queue 10/04/2011End-To-End Solution: GlideinWMS & GlobusOnline10