1 Development of a High-Throughput Computing Cluster at Florida Tech P. FORD, R. PENA, J. HELSBY, R. HOCH, M. HOHLMANN Physics and Space Sciences Dept,

Slides:



Advertisements
Similar presentations
Florida Tech Grid Cluster P. Ford 2 * X. Fave 1 * M. Hohlmann 1 High Energy Physics Group 1 Department of Physics and Space Sciences 2 Department of Electrical.
Advertisements

CHEPREO Tier-3 Center Achievements. FIU Tier-3 Center Tier-3 Centers in the CMS computing model –Primarily employed in support of local CMS physics community.
Dynamic Load Balancing for VORPAL Viktor Przebinda Center for Integrated Plasma Studies.
User Documentation.  You cannot build a system for a client and leave them without adequate documentation  Computer systems are complex, require configuration.
Overview of Wisconsin Campus Grid Dan Bradley Center for High-Throughput Computing.
ANTHONY TIRADANI AND THE GLIDEINWMS TEAM glideinWMS in the Cloud.
Setting up of condor scheduler on computing cluster Raman Sehgal NPD-BARC.
Duke Atlas Tier 3 Site Doug Benjamin (Duke University)
Information Technology Center Introduction to High Performance Computing at KFUPM.
S. Gadomski, "ATLAS computing in Geneva", journee de reflexion, 14 Sept ATLAS computing in Geneva Szymon Gadomski description of the hardware the.
DataGrid Kimmo Soikkeli Ilkka Sormunen. What is DataGrid? DataGrid is a project that aims to enable access to geographically distributed computing power.
Lesson 7: Creating and Configuring Virtual Machine Settings
High Performance Computing (HPC) at Center for Information Communication and Technology in UTM.
Rocks cluster : a cluster oriented linux distribution or how to install a computer cluster in a day.
Computing at COSM by Lawrence Sorrillo COSM Center.
Open Science Grid Software Stack, Virtual Data Toolkit and Interoperability Activities D. Olson, LBNL for the OSG International.
Open Science Grid For CI-Days Internet2: Fall Member Meeting, 2007 John McGee – OSG Engagement Manager Renaissance Computing Institute.
Multi-media Computers and Computer Networks. Questions ? Media is used for ………………… Multimedia computer is capable of integrating ………………………………….. OCR stands.
Cluster currently consists of: 1 Dell PowerEdge Ghz Dual, quad core Xeons (8 cores) and 16G of RAM Original GRIDVM - SL4 VM-Ware host 1 Dell PowerEdge.
So, Jung-ki Distributed Computing System LAB School of Computer Science and Engineering Seoul National University Implementation of Package Management.
Rocks ‘n’ Rolls An Introduction to Programming Clusters using Rocks © 2008 UC Regents Anoop Rajendra.
OSG Site Provide one or more of the following capabilities: – access to local computational resources using a batch queue – interactive access to local.
การติดตั้งและทดสอบการทำคลัสเต อร์เสมือนบน Xen, ROCKS, และไท ยกริด Roll Implementation of Virtualization Clusters based on Xen, ROCKS, and ThaiGrid Roll.
Open Science Grid For CI-Days Elizabeth City State University Jan-2008 John McGee – OSG Engagement Manager Manager, Cyberinfrastructure.
Computing at the High Energy Physics Lab at FIT Patrick Ford, Jen Helsby, Richard Hoch, David Pena Dr. Hohlmann, Dr. Mitra.
QCDGrid Progress James Perry, Andrew Jackson, Stephen Booth, Lorna Smith EPCC, The University Of Edinburgh.
CHEP'07 September D0 data reprocessing on OSG Authors Andrew Baranovski (Fermilab) for B. Abbot, M. Diesburg, G. Garzoglio, T. Kurca, P. Mhashilkar.
Large Scale Sky Computing Applications with Nimbus Pierre Riteau Université de Rennes 1, IRISA INRIA Rennes – Bretagne Atlantique Rennes, France
Grid MP at ISIS Tom Griffin, ISIS Facility. Introduction About ISIS Why Grid MP? About Grid MP Examples The future.
14 Aug 08DOE Review John Huth ATLAS Computing at Harvard John Huth.
6/26/01High Throughput Linux Clustering at Fermilab--S. Timm 1 High Throughput Linux Clustering at Fermilab Steven C. Timm--Fermilab.
José D. Zamora, Sean R. Morriss and Manuela Campanelli.
D0SAR - September 2005 Andre Sznajder 1 Rio GRID Initiatives : T2-HEPGRID Andre Sznajder UERJ(Brazil)
SouthGrid SouthGrid SouthGrid is a distributed Tier 2 centre, one of four setup in the UK as part of the GridPP project. SouthGrid.
RAL Site Report John Gordon IT Department, CLRC/RAL HEPiX Meeting, JLAB, October 2000.
26SEP03 2 nd SAR Workshop Oklahoma University Dick Greenwood Louisiana Tech University LaTech IAC Site Report.
São Paulo Regional Analysis Center SPRACE Status Report 22/Aug/2006 SPRACE Status Report 22/Aug/2006.
Tool Integration with Data and Computation Grid GWE - “Grid Wizard Enterprise”
Issues Autonomic operation (fault tolerance) Minimize interference to applications Hardware support for new operating systems Resource management (global.
OSG Tier 3 support Marco Mambelli - OSG Tier 3 Dan Fraser - OSG Tier 3 liaison Tanya Levshina - OSG.
CEOS WGISS-21 CNES GRID related R&D activities Anne JEAN-ANTOINE PICCOLO CEOS WGISS-21 – Budapest – 2006, 8-12 May.
Status of Florida Tier2 Center A mini tutorial on ROCKS appliances Jorge L. Rodriguez February 2003.
INFSO-RI Enabling Grids for E-sciencE OSG-LCG Interoperability Activity Author: Laurence Field (CERN)
T3 analysis Facility V. Bucard, F.Furano, A.Maier, R.Santana, R. Santinelli T3 Analysis Facility The LHCb Computing Model divides collaboration affiliated.
EVGM081 Multi-Site Virtual Cluster: A User-Oriented, Distributed Deployment and Management Mechanism for Grid Computing Environments Takahiro Hirofuchi,
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
LCG LCG-1 Deployment and usage experience Lev Shamardin SINP MSU, Moscow
Grid Appliance The World of Virtual Resource Sharing Group # 14 Dhairya Gala Priyank Shah.
Testing and integrating the WLCG/EGEE middleware in the LHC computing Simone Campana, Alessandro Di Girolamo, Elisa Lanciotti, Nicolò Magini, Patricia.
Doug Benjamin Duke University. 2 ESD/AOD, D 1 PD, D 2 PD - POOL based D 3 PD - flat ntuple Contents defined by physics group(s) - made in official production.
Final Implementation of a High Performance Computing Cluster at Florida Tech P. FORD, X. FAVE, K. GNANVO, R. HOCH, M. HOHLMANN, D. MITRA Physics and Space.
Miron Livny Computer Sciences Department University of Wisconsin-Madison Condor and (the) Grid (one of.
15-Feb-02Steve Traylen, RAL WP6 Test Bed Report1 RAL/UK WP6 Test Bed Report Steve Traylen, WP6 PPGRID/RAL, UK
Ole’ Miss DOSAR Grid Michael D. Joy Institutional Analysis Center.
Automatic testing and certification procedure for IGI products in the EMI era and beyond Sara Bertocco INFN Padova on behalf of IGI Release Team EGI Community.
A. Mohapatra, T. Sarangi, HEPiX-Lincoln, NE1 University of Wisconsin-Madison CMS Tier-2 Site Report D. Bradley, S. Dasu, A. Mohapatra, T. Sarangi, C. Vuosalo.
Evangelos Markatos and Charalampos Gkikas FORTH-ICS Athens, th Mar Institute of Computer Science - FORTH Christos.
INRNE's participation in LCG Elena Puncheva Preslav Konstantinov IT Department.
Tier 3 Support and the OSG US ATLAS Tier2/Tier3 Workshop at UChicago August 20, 2009 Marco Mambelli –
Group # 14 Dhairya Gala Priyank Shah. Introduction to Grid Appliance The Grid appliance is a plug-and-play virtual machine appliance intended for Grid.
Simulation Production System Science Advisory Committee Meeting UW-Madison March 1 st -2 nd 2007 Juan Carlos Díaz Vélez.
Instituto de Biocomputación y Física de Sistemas Complejos Cloud resources and BIFI activities in JRA2 Reunión JRU Española.
Building on virtualization capabilities for ExTENCI Carol Song and Preston Smith Rosen Center for Advanced Computing Purdue University ExTENCI Kickoff.
CERN LCG1 to LCG2 Transition Markus Schulz LCG Workshop March 2004.
CNAF - 24 September 2004 EGEE SA-1 SPACI Activity Italo Epicoco.
Matt Lemons Nate Mayotte
Leanne Guy EGEE JRA1 Test Team Manager
US CMS Testbed.
Composition and Operation of a Tier-3 cluster on the Open Science Grid
Presentation transcript:

1 Development of a High-Throughput Computing Cluster at Florida Tech P. FORD, R. PENA, J. HELSBY, R. HOCH, M. HOHLMANN Physics and Space Sciences Dept, Florida Institute of Technology, 150 W. University Blvd, Melbourne, FL Abstract Open Science Grid Hardware Conclusions and Future Integration onto the OSG requires the installation and careful configuration of many packages and services. Our group attended the Florida International Grid School before making our second installation attempt. The first installation took three months, the second took three weeks. With all software working as intended, future plans are to expand the cluster resources as much as possible, and to move to the OSG Production grid once we can ensure maximum uptime. A new frontend will be arriving soon along with nodes featuring 2GB memory for each CPU - which is the required hardware for data processing in the CMS experiment. For further information, contact Visit to follow this project. The OSG is a collaboration of many virtual organizations (VOs) ranging from biomedical research to particle physics to software development. Software ROCKS - The ROCKS operating system we are using is version It is based on a community enterprise version of Red Hat Linux called CentOS. Since the first concept and implementation of the computing cluster at Florida Tech, we have increased its size and developed the cluster software significantly. We have implemented the Linux-based ROCKS OS as the central controller of all cluster resources. The cluster now uses the Condor high-throughput batch-job system and has been fully integrated into the Open Science Grid test-bed. In addition to contributing to the data-handling capabilities of worldwide scientific grids, the cluster is being used to process and model high-energy particle simulations such as in Muon radiography. The FLTECH cluster has 20 nodes (40 CPUs) operational with functioning simulation software packages (Geant4). We run all essential non-compute elements (Frontend, NAS, Switches) on a 4 kilowatt uninterruptible power supply that has been programmed to perform automatic shutdowns in the case of an extended power outage. In addition to this, a NAS featuring 10 terabytes of available storage is being installed. condor-computecompute-appliance compute Figure 6: The kickstart graph is the cornerstone of the ROCKS OS. It ensures that packages are installed on all cluster machines in the correct order. The bubble shows our addition of the condor-compute appliance (created with an XML file) to the graph, effectively interfacing condor with ROCKS. Figure 2: The current topology of the cluster (left), and hardware (below) Figure 4: Simulations running on Condor Figure 5: Machines available to Condor References and Acknowledgments Rocks Clusters User Guide: documentation/4.2.1/ Accessed March 2008 Open Science Grid: Accessed March 2008 Condor v Manual: Accessed March 2008 Thanks to Dr J. Rodriguez (FIU) and Micha Niskin (FIU) for their guidance. All recent additions are installed in a second 50U rack from the same manufacturer as the one loaned to us from UF. Our goal is to make all current and future hardware rackmount. Figure 7: A map of Open Science Grid sites, provided by VORS and the grid operations center. Our site is located on the east coast of Florida. Figure 8: A map of OSG sites provided by the MonALISA grid monitoring system. Condor - Condor is the job manager that sends data to be computed to any available machines in the cluster. We use it to delegate our own resource intensive muon tomography simulations so that more work can be accomplished in the same amount of time. Condor now stands as our primary job manager serving the Open Science Grid. Figure 3: Ganglia Cluster Monitoring Cluster communication is in the form of a star topology using a high-end Cisco switch as the central manager, and several Linksys switches as node carriers. Figure 1: New High- end Cluster Hardware (NAS) - (above)