The Sharing and Training of HPC Resources at the University of Arkansas Amy Apon, Ph.D. Oklahoma Supercomputing Symposium October 4, 2006.

Slides:



Advertisements
Similar presentations
Building a CFD Grid Over ThaiGrid Infrastructure Putchong Uthayopas, Ph.D Department of Computer Engineering, Faculty of Engineering, Kasetsart University,
Advertisements

C3.ca in Atlantic Canada Virendra Bhavsar Director, Advanced Computational Research Laboratory (ACRL) Faculty of Computer Science University of New Brunswick.
Oklahoma Center for High Energy Physics A DOE EPSCoR Proposal Submitted by a consortium of University of Oklahoma, Oklahoma State University, Langston.
S.Chechelnitskiy / SFU Simon Fraser Running CE and SE in a XEN virtualized environment S.Chechelnitskiy Simon Fraser University CHEP 2007 September 6 th.
DOSAR Workshop VI April 17, 2008 Louisiana Tech Site Report Michael Bryant Louisiana Tech University.
6/2/20071 Grid Computing Sun Grid Engine (SGE) Manoj Katwal.
NICLS: Development of Biomedical Computing and Information Technology Infrastructure Presented by Simon Sherman August 15, 2005.
1 Workshop 20: Teaching a Hands-on Undergraduate Grid Computing Course SIGCSE The 41st ACM Technical Symposium on Computer Science Education Friday.
1 Int System Introduction to Systems and Networking Department Faculty of Computer Science and Engineering Ho Chi Minh City University of Technology.
Arkansas Research and Education Optical Network Arkansas Association of Public Universities Little Rock, Arkansas April 10, 2008 Dr. Robert Zimmerman ARE-ON.
Jenny Jopling Texas Computer-based Testing Collaborative.
December 8 & 9, 2005, Austin, TX SURA Cyberinfrastructure Workshop Series: Grid Technology: The Rough Guide Configuring Resources for the Grid Jerry Perez.
Open Science Grid For CI-Days Internet2: Fall Member Meeting, 2007 John McGee – OSG Engagement Manager Renaissance Computing Institute.
National LambdaRail, Higher Education and the Future of Regional Optical Networks Dr. Robert Zimmerman, Director Computing Services University of Arkansas.
The following 3 Alabama slides were lifted from Alabama’s Cyberinfrastructure presentation at EPSCOR meeting in KY in October. There was no one from Alabama.
and beyond Office of Vice President for Information Technology.
November 18, 1999 Internet 2 Mike Rackley Head of Information Technology Services
SURA Regional HPC Grid Proposal Ed Seidel LSU With Barbara Kucera, Sara Graves, Henry Neeman, Otis Brown, others.
The Cyberinfrastructure Strategic Plan for the State of Arkansas Amy Apon, Ph.D., Director Arkansas High Performance Computing Center Professor, CSCE University.
DynamicBLAST on SURAgrid: Overview, Update, and Demo John-Paul Robinson Enis Afgan and Purushotham Bangalore University of Alabama at Birmingham SURAgrid.
Chao “Bill” Xie, Victor Bolet, Art Vandenberg Georgia State University, Atlanta, GA 30303, USA February 22/23, 2006 SURA, Washington DC Memory Efficient.
PCGRID ‘08 Workshop, Miami, FL April 18, 2008 Preston Smith Implementing an Industrial-Strength Academic Cyberinfrastructure at Purdue University.
Kento Aida, Tokyo Institute of Technology Grid Challenge - programming competition on the Grid - Kento Aida Tokyo Institute of Technology 22nd APAN Meeting.
Principles of Scalable HPC System Design March 6, 2012 Sue Kelly Sandia National Laboratories Abstract: Sandia National.
Cyberinfrastructure for Distributed Rapid Response to National Emergencies Henry Neeman, Director Horst Severini, Associate Director OU Supercomputing.
The Red Storm High Performance Computer March 19, 2008 Sue Kelly Sandia National Laboratories Abstract: Sandia National.
17-April-2007 High Performance Computing Basics April 17, 2007 Dr. David J. Haglin.
LONI Overview State-wide IT initiative: $25M – Gov. Mike Foster, present LONI - $40M, Gov. Kathleen Blanco, LONI - $10M, Gov. Kathleen.
Common Practices for Managing Small HPC Clusters Supercomputing 12
Top Issues Facing Information Technology at UAB Sheila M. Sanders UAB Vice President Information Technology February 8, 2007.
SURA GridPlan Infrastructure Working Group Art Vandenberg Georgia State University Mary Fran Yafchak SURA Working.
Amy Apon, Pawel Wolinski, Dennis Reed Greg Amerson, Prathima Gorjala University of Arkansas Commercial Applications of High Performance Computing Massive.
Edinburgh Investment in e-Science Infrastructure Dr Arthur Trew.
Looking Ahead: A New PSU Research Cloud Architecture Chuck Gilbert - Systems Architect and Systems Team Lead Research CI Coordinating Committee Meeting.
+ Successes, Opportunities, Challenges PRAGMA19 Putchong Uthayopas, Kasetsart University.
Enhancing Networking Expertise Across the Great Plains Greg Monaco, Ph.D. Director for Research & Cyberinfrastructure Initiatives Great Plains Network.
National Science Foundation CI-TEAM Proposal: Blast on Condor How Will This Help [InstAbbrev]? Your Name Here Your Job Title Here Your Department Here.
CCS Overview Rene Salmon Center for Computational Science.
Institute For Digital Research and Education Implementation of the UCLA Grid Using the Globus Toolkit Grid Center’s 2005 Community Workshop University.
Supercomputing at the University of Arkansas Amy Apon, Ph.D. Oklahoma Supercomputing Symposium October 5, 2005.
Obtaining Computer Allocations and Monitoring Use SCD User Meeting at AMS January 11, 2005 Ginger Caldwell, SCD.
Authors: Ronnie Julio Cole David
SURAGrid Project Meeting Washington, DC Wednesday, February 22, 2006 Barry Wilkinson Department of Computer Science UNC-Charlotte SURAGrid and Grid Computing.
ISU DOSAR WORKSHOP Dick Greenwood DOSAR/OSG Statement of Work (SoW) Dick Greenwood Louisiana Tech University April 5, 2007.
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
August 3, March, The AC3 GRID An investment in the future of Atlantic Canadian R&D Infrastructure Dr. Virendra C. Bhavsar UNB, Fredericton.
Cluster Software Overview
Network Status Update AREON Planning Meeting September 15, 2006 Little Rock, Arkansas David Merrifield University of Arkansas AREON Planning Meeting September.
11 January 2005 High Performance Computing at NCAR Tom Bettge Deputy Director Scientific Computing Division National Center for Atmospheric Research Boulder,
Launching a great program! SURAgrid All Hands Meeting – September 21, 2006.
1 NSF/TeraGrid Science Advisory Board Meeting July 19-20, San Diego, CA Brief TeraGrid Overview and Expectations of Science Advisory Board John Towns TeraGrid.
Arkansas Research and Education Optical Network (ARE_ON) What is it? Why do we need it? What will it do?
ACI-REF Virtual Residency Henry Neeman, University of Oklahoma Director, OU Supercomputing Center for Education & Research (OSCER) Assistant Vice President,
Regional Cyberinfrastructure Planning Great Plains Network Greg Monaco, Ph.D. Director for Research and Cyberinfrastructure Initiatives
© Copyright AARNet Pty Ltd PRAGMA Update & some personal observations James Sankar Network Engineer - Middleware.
National Computational Science Ky PACS at the University of Kentucky April 2000 –Advanced Computing Resources –EPSCoR Outreach –SURA Liaison –John.
OKLAHOMA Supercomputing Symposium 2011 University of Oklahoma October 11, 2011 James Wicksted, RII Project Director Associate Director, Oklahoma EPSCoR.
AT LOUISIANA STATE UNIVERSITY CCT: Center for Computation & LSU Condor in Louisiana Tevfik Kosar Center for Computation & Technology Louisiana.
DutchGrid KNMI KUN Delft Leiden VU ASTRON WCW Utrecht Telin Amsterdam Many organizations in the Netherlands are very active in Grid usage and development,
Scheduling systems Carsten Preuß
Deploying Regional Grids Creates Interaction, Ideas, and Integration
Introduction to XSEDE Resources HPC Workshop 08/21/2017
Building a Cyberinfrastructure Culture: IT as a Partner in Research
SA1 ROC Meeting Bologna, October 2004
XSEDE’s Campus Bridging Project
Regional Cyberinfrastructure Planning
Introduce yourself Presented by
Henry Neeman, University of Oklahoma
Overview of the Great Plains Network
Presentation transcript:

The Sharing and Training of HPC Resources at the University of Arkansas Amy Apon, Ph.D. Oklahoma Supercomputing Symposium October 4, 2006

Amy Apon, Ph.D. ● University of Arkansas ● October 4, Outline of Talk HPC at the University of Arkansas –Current status A New Mechanism for Sharing Resources –AREON HPC Training –New course delivery via HDTV collaboration with LSU Collaboration opportunities and challenges –GPNGrid and SURAGrid –Resource allocation issues

Amy Apon, Ph.D. ● University of Arkansas ● October 4, High Performance Computing Resources at the University of Arkansas Red Diamond supercomputer –NSF MRI grant, August, 2004 Substantial University match Substantial gift from Dell –First supercomputer in Arkansas Number 379 on the Top 500 list, June, node (256 processor), TFlops

Amy Apon, Ph.D. ● University of Arkansas ● October 4, More Resources Prospero cluster –30 dual processor PIII nodes –SURAGrid resource Ace cluster –4 dual processor Opteron –Our entry point to the GPNGrid/Open Science Grid Trillion cluster –48 dual processor Opteron –Owned by Mechanical Engineering –About 1TFlop

Amy Apon, Ph.D. ● University of Arkansas ● October 4, How are we doing

Amy Apon, Ph.D. ● University of Arkansas ● October 4, We are seeing research results Computational Chemistry and Materials Science (NSF) –New formulas for new drugs –Nanomaterials –Chemistry, Physics, Mechanical Engineering –Over 95% of our usage is in these areas

Amy Apon, Ph.D. ● University of Arkansas ● October 4, Research results in other areas, also Multiscale Modeling DNA Computing Middleware and HPC Infrastructure –Tools for managing data for large-scale applications (NSF) –Performance modeling of grid systems (Acxiom)

Amy Apon, Ph.D. ● University of Arkansas ● October 4, We have done some significant upgrades For the first year we used SGE on half the computer and half of the computer was self- scheduled PVM jobs LSF scheduler installed May 2006 About 60 users, about 10 very active users

Amy Apon, Ph.D. ● University of Arkansas ● October 4, Thanks to LSF, we are busy LSF Daily Pending Parallel Job Statistics by Queue (jobs waiting)

Amy Apon, Ph.D. ● University of Arkansas ● October 4, And jobs have to wait LSF Hourly Turnaround Time of Normal Queue

Amy Apon, Ph.D. ● University of Arkansas ● October 4, We have something exciting to share

12 Fayetteville Magnolia Russellville Fort Smith Arkadelphia Conway Monticello Pine Bluff Jonesboro Little Rock TULSA DALLAS MEMPHIS MONROE 25-July-2006 AREON Arkansas Research and Education Optical Network We ARE ON !

13 The first bond issue (last fall) failed Governor Huckabee of Arkansas granted $6.4M (PI Zimmerman) MBO loop between Tulsa and Fayetteville fiber is in place, network hardware is being shipped The campus (last mile) connections are in progress All is on target for a demo to the Governor on 12/5/06! Fayetteville Magnolia Russellville Fort Smith Arkadelphia Conway Monticello Pine Bluff Jonesboro Little Rock TULSA DALLAS MEMPHIS MONROE AREON Arkansas Research and Education Optical Network

Amy Apon, Ph.D. ● University of Arkansas ● October 4, AREON Arkansas Research and Education Optical Network This fall, Uark will have connectivity to Internet2 and the National Lambda Rail The bond issue is on the ballot again this coming fall If it passes then the other research institutions will be connected to AREON We hope this happens! The timeframe for this is about a year and a half

Amy Apon, Ph.D. ● University of Arkansas ● October 4, Opportunities for collaboration with OneNet, LEARN, LONI, GPN, and others

Amy Apon, Ph.D. ● University of Arkansas ● October 4, A Demonstration Application High Performance Computing New course in Spring 2007 In collaboration with LSU and Dr. Thomas Sterling –We are exploring new methods of course delivery using streaming high-definition TV –We expect about 40 students at five locations this time –Taught live via Access Grid and HDTV over AREON and LONI, … –A test run for future delivery of HPC education

Amy Apon, Ph.D. ● University of Arkansas ● October 4, Collaboration via GPN Grid –Active middleware collaboration for almost 3 years –GPNGrid is in the process of making application as a new Virtual Organization in Open Science Grid –Sponsored by University of Nebraska – Lincoln, includes participants from Arkansas, UNL, Missouri, KU, KSU, OU –Hardware grant from Sun and NSF provide 4 small Opteron clusters for the starting grid environment –Applications are in the process of being defined

Amy Apon, Ph.D. ● University of Arkansas ● October 4, Collaboration via SURA Grid Uark has a 30-node Pentium cluster in SURAGrid Some differences with GPN –CA is different –Account management, discovery stacks are different –AUP policy is different - SURA Grid applications are increasing. Uark can run coastal modeling and is open to running other SURA applications

Amy Apon, Ph.D. ● University of Arkansas ● October 4, More Collaboration Mechanisms Arkansas is participating with the recently awarded CI-TEAM award to OU, PI Neeman –Will deploy Condor across Oklahoma and with participating collaborators LSF Multicluster provides another mechanism for collaboration AREON will give the University of Arkansas great bandwidth

Amy Apon, Ph.D. ● University of Arkansas ● October 4, UofA Current HPC Challenges We have some I/O infrastructure challenges –The system was designed to have a large amount of storage, but it is not fast Supercomputing operations –AC, power, and UPS need to be upgraded Funding models for on-going operations –How will basic systems administration and project director be funded?

Amy Apon, Ph.D. ● University of Arkansas ● October 4, Collaboration and sharing bring a challenge Usage policies –How do you partition usage fairly among existing users? –How do you incorporate usage from new faculty? Current policy uses a fair-share scheduling policy. Dynamic Priority = (# shares) / (#slots*F1 + cpu_time*F2 + run_time*F3); Shares divided among largest users groups: chem, phys, others

Amy Apon, Ph.D. ● University of Arkansas ● October 4, Collaboration and sharing bring a challenge Are max run times needed? –Almost everyone has them –Requires checkpointing of jobs which is hard to do with our current I/O infrastructure –Requires user education and a change of culture Are user allocations and accounting of usage needed? Your suggestions here

Amy Apon, Ph.D. ● University of Arkansas ● October 4, Questions? Contact information: