TG Quarterly Meeting Breckenridge, CO Apr 11, 2007 NCSA TG RP Update 1Q07.

Slides:



Advertisements
Similar presentations
TeraGrid Community Software Areas (CSA) JP (John-Paul) Navarro TeraGrid Grid Infrastructure Group Software Integration University of Chicago and Argonne.
Advertisements

Test harness and reporting framework Shava Smallen San Diego Supercomputer Center Grid Performance Workshop 6/22/05.
Statewide IT Conference30-September-2011 HPC Cloud Penguin on David Hancock –
SAN DIEGO SUPERCOMPUTER CENTER at the UNIVERSITY OF CALIFORNIA; SAN DIEGO SDSC RP Update October 21, 2010.
NSF Site Visit HYDRA Using Windows Desktop Systems in Distributed Parallel Computing.
Information Technology Center Introduction to High Performance Computing at KFUPM.
ASKAP Central Processor: Design and Implementation Calibration and Imaging Workshop 2014 ASTRONOMY AND SPACE SCIENCE Ben Humphreys | ASKAP Software and.
SUMS Storage Requirement 250 TB fixed disk cache 130 TB annual increment for permanently on- line data 100 TB work area (not controlled by SUMS) 2 PB near-line.
1 PSC update July 3, 2008 Ralph Roskies, Scientific Director Pittsburgh Supercomputing Center
NWfs A ubiquitous, scalable content management system with grid enabled cross site data replication and active storage. R. Scott Studham.
Academic and Research Technology (A&RT)
1 Maui High Performance Computing Center Open System Support An AFRL, MHPCC and UH Collaboration December 18, 2007 Mike McCraney MHPCC Operations Director.
Site Report US CMS T2 Workshop Samir Cury on behalf of T2_BR_UERJ Team.
Administration and management of Windows-based clusters Windows HPC Server 2008 Matej Ciesko HPC Consultant, PM
Research Computing with Newton Gerald Ragghianti Newton HPC workshop Sept. 3, 2010.
TeraGrid Information Services December 1, 2006 JP Navarro GIG Software Integration.
Purdue RP Highlights TeraGrid Round Table September 23, 2010 Carol Song Purdue TeraGrid RP PI Rosen Center for Advanced Computing Purdue University.
Progress on TeraGrid Stability for the LEAD project.
National Center for Supercomputing Applications The Computational Chemistry Grid: Production Cyberinfrastructure for Computational Chemistry PI: John Connolly.
Kento Aida, Tokyo Institute of Technology Grid Challenge - programming competition on the Grid - Kento Aida Tokyo Institute of Technology 22nd APAN Meeting.
Site Lightning Report: MWT2 Mark Neubauer University of Illinois at Urbana-Champaign US ATLAS Facilities UC Santa Cruz Nov 14, 2012.
Big Red II & Supporting Infrastructure Craig A. Stewart, Matthew R. Link, David Y Hancock Presented at IUPUI Faculty Council Information Technology Subcommittee.
August 2007 Advancing Scientific Discovery through TeraGrid Adapted from S. Lathrop’s talk in SC’07
Introduction to the HPCC Dirk Colbry Research Specialist Institute for Cyber Enabled Research.
TeraGrid Overview Cyberinfrastructure Days Internet2 10/9/07 Mark Sheddon Resource Provider Principal Investigator San Diego Supercomputer Center
1 Preparing Your Application for TeraGrid Beyond 2010 TG09 Tutorial June 22, 2009.
1 PY4 Project Report Summary of incomplete PY4 IPP items.
Kelly Gaither Visualization Area Report. Efforts in 2008 Focused on providing production visualization capabilities (software and hardware) Focused on.
14 Aug 08DOE Review John Huth ATLAS Computing at Harvard John Huth.
National Computational Science National Center for Supercomputing Applications National Computational Science NCSA Terascale Clusters Dan Reed Director,
TeraGrid Privacy Policy: What is it and why are we doing it… Von Welch TeraGrid Quarterly Meeting March 6, 2008.
Headline in Arial Bold 30pt HPC User Forum, April 2008 John Hesterberg HPC OS Directions and Requirements.
TeraGrid CTSS Plans and Status Dane Skow for Lee Liming and JP Navarro OSG Consortium Meeting 22 August, 2006.
The LBNL Perceus Cluster Infrastructure Next Generation Cluster Provisioning and Management October 10, 2007 Internet2 Fall Conference Gary Jung, SCS Project.
TeraGrid Quarterly Meeting Arlington, VA Sep 6-7, 2007 NCSA RP Status Report.
NICS RP Update TeraGrid Round Table March 10, 2011 Ryan Braby NICS HPC Operations Group Lead.
Sergiu April 2006June 2006 Overview of TeraGrid Resources and Services Sergiu Sanielevici, TeraGrid Area Director for User.
NOS Report Jeff Koerner Feb 10 TG Roundtable. Security-wg In Q a total of 11 user accounts and one login node were compromised. The Security team.
User-Facing Projects Update David Hart, SDSC April 23, 2009.
Data, Visualization and Scheduling (DVS) TeraGrid Annual Meeting, April 2008 Kelly Gaither, GIG Area Director DVS.
GridChem Developers Conference Focus For Final Year Sudhakar Pamidighantam NCSA 25 August 2006.
GridChem Sciene Gateway and Challenges in Distributed Services Sudhakar Pamidighantam NCSA, University of Illinois at Urbaba- Champaign
NICS Update Bruce Loftis 16 December National Institute for Computational Sciences University of Tennessee and ORNL partnership  NICS is the 2.
TeraGrid-Wide Operations Von Welch Area Director for Networking, Operations and Security NCSA, University of Illinois April, 2009.
TeraGrid-Wide Operations DRAFT #2 Mar 31 Von Welch.
Grid Remote Execution of Large Climate Models (NERC Cluster Grid) Dan Bretherton, Jon Blower and Keith Haines Reading e-Science Centre
User Champion Field Report 12/11/2008 Chris Hempel, TACC
Attribute-based Authentication for Gateways Jim Basney Terry Fleury Stuart Martin JP Navarro Tom Scavo Nancy Wilkins-Diehr.
Gateway Security Summit, January 28-30, 2008 Welcome to the Gateway Security Summit Nancy Wilkins-Diehr Science Gateways Area Director.
Quality Assurance (QA) Working Group Update July 1, 2010 Kate Ericson (SDSC) Shava Smallen (SDSC)
CTSS Rollout update Mike Showerman JP Navarro April
AT LOUISIANA STATE UNIVERSITY CCT: Center for Computation & Technology Introduction to the TeraGrid Daniel S. Katz Lead, LONI as a TeraGrid.
TeraGrid’s Common User Environment: Status, Challenges, Future Annual Project Review April, 2008.
Software Integration Highlights CY2008 Lee Liming, JP Navarro GIG Area Directors for Software Integration University of Chicago, Argonne National Laboratory.
Evangelos Markatos and Charalampos Gkikas FORTH-ICS Athens, th Mar Institute of Computer Science - FORTH Christos.
Getting Started: XSEDE Comet Shahzeb Siddiqui - Software Systems Engineer Office: 222A Computer Building Institute of CyberScience May.
Scheduling a 100,000 Core Supercomputer for Maximum Utilization and Capability September 2010 Phil Andrews Patricia Kovatch Victor Hazlewood Troy Baer.
Computational Sciences at Indiana University an Overview Rob Quick IU Research Technologies HTC Manager.
TG ’08, June 9-13, State of TeraGrid John Towns Co-Chair, TeraGrid Forum Director, Persistent Infrastructure National Center for Supercomputing.
TeraGrid Capability Discovery John-Paul “JP” Navarro TeraGrid Area Co-Director for Software Integration University of Chicago/Argonne National Laboratory.
Creating Grid Resources for Undergraduate Coursework John N. Huffman Brown University Richard Repasky Indiana University Joseph Rinkovsky Indiana University.
Quarterly Meeting Spring 2007 NSTG: Some Notes of Interest Adapting Neutron Science community codes for TeraGrid use and deployment. (Lynch, Chen) –Geared.
TeraGrid Software Integration: Area Overview (detailed in 2007 Annual Report Section 3) Lee Liming, JP Navarro TeraGrid Annual Project Review April, 2008.
Slide 1 Cluster Workload Analytics Revisited Saurabh Bagchi Purdue University Joint work with: Subrata Mitra, Suhas Javagal, Stephen Harrell (Purdue),
Quality Assurance Working Group Doru Marcusiu, NCSA QA Working Group Lead TeraGrid Annual Review April, 2009.
OpenPBS – Distributed Workload Management System
Heterogeneous Computation Team HybriLIT
Porting MM5 and BOLAM codes to the GRID
Stallo: First impressions
System G And CHECS Cal Ribbens
Presentation transcript:

TG Quarterly Meeting Breckenridge, CO Apr 11, 2007 NCSA TG RP Update 1Q07

TG Quarterly Meeting Breckenridge, CO Apr 11, 2007 CSE-Online Science Gateway  Production Date: Mar 9, 2007  Developed under ITR program  DAC Community Allocation  MRAC Community Allocation just awarded  Dedicated 4 nodes on Mercury  Results from first 30 days (next slide) – Gaussian jobs running in restricted shell  Changing reservation to 1 node based on results, will continue to monitor usage

TG Quarterly Meeting Breckenridge, CO Apr 11, 2007 CSE Online Utilization Dedicated 4 nodes initially, now one node Goal: improved turnaround for a large number of small jobs submitted through the gateway.

TG Quarterly Meeting Breckenridge, CO Apr 11, 2007 LEAD Science Gateway  Supported Spring Weather Challenge ( forecasting contest for undergraduate atmospheric science studentswww.wxchallenge.com  Feb 19-26: daily testing, 80 processors, 12pm-5pm  Feb 26-April 27th: 160 processors; 12pm-5pm Monday through Thursday.  Actual contest submissions started week of March 26

TG Quarterly Meeting Breckenridge, CO Apr 11, 2007 LEAD Gateway Statistics  250 jobs per week, consuming 1800 SUs/week  Each workflow is 5 jobs –  250 jobs corresponds to 50 workflows  Expect this to increase once issues are resolved/reliability improves  LEAD Gateway typically the most or 2 nd most active gateway in terms of resources used  (BIRN or GridChem are often ahead)

TG Quarterly Meeting Breckenridge, CO Apr 11, 2007 Issues Uncovered by both Science Gateways  Remote job submission – great when jobs run – hard to know problems – even simple things such as planned downtime  Reservation Issues – can’t overflow end of reservation when many jobs stack up (LEAD)  If user assigns an obsolete project, don’t get useful error message back  GridFTP striped server – one fails, all fail

TG Quarterly Meeting Breckenridge, CO Apr 11, 2007 SG Next Steps  Meetings with teams to understand usage modes and issues  CSE Online NCSA contingent visiting CSE Online group at Univ of Utah Apr 23 – 25  LEAD NCSA and IU RP’s setting up a date to visit LEAD group at IU

TG Quarterly Meeting Breckenridge, CO Apr 11, 2007 New Resource - Abe  Abe: 1955 blade cluster  2.33 GHz Cloverton Quad-Core 1,200 blades/9,600 cores 89.5 TF; 9.6 TB RAM; 120 TB disk Perceus management; diskless boot  Cisco Infiniband 2 to 1 oversubscribed  Lustre over IB 8.4GB/s sustained  Power/Cooling 500 KW / 140 tons  TG Software deployment CTSS Inca  Production date: May 2007 (anticipated)  User Environment Torque/Moab Softenv Intel Compiler MPI: evaluating Intel MPI, MPICH, MVAPICH, VMI-2, etc.

TG Quarterly Meeting Breckenridge, CO Apr 11, 2007 March Allocations  25.1 M SUs (672M NUs) awarded to NCSA systems  34% of allocated resources  Several large supplements coming in after the meeting  Several 1M+ SU NCSA  Silas Beane: 2.0M on Tungsten  Ali Uzun: 2.0M on Abe  Adrian Roitberg: 1.5M on Abe  Thom Cheatham: 1.0M on Abe