SAN DIEGO SUPERCOMPUTER CENTER Advanced User Support Project Outline October 9th 2008 Ross C. Walker.

Slides:



Advertisements
Similar presentations
SAN DIEGO SUPERCOMPUTER CENTER UNIVERSITY OF CALIFORNIA, SAN DIEGO Scalable Spectral Transforms at Petascale Dmitry Pekurovsky San Diego Supercomputer.
Advertisements

National Center for Supercomputing Applications University of Illinois at Urbana-Champaign Porting, Benchmarking, and Optimizing Computational Material.
SAN DIEGO SUPERCOMPUTER CENTER at the UNIVERSITY OF CALIFORNIA, SAN DIEGO Performance of Applications Using Dual-Rail InfiniBand 3D Torus Network on the.
Lecture 7: 9/17/2002CS170 Fall CS170 Computer Organization and Architecture I Ayman Abdel-Hamid Department of Computer Science Old Dominion University.
Chapter 1 CSF 2009 Computer Performance. Defining Performance Which airplane has the best performance? Chapter 1 — Computer Abstractions and Technology.
Parallel Programming Henri Bal Rob van Nieuwpoort Vrije Universiteit Amsterdam Faculty of Sciences.
Advanced Topics in Algorithms and Data Structures An overview of the lecture 2 Models of parallel computation Characteristics of SIMD models Design issue.
Parallel Programming Henri Bal Vrije Universiteit Faculty of Sciences Amsterdam.
Instruction Level Parallelism (ILP) Colin Stevens.
NPACI: National Partnership for Advanced Computational Infrastructure Supercomputing ‘98 Mannheim CRAY T90 vs. Tera MTA: The Old Champ Faces a New Challenger.
National Partnership for Advanced Computational Infrastructure San Diego Supercomputer Center Evaluating the Tera MTA Allan Snavely, Wayne Pfeiffer et.
Parallelizing Compilers Presented by Yiwei Zhang.
Copyright © 1998 Wanda Kunkle Computer Organization 1 Chapter 2.1 Introduction.
Lecture 3: Computer Performance
Multicore experiment: Plurality Hypercore Processor Performed by: Anton Fulman Ze’ev Zilberman Supervised by: Mony Orbach Characterization presentation.
Parallel Programming Henri Bal Vrije Universiteit Amsterdam Faculty of Sciences.
Parallel Programming Using Basic MPI Presented by Timothy H. Kaiser, Ph.D. San Diego Supercomputer Center Presented by Timothy H. Kaiser, Ph.D. San Diego.
Joram Benham April 2,  Introduction  Motivation  Multicore Processors  Overview, CELL  Advantages of CMPs  Throughput, Latency  Challenges.
CC02 – Parallel Programming Using OpenMP 1 of 25 PhUSE 2011 Aniruddha Deshmukh Cytel Inc.
1 Developing Native Device for MPJ Express Advisor: Dr. Aamir Shafi Co-advisor: Ms Samin Khaliq.
ExTASY 0.1 Beta Testing 1 st April 2015
Statistical Performance Analysis for Scientific Applications Presentation at the XSEDE14 Conference Atlanta, GA Fei Xing Haihang You Charng-Da Lu July.
Descriptive Data Analysis of File Transfer Data Sudarshan Srinivasan Victor Hazlewood Gregory D. Peterson.
ICOM 5995: Performance Instrumentation and Visualization for High Performance Computer Systems Lecture 7 October 16, 2002 Nayda G. Santiago.
EET 4250: Chapter 1 Computer Abstractions and Technology Acknowledgements: Some slides and lecture notes for this course adapted from Prof. Mary Jane Irwin.
San Diego Supercomputer Center National Partnership for Advanced Computational Infrastructure San Diego Supercomputer Center National Partnership for Advanced.
1 Preparing Your Application for TeraGrid Beyond 2010 TG09 Tutorial June 22, 2009.
C OMPUTER O RGANIZATION AND D ESIGN The Hardware/Software Interface 5 th Edition Chapter 1 Computer Abstractions and Technology Sections 1.5 – 1.11.
Planned AlltoAllv a clustered approach Stephen Booth (EPCC) Adrian Jackson (EPCC)
NIH Resource for Biomolecular Modeling and Bioinformatics Beckman Institute, UIUC NAMD Development Goals L.V. (Sanjay) Kale Professor.
NIH Resource for Biomolecular Modeling and Bioinformatics Beckman Institute, UIUC NAMD Development Goals L.V. (Sanjay) Kale Professor.
Advanced User Support Amit Majumdar 5/7/09. Outline  Three categories of AUS  Update on Operational Activities  AUS.ASTA  AUS.ASP  AUS.ASEOT.
HPCMP Benchmarking Update Cray Henry April 2008 Department of Defense High Performance Computing Modernization Program.
Chapter 1 Computer Abstractions and Technology. Chapter 1 — Computer Abstractions and Technology — 2 The Computer Revolution Progress in computer technology.
1/20 Study of Highly Accurate and Fast Protein-Ligand Docking Method Based on Molecular Dynamics Reporter: Yu Lun Kuo
CS 460/660 Compiler Construction. Class 01 2 Why Study Compilers? Compilers are important – –Responsible for many aspects of system performance Compilers.
GPUs: Overview of Architecture and Programming Options Lee Barford firstname dot lastname at gmail dot com.
Member of the Helmholtz Association 03/10/2015 | RDA Fifth Plenary Meeting | San Diego, USA | Paradise Point Resort Markus Götz Jülich Supercomputing Center.
Morgan Kaufmann Publishers
Parallelization of CC Workshop Benchmark Suggestion Sudhakar Pamidighantam NCSA.
Cs 147 Spring 2010 Meg Genoar. History Started to emerge in mid-1970s 1988 – RISC took over workstation market.
SAN DIEGO SUPERCOMPUTER CENTER Advanced User Support Project Overview Adrian E. Roitberg University of Florida July 2nd 2009 By Ross C. Walker.
Advanced User Support -Update Amit Majumdar SDSC.
SAN DIEGO SUPERCOMPUTER CENTER at the UNIVERSITY OF CALIFORNIA, SAN DIEGO Hybrid MPI/Pthreads Parallelization of the RAxML Phylogenetics Code Wayne Pfeiffer.
Linear Algebra Libraries: BLAS, LAPACK, ScaLAPACK, PLASMA, MAGMA
1  1998 Morgan Kaufmann Publishers How to measure, report, and summarize performance (suorituskyky, tehokkuus)? What factors determine the performance.
SAN DIEGO SUPERCOMPUTER CENTER at the UNIVERSITY OF CALIFORNIA, SAN DIEGO Advanced User Support for MPCUGLES code at University of Minnesota October 09,
SAN DIEGO SUPERCOMPUTER CENTER Advanced User Support Project Overview Thomas E. Cheatham III University of Utah Jan 14th 2010 By Ross C. Walker.
Advanced User Support Amit Majumdar 8/13/09. Outline  Three categories of AUS  Operational Activities  AUS.ASTA  AUS.ASP  ASTA example.
Performance Computer Organization II 1 Computer Science Dept Va Tech January 2009 © McQuain & Ribbens Defining Performance Which airplane has.
Introduction Goal: connecting multiple computers to get higher performance – Multiprocessors – Scalability, availability, power efficiency Job-level (process-level)
TeraGrid Advanced User Support (AUS) Amit Majumdar, SDSC Area Director – AUS TeraGrid Annual Review April 6-7,
SEAMCAT European Communications Office José Carrascosa - SEAMCAT Manager 5 April 2016.
Hybrid Parallel Implementation of The DG Method Advanced Computing Department/ CAAM 03/03/2016 N. Chaabane, B. Riviere, H. Calandra, M. Sekachev, S. Hamlaoui.
TEMPLATE DESIGN © H. Che 2, E. D’Azevedo 1, M. Sekachev 3, K. Wong 3 1 Oak Ridge National Laboratory, 2 Chinese University.
© 2008 Pittsburgh Supercomputing Center Proposed ideas for consideration under AUS.
INTRODUCTION TO HIGH PERFORMANCE COMPUTING AND TERMINOLOGY.
SCEC Capability Simulations on TeraGrid
Measuring Performance II and Logic Design
Introduction to Parallel Computing: MPI, OpenMP and Hybrid Programming
Defining Performance Which airplane has the best performance?
Morgan Kaufmann Publishers
Modern Processor Design: Superscalar and Superpipelining
5.2 Eleven Advanced Optimizations of Cache Performance
משרד התעשייה, המסחר והתעסוקה פעולות המשרד לעידוד מגזר המיעוטים
Numerical Algorithms Quiz questions
Times.
Gengbin Zheng, Esteban Meneses, Abhinav Bhatele and Laxmikant V. Kale
Vrije Universiteit Amsterdam
What Are Performance Counters?
Presentation transcript:

SAN DIEGO SUPERCOMPUTER CENTER Advanced User Support Project Outline October 9th 2008 Ross C. Walker

SAN DIEGO SUPERCOMPUTER CENTER Project Listing 1) Adrian Roitberg (UFL) Ross Walker (SDSC) [25 %] Sudhakar Pamidighantam (NCSA) [10 %] Christian Halloy (NICS) [5 %] 2) Markus Buehler (MIT) Ross Walker (SDSC) [15 %] Dodi Heryadi (NCSA) [10 %]

SAN DIEGO SUPERCOMPUTER CENTER Adrian Roitberg Improvements to the AMBER MD Code Does NOT just cover simple MD simulations but also more complex approaches such as QM/MM etc. Amber QM/MM performance in parallel / serial Focused on improving existing codes. Looking at new / novel approaches for QM/MM in parallel. –predictive scf. –parallel Fock matrix prediction. –Faster distance algorithms for variable QM solvent models. PMEMD Performance on multicore machines. Improvements to Replica Exchange Simulations. Parallelization of Ptraj Analysis Tools.

SAN DIEGO SUPERCOMPUTER CENTER PMEMD 10 Performance Currently performs exceptionally well in parallel. E.g. As fast or faster than NAMD for equivalent simulations. Currently suffers on multi-core machines. It can be cheaper to leave cores idle due to the improved performance. Performance on Kraken and Ranger is equivalent at lower processor counts. At higher counts Kraken seems to perform better on these benchmarks. Can we improve on this? –Multiple compilers / MPI Libraries - which combination works best. –Can we improve TACC_AFFINITY etc? Ultimately publish the benchmarks and instructions for compiling etc on the AMBER website.

SAN DIEGO SUPERCOMPUTER CENTER

Markus Buehler Works on nano-material simulations. Spider silk Carbon nano tubes... Use modified versions of Lammps and NAMD Have their own code internally. Do not have experience in coding within Lammps / NAMD for parallel. Transmitting arrays around etc... Hopefully we can educate his students / postdocs to be self suffcient.