CHTC working w/ Chemistry Dept

Slides:



Advertisements
Similar presentations
Jaime Frey Computer Sciences Department University of Wisconsin-Madison OGF 19 Condor Software Forum Routing.
Advertisements

CamGrid Mark Calleja Cambridge eScience Centre. What is it? A number of like minded groups and departments (10), each running their own Condor pool(s),
Dan Bradley Computer Sciences Department University of Wisconsin-Madison Schedd On The Side.
Overview of Wisconsin Campus Grid Dan Bradley Center for High-Throughput Computing.
Greg Thain Computer Sciences Department University of Wisconsin-Madison Condor Parallel Universe.
Computational methods in molecular biophysics (examples of solving real biological problems) EXAMPLE I: THE PROTEIN FOLDING PROBLEM Alexey Onufriev, Virginia.
More HTCondor 2014 OSG User School, Monday, Lecture 2 Greg Thain University of Wisconsin-Madison.
Post ASP 2012: Protein Docking at the IU School of Medicine.
Open Science Grid: More compute power Alan De Smet
Condor Project Computer Sciences Department University of Wisconsin-Madison Running Map-Reduce Under Condor.
Zach Miller Condor Project Computer Sciences Department University of Wisconsin-Madison Flexible Data Placement Mechanisms in Condor.
When and How to Use Large-Scale Computing: CHTC and HTCondor Lauren Michael, Research Computing Facilitator Center for High Throughput Computing STAT 692,
Utilizing Condor and HTC to address archiving online courses at Clemson on a weekly basis Sam Hoover 1 Project Blackbird Computing,
HTPC - High Throughput Parallel Computing (on the OSG) Dan Fraser, UChicago OSG Production Coordinator Horst Severini, OU (Greg Thain, Uwisc) OU Supercomputing.
High Throughput Parallel Computing (HTPC) Dan Fraser, UChicago Greg Thain, Uwisc.
Parallel Computing The Bad News –Hardware is not getting faster fast enough –Too many architectures –Existing architectures are too specific –Programs.
Ana Damjanovic (JHU, NIH) JHU: Petar Maksimovic Bertrand Garcia-Moreno NIH: Tim Miller Bernard Brooks OSG: Torre Wenaus and team.
1 Integrating GPUs into Condor Timothy Blattner Marquette University Milwaukee, WI April 22, 2009.
High Throughput Computing with Condor at Purdue XSEDE ECSS Monthly Symposium Condor.
Prof. Heon Y. Yeom Distributed Computing Systems Lab. Seoul National University FT-MPICH : Providing fault tolerance for MPI parallel applications.
HTCondor workflows at Utility Supercomputing Scale: How? Ian D. Alderman Cycle Computing.
Building a Real Workflow Thursday morning, 9:00 am Lauren Michael Research Computing Facilitator University of Wisconsin - Madison.
IntroductiontotoHTCHTC 2015 OSG User School, Monday, Lecture1 Greg Thain University of Wisconsin– Madison Center For High Throughput Computing.
Condor Tugba Taskaya-Temizel 6 March What is Condor Technology? Condor is a high-throughput distributed batch computing system that provides facilities.
ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.
1 Evolution of OSG to support virtualization and multi-core applications (Perspective of a Condor Guy) Dan Bradley University of Wisconsin Workshop on.
High Throughput Parallel Computing (HTPC) Dan Fraser, UChicago Greg Thain, UWisc Condor Week April 13, 2010.
Compiled Matlab on Condor: a recipe 30 th October 2007 Clare Giacomantonio.
Workflows: from Development to Production Thursday morning, 10:00 am Lauren Michael Research Computing Facilitator University of Wisconsin - Madison.
Turning science problems into HTC jobs Wednesday, July 20 th 9am Zach Miller & Greg Thain Condor Team University of Wisconsin-Madison.
Greg Thain Computer Sciences Department University of Wisconsin-Madison cs.wisc.edu Interactive MPI on Demand.
Intermediate Condor Rob Quick Open Science Grid HTC - Indiana University.
Condor: High-throughput Computing From Clusters to Grid Computing P. Kacsuk – M. Livny MTA SYTAKI – Univ. of Wisconsin-Madison
Building a Real Workflow Thursday morning, 9:00 am Greg Thain University of Wisconsin - Madison.
Workflows: from development to Production Thursday morning, 10:00 am Greg Thain University of Wisconsin - Madison.
1 Condor Team 2009 Established 1985.
Building a Real Workflow Thursday morning, 9:00 am Lauren Michael Research Computing Facilitator University of Wisconsin - Madison.
Condor Week 2007 Windows Condor Pool at Clemson University Sebastien Goasguen School of Computing and Clemson Computing Information Technology (CCIT) May.
Condor Project Computer Sciences Department University of Wisconsin-Madison Grids and Condor Barcelona,
What’s Coming? What are we Planning?. › Better docs › Goldilocks – This slot size is just right › Storage › New.
How High Throughput was my cluster? Greg Thain Center for High Throughput Computing.
Condor Project Computer Sciences Department University of Wisconsin-Madison Running Interpreted Jobs.
Five todos when moving an application to distributed HTC.
Job submission overview Marco Mambelli – August OSG Summer Workshop TTU - Lubbock, TX THE UNIVERSITY OF CHICAGO.
Greg Thain Computer Sciences Department University of Wisconsin-Madison HTPC on the OSG.
Intermediate Condor Monday morning, 10:45am Alain Roy OSG Software Coordinator University of Wisconsin-Madison.
Introduction to High Throughput Computing and HTCondor Monday AM, Lecture 1 Ian Ross Center for High Throughput Computing University of Wisconsin-Madison.
Condor Week May 2012No user requirements1 Condor Week 2012 An argument for moving the requirements out of user hands - The CMS experience presented.
Christina Koch Research Computing Facilitators
CHTC Policy and Configuration
Auburn University
Checkpoint/restart in Slurm: current status and new developments
Pegasus WMS Extends DAGMan to the grid world
Introduction to programming
Examples Example: UW-Madison CHTC Example: Global CMS Pool
Applications of molecular simulation in materials science and biology
Harnessing the Power of Condor for Human Genetics
Condor: Job Management
Understanding Supernovae with Condor
Job Matching, Handling, and Other HTCondor Features
Condor and Multi-core Scheduling
Atomic Level Study of Adsorption, Conformational Change and Dimerization of an α-helical Peptide at Graphene Surface Luchun Ou, Yin Luo, and Guanghong.
Introduction to High Throughput Computing and HTCondor
The Condor JobRouter.
rvGAHP – Push-Based Job Submission Using Reverse SSH Connections
Condor-G Making Condor Grid Enabled
GLOW A Campus Grid within OSG
PU. Setting up parallel universe in your pool and when (not
Condor-G: An Update.
Presentation transcript:

CHTC working w/ Chemistry Dept

Gromacs

30+ day runtime Too long, even as HTC Step one – compile with SSE support 10x improvement Just a Gromacs compile-time option Hand-coded assembly, not gcc option

3 days still too long… Gromacs also support MPI CHTC doesn’t have infiniband What do to?

Whole machine jobs Submit file magic to claim all 8 slots universe = vanilla requirements = (CAN_RUN_WHOLE_MACHINE =?= TRUE) +RequiresWholeMachine=true executable = some job arguments = arguments should_transfer_files = yes when_to_transfer_output = on_exit transfer_input_files = inputs queue

MPI on Whole machine jobs Whole machine mpi submit file universe = vanilla requirements = (CAN_RUN_WHOLE_MACHINE =?= TRUE) +RequiresWholeMachine=true executable = mpiexec arguments = -np 8 real_exe should_transfer_files = yes when_to_transfer_output = on_exit transfer_input_files = real_exe queue Condor Motto: If you want it, Bring it yourself

Advantages Condor is parallel agnostic: High-bandwith memory transport MPICH, OpenMPI, pthreads, fork, etc. High-bandwith memory transport Easy to debug Ssh-to-job still works Access to all machine’s memory

Disadvantages Still need to debug parallel program helps if others already have Fewer full-machine slots Currently 30, more coming Lots of competition for 30 machines

30 machines not enough: OSG to the rescue

Glideing-in MPI to OSG CHTC Pool CHTC schedd Glide-in Pool Job queue

OSG Submit File +WantGlidein = true universe = vanilla requirements = (CAN_RUN_WHOLE_MACHINE =?= TRUE) +RequiresWholeMachine=true +WantGlidein = true executable = some job arguments = arguments should_transfer_files = yes when_to_transfer_output = on_exit transfer_input_files = inputs queue

Restrictions of glidein More diverse hardware resources No pre-staged, some AMD machines… Less guaranteed memory (1.7 Gb) Preemption at any time Restartable or checkpointable a big plus (Gromacs, others has checkpointing!)

Computational Results

Real Results Iron-Catalyzed Oxidation Intermediates Captured in A DNA Repair Monooxygenase, C. Yi, G. Jia, G. Hou, Q. Dai, G. Zheng, X. Jian, C. G. Yang, Q. Cui, and C. He, {\it Science}, Submitted

Real Results Disruption and formation of surface salt bridges are coupled to DNA binding in integration host factor (IHF): acomputational analysis, L. Ma, M. T. Record, Jr., N. Sundlass, R. T. Raines and Q. Cui, {\it J. Mol. Biol.}, Submitted

Real Results An implicit solvent model for SCC-DFTB with Charge-Dependent Radii, G. Hou, X. Zhu and Q. Cui, {\it J. Chem. Theo. Comp.}, Submitted

Real Results Sequence-dependent interaction of $\beta$-peptides with membranes, J. Mondal, X. Zhu, Q. Cui and A. Yethiraj, {\it J. Am. Chem. Soc.}, Submitted

Real Results A new coarse-grained model for water: The importance of electrostatic interactions, Z. Wu, Q. Cui and A. Yethiraj, {\it J. Phys. Chem. B} Submitted

Real Results How does bone sialoprotein promote the nucleation of hydroxyapatite? A molecular dynamics study using model peptides of different conformations, Y. Yang, Q. Cui, and N. Sahai, {\it Langmuir}, Submitted

Real Results Preferential interactions between small solutes and the protein backbone: A computational analysis, L. Ma, L. Pegram, M. T. Record, Jr., Q. Cui, {\it Biochem.}, 49, 1954-1962 (2010)

Real Results Establishing effective simulation protocols for $\beta$- and$\alpha/\beta$-peptides. III. Molecular Mechanical (MM) model for a non-cyclic $\beta$-residue, X. Zhu, P. K\"onig, M. Hoffman, A. Yethiraj and Q. Cui, {\it J. Comp. Chem.}, In press (DOI: 10.1002/jcc.21493)

Real Results Curvature Generation and Pressure Profile in Membrane with lysolipids: Insights from coarse-grained simulations, J. Yoo and Q. Cui, {\it Biophys. J.} 97, 2267-2276 (2009)

Summary CHTC can deliver a lot of CPU, if jobs are “shaped” appropriately CHTC supports whole machine jobs Consider “small mpi”, or High Throughput of High performance OSG Record: 50,000 CPU hours in a day! (this could be you!)