Support for Adaptive Computations Applied to Simulation of Fluids in Biological Systems Kathy Yelick U.C. Berkeley.

Slides:



Advertisements
Similar presentations
Barcelona Supercomputing Center. The BSC-CNS objectives: R&D in Computer Sciences, Life Sciences and Earth Sciences. Supercomputing support to external.
Advertisements

School of Computing University of Leeds Computational PDEs Unit A Grid-based approach to the validation and testing of lubrication models Christopher Goodyer.
Weather Research & Forecasting: A General Overview
Unified Parallel C at LBNL/UCB Implementing a Global Address Space Language on the Cray X1 Christian Bell and Wei Chen.
Thoughts on Shared Caches Jeff Odom University of Maryland.
Parallel Programming Laboratory1 Fault Tolerance in Charm++ Sayantan Chakravorty.
Parallelizing stencil computations Based on slides from David Culler, Jim Demmel, Bob Lucas, Horst Simon, Kathy Yelick, et al., UCB CS267.
Languages and Compilers for High Performance Computing Kathy Yelick EECS Department U.C. Berkeley.
Cactus in GrADS (HFA) Ian Foster Dave Angulo, Matei Ripeanu, Michael Russell.
Towards a Digital Human Kathy Yelick EECS Department U.C. Berkeley.
Unified Parallel C at LBNL/UCB UPC at LBNL/U.C. Berkeley Overview Kathy Yelick U.C. Berkeley, EECS LBNL, Future Technologies Group.
Support for Adaptive Computations Applied to Simulation of Fluids in Biological Systems Kathy Yelick U.C. Berkeley.
The Virtual Microscope Umit V. Catalyurek Department of Biomedical Informatics Division of Data Intensive and Grid Computing.
Simulating the Cochlea With Titanium Generic Immersed Boundary Software (TiGIBS) Contractile Torus: (NYU) Oval Window of Cochlea: (CACR) Mammalian Heart:
Steady Aeroelastic Computations to Predict the Flying Shape of Sails Sriram Antony Jameson Dept. of Aeronautics and Astronautics Stanford University First.
Scripting Languages For Virtual Worlds. Outline Necessary Features Classes, Prototypes, and Mixins Static vs. Dynamic Typing Concurrency Versioning Distribution.
Impact of the Cardiac Heart Flow Alpha Project Kathy Yelick EECS Department U.C. Berkeley.
Programming Systems for a Digital Human Kathy Yelick EECS Department U.C. Berkeley.
Support for Adaptive Computations Applied to Simulation of Fluids in Biological Systems Immersed Boundary Method Simulation in Titanium Siu Man Yau, Katherine.
PARALLEL PROCESSING The NAS Parallel Benchmarks Daniel Gross Chen Haiout.
Support for Adaptive Computations Applied to Simulation of Fluids in Biological Systems Kathy Yelick U.C. Berkeley.
Use of a High Level Language in High Performance Biomechanics Simulations Katherine Yelick, Armando Solar-Lezama, Jimmy Su, Dan Bonachea, Amir Kamil U.C.
UPC at CRD/LBNL Kathy Yelick Dan Bonachea, Jason Duell, Paul Hargrove, Parry Husbands, Costin Iancu, Mike Welcome, Christian Bell.
Support for Adaptive Computations Applied to Simulation of Fluids in Biological Systems Immersed Boundary Method Simulation in Titanium.
Kathy Yelick, 1 Advanced Software for Biological Simulations Elastic structures in an incompressible fluid. Blood flow, clotting, inner ear, embryo growth,
Copyright Arshi Khan1 System Programming Instructor Arshi Khan.
1 Titanium Review: Ti Parallel Benchmarks Kaushik Datta Titanium NAS Parallel Benchmarks Kathy Yelick U.C. Berkeley September.
Global Address Space Applications Kathy Yelick NERSC/LBNL and U.C. Berkeley.
1 Developing Native Device for MPJ Express Advisor: Dr. Aamir Shafi Co-advisor: Ms Samin Khaliq.
CompuCell Software Current capabilities and Research Plan Rajiv Chaturvedi Jesús A. Izaguirre With Patrick M. Virtue.
Processing of a CAD/CAE Jobs in grid environment using Elmer Electronics Group, Physics Department, Faculty of Science, Ain Shams University, Mohamed Hussein.
LLNL-PRES-XXXXXX This work was performed under the auspices of the U.S. Department of Energy by Lawrence Livermore National Laboratory under Contract DE-AC52-07NA27344.
ICOM 5995: Performance Instrumentation and Visualization for High Performance Computer Systems Lecture 7 October 16, 2002 Nayda G. Santiago.
Threads, Thread management & Resource Management.
Center for Programming Models for Scalable Parallel Computing: Project Meeting Report Libraries, Languages, and Execution Models for Terascale Applications.
CCGrid 2014 Improving I/O Throughput of Scientific Applications using Transparent Parallel Compression Tekin Bicer, Jian Yin and Gagan Agrawal Ohio State.
Loosely Coupled Parallelism: Clusters. Context We have studied older archictures for loosely coupled parallelism, such as mesh’s, hypercubes etc, which.
Simulated vascular reconstruction in a virtual operating theatre Robert G. Belleman, Peter M.A. Sloot Section Computational Science, University of Amsterdam,
MATRIX MULTIPLY WITH DRYAD B649 Course Project Introduction.
Boundary Assertion in Behavior-Based Robotics Stephen Cohorn - Dept. of Math, Physics & Engineering, Tarleton State University Mentor: Dr. Mircea Agapie.
Issues Autonomic operation (fault tolerance) Minimize interference to applications Hardware support for new operating systems Resource management (global.
NIH Resource for Biomolecular Modeling and Bioinformatics Beckman Institute, UIUC NAMD Development Goals L.V. (Sanjay) Kale Professor.
Numerical Libraries Project Microsoft Incubation Group Mary Beth Hribar Microsoft Corporation CSCAPES Workshop June 10, 2008 Copyright Microsoft Corporation,
CCGrid 2014 Improving I/O Throughput of Scientific Applications using Transparent Parallel Compression Tekin Bicer, Jian Yin and Gagan Agrawal Ohio State.
Improving I/O with Compiler-Supported Parallelism Why Should We Care About I/O? Disk access speeds are much slower than processor and memory access speeds.
1 Qualifying ExamWei Chen Unified Parallel C (UPC) and the Berkeley UPC Compiler Wei Chen the Berkeley UPC Group 3/11/07.
I/O for Structured-Grid AMR Phil Colella Lawrence Berkeley National Laboratory Coordinating PI, APDEC CET.
Enabling Self-management of Component-based High-performance Scientific Applications Hua (Maria) Liu and Manish Parashar The Applied Software Systems Laboratory.
Domain Decomposition in High-Level Parallelizaton of PDE codes Xing Cai University of Oslo.
An Evaluation of Partitioners for Parallel SAMR Applications Sumir Chandra & Manish Parashar ECE Dept., Rutgers University Submitted to: Euro-Par 2001.
Cardiac Blood Flow Alpha Project Impact Kathy Yelick EECS Department U.C. Berkeley.
Pre-calculated Fluid Simulator States Tree Marek Gayer and Pavel Slavík C omputer G raphics G roup Department of Computer Science and Engineering Faculty.
Comprehensive Scientific Support Of Large Scale Parallel Computation David Skinner, NERSC.
The Performance Evaluation Research Center (PERC) Participating Institutions: Argonne Natl. Lab.Univ. of California, San Diego Lawrence Berkeley Natl.
SDM Center High-Performance Parallel I/O Libraries (PI) Alok Choudhary, (Co-I) Wei-Keng Liao Northwestern University In Collaboration with the SEA Group.
1 Rocket Science using Charm++ at CSAR Orion Sky Lawlor 2003/10/21.
C OMPUTATIONAL R ESEARCH D IVISION 1 Defining Software Requirements for Scientific Computing Phillip Colella Applied Numerical Algorithms Group Lawrence.
ECG Simulation NCRR Overview Technology for the ECG Simulation project CardioWave BioPSE project background Tools developed to date Tools for the next.
Background Computer System Architectures Computer System Software.
1 Titanium Review: Immersed Boundary Armando Solar-Lezama Biological Simulations Using the Immersed Boundary Method in Titanium Ed Givelberg, Armando Solar-Lezama,
Introduction to Performance Tuning Chia-heng Tu PAS Lab Summer Workshop 2009 June 30,
Unified Parallel C at LBNL/UCB UPC at LBNL/U.C. Berkeley Overview Kathy Yelick LBNL and U.C. Berkeley.
POE Parallel Operating Environment. Cliff Montgomery.
UPC at NERSC/LBNL Kathy Yelick, Christian Bell, Dan Bonachea,
Xing Cai University of Oslo
Programming Models for SimMillennium
Abstract Machine Layer Research in VGrADS
Immersed Boundary Method Simulation in Titanium Objectives
Support for Adaptivity in ARMCI Using Migratable Objects
L. Glimcher, R. Jin, G. Agrawal Presented by: Leo Glimcher
Presentation transcript:

Support for Adaptive Computations Applied to Simulation of Fluids in Biological Systems Kathy Yelick U.C. Berkeley

Project Summary Provide easy-to-use, high performance tool for simulation of fluid flow in biological systems. –Using the Immersed Boundary Method Enable simulations on large-scale parallel machines. –Distributed memory machine including SMP clusters Using Titanium, ADR, and KeLP with AMR Specific demonstration problem: Simulation of the heart model on Blue Horizon.

Overview of FY03 Plan Extend to new applications of the method –Support for non-fiber boundaries –Improved accuracy –Scalable solvers Heart simulation –Improved visualization –Large application run Data analysis –Use in model construction –Data compression for results

Support for New Applications The heart model is based on –Fibers representing muscle fibers –Also used for blood clotting and other apps Some problems require other structures –Plates, shells, sets of fibers, etc. –Used in cochlea model, insect flight,… –Initial prototype of plate in TIBM code FY03 Plan –Made code more general (UCB) –Specific support for the cochlea model (UCB)

Cochlea Model Model of the inner ear –Developed by Julian Bunn and Ed Givelberg Contains new features, e.g. membranes Implemented one of these last fall Givelberg will be at UCB next year

Supporting Applications: Accuracy The heart simulation uses –First order accurate method –Second order method known Demonstrated in Fortran code –Necessary for fluids with high Reynold’s numbers For example, air flow around insect wing FY03 Plan –Implement the second order accurate method in the TIBM code. (NYU + UCB)

Support Applications: Scalability Replacing 3D FFT with multigrid –Scallop: 3D Poisson solver using the “method of local corrections.” –Scallop algorithm is more scalable than traditional multigrid (fewer messages) –Complete by end of FY02 FY03 Plan –Complete integration in TIBM (UCB, UCSD) –Evaluate solvers (UCB, UCSD)

Heart Simulation Recent improvements –Support for heart input files –Generate data for NYU visualization –Basic visualization is now OpenGL –Checkpoint/restart (underway) FY03 plans –Large application run (NYU) –Extend vis support in “open” code (NYU) –Validation (OS and NYU)

Data Analysis: Cardiac Simulations Methods and tools to analyze 3D datasets from cardiac blood flow. –Outputs are velocity and pressure values on a 3D grid (128 3 ) over many time steps. Characterize behavior –Natural and artificial heart valves –Vary viscosity, density, fiber stiffness, … FY03 Plan – Grid support for DataCutter –Performance tuning –Use of FASTR to reduce data output size

FY03 Summary Demonstration of TIBM on large problem –Large heart simulation run –Improve visualization software Extend community through more applications –Immersed boundaries other than fibers –Better accuracy –Scallop for improved scalability Data analysis –Close the loop: Build input data & analyze results –Compress output

Backup Slides These came from the review, and have more details on the current status

Alpha Project Plans Several categories –Application development Heart and cochlea-component simulation –Application-level package Generic immersed boundary method Parallel for shared and distributed memory Enables new larger-scale simulations; finer grid –Solver libraries Method of Local Corrections Improved scalability and load balance expected –Data analysis Building input data and analysis of results application data software systems

Immersed Boundary Method Recent Performance Improvements –Use of FFTW in Spectral solver 10x performance improvement on t3e Use on BH still pending –Use of scatter/gather communication Copying bounding boxes is still faster Depends on application and machine –Load balancing Alignment of fluid grid (in slabs) and fiber Multigrid solver might offer more possibilities

Titanium on Blue Horizon Recent improvements: –Support for PAPI (performance analysis) –Cache optimizations –Portable runtime layer (maintainable) –Faster LAPI-based implementation LAPI is IBM’s “active message” layer FY03 Plans –Communication optimizations –Common runtime with UPC (possibly CAF)

MPI vs. LAPI on the IBM SP LAPI bandwidth higher than MPI Also better small-message overhead –9usec vs. 11usec Latest Titanium release leverages this

Immersed Boundary Method Structure Fiber activation & force calculation Interpolate Velocity Navier-Stokes Solver Spread Force 4 steps in each timestep Fiber Points Interaction Fluid Lattice

Scallop: Multigrid Poisson Solver A latency tolerant elliptical solver library –Will be used to build Navier-Stokes Solver –Implemented in KeLP, with a simple interface Work by Scott Baden and Greg Balls –Based on Balls/Colella algorithm –2D implementation in both KeLP and Titanium 3D Solver –Algorithm is complete –Implementation running, but performance tuning is ongoing –Interface between Titanium and KeLP developed

Elliptic solvers A finite-difference based solvers –Good for regular, block-structured domains Method of Local Corrections –Local solutions corrected by a coarse solution –Good accuracy, well-conditioned solutions Limited communication –Once to generate coarse grid values –Once to correct local solutions –Trades off extra computation for fewer messages

KeLP implementation Advantages –abstractions available in C++ –built in domain calculus –communication management –numerical kernels written in Fortran Simple interface –callable from other languages –no KeLP required in user code

Load Balancing Egg slicer Pizza cutter Fluid grid is divided in slabs for 3D FFT

Application: Heart Simulation Performance improvements over the last year -64 node t3e ~= 2 node C90 ~= 1-node (8p) BH (probably) Heart simulation on a Cray T3E

Improved Heart Structure Model Current model is –Based on dog heart, textbook anatomy –Approximation by composing cones Building a more accurate model –Use modern imaging on human heart for model –Need to see individual fibers –Collaboration between Joel Saltz’s group and Dr. Robert DePhilip in Anatomy Division of Biomedical Informatics Dept. at Ohio State University Long term goal –Specialize model to patient using MRI data