Immersed Boundary Method Simulation in Titanium Objectives

Slides:



Advertisements
Similar presentations
Parallel Processing with OpenMP
Advertisements

IBM’s X10 Presentation by Isaac Dooley CS498LVK Spring 2006.
Languages and Compilers for High Performance Computing Kathy Yelick EECS Department U.C. Berkeley.
The Poisson operator, aka the Laplacian, is a second order elliptic differential operator and defined in an n-dimensional Cartesian space by: The Poisson.
Support for Adaptive Computations Applied to Simulation of Fluids in Biological Systems Kathy Yelick U.C. Berkeley.
Simulating the Cochlea With Titanium Generic Immersed Boundary Software (TiGIBS) Contractile Torus: (NYU) Oval Window of Cochlea: (CACR) Mammalian Heart:
National Partnership for Advanced Computational Infrastructure San Diego Supercomputer Center Evaluating the Tera MTA Allan Snavely, Wayne Pfeiffer et.
Java for High Performance Computing Jordi Garcia Almiñana 14 de Octubre de 1998 de la era post-internet.
Impact of the Cardiac Heart Flow Alpha Project Kathy Yelick EECS Department U.C. Berkeley.
Programming Systems for a Digital Human Kathy Yelick EECS Department U.C. Berkeley.
Support for Adaptive Computations Applied to Simulation of Fluids in Biological Systems Immersed Boundary Method Simulation in Titanium Siu Man Yau, Katherine.
Support for Adaptive Computations Applied to Simulation of Fluids in Biological Systems Kathy Yelick U.C. Berkeley.
Support for Adaptive Computations Applied to Simulation of Fluids in Biological Systems Kathy Yelick U.C. Berkeley.
Use of a High Level Language in High Performance Biomechanics Simulations Katherine Yelick, Armando Solar-Lezama, Jimmy Su, Dan Bonachea, Amir Kamil U.C.
UPC at CRD/LBNL Kathy Yelick Dan Bonachea, Jason Duell, Paul Hargrove, Parry Husbands, Costin Iancu, Mike Welcome, Christian Bell.
Support for Adaptive Computations Applied to Simulation of Fluids in Biological Systems Immersed Boundary Method Simulation in Titanium.
Yelick 1 ILP98, Titanium Titanium: A High Performance Java- Based Language Katherine Yelick Alex Aiken, Phillip Colella, David Gay, Susan Graham, Paul.
Kathy Yelick, 1 Advanced Software for Biological Simulations Elastic structures in an incompressible fluid. Blood flow, clotting, inner ear, embryo growth,
1 Parallel computing and its recent topics. 2 Outline 1. Introduction of parallel processing (1)What is parallel processing (2)Classification of parallel.
KUAS.EE Parallel Computing at a Glance. KUAS.EE History Parallel Computing.
Efficient Parallel Implementation of Molecular Dynamics with Embedded Atom Method on Multi-core Platforms Reporter: Jilin Zhang Authors:Changjun Hu, Yali.
Global Address Space Applications Kathy Yelick NERSC/LBNL and U.C. Berkeley.
PuReMD: Purdue Reactive Molecular Dynamics Package Hasan Metin Aktulga and Ananth Grama Purdue University TST Meeting,May 13-14, 2010.
CSE 260 – Parallel Processing UCSD Fall 2006 A Performance Characterization of UPC Presented by – Anup Tapadia Fallon Chen.
Case Study in Computational Science & Engineering - Lecture 2 1 Parallel Architecture Models Shared Memory –Dual/Quad Pentium, Cray T90, IBM Power3 Node.
Supercomputing ‘99 Parallelization of a Dynamic Unstructured Application using Three Leading Paradigms Leonid Oliker NERSC Lawrence Berkeley National Laboratory.
1CPSD Software Infrastructure for Application Development Laxmikant Kale David Padua Computer Science Department.
Spring 2003CSE P5481 Issues in Multiprocessors Which programming model for interprocessor communication shared memory regular loads & stores message passing.
October 11, 2007 © 2007 IBM Corporation Multidimensional Blocking in UPC Christopher Barton, Călin Caşcaval, George Almási, Rahul Garg, José Nelson Amaral,
1 Qualifying ExamWei Chen Unified Parallel C (UPC) and the Berkeley UPC Compiler Wei Chen the Berkeley UPC Group 3/11/07.
Gtb 1 Titanium Titanium: Language and Compiler Support for Scientific Computing Gregory T. Balls University of California - Berkeley Alex Aiken, Dan Bonachea,
Unified Parallel C Kathy Yelick EECS, U.C. Berkeley and NERSC/LBNL NERSC Team: Dan Bonachea, Jason Duell, Paul Hargrove, Parry Husbands, Costin Iancu,
Cardiac Blood Flow Alpha Project Impact Kathy Yelick EECS Department U.C. Berkeley.
+ Clusters Alternative to SMP as an approach to providing high performance and high availability Particularly attractive for server applications Defined.
C OMPUTATIONAL R ESEARCH D IVISION 1 Defining Software Requirements for Scientific Computing Phillip Colella Applied Numerical Algorithms Group Lawrence.
1 Titanium Review: Immersed Boundary Armando Solar-Lezama Biological Simulations Using the Immersed Boundary Method in Titanium Ed Givelberg, Armando Solar-Lezama,
Parallel OpenFOAM CFD Performance Studies Student: Adi Farshteindiker Advisors: Dr. Guy Tel-Zur,Prof. Shlomi Dolev The Department of Computer Science Faculty.
Lecture 5. Example for periority The average waiting time : = 41/5= 8.2.
Flexibility and Interoperability in a Parallel MD code Robert Brunner, Laxmikant Kale, Jim Phillips University of Illinois at Urbana-Champaign.
Introduction to threads
These slides are based on the book:
UPC at NERSC/LBNL Kathy Yelick, Christian Bell, Dan Bonachea,
Chapter 4: Threads.
SparkBWA: Speeding Up the Alignment of High-Throughput DNA Sequencing Data - Aditi Thuse.
Xing Cai University of Oslo
An Emerging, Portable Co-Array Fortran Compiler for High-Performance Computing Daniel Chavarría-Miranda, Cristian Coarfa, Yuri.
CS427 Multicore Architecture and Parallel Computing
MPI: Portable Parallel Programming for Scientific Computing
Processes and Threads Processes and their scheduling
Programming Models for SimMillennium
Team 1 Aakanksha Gupta, Solomon Walker, Guanghong Wang
Parallelizing Incremental Bayesian Segmentation (IBS)
University of Technology
Implementing Simplified Molecular Dynamics Simulation in Different Parallel Paradigms Chao Mei April 27th, 2006 CS498LVK.
Chapter 4: Threads.
Chapter 4: Threads.
Efficient Parallel Simulation of Fluid Dynamics on Cartesian Grids
Milind A. Bhandarkar Adaptive MPI Milind A. Bhandarkar
UPC and Titanium Kathy Yelick University of California, Berkeley and
MATLAB HPCS Extensions
GENERAL VIEW OF KRATOS MULTIPHYSICS
What is Concurrent Programming?
What is Concurrent Programming?
Introduction to CUDA.
Chapter 4 Multiprocessors
Gary M. Zoppetti Gagan Agrawal Rishi Kumar
Support for Adaptivity in ARMCI Using Migratable Objects
The George Washington University
Programming Parallel Computers
Presentation transcript:

Support for Adaptive Computations Applied to Simulation of Fluids in Biological Systems Immersed Boundary Method Simulation in Titanium Objectives Immersed Boundary Method Titanium High performance Java dialect Explicitly parallel SPMD model Bulk synchronous style Global address space Language level enhancements to support parallelism and performance Immutable “value” classes Multi-dimensional arrays Region-based memory management Compiled into C with light weight message passing Developed by Peskin and McQueen Models elastic fibers in an incompressible fluid. Used for heart, blood clotting, embryos, ciliated cells, and more. The heart model used to design artificial heart valves. NYU software runs on vector and shared memory machines 4 phase algorithm: Provide easy-to-use, high performance tool for simulation of fluid flow in biological systems. Two specific goals: Scalable, parallel version of heart simulation to run on machines like Blue Horizon. Toolkit for general Immersed Boundary Method Simulation. Fiber activation & force calculation Interpolate Velocity Spread Force Navier Stokes Solver Titanium Implementation Status Immersed Boundary Software in Titanium Compiler and runtime systems for: Uniprocessors SMPs running POSIX threads Clusters with: Shared memory -- SGI Origin (ANL), Tera MTA (SDSC) Global Address space – T3E (NERSC) Active Messages -- NOW & Millennium (UCB) LAPI -- IBM SP2, 3 (SDSC) Immersed Boundary Method Software package written in Titanium Based on “Generic” software from NYU Sufficient for the heart model Functionally complete as of 10/2000 A contractile torus simulation run on the Berkeley NOW and ANL Origin. Challenges to Parallelization Scalability of the current algorithm: Irregular fiber points interact with regular fluid lattice Elliptic solver is communication-intensive Scalability of future algorithm: Add adaptivity to fluid solver for accuracy Further complicates communication Conclusions Performance tuning started Uniprocessor performance: Compiler enhancements underway Data collected with C code for 2 kernels Communication: Alternate data layouts Extension for irregular communication Titanium support aided development ~7 months to develop running version Indication for small language extensions Future plans: Add Adaptivity in Fluid Solver Performance tuning, especially for the IBM SP Add IB functionality for other biological systems (bending angles, anchorage points, source & sink)