Simulation at NASA for the Space Radiation Effort

Slides:



Advertisements
Similar presentations
Simulation at NASA for the Space Radiation Effort Dr. Robert C. Singleterry Jr. NASA Administrator's Fellow (Cohort 6) NASA Langley Research Center HPC.
Advertisements

Refining High Performance FORTRAN Code from Programming Model Dependencies Ferosh Jacob University of Alabama Department of Computer Science
System Simulation Of 1000-cores Heterogeneous SoCs Shivani Raghav Embedded System Laboratory (ESL) Ecole Polytechnique Federale de Lausanne (EPFL)
Ionization of the Hydrogen Molecular Ion by Ultrashort Intense Elliptically Polarized Laser Radiation Ryan DuToit Xiaoxu Guan (Mentor) Klaus Bartschat.
GPGPU Introduction Alan Gray EPCC The University of Edinburgh.
HPC - High Performance Productivity Computing and Future Computational Systems: A Research Engineer’s Perspective Dr. Robert C. Singleterry Jr. NASA Langley.
NASA High Performance Computing (HPC) Directions, Issues, and Concerns: A User’s Perspective Dr. Robert C. Singleterry Jr. NASA Langley Research Center.
Introduction CS 524 – High-Performance Computing.
Software Group © 2006 IBM Corporation Compiler Technology Task, thread and processor — OpenMP 3.0 and beyond Guansong Zhang, IBM Toronto Lab.
Parallel Algorithms - Introduction Advanced Algorithms & Data Structures Lecture Theme 11 Prof. Dr. Th. Ottmann Summer Semester 2006.
Lecture 29 Fall 2006 Lecture 29: Parallel Programming Overview.
A COMPARISON MPI vs POSIX Threads. Overview MPI allows you to run multiple processes on 1 host  How would running MPI on 1 host compare with POSIX thread.
ICOM 5995: Performance Instrumentation and Visualization for High Performance Computer Systems Lecture 7 October 16, 2002 Nayda G. Santiago.
1b.1 Types of Parallel Computers Two principal approaches: Shared memory multiprocessor Distributed memory multicomputer ITCS 4/5145 Parallel Programming,
Results Matter. Trust NAG. Numerical Algorithms Group Mathematics and technology for optimized performance Alternative Processors Panel IDC, Tucson, Sept.
Loosely Coupled Parallelism: Clusters. Context We have studied older archictures for loosely coupled parallelism, such as mesh’s, hypercubes etc, which.
GPU Architecture and Programming
Alternative ProcessorsHPC User Forum Panel1 HPC User Forum Alternative Processor Panel Results 2008.
Sep 08, 2009 SPEEDUP – Optimization and Porting of Path Integral MC Code to New Computing Architectures V. Slavnić, A. Balaž, D. Stojiljković, A. Belić,
Lesson 4: Computer method overview
Experts in numerical algorithms and HPC services Compiler Requirements and Directions Rob Meyer September 10, 2009.
FOUNDATION IN INFORMATION TECHNOLOGY (CS-T-101) TOPIC : INFORMATION SYSTEM – SOFTWARE.
Parallel Computing Presented by Justin Reschke
Background Computer System Architectures Computer System Software.
Is MPI still part of the solution ? George Bosilca Innovative Computing Laboratory Electrical Engineering and Computer Science Department University of.
1 A simple parallel algorithm Adding n numbers in parallel.
Building on virtualization capabilities for ExTENCI Carol Song and Preston Smith Rosen Center for Advanced Computing Purdue University ExTENCI Kickoff.
Multicore Applications in Physics and Biochemical Research Hristo Iliev Faculty of Physics Sofia University “St. Kliment Ohridski” 3 rd Balkan Conference.
Computer Security: Chapter 5 Operating Systems Security.
Embedded Systems. What is Embedded Systems?  Embedded reflects the facts that they are an integral.
Our Graphics Environment Landscape Rendering. Hardware  CPU  Modern CPUs are multicore processors  User programs can run at the same time as other.
Parallel OpenFOAM CFD Performance Studies Student: Adi Farshteindiker Advisors: Dr. Guy Tel-Zur,Prof. Shlomi Dolev The Department of Computer Science Faculty.
Application of Design Patterns to Geometric Decompositions V. Balaji, Thomas L. Clune, Robert W. Numrich and Brice T. Womack.
Group Members Hamza Zahid (131391) Fahad Nadeem khan Abdual Hannan AIR UNIVERSITY MULTAN CAMPUS.
Simulation Modeling.
Introduction to Parallel Computing: MPI, OpenMP and Hybrid Programming
Chapter 1 Introduction.
COMBINED PAGING AND SEGMENTATION
Development and deployment trends
Chapter 1 Introduction.
AWS Batch Overview A highly-efficient, dynamically-scaled, batch computing service May 2017.
Ray-Cast Rendering in VTK-m
Experience with Maintaining the GPU Enabled Version of COSMO
The Future of Fortran is Bright …
Introduction.
Dycore Rewrite Tobias Gysi.
Compiler Back End Panel
Objective of This Course
CSCE569 Parallel Computing
Lesson Objectives Aims Understand how machine code is generated
Compiler Back End Panel
GENERAL VIEW OF KRATOS MULTIPHYSICS
Alternative Processor Panel
Alternative Processor Panel Results 2008
1.1 The Characteristics of Contemporary Processors, Input, Output and Storage Devices Types of Processors.
Back End Compiler Panel
Introduction to CUDA.
Java Programming Introduction
The Challenge of Cross - Language Interoperability
Tonga Institute of Higher Education IT 141: Information Systems
Chapter 4 Multiprocessors
Parallel Computing Explained How to Parallelize a Code
Tonga Institute of Higher Education IT 141: Information Systems
Defining the Grid Fabrizio Gagliardi EMEA Director Technical Computing
The Transport Equation
EE 155 / Comp 122 Parallel Computing
Types of Parallel Computers
Research: Past, Present and Future
Question 1 How are you going to provide language and/or library (or other?) support in Fortran, C/C++, or another language for massively parallel programming.
Workflow and HPC erhtjhtyhy Doug Benjamin Argonne National Lab.
Presentation transcript:

Simulation at NASA for the Space Radiation Effort Dr. Robert C. Singleterry Jr. NASA Administrator's Fellow (Cohort 6) NASA Langley Research Center HPC User’s Forum, Stuttgart, GM

Overview Simulation at NASA for Space Radiation What are Our problems? Am I (we) Unique? Possible Solutions? No Real Conclusion!

Simulation at NASA for Space Radiation What is Space Radiation? One of the top 5 problems that must be solved for extended space travel

Simulation at NASA for Space Radiation Just solve the Boltzmann Equations: Easy as pie …… there’s pie?

Simulation at NASA for Space Radiation Not so Easy!! Stochastic Methods Monte Carlo is the most prevalent Deterministic Methods Discrete ordinates is the most prevalent Now can use finite elements for the geometry solution Straight ahead method (what we use now) Dozens if not hundreds of other methods exists also Possible solutions for space radiation? Physics allows the straight ahead method Many ways to even solve with this method Interpolation Ray-by-ray

Simulation at NASA for Space Radiation Today’s method of choice for us

All is not lost, just limited What are Our Problems? As vendors move towards a multi/many core environment, the individual cores slow down to beat thermal limits The NASA Space Radiation code is a physics research code with a web based front end Last thing thought about was execution time Physics code was written from 1970 to present Legacy does not begin to explain it It is serial!!! (We have added dynamic memory allocation) All is not lost, just limited

Interpolation – not much to do here What are Our Problems? Interpolation – not much to do here Thread the mathematics Course parallel over interpolation points Ray-by-Ray Since the rays are independent, each ray can go on a core Still hit a wall at about 1000-2000 cores Would take more money than we have to rewrite our code for a cluster environment Latest NASA machine: 43,008 cores!!!

-------------- SUMMARY ------------- What are Our Problems? As cores go from 43,008 to 1,000,000 Our algorithm is stuck at 1000-2000 cores without major rework that we cannot afford Not sure how many more cores we could utilize if we could afford a rework but << 1,000,000!!! -------------- SUMMARY ------------- As the cores get slower, our execution time gets longer as we cannot use more cores Yet users want more and better answers as computers get more powerful OK, as user’s needs become more demanding

HPCWire, 9/24/08, “Intel: CPUs Will Prevail Over Accelerators in HPC” What are Our Problems? HPCWire, 9/24/08, “Intel: CPUs Will Prevail Over Accelerators in HPC” “What we're finding is that if someone is going to go to the effort of optimizing an application to take advantage of an offload engine, whatever it may be, the first thing they have to do is parallelize their code” Richard Dracott, General Manager, HPC Business Unit

There are many small and large ISV’s Am I (we) Unique? There are many small and large ISV’s Abaqus < 128 cores Few open source packages can >>128 cores Most (if not all) engineering, day-to-day packages cannot use more than 1000 cores MCNPX can use <2000-4000 cores for certain types of problems Most everybody needs help! Those that do not need help can afford To rewrite code when new architectures arrive To write code from scratch to fit an architecture

Smarter compilers – User’s Point of View! Possible Solutions? Smarter compilers – User’s Point of View! No new language, just amend Fortran and C Like MP but with MPI – Programming Environment Nice if housed in current environments Intel PGI Absoft etc… Do not care if production compile takes days Enable non-x86 hardware in current compilers ASICs GPGPU/Cell FPGA

Nothing but the possible solutions No Real Conclusions Nothing but the possible solutions Develop new algorithms to solve the Boltzmann Equation so >1,000,000 cores can be utilized Over 10M$ and 3 years to parallelize and V&V what we already have and must be done first!! Over 100M$ and 10 years to develop new methods of solution to fit the vision of chip makers and then V&V the methods Space Radiation is a small but unique solution domain of the total radiation analysis world