Neutron Science TeraGrid Gateway Update Vickie Lynch, Meili Chen, John Cobb Oak Ridge National Laboratory AUS Technical Presentation October 8, 2009.

Slides:



Advertisements
Similar presentations
INTRODUCTION TO SIMULATION WITH OMNET++ José Daniel García Sánchez ARCOS Group – University Carlos III of Madrid.
Advertisements

Parallel ISDS Chris Hans 29 November 2004.
SAN DIEGO SUPERCOMPUTER CENTER Blue Gene for Protein Structure Prediction (Predicting CASP Targets in Record Time) Ross C. Walker.
1 DISTRIBUTION STATEMENT XXX– Unclassified, Unlimited Distribution Laser Propagation Modeling Making Large Scale Ultrashort Pulse Laser Simulations Possible.
Parallel Computing Majid Almeshari John Conklin. Outline The Challenge Available Parallelization Resources Status of Parallelization Plan & Next Step.
Benchmarking Parallel Code. Benchmarking2 What are the performance characteristics of a parallel code? What should be measured?
Information Technology Center Introduction to High Performance Computing at KFUPM.
Cyberinfrastructure for Scalable and High Performance Geospatial Computation Xuan Shi Graduate assistants supported by the CyberGIS grant Fei Ye (2011)
SKELETON BASED PERFORMANCE PREDICTION ON SHARED NETWORKS Sukhdeep Sodhi Microsoft Corp Jaspal Subhlok University of Houston.
Summary Role of Software (1 slide) ARCS Software Architecture (4 slides) SNS -- Caltech Interactions (3 slides)
Claude TADONKI Mines ParisTech – LAL / CNRS / INP 2 P 3 University of Oujda (Morocco) – October 7, 2011 High Performance Computing Challenges and Trends.
The Role of DANSE at SNS Steve Miller Scientific Computing Group Leader January 22, 2007.
Brad Whitlock October 14, 2009 Brad Whitlock October 14, 2009 Porting VisIt to BG/P.
1 Aug 7, 2004 GPU Req GPU Requirements for Large Scale Scientific Applications “Begin with the end in mind…” Dr. Mark Seager Asst DH for Advanced Technology.
Introduction to Scientific Computing Doug Sondak Boston University Scientific Computing and Visualization.
Science Advisory Committee Meeting - 20 September 3, 2010 Stanford University 1 04_Parallel Processing Parallel Processing Majid AlMeshari John W. Conklin.
Homework 2 In the docs folder of your Berkeley DB, have a careful look at documentation on how to configure BDB in main memory. In the docs folder of your.
Genetic Art. Network Record Using Network Information Chicago Evanston ParisMoscow.
Cliff Rhyne and Jerry Fu June 5, 2007 Parallel Image Segmenter CSE 262 Spring 2007 Project Final Presentation.
CS 732: Advance Machine Learning Usman Roshan Department of Computer Science NJIT.
Parallel K-Means Clustering Based on MapReduce The Key Laboratory of Intelligent Information Processing, Chinese Academy of Sciences Weizhong Zhao, Huifang.
Plans for Exploitation of the ORNL Titan Machine Richard P. Mount ATLAS Distributed Computing Technical Interchange Meeting May 17, 2013.
1 Reasons for parallelization Can we make GA faster? One of the most promising choices is to use parallel implementations. The reasons for parallelization.
TeraGrid Gateway User Concept – Supporting Users V. E. Lynch, M. L. Chen, J. W. Cobb, J. A. Kohl, S. D. Miller, S. S. Vazhkudai Oak Ridge National Laboratory.
Parallel Communications and NUMA Control on the Teragrid’s New Sun Constellation System Lars Koesterke with Kent Milfeld and Karl W. Schulz AUS Presentation.
Exercise problems for students taking the Programming Parallel Computers course. Janusz Kowalik Piotr Arlukowicz Tadeusz Puzniakowski Informatics Institute.
Incremental Support Vector Machine Classification Second SIAM International Conference on Data Mining Arlington, Virginia, April 11-13, 2002 Glenn Fung.
TeraGrid Arch Meeting RP Update: ORNL July 16, 2009 (This day in history: July 16, 1945: Alamogorda: Trinity) John W. Cobb.
DANSE Review – SNS/HFIR Update Steve Miller Scientific Computing Group Leader Neutron Scattering Science Division (NSSD) Spallation Neutron Source May.
1 Hybrid methods for solving large-scale parameter estimation problems Carlos A. Quintero 1 Miguel Argáez 1 Hector Klie 2 Leticia Velázquez 1 Mary Wheeler.
Test Of Distributed Data Quality Monitoring Of CMS Tracker Dataset H->ZZ->2e2mu with PileUp - 10,000 events ( ~ 50,000 hits for events) The monitoring.
Scientific Computing Topics for Final Projects Dr. Guy Tel-Zur Version 2,
Nick Draper Teswww.mantidproject.orgwww.mantidproject.org Instrument Independent Reduction and Analysis at ISIS and SNS.
A Metadata Based Approach For Supporting Subsetting Queries Over Parallel HDF5 Datasets Vignesh Santhanagopalan Graduate Student Department Of CSE.
Monte Carlo Instrument Simulation Activity at ISIS Dickon Champion, ISIS Facility.
Distributed Monte Carlo Instrument Simulations at ISIS Tom Griffin, ISIS Facility & University of Manchester.
17-April-2007 High Performance Computing Basics April 17, 2007 Dr. David J. Haglin.
Jaguar Super Computer Topics Covered Introduction Architecture Location & Cost Bench Mark Results Location & Manufacturer Machines in top 500 Operating.
CCGrid 2014 Improving I/O Throughput of Scientific Applications using Transparent Parallel Compression Tekin Bicer, Jian Yin and Gagan Agrawal Ohio State.
GU Junli SUN Yihe 1.  Introduction & Related work  Parallel encoder implementation  Test results and Analysis  Conclusions 2.
CMAQ Runtime Performance as Affected by Number of Processors and NFS Writes Patricia A. Bresnahan, a * Ahmed Ibrahim b, Jesse Bash a and David Miller a.
Grid MP at ISIS Tom Griffin, ISIS Facility. Introduction About ISIS Why Grid MP? About Grid MP Examples The future.
Nuclear structure and reactions Nicolas Michel University of Tennessee.
NIH Resource for Biomolecular Modeling and Bioinformatics Beckman Institute, UIUC NAMD Development Goals L.V. (Sanjay) Kale Professor.
Parallelization of 2D Lid-Driven Cavity Flow
Blue Brain Project Carlos Osuna, Carlos Aguado, Fabien Delalondre.
CCGrid 2014 Improving I/O Throughput of Scientific Applications using Transparent Parallel Compression Tekin Bicer, Jian Yin and Gagan Agrawal Ohio State.
The Research Alliance in Math and Science program is sponsored by the Office of Advanced Scientific Computing Research, Office of Science, U.S. Department.
Introduction to Research 2011 Introduction to Research 2011 Ashok Srinivasan Florida State University Images from ORNL, IBM, NVIDIA.
Nick Draper Tessella Instrument Independent Reduction and Analysis at ISIS and SNS.
1 Cluster Development at Fermilab Don Holmgren All-Hands Meeting Jefferson Lab June 1-2, 2005.
TeraGrid Quarterly Meeting Arlington, VA Sep 6-7, 2007 NCSA RP Status Report.
1 Spallation Neutron Source Data Analysis Jessica Travierso Research Alliance in Math and Science Program Austin Peay State University Mentor: Vickie E.
Distributed simulation with MPI in ns-3 Joshua Pelkey and Dr. George Riley Wns3 March 25, 2011.
TeraGrid Gateway User Concept – Supporting Users V. E. Lynch, M. L. Chen, J. W. Cobb, J. A. Kohl, S. D. Miller, S. S. Vazhkudai Oak Ridge National Laboratory.
SAN DIEGO SUPERCOMPUTER CENTER at the UNIVERSITY OF CALIFORNIA, SAN DIEGO Advanced User Support for MPCUGLES code at University of Minnesota October 09,
Image Processing A Study in Pixel Averaging Building a Resolution Pyramid With Parallel Computing Denise Runnels and Farnaz Zand.
CS 732: Advance Machine Learning
The Integrated Spectral Analysis Workbench (ISAW) DANSE Kickoff Meeting, Aug. 15, 2006, D. Mikkelson, T. Worlton, Julian Tao.
TeraGrid Arch Meeting RP Update: ORNL/NSTG 1 ORNL_RP_Update_ TeraGrid Arch Meeting RP Update: ORNL/NSTG August 1, 2008 John W. Cobb.
Large-scale accelerator simulations: Synergia on the Grid turn 1 turn 27 turn 19 turn 16 C++ Synergia Field solver (FFT, multigrid) Field solver (FFT,
Simulation Production System Science Advisory Committee Meeting UW-Madison March 1 st -2 nd 2007 Juan Carlos Díaz Vélez.
Mantid Scientific Steering Committee Nick Draper 18/06/2010.
Multicore Applications in Physics and Biochemical Research Hristo Iliev Faculty of Physics Sofia University “St. Kliment Ohridski” 3 rd Balkan Conference.
HPC need and potential of ANSYS CFD and mechanical products at CERN A. Rakai EN-CV-PJ2 5/4/2016.
Cooperative International Simulations with McStas Vickie Lynch, Meili Chen, John Cobb, Mark Hagen, James Kohl, Stephen Miller, Michael Reuter, Sudharshan.
Advanced Computing Facility Introduction
Compute and Storage For the Farm at Jlab
Introduction to Parallel Computing: MPI, OpenMP and Hybrid Programming
Hybrid Programming with OpenMP and MPI
Presentation transcript:

Neutron Science TeraGrid Gateway Update Vickie Lynch, Meili Chen, John Cobb Oak Ridge National Laboratory AUS Technical Presentation October 8, 2009

2Managed by UT-Battelle for the U.S. Department of Energy Outline Sinergia Startup Project Fitting Service Transition to SNS

3Managed by UT-Battelle for the U.S. Department of Energy Sinergia Startup Project Collaboration between the University of Zürich, ETH Zürich and Oak Ridge National Laboratory (ORNL) Swiss Sinergia grant supports PostDoc at Zürich and graduate student at ORNL

4Managed by UT-Battelle for the U.S. Department of Energy Diffuse Scattering Modeling 3D fitting of large datasets Needed for Single Crystal instruments – Topaz (SNS) – Snap (SNS) – MaNDi (SNS) – Corelli (SNS) – Four-circle Diffractometer (HFIR) – Imagine (HFIR)

5Managed by UT-Battelle for the U.S. Department of Energy Sinergia July 09 Kickoff in Zurich Organized by Hans-Beat Bürgi Michel Chodkiewicz (PostDoc at Zürich) will develop new C++ diffuse scattering code over 3 years Tara Clark (Grad student at ORNL) working with Vickie Lynch will get previous code running on TeraGrid and test Jürg Hauser from University of Bern will provide a reference dataset with less noise

6Managed by UT-Battelle for the U.S. Department of Energy Original diffuse scattering code Used for publications: – Weber T.,Bürgi H.B., Determination and refinement of disordered crystal structures using evolutionary algorithms in combination with Monte Carlo methods, Acta Crystallographica A 58, (2002) – H.B. Bürgi, J. Hauser, T. Weber, R.B. Neder, Supramolecular Architecture in a Disordered Perhydrotriphenylene Inclusion Compound from Diffuse X-ray Diffraction Data. Crystal Growth & Design, (2005) Master/slave code was written mainly in Perl with telnet to communicate between workstations (Calls Fortran and C codes for FFTs) Much file communication between master and slave – Code assumed unique disk space for each slave One run for over 200 generations took ~30 days on ~10 workstations

7Managed by UT-Battelle for the U.S. Department of Energy VOP code by Phil Bentley Vitess Optimisation Program written for optimizing parameters in instrument simulations Genetic algorithm with tournament selection; Swarms Written in C++ using MPI for parallel computing (Allreduce to find min; Bcast to sent best parameters) Command line input for Vitess with parameters to be optimized preceded Limits file read by code Written to maximize flux from Vitess, but easily modified to minimize  2

8Managed by UT-Battelle for the U.S. Department of Energy Code port to TeraGrid Replaced master telnet code with VOP MPI code VOP calls Perl slaves with pipes Unique directories created for each slave Ported and ran successfully on NSTG Needed faster turnaround and 1048 clones Got Kraken Startup allocation Popen for Perl slaves did not work with CNL

9Managed by UT-Battelle for the U.S. Department of Energy Code port to Ranger Tested that popen worked on Ranger – Transferred part of Startup to Ranger Too much file communication in Perl slaves for 16 cores/node Put OpenMP directives in FFT codes Using 4 slaves/node with 4 OpenMP threads/slave

10Managed by UT-Battelle for the U.S. Department of Energy VOP results with Perl slaves

11Managed by UT-Battelle for the U.S. Department of Energy Sinergia accomplishments Add differential evolution option to vop master Compared differential evolution, genetic algorithm and swarm for disordered crystal data. Genetic algorithm converged fasted, but each generation takes twice as long using tournament selection Added hybrid OpenMP directives Have code running on NSTG and Ranger Have Levenberg-Marquardt nonlinear least squares driver for slaves for further optimization Running using reference dataset on Ranger now (160 cores)

12Managed by UT-Battelle for the U.S. Department of Energy Future Sinergia work Install PGPLOT Perl module and helper scripts Testing with reference dataset – Number of individuals – Number of clones – Optimize mutation and crossover constants Scale to thousands of cores Automatically switch between genetic algorithms and least squares using history of chisq Write slave code in C++, not Perl Communicate between modules with memory, not files

13Managed by UT-Battelle for the U.S. Department of Energy Fitting Service Fast data fitting, the major and the most time consuming procedure of the data analysis, is the key of high performance data analysis and real-time data processing. Uses NL2SOL or Dakota to fit experimental data – Instrument scientists writes functions – Read data – Model to fit data –

14Managed by UT-Battelle for the U.S. Department of Energy Fitting Service results Fitting service used for experimental reflectometer data is shown NL2SOL results show 5 times smaller Chisq than previous fitting code used for this data

15Managed by UT-Battelle for the U.S. Department of Energy Fitting Service uses TeraGrid New fitting service – going into production – available from the development portal – will be “monitored” for health in the future Calculation scales linearly on TeraGrid on Mercury at NCSA See TG09 slides by Meili Chen for more details.

16Managed by UT-Battelle for the U.S. Department of Energy Fitting Service GUI in Portal

17Managed by UT-Battelle for the U.S. Department of Energy NSTG Cluster Transition to SNS Cluster: dual processor 3.06 GHz Intel Xeon nodes 14 nodes with 2.5GB of memory are compute nodes 4 nodes with 4GB memory are dedicated to GridFTP Transitioning to Scientific Linux 5.2 Move planned from ORNL to SNS machine room (~ 3 miles) Cluster will link SNS with TeraGrid