Visualization as a Science Discovery Tool Issues and Concerns Kelly Gaither Director of Visualization/ Sr. Research Scientist Texas Advanced Computing.

Slides:



Advertisements
Similar presentations
Founded in 2010: UCL, Southampton, Oxford and Bristol Key Objectives of the Consortium: Prove the concept of shared, regional e-infrastructure services.
Advertisements

Computing Infrastructure
CURRENT AND FUTURE HPC SOLUTIONS. T-PLATFORMS  Russia’s leading developer of turn-key solutions for supercomputing  Privately owned  140+ employees.
National Center for Atmospheric Research John Clyne 4/27/11 4/26/20111.
Parallel Visualization At TACC Greg Abram. Visualization Problems Small problems: Data are small and easily moved Office machines and laptops are adequate.
HPCC Mid-Morning Break High Performance Computing on a GPU cluster Dirk Colbry, Ph.D. Research Specialist Institute for Cyber Enabled Discovery.
WEST VIRGINIA UNIVERSITY HPC and Scientific Computing AN OVERVIEW OF HIGH PERFORMANCE COMPUTING RESOURCES AT WVU.
LinkSCEEM-2: A computational resource for the development of Computational Sciences in the Eastern Mediterranean Mostafa Zoubi SESAME SESAME – LinkSCEEM.
Advanced Scientific Visualization Paul Navrátil 28 May 2009.
Academic and Research Technology (A&RT)
Mike Smorul Saurabh Channan Digital Preservation and Archiving at the Institute for Advanced Computer Studies University of Maryland, College Park.
April 2009 OSG Grid School - RDU 1 Open Science Grid John McGee – Renaissance Computing Institute University of North Carolina, Chapel.
Research Computing with Newton Gerald Ragghianti Nov. 12, 2010.
Illinois Campus Cluster Program User Forum October 24, 2012 Illini Union Room 210 2:00PM – 3:30PM.
IPlant Collaborative Tools and Services Workshop iPlant Collaborative Tools and Services Workshop Overview of Atmosphere.
HPCC Mid-Morning Break Dirk Colbry, Ph.D. Research Specialist Institute for Cyber Enabled Discovery Introduction to the new GPU (GFX) cluster.
Digital Graphics and Computers. Hardware and Software Working with graphic images requires suitable hardware and software to produce the best results.
TeraGrid Gateway User Concept – Supporting Users V. E. Lynch, M. L. Chen, J. W. Cobb, J. A. Kohl, S. D. Miller, S. S. Vazhkudai Oak Ridge National Laboratory.
QCDgrid Technology James Perry, George Beckett, Lorna Smith EPCC, The University Of Edinburgh.
Project Overview:. Longhorn Project Overview Project Program: –NSF XD Vis Purpose: –Provide remote interactive visualization and data analysis services.
Research Support Services Research Support Services.
Introduction to HPC resources for BCB 660 Nirav Merchant
IPlant Collaborative Tools and Services Workshop iPlant Collaborative Tools and Services Workshop Overview of Atmosphere.
VO Sandpit, November 2009 e-Infrastructure to enable EO and Climate Science Dr Victoria Bennett Centre for Environmental Data Archival (CEDA)
IPlant Collaborative Tools and Services Workshop iPlant Collaborative Tools and Services Workshop Collaborating with iPlant.
Open Science Grid For CI-Days Elizabeth City State University Jan-2008 John McGee – OSG Engagement Manager Manager, Cyberinfrastructure.
David S. Ebert David S. Ebert Visual Analytics to Enable Discovery and Decision Making: Potential, Challenges, and.
IPlant Collaborative Tools and Services Workshop iPlant Collaborative Tools and Services Workshop Collaborating with iPlant.
INVITATION TO COMPUTER SCIENCE, JAVA VERSION, THIRD EDITION Chapter 6: An Introduction to System Software and Virtual Machines.
The Future of the iPlant Cyberinfrastructure: Coming Attractions.
National Energy Research Scientific Computing Center (NERSC) Visualization Tools and Techniques on Seaborg and Escher Wes Bethel & Cristina Siegerist NERSC.
Kelly Gaither Visualization Area Report. Efforts in 2008 Focused on providing production visualization capabilities (software and hardware) Focused on.
Looking Ahead: A New PSU Research Cloud Architecture Chuck Gilbert - Systems Architect and Systems Team Lead Research CI Coordinating Committee Meeting.
14 Aug 08DOE Review John Huth ATLAS Computing at Harvard John Huth.
NML Bioinformatics Service— Licensed Bioinformatics Tools High-throughput Data Analysis Literature Study Data Mining Functional Genomics Analysis Vector.
ARGONNE NATIONAL LABORATORY Climate Modeling on the Jazz Linux Cluster at ANL John Taylor Mathematics and Computer Science & Environmental Research Divisions.
A Framework for Visualizing Science at the Petascale and Beyond Kelly Gaither Research Scientist Associate Director, Data and Information Analysis Texas.
GVis: Grid-enabled Interactive Visualization State Key Laboratory. of CAD&CG Zhejiang University, Hangzhou
Scientific Advisory Committee – September 2011COLA Information Systems COLA’s Information Systems 2011.
VAPoR: A Discovery Environment for Terascale Scientific Data Sets Alan Norton & John Clyne National Center for Atmospheric Research Scientific Computing.
TeraGrid Quarterly Meeting Arlington, VA Sep 6-7, 2007 NCSA RP Status Report.
1 NSF/TeraGrid Science Advisory Board Meeting July 19-20, San Diego, CA Brief TeraGrid Overview and Expectations of Science Advisory Board John Towns TeraGrid.
Using the Weizmann Cluster Nov Overview Weizmann Cluster Connection Basics Getting a Desktop View Working on cluster machines GPU For many more.
TeraGrid Gateway User Concept – Supporting Users V. E. Lynch, M. L. Chen, J. W. Cobb, J. A. Kohl, S. D. Miller, S. S. Vazhkudai Oak Ridge National Laboratory.
NICS RP Update TeraGrid Round Table March 10, 2011 Ryan Braby NICS HPC Operations Group Lead.
National Center for Supercomputing Applications University of Illinois at Urbana–Champaign Visualization Support for XSEDE and Blue Waters DOE Graphics.
Open XDMoD Overview Tom Furlani, Center for Computational Research
NICS Update Bruce Loftis 16 December National Institute for Computational Sciences University of Tennessee and ORNL partnership  NICS is the 2.
Remote & Collaborative Visualization. TACC Remote Visualization Systems Longhorn – Dell XD Visualization Cluster –256 nodes, each with 48 GB (or 144 GB)
What’s Coming? What are we Planning?. › Better docs › Goldilocks – This slot size is just right › Storage › New.
Locate Potential Support Vectors for Faster
Visualization Update June 18, 2009 Kelly Gaither, GIG Area Director DV.
PEER 2003 Meeting 03/08/031 Interdisciplinary Framework Major focus areas Structural Representation Fault Systems Earthquake Source Physics Ground Motions.
TG ’08, June 9-13, State of TeraGrid John Towns Co-Chair, TeraGrid Forum Director, Persistent Infrastructure National Center for Supercomputing.
The New Lonestar Dan Stanzione Deputy Director, TACC 3/24/2011.
Introduction to Data Analysis with R on HPC Texas Advanced Computing Center Feb
INTRODUCTION TO XSEDE. INTRODUCTION  Extreme Science and Engineering Discovery Environment (XSEDE)  “most advanced, powerful, and robust collection.
An Brief Introduction Charlie Taylor Associate Director, Research Computing UF Research Computing.
Advanced Computing Facility Introduction
Extreme Scale Infrastructure
Workstations & Thin Clients
What is HPC? High Performance Computing (HPC)
Jay Boisseau, Director Texas Advanced Computing Center
Brad Sutton Assoc Prof, Bioengineering Department
Working With Azure Batch AI
VirtualGL.
Architecture & System Overview
Introduction to High Performance Computing Using Sapelo2 at GACRC
H2020 EU PROJECT | Topic SC1-DTH | GA:
Presentation transcript:

Visualization as a Science Discovery Tool Issues and Concerns Kelly Gaither Director of Visualization/ Sr. Research Scientist Texas Advanced Computing Center September 8, 2011

Issues and Concerns –Maximizing Scientific Impact –Managing Data at Scale –Providing Resources at Scale –Ensuring Broad Accessibility/Developing Ubiquitous Tools

Visualization Mantra “The purpose of computing is insight not numbers.” -- R. W. Hamming (1961) “The purpose of computing is insight not numbers.” -- R. W. Hamming (1961) “The purpose of visualization is insight not pictures.” -- Ben Schneiderman (2005) “The purpose of visualization is insight not pictures.” -- Ben Schneiderman (2005)

Maximizing Scientific Impact Image: Greg P. Johnson, Romy Schneider, TACCImage: Adam Kubach, Karla Vega, Clint Dawson Image: Karla Vega, Shaolie Hossain, Thomas J.R., Hughes Greg Abram, Carsten Burstedde, Georg Stadler, Lucas C. Wilcox, James R. Martin, Tobin Isaac, Tan Bui-Thanh,and Omar Ghattas

Managing Data at Scale Large-Scale Visualization Resource Large-Scale Visualization Resource HPC System HPC System Data Archive Data Archive Display Remote Site Wide-Area Network Wide-Area Network Local Site Pixels Mouse

Where Does Technology Fit In? We have always used technology to create visualizations of what we see in our minds eye. What changes over time is the technology we use to do the visualization.

Longhorn First NSF XD Visualization Resource 256 Dell Dual Socket, Quad Core Intel Nehalem Nodes –240 with 48 GB shared memory/node (6 GB/core) –16 with 144 GB shared memory/node (18 GB/core) –73 GB Local Disk –2 Nvidia GPUs/Node (FX 5800 – 4GB RAM) ~13.5 TB aggregate memory QDR InfiniBand Interconnect Jobs launched through SGE ~6GB/s to scratch filesystem ~6GB/s to Ranger filesystem 256 Nodes, 2048 Cores, 512 GPUs, 14.5 TB Memory Kelly Gaither (PI), Valerio Pascucci, Chuck Hansen, David Ebert, John Clyne (Co-PI), Hank Childs

Hadoop on Longhorn PI: Weijia Xu (UT LIFT Grant) Local Storage Expansion – GB 7.2k drives are installed on 48 R610 nodes on Longhorn (96 usable TB) – GB 15k drives are installed on 16 R710 nodes on Longhorn. (16 usable TB) /hadoop file system Went in to production December early user projects (text mining and information retrieval) 142k core hours from Dec to Feb

Longhorn Usage Modalities: Remote/Interactive Visualization –Highest priority jobs –Remote/Interactive capabilities facilitated through VNC –Run on 3 hour queue limit boundary GPGPU jobs –Run on a lower priority than the remote/interactive jobs –Run on a 12 hour queue limit boundary CPU jobs with higher memory requirements –Run on lowest priority when neither remote/interactive nor GPGPU jobs are waiting in the queue –Run on a 12 hour queue limit boundary

Longhorn Queue Structure Example: qsub -q normal -P vis

Software Available on Longhorn Programming APIs: OpenGL, vtk (Not natively parallel) –OpenGL – low level primitives, useful for programming at a relatively low level with respect to graphics –VTK (Visualization Toolkit) – open source software system for 3D computer graphics, image processing, and visualization –IDL Visualization Turnkey Systems –VisIt – free open source parallel visualization and graphical analysis tool –ParaView – free open source general purpose parallel visualization system –VAPOR – free flow visualization package developed out of NCAR –EnSight – commercial turnkey parallel visualization package targeted at CFD visualization –Amira – commercial turnkey visualization package targeted at visualizing scanned medical data (CAT scan, MRI, etc..)

Longhorn Visualization Portal portal.longhorn.tacc.utexas.edu Developed to provide easy access to Longhorn and abstract away complexities involved with command line access Leverages TeraGrid user portal codebase and employs a fraction of TGUP developers to ensure continuity Used for all in-person Longhorn training

Longhorn Visualization Portal portal.longhorn.tacc.utexas.edu >3000 jobs submitted through the portal

Longhorn Visualization Portal portal.longhorn.tacc.utexas.edu Specify type of session Specify resolution of vnc session Specify number of nodes needed and the wayness of the nodes Provides graphic of machine load

Longhorn Visualization Portal portal.longhorn.tacc.utexas.edu Vnc session opens up in java enabled browser Behaves as if user had gotten a remote desktop into Longhorn.

EnVision Greg Johnson, Brandt Westing Web-based visualization software that allows researchers to develop interactive visualizations intuitively. Currently integrated into the Longhorn Visualization Portal but can run independently. Began collaborations with ParaView team.

Summary Close collaborations with the science partners are key –User support Minimize data transfers if possible –Data stays in single location Scale resources effectively based on use cases Easy accessibility to and interaction with technologies encourages diverse communities participation

Questions?