ICALEPCS 2005 10/13/2005 M. Greenwald Visions for Data Management and Remote Collaboration on ITER M. Greenwald, D. Schissel, J. Burruss, T. Fredian, J.

Slides:



Advertisements
Similar presentations
Implementing Tableau Server in an Enterprise Environment
Advertisements

A Workflow Engine with Multi-Level Parallelism Supports Qifeng Huang and Yan Huang School of Computer Science Cardiff University
Distributed Data Processing
ASCR Data Science Centers Infrastructure Demonstration S. Canon, N. Desai, M. Ernst, K. Kleese-Van Dam, G. Shipman, B. Tierney.
High Performance Computing Course Notes Grid Computing.
MDSplus Tom Fredian MIT Plasma Science and Fusion Center.
SWIM WEB PORTAL by Dipti Aswath SWIM Meeting ORNL Oct 15-17, 2007.
Summary Role of Software (1 slide) ARCS Software Architecture (4 slides) SNS -- Caltech Interactions (3 slides)
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
Building Enterprise Applications Using Visual Studio ®.NET Enterprise Architect.
Presented by Scalable Systems Software Project Al Geist Computer Science Research Group Computer Science and Mathematics Division Research supported by.
Network Management Overview IACT 918 July 2004 Gene Awyzio SITACS University of Wollongong.
Notes to the presenter. I would like to thank Jim Waldo, Jon Bostrom, and Dennis Govoni. They helped me put this presentation together for the field.
David Adams ATLAS DIAL Distributed Interactive Analysis of Large datasets David Adams BNL March 25, 2003 CHEP 2003 Data Analysis Environment and Visualization.
Workload Management Workpackage Massimo Sgaravatto INFN Padova.
Workload Management Massimo Sgaravatto INFN Padova.
Chapter 9: Moving to Design
Course Instructor: Aisha Azeem
Architectural Design Establishing the overall structure of a software system Objectives To introduce architectural design and to discuss its importance.
Project Proposal: Academic Job Market and Application Tracker Website Project designed by: Cengiz Gunay Client: Cengiz Gunay Audience: PhD candidates and.
V. Chandrasekar (CSU), Mike Daniels (NCAR), Sara Graves (UAH), Branko Kerkez (Michigan), Frank Vernon (USCD) Integrating Real-time Data into the EarthCube.
1 Autonomic Computing An Introduction Guenter Kickinger.
1 Dr. Markus Hillenbrand, ICSY Lab, University of Kaiserslautern, Germany A Generic Database Web Service for the Venice Service Grid Michael Koch, Markus.
Current Job Components Information Technology Department Network Systems Administration Telecommunications Database Design and Administration.
Rational Unified Process Fundamentals Module 4: Disciplines II.
Microsoft Active Directory(AD) A presentation by Robert, Jasmine, Val and Scott IMT546 December 11, 2004.
Scalable Systems Software Center Resource Management and Accounting Working Group Face-to-Face Meeting June 13-14, 2002.
DISTRIBUTED COMPUTING
ICALEPCS /13/2005 M. Greenwald Visions for Data Management and Remote Collaboration on ITER M. Greenwald, D. Schissel, J. Burruss, T. Fredian, J.
Database structure for the European Integrated Tokamak Modelling Task Force F. Imbeaux On behalf of the Data Coordination Project.
SAMANVITHA RAMAYANAM 18 TH FEBRUARY 2010 CPE 691 LAYERED APPLICATION.
1 COMPSCI 110 Operating Systems Who - Introductions How - Policies and Administrative Details Why - Objectives and Expectations What - Our Topic: Operating.
Scalable Systems Software Center Resource Management and Accounting Working Group Face-to-Face Meeting October 10-11, 2002.
CSE 219 Computer Science III Program Design Principles.
OOI CI LCA REVIEW August 2010 Ocean Observatories Initiative OOI Cyberinfrastructure Architecture Overview Michael Meisinger Life Cycle Architecture Review.
Grid Workload Management Massimo Sgaravatto INFN Padova.
© 2012 xtUML.org Bill Chown – Mentor Graphics Model Driven Engineering.
August 2003 At A Glance VMOC-CE is an application framework that facilitates real- time, remote cooperative work among geographically dispersed mission.
The european ITM Task Force data structure F. Imbeaux.
NOVA Networked Object-based EnVironment for Analysis P. Nevski, A. Vaniachine, T. Wenaus NOVA is a project to develop distributed object oriented physics.
Systems Analysis and Design in a Changing World, Fourth Edition
1 Computing Challenges for the Square Kilometre Array Mathai Joseph & Harrick Vin Tata Research Development & Design Centre Pune, India CHEP Mumbai 16.
Ames Research CenterDivision 1 Information Power Grid (IPG) Overview Anthony Lisotta Computer Sciences Corporation NASA Ames May 2,
Framework for MDO Studies Amitay Isaacs Center for Aerospace System Design and Engineering IIT Bombay.
NA-MIC National Alliance for Medical Image Computing UCSD: Engineering Core 2 Portal and Grid Infrastructure.
Presented by Scientific Annotation Middleware Software infrastructure to support rich scientific records and the processes that produce them Jens Schwidder.
9 Systems Analysis and Design in a Changing World, Fourth Edition.
GRID Overview Internet2 Member Meeting Spring 2003 Sandra Redman Information Technology and Systems Center and Information Technology Research Center National.
Scalable Systems Software for Terascale Computer Centers Coordinator: Al Geist Participating Organizations ORNL ANL LBNL.
Presented by Jens Schwidder Tara D. Gibson James D. Myers Computing & Computational Sciences Directorate Oak Ridge National Laboratory Scientific Annotation.
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
Introduction to Grids By: Fetahi Z. Wuhib [CSD2004-Team19]
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
Architecture View Models A model is a complete, simplified description of a system from a particular perspective or viewpoint. There is no single view.
10/041 Remote Collaboration at the PSFC and in the Fusion Community Joshua Stillerman MIT Plasma Science and Fusion Center National Fusion Collaboratory.
GRID ANATOMY Advanced Computing Concepts – Dr. Emmanuel Pilli.
Douglas Thain, John Bent Andrea Arpaci-Dusseau, Remzi Arpaci-Dusseau, Miron Livny Computer Sciences Department, UW-Madison Gathering at the Well: Creating.
Cyberinfrastructure Overview of Demos Townsville, AU 28 – 31 March 2006 CREON/GLEON.
INFSO-RI Enabling Grids for E-sciencE File Transfer Software and Service SC3 Gavin McCance – JRA1 Data Management Cluster Service.
System Architecture CS 560. Project Design The requirements describe the function of a system as seen by the client. The software team must design a system.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Introduction Salma Saber Electronic.
Grid Services for Digital Archive Tao-Sheng Chen Academia Sinica Computing Centre
Workload Management Workpackage
Clouds , Grids and Clusters
Grid Computing.
Integration of Network Services Interface version 2 with the JUNOS Space SDK
Design and Implementation
The Improvement of PaaS Platform ZENG Shu-Qing, Xu Jie-Bin 2010 First International Conference on Networking and Distributed Computing SQUARE.
LQCD Computing Operations
PLANNING A SECURE BASELINE INSTALLATION
Presentation transcript:

ICALEPCS /13/2005 M. Greenwald Visions for Data Management and Remote Collaboration on ITER M. Greenwald, D. Schissel, J. Burruss, T. Fredian, J. Lister, J. Stillerman MIT, GA, CRPP Presented by Martin Greenwald MIT – Plasma Science & Fusion Center CRPP, Lausanne, 2005

ICALEPCS /13/2005 M. Greenwald ITER is Clearly the Next Big Thing in Magnetic Fusion Research It will be the largest and most expensive scientific instrument ever built for fusion research Built and operated as an international collaboration To ensure its scientific productivity, systems for data management and remote collaboration must be done well.

ICALEPCS /13/2005 M. Greenwald What Challenges Will ITER Present? Fusion experiments require extensive “near” real-time data visualization and analysis in support of between-shot decision making. –For ITER, shots are: à~400 seconds each, maybe 1 hour apart à2,000 per year for 15 years –Average cost per shot is very high (order $1M) Today, teams of ~ work together closely during operation of experiments. Real-time remote participation is standard operating procedure.

ICALEPCS /13/2005 M. Greenwald Challenges: Experimental Fusion Science is a Demanding Real-Time Activity Run-time goals: –Optimize fusion performance –Ensure conditions are fully documented before moving on Drives need to assimilate, analyze and visualize large quantity of data between shots.

ICALEPCS /13/2005 M. Greenwald Challenge: Long Pulse Length Requires concurrent writing, reading, analysis –(We don’t think this will be too hard) Data sets will be larger than they are today –Perhaps 1 TB per shot, > 1 PB per year –(We think this will be manageable when needed) More challenging – integration across time scales Data will span range > 10 9 in significant time scales –Fluctuation time scale  pulse length Will require efficient tools –To browse very long records –To locate and describe specific events or intervals

ICALEPCS /13/2005 M. Greenwald Challenge: Long Life of Project 10 years construction; 15+ years operation Systems must adapt to decades of information technology evolution –Software, protocols, hardware will all change. –Think back 25 years! We need to anticipate a complete changeover in workforce. Backward compatibility must be maintained

ICALEPCS /13/2005 M. Greenwald Challenges: International, Remote Participation Scientists will want to participate in experiments from their home institutions dispersed around the world. –View and analyze data during operations –Manage ITER diagnostics –Lead experimental sessions –Participate in international task forces Collaborations span many administrative domains (more on this later) Cyber-security must be maintained, plant security must be inviolable.

ICALEPCS /13/2005 M. Greenwald Challenges: International, Remote Participation Scientists will want to participate in experiments from their home institutions dispersed around the world. –View and analyze data during operations –Manage ITER diagnostics –Lead experimental sessions –Participate in international task forces Collaborations span many administrative domains (more on this later) Cyber-security must be maintained, plant security must be inviolable.

ICALEPCS /13/2005 M. Greenwald Challenges: International, Remote Participation Scientists will want to participate in experiments from their home institutions dispersed around the world. –View and analyze data during operations –Manage ITER diagnostics –Lead experimental sessions –Participate in international task forces Collaborations span many administrative domains (more on this later) Cyber-security must be maintained, plant security must be inviolable.

ICALEPCS /13/2005 M. Greenwald We Are Beginning the Dialogue About How to Proceed This is not yet an “official” ITER activity What follows is our vision for data management and remote participation systems Opinions expressed here are the authors alone.

ICALEPCS /13/2005 M. Greenwald Strategy: Design, Prototype and Demo With 10 years before first operation, it is too early to choose specific implementations – software or hardware Begin now on enduring features –Define requirements, scope of effort, approach –Decide on general principles and features of architecture Within 2 years: start on prototypes:  Part of conceptual design Within 4 years: demonstrate: –Test, especially on current facilities –Simulation codes could provide testbed for long-pulse features In 6 years: proven implementations expanded and elaborated to meet requirements

ICALEPCS /13/2005 M. Greenwald Strategy: Design, Prototype and Demo With 10 years before first operation, it is too early to choose specific implementations – software or hardware Begin now on enduring features –Define requirements, scope of effort, approach –Decide on general principles and features of architecture Within 2 years: start on prototypes:  Part of conceptual design Within 4 years: demonstrate: –Test, especially on current facilities –Simulation codes could provide testbed for long-pulse features In 6 years: proven implementations expanded and elaborated to meet requirements

ICALEPCS /13/2005 M. Greenwald Strategy: Design, Prototype and Demo With 10 years before first operation, it is too early to choose specific implementations – software or hardware Begin now on enduring features –Define requirements, scope of effort, approach –Decide on general principles and features of architecture Within 2 years: start on prototypes:  Part of conceptual design Within 4 years: demonstrate: –Test, especially on current facilities –Simulation codes could provide testbed for long-pulse features In 6 years: proven implementations expanded and elaborated to meet requirements

ICALEPCS /13/2005 M. Greenwald Strategy: Design, Prototype and Demo With 10 years before first operation, it is too early to choose specific implementations – software or hardware Begin now on enduring features –Define requirements, scope of effort, approach –Decide on general principles and features of architecture Within 2 years: start on prototypes:  Part of conceptual design Within 4 years: demonstrate: –Test, especially on current facilities –Simulation codes could provide testbed for long-pulse features In 6 years: proven implementations expanded and elaborated to meet requirements

ICALEPCS /13/2005 M. Greenwald Strategy: Design, Prototype and Demo With 10 years before first operation, it is too early to choose specific implementations – software or hardware Begin now on enduring features –Define requirements, scope of effort, approach –Decide on general principles and features of architecture Within 2 years: start on prototypes:  Part of conceptual design Within 4 years: demonstrate: –Test, especially on current facilities –Simulation codes could provide testbed for long-pulse features In 6 years: proven implementations expanded and elaborated to meet requirements

ICALEPCS /13/2005 M. Greenwald General Features Extensible, flexible, scalable –We won’t be able to predict all future needs –Capable of continuous and incremental improvement –Requires robust underlying abstraction Data Accessible from wide range of languages, software frameworks and hardware platforms –The international collaboration will be heterogeneous Built-in security –Must protect plant without endangering science mission

ICALEPCS /13/2005 M. Greenwald Proposed Top Level Data Architecture Data Acquisition Control Service Oriented API Analysis Applications Visualization Applications Relational Database Main Repository Data Acquisition Systems Contains data searchable by their contents Contains multi- dimensional data indexed by their independent parameters

ICALEPCS /13/2005 M. Greenwald Data System – Contents & Structure Provide coherent, complete, integrated, self-descriptive view of all data through simple interfaces. –All  Raw, processed, analyzed data, configuration, geometry calibrations, data acquisition setup, code parameters, labels, comments… –No data in applications or private files Metadata stored for each data element Logical relationships and associations among data elements are made explicit by structure (probably multiple hierarchies). Data structures can be traversed independent of reading data. Powerful data directories (10 5 – 10 6 named items)

ICALEPCS /13/2005 M. Greenwald Data System – Contents & Structure Provide coherent, complete, integrated, self-descriptive view of all data through simple interfaces. –All  Raw, processed, analyzed data, configuration, geometry calibrations, data acquisition setup, code parameters, labels, comments… –No data in applications or private files Metadata stored for each data element Logical relationships and associations among data elements are made explicit by structure (probably multiple hierarchies). Data structures can be traversed independent of reading data. Powerful data directories (10 5 – 10 6 named items)

ICALEPCS /13/2005 M. Greenwald Data System – Contents & Structure Provide coherent, complete, integrated, self-descriptive view of all data through simple interfaces. –All  Raw, processed, analyzed data, configuration, geometry calibrations, data acquisition setup, code parameters, labels, comments… –No data in applications or private files Metadata stored for each data element Logical relationships and associations among data elements are made explicit by structure (probably multiple hierarchies). Data structures can be traversed independent of reading data. Powerful data directories (10 5 – 10 6 named items)

ICALEPCS /13/2005 M. Greenwald Data System – Contents & Structure Provide coherent, complete, integrated, self-descriptive view of all data through simple interfaces. –All  Raw, processed, analyzed data, configuration, geometry calibrations, data acquisition setup, code parameters, labels, comments… –No data in applications or private files Metadata stored for each data element Logical relationships and associations among data elements are made explicit by structure (probably multiple hierarchies). Data structures can be traversed independent of reading data. Powerful data directories (10 5 – 10 6 named items)

ICALEPCS /13/2005 M. Greenwald Data System – Contents & Structure Provide coherent, complete, integrated, self-descriptive view of all data through simple interfaces. –All  Raw, processed, analyzed data, configuration, geometry calibrations, data acquisition setup, code parameters, labels, comments… –No data in applications or private files Metadata stored for each data element Logical relationships and associations among data elements are made explicit by structure (probably multiple hierarchies). Data structures can be traversed independent of reading data. Powerful data directories (10 5 – 10 6 named items)

ICALEPCS /13/2005 M. Greenwald Data System – Contents & Structure Provide coherent, complete, integrated, self-descriptive view of all data through simple interfaces. –All  Raw, processed, analyzed data, configuration, geometry calibrations, data acquisition setup, code parameters, labels, comments… –No data in applications or private files Metadata stored for each data element Logical relationships and associations among data elements are made explicit by structure (probably multiple hierarchies). Data structures can be traversed independent of reading data. Powerful data directories (10 5 – 10 6 named items)

ICALEPCS /13/2005 M. Greenwald Data System – Service Oriented Architectures Service oriented –Loosely coupled applications, running on distributed servers –Interfaces simple and generic, implementation details hidden àTransparency and ease-of-use are crucial àApplications specify what is to be done, not how –Data structures shared –Service discovery supported We’re already headed in this direction –MDSplus –TRANSP “FusionGrid” service

ICALEPCS /13/2005 M. Greenwald Resources Accessible Via Network Services Resources = Computers, Codes, Data, Analysis Routines, Visualization tools, Experimental Status and Operations Access is stressed rather than portability Users are shielded from implementation details. Transparency and ease-of-use are crucial elements Shared toolset enables collaboration between sites and across sub- disciplines. Knowledge of relevant physics is still required of course.

ICALEPCS /13/2005 M. Greenwald Case Study: TRANSP CODE – “FusionGrid Service” Over 20 years of development by PPPL (+ others) –>1,000,000 lines of Fortran, C, C++ –>3,000 program modules –10,000s lines of supporting script code: perl, python, shell-script –Used internationally for most tokamak experiments –Local maintenance has been very manpower intensive Now fully integrated with MDSplus data system –Standard data “trees” developed for MIT, GA, PPPL, JET, … –Standard toolset for run preparation, visualization

ICALEPCS /13/2005 M. Greenwald Production TRANSP System User (anywhere) Experimental Site PPPL Authorization server may be consulted at each stage

ICALEPCS /13/2005 M. Greenwald TRANSP Service Has Had Immediate Payoff Remote sites avoid costly installation and code maintenance –Was ~1 man-month per year, per site Users always have access to latest code version PPPL maintains and supports a single production version of code on well characterized platform –Improves user support at reduced costs Users have access to high-performance production system –16 processor linux cluster –Dedicated PBS queue –Tools developed for job submission, cancellation, monitoring

ICALEPCS /13/2005 M. Greenwald TRANSP Jobs Tracked by Fusion Grid Monitor Java Servlet derived from GA Data Analysis Monitor User presented with dynamical web display Sits on top of relational database – can feed accounting database Provides information on state of jobs, servers, logs, etc.

ICALEPCS /13/2005 M. Greenwald Usage Continues to Grow As of July: 5,800 runs from 10 different experiments

ICALEPCS /13/2005 M. Greenwald This Is Related To, But Not The Same As “Grid” Computing Traditional computational grids –Arrays of heterogeneous servers –Machines can arrive and leave –Adaptive discovery – problems find resources –Workload balancing and cycle scavenging –Bandwidth diversity – not all machines are well connected This model is not especially suited to our problems Instead, we are aiming to move high-performance distributed computing out onto the wide-area network We are not focusing on “traditional” grid applications – cycle scavenging and dynamically configured server farms

ICALEPCS /13/2005 M. Greenwald Putting Distributed Computing Applications out on the Wide Area Network Presents Significant Challenges Crosses administrative boundaries –Increased concerns and complexity for security model (authentication, authorization, resource management) Resources not owned by a single project or program –Distributed control of resources by owners is essential Needs for end-to-end application performance and problem resolution –Resource monitoring, management and troubleshooting are not straightforward Higher latency challenges network throughput, interactivity People are not in one place for easy communication

ICALEPCS /13/2005 M. Greenwald Data Driven Applications Data driven –All parameters in database not imbedded in applications –Data structure, relations, associations are data themselves –Processing “tree” maintained as data Enable “generic” applications –Processing can be sensitive to data relationships and to position of data within structure –Scope of applications can grow without modifying code

ICALEPCS /13/2005 M. Greenwald Data System - Higher Level Organization All part of database All indexed into main data repository High level physics analysis –Scalar and profile databases Event identification, logging & tracking Integrated and shared workspaces –Electronic logbook –Summaries and status àRuns àTask groups àCampaigns –Presentations & publications

ICALEPCS /13/2005 M. Greenwald Remote Participation Creating an Environment Which Is Equally Productive for Local and Remote Researchers Transparent remote access to data –Secure and timely Real-time info –Machine status –Shot cycle –Data acquisition and analysis monitoring –Announcements Shared applications Provision for Ad Hoc interpersonal communications Provision for Structured communications

ICALEPCS /13/2005 M. Greenwald Remote is Easy, Distributed is Hard Informal interactions in the control room are a crucial part of the research We must extend this into remote and distributed operations Fully engaging remote participants is challenging (Fortunately we have already substantial experience)

ICALEPCS /13/2005 M. Greenwald Remote Participation Ad Hoc Communications Exploit convergence of telecom and internet technologies (eg. SIP) Deploy integrated communications –Voice –Video –Messaging – –Data streaming Advanced directory services –Identification, location, scheduling –“Presence” –Support for “roles”

ICALEPCS /13/2005 M. Greenwald Cyber-Security Needs to Be Built In Must protect plant without endangering science mission Employ best features of identity-based, application and perimeter security models Strong authentication mechanisms Single sign-on – a must if there are many distributed services Distributed authorization and resource management –Allows stakeholders to control their own resources. –Facility owners can protect computers, data and experiments –Code developers can control intellectual property –Fair use of shared resources can be demonstrated and controlled.

ICALEPCS /13/2005 M. Greenwald Cyber-Security Needs to Be Built In Must protect plant without endangering science mission Employ best features of identity-based, application and perimeter security models Strong authentication mechanisms Single sign-on – a must if there are many distributed services Distributed authorization and resource management –Allows stakeholders to control their own resources. –Facility owners can protect computers, data and experiments –Code developers can control intellectual property –Fair use of shared resources can be demonstrated and controlled.

ICALEPCS /13/2005 M. Greenwald Cyber-Security Needs to Be Built In Must protect plant without endangering science mission Employ best features of identity-based, application and perimeter security models Strong authentication mechanisms Single sign-on – a must if there are many distributed services Distributed authorization and resource management –Allows stakeholders to control their own resources. –Facility owners can protect computers, data and experiments –Code developers can control intellectual property –Fair use of shared resources can be demonstrated and controlled.

ICALEPCS /13/2005 M. Greenwald Cyber-Security Needs to Be Built In Must protect plant without endangering science mission Employ best features of identity-based, application and perimeter security models Strong authentication mechanisms Single sign-on – a must if there are many distributed services Distributed authorization and resource management –Allows stakeholders to control their own resources. –Facility owners can protect computers, data and experiments –Code developers can control intellectual property –Fair use of shared resources can be demonstrated and controlled.

ICALEPCS /13/2005 M. Greenwald Cyber-Security Needs to Be Built In Must protect plant without endangering science mission Employ best features of identity-based, application and perimeter security models Strong authentication mechanisms Single sign-on – a must if there are many distributed services Distributed authorization and resource management –Allows stakeholders to control their own resources. –Facility owners can protect computers, data and experiments –Code developers can control intellectual property –Fair use of shared resources can be demonstrated and controlled.

ICALEPCS /13/2005 M. Greenwald Summary While ITER operation is many years in the future, work on the systems for data management and remote participation should begin now We propose: All data into a single, coherent, self-descriptive structure Service-oriented access All applications data driven Remote participation fully supported –Transparent, secure, timely remote data access –Support for ad hoc interpersonal communications –Shared applications enabled