TAT CrossGrid Yearly Review, Brussels, March 12, 2003 CrossGrid After the First Year: A Technical Overview Marian Bubak, Maciej Malawski, and Katarzyna.

Slides:



Advertisements
Similar presentations
Data Management Expert Panel - WP2. WP2 Overview.
Advertisements

Interaction model of grid services in mobile grid environment Ladislav Pesicka University of West Bohemia.
EGC 2005, CrossGrid technical achievements, Amsterdam, Feb. 16th, 2005 WP2-3 New Generation Environment for Grid Interactive MPI Applications M igrating.
FP7-INFRA Enabling Grids for E-sciencE EGEE Induction Grid training for users, Institute of Physics Belgrade, Serbia Sep. 19, 2008.
High Performance Computing Course Notes Grid Computing.
This product includes material developed by the Globus Project ( Introduction to Grid Services and GT3.
EU-GRID Work Program Massimo Sgaravatto – INFN Padova Cristina Vistoli – INFN Cnaf as INFN members of the EU-GRID technical team.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
Cracow Grid Workshop, November 5-6, 2001 Towards the CrossGrid Architecture Marian Bubak, Marek Garbacz, Maciej Malawski, and Katarzyna Zając.
Workload Management Workpackage Massimo Sgaravatto INFN Padova.
The CrossGrid project Juha Alatalo Timo Koivusalo.
5 th EU DataGrid Conference, Budapest, September 2002 The European CrossGrid Project Marcel Kunze Abteilung Grid-Computing und e-Science Forschungszentrum.
DataGrid Kimmo Soikkeli Ilkka Sormunen. What is DataGrid? DataGrid is a project that aims to enable access to geographically distributed computing power.
Dagstuhl Seminar 02341: Performance Analysis and Distributed Computing, August 18 – 23, 2002, Germany Monitoring of Interactive Grid Applications Marian.
CrossGrid Task 3.3 Grid Monitoring Trinity College Dublin (TCD) Brian Coghlan Paris MAR-2002.
TAT Cracow Grid Workshop, October 27 – 29, 2003 Marian Bubak, Michal Turala and the CrossGrid Collaboration CrossGrid in Its Halfway:
Workload Management Massimo Sgaravatto INFN Padova.
M.Kunze, NEC2003, Varna The European CrossGrid Project Marcel Kunze Institute for Scientific Computing (IWR) Forschungszentrum Karlsruhe GmbH
Web-based Portal for Discovery, Retrieval and Visualization of Earth Science Datasets in Grid Environment Zhenping (Jane) Liu.
The SAM-Grid Fabric Services Gabriele Garzoglio (for the SAM-Grid team) Computing Division Fermilab.
SUN HPC Consortium, Heidelberg 2004 Grid(Lab) Resource Management System (GRMS) and GridLab Services Krzysztof Kurowski Poznan Supercomputing and Networking.
QCDgrid Technology James Perry, George Beckett, Lorna Smith EPCC, The University Of Edinburgh.
GRACE Project IST EGAAP meeting – Den Haag, 25/11/2004 Giuseppe Sisto – Telecom Italia Lab.
Workload Management WP Status and next steps Massimo Sgaravatto INFN Padova.
WP9 Resource Management Current status and plans for future Juliusz Pukacki Krzysztof Kurowski Poznan Supercomputing.
GT Components. Globus Toolkit A “toolkit” of services and packages for creating the basic grid computing infrastructure Higher level tools added to this.
Integrating HPC into the ATLAS Distributed Computing environment Doug Benjamin Duke University.
QCDGrid Progress James Perry, Andrew Jackson, Stephen Booth, Lorna Smith EPCC, The University Of Edinburgh.
Grid Workload Management & Condor Massimo Sgaravatto INFN Padova.
DORII Joint Research Activities DORII Joint Research Activities Status and Progress 6 th All-Hands-Meeting (AHM) Alexey Cheptsov on.
DataGrid WP1 Massimo Sgaravatto INFN Padova. WP1 (Grid Workload Management) Objective of the first DataGrid workpackage is (according to the project "Technical.
1 DIRAC – LHCb MC production system A.Tsaregorodtsev, CPPM, Marseille For the LHCb Data Management team CHEP, La Jolla 25 March 2003.
The PROGRESS Grid Service Provider Maciej Bogdański Portals & Portlets 2003 Edinburgh, July 14th-17th.
Grid Workload Management Massimo Sgaravatto INFN Padova.
Virtual Data Grid Architecture Ewa Deelman, Ian Foster, Carl Kesselman, Miron Livny.
Resource Brokering in the PROGRESS Project Juliusz Pukacki Grid Resource Management Workshop, October 2003.
Grid Execution Management for Legacy Code Applications Grid Enabling Legacy Code Applications Tamas Kiss Centre for Parallel.
High Level Architecture (HLA)  used for building interactive simulations  connects geographically distributed nodes  time management (for time- and.
Middleware for Grid Computing and the relationship to Middleware at large ECE 1770 : Middleware Systems By: Sepehr (Sep) Seyedi Date: Thurs. January 23,
Ames Research CenterDivision 1 Information Power Grid (IPG) Overview Anthony Lisotta Computer Sciences Corporation NASA Ames May 2,
9 Systems Analysis and Design in a Changing World, Fourth Edition.
NA-MIC National Alliance for Medical Image Computing UCSD: Engineering Core 2 Portal and Grid Infrastructure.
GO-ESSP Workshop, LLNL, Livermore, CA, Jun 19-21, 2006, Center for ATmosphere sciences and Earthquake Researches Construction of e-science Environment.
What is SAM-Grid? Job Handling Data Handling Monitoring and Information.
GRIDS Center Middleware Overview Sandra Redman Information Technology and Systems Center and Information Technology Research Center National Space Science.
GRID Overview Internet2 Member Meeting Spring 2003 Sandra Redman Information Technology and Systems Center and Information Technology Research Center National.
Grid Execution Management for Legacy Code Applications Grid Enabling Legacy Applications.
Kraków Kick-off meeting Migrating Desktop General concept Intuitive Grid-user’s work environment independent of a hardware.
Introduction to Grids By: Fetahi Z. Wuhib [CSD2004-Team19]
7. Grid Computing Systems and Resource Management
CERN, DataGrid PTB, April 10, 2002 CrossGrid – DataGrid Collaboration (Framework) Marian Bubak and Bob Jones.
EC Review – 01/03/2002 – WP9 – Earth Observation Applications – n° 1 WP9 Earth Observation Applications 1st Annual Review Report to the EU ESA, KNMI, IPSL,
High Level Architecture (HLA)  used for building interactive simulations  connects geographically distributed nodes  time management (for time- and.
Ariel Garcia DataGrid WP6, Heidelberg, 26 th September 2003 Ariel García CrossGrid testbed status Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft.
BOF at GGF5, Edinburgh, Scotland, July 21-24, 2002 CrossGrid Architecture Marian Bubak and TAT Institute of Computer Science & ACC CYFRONET AGH, Cracow,
Marian Bubak 1,2, Włodzimierz Funika 1,2, Roland Wismüller 3, Tomasz Arodź 1,2, Marcin Kurdziel 1,2 1 Institute of Computer Science, AGH, Kraków, Poland.
PROGRESS: GEW'2003 Using Resources of Multiple Grids with the Grid Service Provider Michał Kosiedowski.
Migrating Desktop Uniform Access to the Grid Marcin Płóciennik Poznan Supercomputing and Networking Center Poznan, Poland EGEE’07, Budapest, Oct.
INFSO-RI JRA2 Test Management Tools Eva Takacs (4D SOFT) ETICS 2 Final Review Brussels - 11 May 2010.
Grid Activities in CMS Asad Samar (Caltech) PPDG meeting, Argonne July 13-14, 2000.
Migrating Desktop Uniform Access to the Grid Marcin Płóciennik Poznan Supercomputing and Networking Center Poland EGEE’08 Conference, Istanbul, 24 Sep.
Grid Execution Management for Legacy Code Architecture Exposing legacy applications as Grid services: the GEMLCA approach Centre.
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI EGI Services for Distributed e-Infrastructure Access Tiziana Ferrari on behalf.
Workload Management Workpackage
EO Applications Parallel Session
Grid Computing.
PROCESS - H2020 Project Work Package WP6 JRA3
Report on GLUE activities 5th EU-DataGRID Conference
Wide Area Workload Management Work Package DATAGRID project
Grid Application Programming Environment
Presentation transcript:

TAT CrossGrid Yearly Review, Brussels, March 12, 2003 CrossGrid After the First Year: A Technical Overview Marian Bubak, Maciej Malawski, and Katarzyna Zając X# TAT Institute of Computer Science & ACC CYFRONET AGH, Kraków, Poland

TAT CrossGrid Yearly Review, Brussels, March 12, 2003 Main Objectives  A new category of Grid-enabled applications Compute- and data-intensive distributed near real-time response (person in a loop) layered  New programming tools  Grid more user-friendly, secure and efficient  Interoperability with other Grids  Implementation of standards

TAT CrossGrid Yearly Review, Brussels, March 12, 2003 CrossGrid in a Nutshell Interactive, Compute and Data Intensive Applications  I nteractive simulation and visualization of a biomedical system  Flooding crisis team support  Distributed data analysis in HEP  Weather forecasting and air pollution modeling Tool Environment  MPI code debugging and verification  Metrics and benchmarks  Interactive and semiautomatic performance evaluation tools New Generic Grid Services Globus Middleware Fabric DataGrid Services  Portals and roaming access  Scheduling agents  Application and Grid monitoring  Optimization of data access Application Specific Services  User Interactive Services  Grid Visualization Kernel

TAT CrossGrid Yearly Review, Brussels, March 12, 2003 Key Features of CG Applications  Data Data generators and databases geographically distributed Selected on demand  Processing Interactive Requires large processing capacity; both HPC & HTC  Presentation Complex data requires versatile 3D visualisation Support interaction and feedback to other components

TAT CrossGrid Yearly Review, Brussels, March 12, 2003 Biomedical Application  Adding small modifications to the proposed structure results in immediate changes in the blood flow.  Online presentation of simulation results via a 3D environment.  The progress of the simulation and the estimated time of convergence should be available for inspection. LB flow simulation VE WD PC PDA Visualization Interaction

TAT CrossGrid Yearly Review, Brussels, March 12, 2003 Basic Characteristics of Flood Simulation  Meteorological Intensive simulation (HPC), large input/output data sets, high availability of resources  Hydrological Parametric simulations (HTC) may require different models (heterogeneous simulations)  Hydraulic Many 1-D simulations HTC, 2-D hydraulic simulations require HPC Data sources Meteorological simulations Hydraulic simulations Hydrological simulations Users Output visualization

TAT CrossGrid Yearly Review, Brussels, March 12, 2003 Distributed Data Analysis in HEP  Objectives Distributed data access Distributed data mining techniques with neural networks  Issues Typical interactive requests will run on o(TB) of distributed data Transfer/replication times for the whole data on the order of one hour Data transfers once and in advance of the interactive session. Allocation, installation and setup the corresponding database servers before the interactive session starts Replica Manager Interactive Session Resource Broker DISTRIBUTED PROCESSING DBInstallation Interactive Session Database server Interactive Session Manager Interactive Session Worker Interactive Session Worker Interactive Session Worker Interactive Session Worker Interactive Session Worker Portal XML in/out On-line output

TAT CrossGrid Yearly Review, Brussels, March 12, 2003 Weather Forecasting and Air Pollution Modeling  Distributed/parallel code on Grid Coupled Ocean/Atmosphere Mesoscale Prediction System STEM-II Air Pollution Code Integration of distributed databases  Data mining applied to downscaling weather forecasts

TAT CrossGrid Yearly Review, Brussels, March 12, 2003 Supporting Tools 1.4 Meteo Pollution 1.4 Meteo Pollution 3.1 Portal & Migrating Desktop Applications Development Support 2.4 Performance Analysis 2.4 Performance Analysis 2.2 MPI Verification 2.3 Metrics and Benchmarks 2.3 Metrics and Benchmarks App. Spec Services 1.1 Grid Visualisation Kernel 1.3 Data Mining on Grid (NN) 1.3 Data Mining on Grid (NN) 1.3 Interactive Distributed Data Access 3.1 Roaming Access 3.1 Roaming Access 3.2 Scheduling Agents 3.2 Scheduling Agents 3.3 Grid Monitoring 3.3 Grid Monitoring MPICH-G Fabric 1.1, 1.2 HLA and others 3.4 Optimization of Grid Data Access 3.4 Optimization of Grid Data Access 1.2 Flooding 1.2 Flooding 1.1 BioMed 1.1 BioMed Applications Generic Services 1.3 Interactive Session Services 1.3 Interactive Session Services GRAM GSI Replica Catalog GIS / MDS GridFTP Globus-IO DataGrid Replica Manager DataGrid Replica Manager DataGrid Job Submission Service Resource Manager (CE) Resource Manager (CE) CPU Resource Manager Resource Manager Resource Manager (SE) Resource Manager (SE) Secondary Storage Resource Manager Resource Manager Instruments ( Satelites, Radars) Instruments ( Satelites, Radars) 3.4 Optimization of Local Data Access 3.4 Optimization of Local Data Access Tertiary Storage Replica Catalog Globus Replica Manager Globus Replica Manager 1.1 User Interaction Services 1.1 User Interaction Services Initial version of X# architecture

TAT CrossGrid Yearly Review, Brussels, March 12, 2003 Project Phases M 1 - 3: requirements definition and merging M : first development phase: design, 1st prototypes, refinement of requirements M : second development phase: integration of components, 2nd prototypes M : third development phase: complete integration, final code versions M : final phase: demonstration and documentation

TAT CrossGrid Yearly Review, Brussels, March 12, 2003 Tools  MPI code debugging and verification  Metrics and benchmarks for the Grid environment  Grid-enabled Performance Measurement  Performance Prediction Component Grid Monitoring Performance Prediction Component High Level Analysis Component User Interface and Visualization Component Performance Measurement Component Applications executing on Grid testbed Application source code G-PM RMD PMD MPI Verification MARMOT Benchmarks

TAT CrossGrid Yearly Review, Brussels, March 12, 2003 MPI Verification  verifies the correctness of parallel, distributed Grid applications (MPI)  technical basis: MPI profiling interface which allows a detailed analysis of the MPI application Core Tool Application or Test Tool MPI Additional Process (Debug Server) Client Side Profiling Interface Server Side

TAT CrossGrid Yearly Review, Brussels, March 12, 2003 Benchmark Categories  Micro-benchmarks For identifying basic performance properties of Grid services, sites, and constellations  Micro-kernels Generic HPC/HTC kernels, including general and often- used kernels in Grid environments  Application kernels Characteristic of representative CG applications Portal gbView gbARC gbControl gbRMP Grid Bench suite SE storage Embedding Retrieval Invocation Invocation/ Collection through GPM Direct Invocation Storage/ Retrieval

TAT CrossGrid Yearly Review, Brussels, March 12, 2003 Performance Measurement Tool G-PM  Components: performance measurement component (PMC), component for high-level analysis (HLAC), component for performance prediction (PPC) based on analytical performance models of application kernels, user interface and visualization component UIVC. Interface Measurement OCM-G Interface UIVC HLAC OCM-G PMC

TAT CrossGrid Yearly Review, Brussels, March 12, 2003 User Interactive Service Interaction GidService RTIExec GridService Simulation GridService Registry OGSA WSDL RTI Tuple Space functionality description +Dynamic discovery of OGSA Services Large On-line Data transferShort Messages and Events GridFTPSOAP/IIOP TCP or UDP/IP Visualisation GridService  enables end users to run distributed simulations in the Grid environment and to steer those simulations in near real time  uses OGSA mechanisms to call external resource brokers, job submission services (efficient and transparent execution of the simulation on the Grid).

TAT CrossGrid Yearly Review, Brussels, March 12, 2003 Grid Visualization Kernel  addresses the problems of distributed visualization on heterogeneous devices  allows easily and transparently interconnect Grid applications with existing visualisation tools (AVS, OpenDX, VTK,...)  handles multiple concurrent input data streams  multiplexes compressed data and images efficiently across long- distance networks GVK Portal Server GVK Visualization Planner Simulation Init Visualization Update Visualization GRAMGASSMDS GVK Visualization pipeline Simulation Data

TAT CrossGrid Yearly Review, Brussels, March 12, 2003 New Grid Services  Portals and roaming access  Grid resource management  Grid monitoring  Optimization of data access

TAT CrossGrid Yearly Review, Brussels, March 12, 2003 Roaming Access – Current Design Web Browser Line Roaming Access Server LDAP DataBase Manager Application Portal Server Desktop Portal Server Web Browser Benchmarks  Portal - easier access and use of the Grid by applications  Migrating Desktop - a transparent, independent user environment  Roaming Access Server - responsible for managing user profiles, job submission, file transfers and Grid monitoring Command Replica Agent Scheduling

TAT CrossGrid Yearly Review, Brussels, March 12, 2003 Scheduling Agents - Current Design JSS commands Job monitoring Scheduling Agent Web Portal JSS / CondorG Resource list Logging & Bookkeping Resource Broker CE  scheduling user jobs over the CrossGrid testbed infrastructure,  submition based on Condor-G,  support for sequential and MPI parallel jobs, batch jobs and interactive jobs,  priorities and preferences determined by the user for each job

TAT CrossGrid Yearly Review, Brussels, March 12, 2003 Application Monitoring  OCM-G Components Service Managers Local Monitors  Application processes  Tool(s)  External name service Component discovery ServiceManager LocalMonitor Tool SharedMemory OMIS ExternalLocalization ApplicationProcess

TAT CrossGrid Yearly Review, Brussels, March 12, 2003 Jiro Services Information DB System Non-invasive Monitoring MDS Globus MDS info Jiro info Instruments Infrastructure Static info Performance Information Post-processing  Infrastructure monitoring Invasive monitoring (based on Jiro technology) Non-invasive monitoring (Santa-G) Infrastructure Monitoring

TAT CrossGrid Yearly Review, Brussels, March 12, 2003  Selection of specialized components best suited for data access operations  Estimation of data access latency and bandwidth inside the storage elements  Faster access to large tape-resident through fragmentation Data Access Design

TAT CrossGrid Yearly Review, Brussels, March 12, 2003 Generic Services Application Specific Services Supporting Tools Applications Current status of CG Architecture

TAT CrossGrid Yearly Review, Brussels, March 12, 2003 Application-centric view

TAT CrossGrid Yearly Review, Brussels, March 12, 2003 The Current Testbed  The current CrossGrid testbed is based on: EDG distribution release and (production) EDG distribution release (validation)  The current infrastructure permits: installation of initial prototypes of CrossGrid software releases (described in M12 Deliverables) testing applications using: Globus and EDG middleware MPI achieving compatibility with DataGrid and therefore extending Grid coverage in Europe

TAT CrossGrid Yearly Review, Brussels, March 12, 2003 Grid Service  Transient, stateful Web Service (created dynamically)  Described by WSDL  Identified by Grid Service Handle (GSH) in the form of URI  Can be queried for configuration and state in standard way – Service Data mechanism

TAT CrossGrid Yearly Review, Brussels, March 12, 2003 Why use OGSA  Standards  „to be part of the Grid = to implement OGSA Grid protocols”  Interoperability in heterogeneous environments  Possible contribution to future Grid activities

TAT CrossGrid Yearly Review, Brussels, March 12, 2003 Grid Services – where?  Dynamic service creation and lifetime management to control the state of some process, e.g.: user session in a portal data transfer running simulation.  Service data model can be applied to monitoring systems that can be used as information providers for other services.  Service discovery – to solve the bootstrap problem: to connect the modules of a distributed simulation to connect the application to a monitoring system

TAT CrossGrid Yearly Review, Brussels, March 12, 2003 Steps towards OGSA  Using Web Service interfaces and XML where possible  Experimenting with prototyping services using OGSA alpha releases  Applying Grid Service extensions to services  Solving GT2 - GT3 transition and compatibility issues

TAT CrossGrid Yearly Review, Brussels, March 12, 2003 Summary  Achievements of the first project year : Software Requirements Specifications together with use cases written CrossGrid Architecture defined Detailed Design documents for tools and new Grid services (OO approach, UML) written First prototype of software running and documented Detailed description of the test and integration procedures created Testbed set up