GriPhyN Status and Project Plan Mike Wilde Mathematics and Computer Science Division Argonne National Laboratory.

Slides:



Advertisements
Similar presentations
Virtual Data and the Chimera System* Ian Foster Mathematics and Computer Science Division Argonne National Laboratory and Department of Computer Science.
Advertisements

A. Arbree, P. Avery, D. Bourilkov, R. Cavanaugh, S. Katageri, G. Graham, J. Rodriguez, J. Voeckler, M. Wilde CMS & GriPhyN Conference in High Energy Physics,
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
USING THE GLOBUS TOOLKIT This summary by: Asad Samar / CALTECH/CMS Ben Segal / CERN-IT FULL INFO AT:
The DOE Science Grid Computing and Data Infrastructure for Large-Scale Science William Johnston, Lawrence Berkeley National Lab Ray Bair, Pacific Northwest.
Presented by Scalable Systems Software Project Al Geist Computer Science Research Group Computer Science and Mathematics Division Research supported by.
Workload Management Workpackage Massimo Sgaravatto INFN Padova.
6th Biennial Ptolemy Miniconference Berkeley, CA May 12, 2005 Distributed Computing in Kepler Ilkay Altintas Lead, Scientific Workflow Automation Technologies.
Robust Tools for Archiving and Preserving Digital Data Joseph JaJa, Mike Smorul, and Mike McGann Institute for Advanced Computer Studies Department of.
Milos Kobliha Alejandro Cimadevilla Luis de Alba Parallel Computing Seminar GROUP 12.
Pegasus: Mapping complex applications onto the Grid Ewa Deelman Center for Grid Technologies USC Information Sciences Institute.
Workload Management Massimo Sgaravatto INFN Padova.
The Software Product Life Cycle. Views of the Software Product Life Cycle  Management  Software engineering  Engineering design  Architectural design.
Assessment of Core Services provided to USLHC by OSG.
Miron Livny Computer Sciences Department University of Wisconsin-Madison From Compute Intensive to Data.
Grappa: Grid access portal for physics applications Shava Smallen Extreme! Computing Laboratory Department of Physics Indiana University.
XCAT Science Portal Status & Future Work July 15, 2002 Shava Smallen Extreme! Computing Laboratory Indiana University.
10/20/05 LIGO Scientific Collaboration 1 LIGO Data Grid: Making it Go Scott Koranda University of Wisconsin-Milwaukee.
HEP Experiment Integration within GriPhyN/PPDG/iVDGL Rick Cavanaugh University of Florida DataTAG/WP4 Meeting 23 May, 2002.
Long Term Ecological Research Network Information System LTER Grid Pilot Study LTER Information Manager’s Meeting Montreal, Canada 4-7 August 2005 Mark.
ARGONNE  CHICAGO Ian Foster Discussion Points l Maintaining the right balance between research and development l Maintaining focus vs. accepting broader.
CoG Kit Overview Gregor von Laszewski Keith Jackson.
1 School of Computer, National University of Defense Technology A Profile on the Grid Data Engine (GridDaEn) Xiao Nong
INFSO-RI Enabling Grids for E-sciencE The US Federation Miron Livny Computer Sciences Department University of Wisconsin – Madison.
Miguel Branco CERN/University of Southampton Enabling provenance on large-scale e-Science applications.
Grid Workload Management & Condor Massimo Sgaravatto INFN Padova.
Scalable Systems Software Center Resource Management and Accounting Working Group Face-to-Face Meeting October 10-11, 2002.
Programming Models & Runtime Systems Breakout Report MICS PI Meeting, June 27, 2002.
K. De UTA Grid Workshop April 2002 U.S. ATLAS Grid Testbed Workshop at UTA Introduction and Goals Kaushik De University of Texas at Arlington.
Grid Status - PPDG / Magda / pacman Torre Wenaus BNL U.S. ATLAS Physics and Computing Advisory Panel Review Argonne National Laboratory Oct 30, 2001.
PPDG and ATLAS Particle Physics Data Grid Ed May - ANL ATLAS Software Week LBNL May 12, 2000.
Miron Livny Computer Sciences Department University of Wisconsin-Madison Welcome and Condor Project Overview.
Ruth Pordes, Fermilab CD, and A PPDG Coordinator Some Aspects of The Particle Physics Data Grid Collaboratory Pilot (PPDG) and The Grid Physics Network.
Grid Workload Management Massimo Sgaravatto INFN Padova.
Pegasus: Mapping Scientific Workflows onto the Grid Ewa Deelman Center for Grid Technologies USC Information Sciences Institute.
Data Grid projects in HENP R. Pordes, Fermilab Many HENP projects are working on the infrastructure for global distributed simulated data production, data.
Virtual Data Grid Architecture Ewa Deelman, Ian Foster, Carl Kesselman, Miron Livny.
Perspectives on Grid Technology Ian Foster Argonne National Laboratory The University of Chicago.
Ames Research CenterDivision 1 Information Power Grid (IPG) Overview Anthony Lisotta Computer Sciences Corporation NASA Ames May 2,
Virtual Batch Queues A Service Oriented View of “The Fabric” Rich Baker Brookhaven National Laboratory April 4, 2002.
NA-MIC National Alliance for Medical Image Computing UCSD: Engineering Core 2 Portal and Grid Infrastructure.
Pegasus: Mapping complex applications onto the Grid Ewa Deelman Center for Grid Technologies USC Information Sciences Institute.
What is SAM-Grid? Job Handling Data Handling Monitoring and Information.
Atlas Grid Status - part 1 Jennifer Schopf ANL U.S. ATLAS Physics and Computing Advisory Panel Review Argonne National Laboratory Oct 30, 2001.
The GriPhyN Planning Process All-Hands Meeting ISI 15 October 2001.
GRIDS Center Middleware Overview Sandra Redman Information Technology and Systems Center and Information Technology Research Center National Space Science.
GRID Overview Internet2 Member Meeting Spring 2003 Sandra Redman Information Technology and Systems Center and Information Technology Research Center National.
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
DGC Paris WP2 Summary of Discussions and Plans Peter Z. Kunszt And the WP2 team.
AliEn AliEn at OSC The ALICE distributed computing environment by Bjørn S. Nilsen The Ohio State University.
Alain Roy Computer Sciences Department University of Wisconsin-Madison Condor & Middleware: NMI & VDT.
Planning Ewa Deelman USC Information Sciences Institute GriPhyN NSF Project Review January 2003 Chicago.
Internet2 Applications Group: Renater Group Presentation T. Charles Yun Internet2 Program Manager, Applications Group 30 October 2001.
Virtual Data Management for CMS Simulation Production A GriPhyN Prototype.
Securing the Grid & other Middleware Challenges Ian Foster Mathematics and Computer Science Division Argonne National Laboratory and Department of Computer.
GriPhyN Project Paul Avery, University of Florida, Ian Foster, University of Chicago NSF Grant ITR Research Objectives Significant Results Approach.
GriPhyN Management Mike Wilde University of Chicago, Argonne Paul Avery University of Florida GriPhyN NSF Project.
U.S. Grid Projects and Involvement in EGEE Ian Foster Argonne National Laboratory University of Chicago EGEE-LHC Town Meeting,
Parag Mhashilkar Computing Division, Fermi National Accelerator Laboratory.
April 25, 2006Parag Mhashilkar, Fermilab1 Resource Selection in OSG & SAM-On-The-Fly Parag Mhashilkar Fermi National Accelerator Laboratory Condor Week.
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
1 CMS Virtual Data Overview Koen Holtman Caltech/CMS GriPhyN all-hands meeting, Marina del Rey April 9, 2001.
Current Globus Developments Jennifer Schopf, ANL.
1 Open Science Grid: Project Statement & Vision Transform compute and data intensive science through a cross- domain self-managed national distributed.
Bob Jones EGEE Technical Director
Pegasus and Condor Gaurang Mehta, Ewa Deelman, Carl Kesselman, Karan Vahi Center For Grid Technologies USC/ISI.
Leigh Grundhoefer Indiana University
Wide Area Workload Management Work Package DATAGRID project
Status of Grids for HEP and HENP
Presentation transcript:

GriPhyN Status and Project Plan Mike Wilde Mathematics and Computer Science Division Argonne National Laboratory

Mike Wilde Planning Goals l Clarify our vision and direction u Know how we can make a difference in science and computing! l Map that vision to each experiment u Design concrete instances of our vision l Coordinate our research programs l Shape toolkit to challenge-problem needs l Coordinate overlapping technologies l Organize as coordinated subteams with specific missions and defined points of interaction

Argonne National LaboratoryMike Wilde Project Approach CS Research VDT Development Application Analysis Infrastructure Development and Deployment Challenge Problem Identification Challenge Problem Solution Development Challenge Problem Solution Integration VDT Development VDT Development Infrastructure Deployment IS Deployment time Process

Argonne National LaboratoryMike Wilde Project Activities l Research u Experiment Analysis l Use cases, statistics, distributions, data flow patterns, tools, data types, HIPO u Vision Refinement u Attacking the “hard problems” l Virtual data identification and manipulation l Advanced resource allocation and execution planning l Scaling this up to Petascale u Architectural Refinement l Toolkit Development l Integration u Identify and Address Challenge Problems u Testbed construction l Support l Evaluation

Argonne National LaboratoryMike Wilde Research Milestone Highlights Y1:Execution framework Virtual data prototypes Y2:Virtual data catalog w/glue language Integ w/ scalable replica catalog service Initial resource usage policy language Y3:Advanced planning, fault recovery Intelligent catalog Advanced policy languages Y4:Knowledge management and location Y5: Transparency and usability Scalability and manageability

Argonne National LaboratoryMike Wilde Research Leadership Centers l Virtual Data: u Chicago (VDC, VDL, KR), ISI (Schema) u Wisconsin (NeST), SDSC (MCAT,SRB) l Request Planning u ISI (algorithms), Chicago (policy), Berkeley (query optimization) l Request Execution u Wisconsin l Fault Tolerance u SDSC l Monitoring u Northwestern l User Interface u Indiana

Argonne National LaboratoryMike Wilde Project Status Overview l Year 1 research fruitful virtual data, planning, execution, integration– demonstrated at SC2001 l Research efforts launched 80% focused – 20% exploratory l VDT effort staffed and launched Yearly major release; VDT1 close; VDT2 planned; VDT3-5 envisioned l Year 2 experiment integrations high level plans done; detailed planning underway l Long term vision refined and unified

Argonne National LaboratoryMike Wilde Milestones: Architecture l Early 2002: u Specify interfaces for new GriPhyN functional modules l Request Planner l Virtual Data Catalog service l Monitoring service u Define how we will connect and integrate our solutions, e.g.: l Virtual data language l Multiple-catalog integration l DAGman graphs l Policy langauge l CAS interaction for policy lookup and enforcement l Year-end 2002: phased migration to a web- services based architecture

Argonne National LaboratoryMike Wilde Status: Virtual Data l Virtual Data u First version of a catalog structure built u Integration language “VDL” developed u Detailed transformation model designed l Replica location service at Chicago & ISI u highly scalable and fault tolerant u soft-state distributed architecture l NeSt at UW u Storage appliance for Condor u treats data transfer as a job step

Argonne National LaboratoryMike Wilde Milestones: Virtual Data l Year 2: u Local Virtual Data Catalog Structures (relational) u Catalog manipulation language (VDL) u Linkage to application metadata l Year 3: Handling multi-modal virtual data u Distributed virtual data catalogs (based on RLS) u advanced transformation signatures u Flat, objects, OODBs, relational u Cross-modal depdendency tracking l Year 4: Knowledge representation u ontologies; data generation paradigms u Fuzzy dependencies and data equivalence l Year 5: Finalize Scalability and Manageability

Argonne National LaboratoryMike Wilde Status: Planning and Execution l Planning and Execution u Major strides in execution environment made with Condor, CondorG, and DAGman u DAGs evolving as pervasive job specification model with the virtual data grid u Large-scale CMS production demonstarted on international wide-area multi-org grid u LIGO demonstrated full GriPhyN integration u Sophisticated policy language for grid-wide resource sharing under design at Chicago u Knowledge representation research underway at Chicago u Research in ClassAds explored in Globus context l Master/worker fault tolerance at UCSD u Design proposed to extend fault tolerance of Condor masters

Argonne National LaboratoryMike Wilde Milestones: Request Planning l Year 2: u Protype planner as a grid service module u Intial CAS and Policy Language Integration u Refinement of DAG language with data flow info l Year 3: u Policy enhancements: dynamic replanning (based on Grid monitoring), cost alternatives and optimizations l Year 4: u Global planning with policy constraints l Year 5: u Incremental global planning u Algorithms evaluated, tuned w/ large-scale simulations

Argonne National LaboratoryMike Wilde Milestones: Request Execution l Year 2: u Request Planning and Execution l Striving for increasingly greater resource leverage with increasing both power AND transparency l Fault tolerance – keeping it all running! u Intial CAS and Policy Language Integration u Refinement of DAG language with data flow info u Resource utiization monitoring to drive planner l Year 3: u Resource co-allocation with recovery u Fault tolerant execution engines l Year 4: u Execution adapts to grid resource availability changes l Year 5: u Simulation-based algorithm eval and tuning

Argonne National LaboratoryMike Wilde Status: Supporting Research l Joint PPDG-GriPhyN Monitoring group u Meeting regularly u Use-case development underway l Research into monitoring, measurement, profiling, and performance predication u Underway at NU and ANL l GRIPE facility for Grid-wide user and host certificate and login management l GRAPPA portal for end-user science access

Argonne National LaboratoryMike Wilde Status – Experiments l ATLAS u 8-site testgrid in place u Year-2 plan well refined l CMS u Working prototypes of production and distributed analysis, both with virtual data u Year-2 plan – simulation production – underway l LIGO u Working prototypes of full VDG demonstrated u Year-2 plan well refined and development underway l SDSS u Year-2 plan well refined u Challenge problem development underway; close collaboration with Chicago on VDC

Argonne National LaboratoryMike Wilde Year 2 Plan: ATLAS

Argonne National LaboratoryMike Wilde Year 2 Plan: CMS Plans

Argonne National LaboratoryMike Wilde Year 2 Plan: LIGO Plans

Argonne National LaboratoryMike Wilde Year 2 Plan: SDSS Plans