GriPhyN Management Mike Wilde University of Chicago, Argonne Paul Avery University of Florida GriPhyN NSF Project.

Slides:



Advertisements
Similar presentations
Virtual Data and the Chimera System* Ian Foster Mathematics and Computer Science Division Argonne National Laboratory and Department of Computer Science.
Advertisements

A conceptual model of grid resources and services Authors: Sergio Andreozzi Massimo Sgaravatto Cristina Vistoli Presenter: Sergio Andreozzi INFN-CNAF Bologna.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
Workload Management Workpackage Massimo Sgaravatto INFN Padova.
The LHC Computing Grid Project Tomi Kauppi Timo Larjo.
Knowledge Environments for Science: Representative Projects Ian Foster Argonne National Laboratory University of Chicago
The Grid as Infrastructure and Application Enabler Ian Foster Mathematics and Computer Science Division Argonne National Laboratory and Department of Computer.
Miron Livny Computer Sciences Department University of Wisconsin-Madison From Compute Intensive to Data.
Overview of the GLUE Project (Grid Laboratory Unified Environment) Author: Piotr Nowakowski, M.Sc. Cyfronet, Kraków.
GriPhyN EAC Meeting (Jan. 7, 2002)Paul Avery1 University of Florida GriPhyN External Advisory Committee.
LCG Milestones for Deployment, Fabric, & Grid Technology Ian Bird LCG Deployment Area Manager PEB 3-Dec-2002.
Grappa: Grid access portal for physics applications Shava Smallen Extreme! Computing Laboratory Department of Physics Indiana University.
XCAT Science Portal Status & Future Work July 15, 2002 Shava Smallen Extreme! Computing Laboratory Indiana University.
Experiment Requirements for Global Infostructure Irwin Gaines FNAL/DOE.
HEP Experiment Integration within GriPhyN/PPDG/iVDGL Rick Cavanaugh University of Florida DataTAG/WP4 Meeting 23 May, 2002.
ARGONNE  CHICAGO Ian Foster Discussion Points l Maintaining the right balance between research and development l Maintaining focus vs. accepting broader.
INFSO-RI Enabling Grids for E-sciencE The US Federation Miron Livny Computer Sciences Department University of Wisconsin – Madison.
LCG and HEPiX Ian Bird LCG Project - CERN HEPiX - FNAL 25-Oct-2002.
Grid Workload Management & Condor Massimo Sgaravatto INFN Padova.
K. De UTA Grid Workshop April 2002 U.S. ATLAS Grid Testbed Workshop at UTA Introduction and Goals Kaushik De University of Texas at Arlington.
Jarek Nabrzyski, Ariel Oleksiak Comparison of Grid Middleware in European Grid Projects Jarek Nabrzyski, Ariel Oleksiak Poznań Supercomputing and Networking.
Grid Status - PPDG / Magda / pacman Torre Wenaus BNL U.S. ATLAS Physics and Computing Advisory Panel Review Argonne National Laboratory Oct 30, 2001.
PPDG and ATLAS Particle Physics Data Grid Ed May - ANL ATLAS Software Week LBNL May 12, 2000.
LIGO- G Z Planning Meeting (Dec 2002)LSC Member Institution (UT Brownsville) 1 Manuela Campanelli The University of Texas at Brownsville
GriPhyN Status and Project Plan Mike Wilde Mathematics and Computer Science Division Argonne National Laboratory.
Miron Livny Computer Sciences Department University of Wisconsin-Madison Welcome and Condor Project Overview.
Ruth Pordes, Fermilab CD, and A PPDG Coordinator Some Aspects of The Particle Physics Data Grid Collaboratory Pilot (PPDG) and The Grid Physics Network.
Pegasus: Mapping Scientific Workflows onto the Grid Ewa Deelman Center for Grid Technologies USC Information Sciences Institute.
Data Grid projects in HENP R. Pordes, Fermilab Many HENP projects are working on the infrastructure for global distributed simulated data production, data.
10/24/2015OSG at CANS1 Open Science Grid Ruth Pordes Fermilab
Virtual Data Grid Architecture Ewa Deelman, Ian Foster, Carl Kesselman, Miron Livny.
José D. Zamora, Sean R. Morriss and Manuela Campanelli.
GriPhyN EAC Meeting (Jan. 7, 2002)Carl Kesselman1 University of Southern California GriPhyN External Advisory Committee Meeting Gainesville,
GriPhyN EAC Meeting (Apr. 12, 2001)Paul Avery1 University of Florida Opening and Overview GriPhyN External.
US ATLAS Grid Projects Rob Gardner Indiana University Mid Year Review of US ATLAS Computing NSF Headquarters, Arlington VA June 20, 2002
October LHCUSA meeting BNL Bjørn S. Nilsen Update on NSF-ITR Proposal Bjørn S. Nilsen The Ohio State University.
SA1/SA2 meeting 28 November The status of EGEE project and next steps Bob Jones EGEE Technical Director EGEE is proposed as.
Authors: Ronnie Julio Cole David
GriPhyN Project Overview Paul Avery University of Florida GriPhyN NSF Project Review January 2003 Chicago.
Atlas Grid Status - part 1 Jennifer Schopf ANL U.S. ATLAS Physics and Computing Advisory Panel Review Argonne National Laboratory Oct 30, 2001.
The GriPhyN Planning Process All-Hands Meeting ISI 15 October 2001.
Middleware Camp NMI (NSF Middleware Initiative) Program Director Alan Blatecky Advanced Networking Infrastructure and Research.
LIGO- G Z EAC Meeting (Jan 2003)LSC Member Institution (UT Brownsville) 1 Manuela Campanelli The University of Texas at Brownsville
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
The Grid Effort at UF Presented by Craig Prescott.
The Particle Physics Data Grid Collaboratory Pilot Richard P. Mount For the PPDG Collaboration DOE SciDAC PI Meeting January 15, 2002.
High Energy Physics and Grids at UF (Dec. 13, 2002)Paul Avery1 University of Florida High Energy Physics.
GriPhyN EAC Meeting (Jan. 7, 2002)Paul Avery1 Integration with iVDGL è International Virtual-Data Grid Laboratory  A global Grid laboratory (US, EU, Asia,
Open Science Grid & its Security Technical Group ESCC22 Jul 2004 Bob Cowles
Alain Roy Computer Sciences Department University of Wisconsin-Madison Condor & Middleware: NMI & VDT.
Planning Ewa Deelman USC Information Sciences Institute GriPhyN NSF Project Review January 2003 Chicago.
Virtual Data Management for CMS Simulation Production A GriPhyN Prototype.
GriPhyN Project Paul Avery, University of Florida, Ian Foster, University of Chicago NSF Grant ITR Research Objectives Significant Results Approach.
U.S. Grid Projects and Involvement in EGEE Ian Foster Argonne National Laboratory University of Chicago EGEE-LHC Town Meeting,
1 Grid Projects: Overview, Status, and Plans LCG Launching Workshop 11 March 2002 Larry Price Argonne National Laboratory.
12 March, 2002 LCG Applications Area - Introduction slide 1 LCG Applications Session LCG Launch Workshop March 12, 2002 John Harvey, CERN LHCb Computing.
G G riPhyN Project Review Criteria l Relevance to Information Technology l Intellectual Merit l Broader Impacts l ITR Evaluation Criteria (innovation in.
Management & Coordination Paul Avery, Rick Cavanaugh University of Florida Ian Foster, Mike Wilde University of Chicago, Argonne
EGEE Project Review Fabrizio Gagliardi EDG-7 30 September 2003 EGEE is proposed as a project funded by the European Union under contract IST
LHC Computing, SPC-FC-CC-C; H F Hoffmann1 CERN/2379/Rev: Proposal for building the LHC computing environment at CERN (Phase 1) Goals of Phase.
1 CMS Virtual Data Overview Koen Holtman Caltech/CMS GriPhyN all-hands meeting, Marina del Rey April 9, 2001.
Planning Session. ATLAS(-CMS) End-to-End Demo Kaushik De is the Demo Czar Need to put team together Atlfast production jobs –Atlfast may be unstable over.
Towards deploying a production interoperable Grid Infrastructure in the U.S. Vicky White U.S. Representative to GDB.
12 March, 2002 LCG Applications Area - Introduction slide 1 LCG Applications Session LCG Launch Workshop March 12, 2002 John Harvey, CERN LHCb Computing.
1 Open Science Grid: Project Statement & Vision Transform compute and data intensive science through a cross- domain self-managed national distributed.
10-Feb-00 CERN HepCCC Grid Initiative ATLAS meeting – 16 February 2000 Les Robertson CERN/IT.
Bob Jones EGEE Technical Director
Ian Bird LCG Project - CERN HEPiX - FNAL 25-Oct-2002
Report on GLUE activities 5th EU-DataGRID Conference
Status of Grids for HEP and HENP
Presentation transcript:

GriPhyN Management Mike Wilde University of Chicago, Argonne Paul Avery University of Florida GriPhyN NSF Project Review January 2003 Chicago

229 Jan 2003 Mike Wilde, University of Chicago GriPhyN Management Management –Paul Avery (Florida)co-Director –Ian Foster (Chicago)co-Director –Mike Wilde (Argonne)Project Coordinator –Rick Cavanaugh (Florida)Deputy Coordinator

329 Jan 2003 Mike Wilde, University of Chicago External Advisory Committee Physics Experiments Project Directors Paul Avery Ian Foster Internet 2DOE Science NSF PACIs Project Coordination Mike Wilde Rick Cavanaugh Outreach/Education Manuela Campanelli Industrial Connections Ian Foster / Paul Avery EDG, LCG, Other Grid Projects Architecture Carl Kesselman VDT Development Coord.: M. Livny Requirements, Definition & Scheduling (Miron Livny) Integration, Testing, Documentation, Support (Alain Roy) Globus Project & NMI Integration (Carl Kesselman) CS Research Coord.: I. Foster Virtual Data (Mike Wilde) Request Planning & Scheduling (Ewa Deelman) Execution Management (Miron Livny) Measurement, Monitoring & Prediction (Valerie Taylor) Applications Coord.: R. Cavanaugh ATLAS (Rob Gardner) CMS (Rick Cavanaugh) LIGO (Albert Lazzarini) SDSS (Alexander Szalay) Inter-Project Coordination: R. Pordes HICB (Larry Price) HIJTB (Carl Kesselman) PPDG (Ruth Pordes) TeraGrid, NMI, etc. (TBD) International (EDG, etc) (Ruth Pordes) GriPhyN Management iVDGL iVDGL Rob Gardner

429 Jan 2003 Mike Wilde, University of Chicago External Advisory Committee Members –Fran Berman (SDSC Director) –Dan Reed (NCSA Director) –Joel Butler (former head, FNAL Computing Division) –Jim Gray (Microsoft) –Bill Johnston (LBNL, DOE Science Grid) –Fabrizio Gagliardi (CERN, EDG Director) –David Williams (former head, CERN IT) –Paul Messina (former CACR Director) –Roscoe Giles (Boston U, NPACI-EOT) Met with us 3 times: 4/2001, 1/2002, 1/2003 –Extremely useful guidance on project scope & goals

529 Jan 2003 Mike Wilde, University of Chicago GriPhyN Project Challenges We balance and coordinate –CS research with “goals, milestones & deliverables” –GriPhyN schedule/priorities/risks with those of the 4 experiments –General tools developed by GriPhyN with specific tools developed by 4 experiments –Data Grid design, architecture & deliverables with those of other Grid projects Appropriate balance requires –Tight management, close coordination, trust We have (so far) met these challenges –But requires constant attention, good will

629 Jan 2003 Mike Wilde, University of Chicago Meetings in GriPhyN/iVDGL meetings –Oct. 2000All-handsChicago –Dec. 2000ArchitectureChicago –Apr. 2001All-hands, EACUSC/ISI –Aug. 2001PlanningChicago –Oct. 2001All-hands, iVDGLUSC/ISI Numerous smaller meetings –CS-experiment –CS research –Liaisons with PPDG and EU DataGrid –US-CMS and US-ATLAS computing reviews –Experiment meetings at CERN

729 Jan 2003 Mike Wilde, University of Chicago Meetings in 2002 GriPhyN/iVDGL meetings –Jan. 2002EAC, Planning, iVDGLFlorida –Mar. 2002Outreach WorkshopBrownsville –Apr. 2002All-handsArgonne –Jul. 2002Reliability WorkshopISI –Oct. 2002Provenance WorkshopArgonne –Dec. 2002Troubleshooting WorkshopChicago –Dec. 2002All-hands technicalISI + Caltech –Jan. 2003EACSDSC Numerous other 2002 meetings –iVDGL facilities workshop (BNL) –Grid activities at CMS, ATLAS meetings –Several computing reviews for US-CMS, US-ATLAS –Demos at IST2002, SC2002 –Meetings with LCG (LHC Computing Grid) project –HEP coordination meetings (HICB)

829 Jan 2003 Mike Wilde, University of Chicago Planning Goals Clarify our vision and direction –Know how to make a difference in science & computing Map that vision to each application –Create concrete realizations of our vision Organize as cooperative subteams with specific missions and defined points of interaction Coordinate our research programs Shape toolkit to meet challenge-problem needs “Stop, Look, and Listen” to each experiment’s need –Excite the customer with our vision –Balance the promotion of our ideas with a solid understanding of the size and nature of the problems

929 Jan 2003 Mike Wilde, University of Chicago Project Approach CS Research VDT Development Application Analysis Infrastructure Development and Deployment Challenge Problem Identification Challenge Problem Solution Development Challenge Problem Solution Integration VDT Development VDT Development Infrastructure Deployment IS Deployment time Process

1029 Jan 2003 Mike Wilde, University of Chicago Project Activities Research –Experiment Analysis >Use cases, statistics, distributions, data flow patterns, tools, data types, HIPO –Vision Refinement –Attacking the “hard problems” >Virtual data identification and manipulation >Advanced resource allocation and execution planning >Scaling this up to Petascale –Architectural Refinement Toolkit Development Integration –Identify and Address Challenge Problems –Testbed construction Support Evaluation

1129 Jan 2003 Mike Wilde, University of Chicago Research Milestone Highlights Y1:Execution framework Virtual data prototypes Y2:Virtual data catalog w/glue language Integ w/ scalable replica catalog service Initial resource usage policy language Y3:Advanced planning, fault recovery Intelligent catalog Advanced policy languages Y4:Knowledge management and location Y5: Transparency and usability Scalability and manageability

1229 Jan 2003 Mike Wilde, University of Chicago Research Leadership Centers Virtual Data: –Chicago (VDC, VDL, KR), ISI (Schema) –Wisconsin (NeST), SDSC (MCAT,SRB) Request Planning –ISI (algorithms), Chicago (policy), Berkeley (query optimization) Request Execution –Wisconsin Fault Tolerance –SDSC Monitoring –Northwestern User interface –Indiana

1329 Jan 2003 Mike Wilde, University of Chicago Project Status Overview Year 1 research fruitful –Virtual data, planning, execution, integration— demonstrated at SC2001 Research efforts launched –80% focused – 20% exploratory VDT effort staffed and launched –Yearly major release; VDT1 close; VDT2 planned; VDT3-5 envisioned Year 2 experiment integrations high level plans done; detailed planning underway Long term vision refined and unified

1429 Jan 2003 Mike Wilde, University of Chicago Milestones: Architecture Early 2002: –Specify interfaces for new GriPhyN functional modules >Request Planner >Virtual Data Catalog service >Monitoring service –Define how we will connect and integrate our solutions, e.g.: >Virtual data language >Multiple-catalog integration >DAGman graphs >Policy langauge >CAS interaction for policy lookup and enforcement Year-end 2002: phased migration to a web- services based architecture

1529 Jan 2003 Mike Wilde, University of Chicago Status: Virtual Data Virtual Data –First version of a catalog structure built –Integration language “VDL” developed –Detailed transformation model designed Replica location service at Chicago & ISI –Highly scalable and fault tolerant –Soft-state distributed architecture NeSt at UW –Storage appliance for the Grid –Treats data transfer as a job step

1629 Jan 2003 Mike Wilde, University of Chicago Milestones: Virtual Data Year 2: –Local Virtual Data Catalog Structures (relational) –Catalog manipulation language (VDL) –Linkage to application metadata Year 3: Handling multi-modal virtual data –Distributed virtual data catalogs (based on RLS) –Advanced transformation signatures –Flat, objects, OODBs, relational –Cross-modal depdendency tracking Year 4: Knowledge representation –Ontologies; data generation paradigms –Fuzzy dependencies and data equivalence Year 5: Finalize Scalability and Manageability

1729 Jan 2003 Mike Wilde, University of Chicago Status: Planning and Execution Planning and Execution –Major strides in execution environment made with Condor, CondorG, and DAGman –DAGs evolving as pervasive job specification model with the virtual data grid –Large-scale CMS production demonstrated on 3-site wide-area multi-organization grid –LIGO demonstrated full GriPhyN integration –Sophisticated policy language for grid-wide resource sharing under design at Chicago –Knowledge representation research underway at Chicago –Research in ClassAds explored in Globus context Master/worker fault tolerance at UCSD –Design proposed to extend fault tolerance of Condor masters

1829 Jan 2003 Mike Wilde, University of Chicago Milestones: Request Planning Year 2: –Protype planner as a grid service module –Intial CAS and Policy Language Integration –Refinement of DAG language with data flow info Year 3: –Policy enhancements: dynamic replanning (based on Grid monitoring), cost alternatives and optimizations Year 4: –Global planning with policy constraints Year 5: –Incremental global planning –Algorithms evaluated, tuned w/ large-scale simulations

1929 Jan 2003 Mike Wilde, University of Chicago Milestones: Request Execution Year 2: –Request Planning and Execution >Striving for increasingly greater resource leverage with increasing both power AND transparency >Fault tolerance – keeping it all running! –Intial CAS and Policy Language Integration –Refinement of DAG language with data flow info –Resource utiization monitoring to drive planner Year 3: –Resource co-allocation with recovery –Fault tolerant execution engines Year 4: –Execution adapts to grid resource availability changes Year 5: –Simulation-based algorithm eval and tuning

2029 Jan 2003 Mike Wilde, University of Chicago Status: Supporting Research Joint PPDG-GriPhyN Monitoring group –Meeting regularly –Use-case development underway Research into monitoring, measurement, profiling, and performance predication –Underway at NU and ANL GRIPE facility for Grid-wide user and host certificate and login management GRAPPA portal for end-user science access

2129 Jan 2003 Mike Wilde, University of Chicago Status – Experiments ATLAS –8-site testgrid in place –data and metadata management prototypes evolving –Ambitious Year-2 plan well refined – will use numerous GriPhyN deliverables CMS –Working prototypes of production and distributed analysis, both with virtual data –Year-2 plan – simulation production – underway LIGO –Working prototypes of full VDG demonstrated –Year-2 plan well refined and development underway SDSS –Year-2 plan well refined –Challenge problem development underway –close collaboration with Chicago on VDC

2229 Jan 2003 Mike Wilde, University of Chicago Year 2 Plan: ATLAS ATLAS-GriPhyN Challenge Problem I –ATLAS DC0: 10M events, O(1000) CPUs –Integration of VDT to provide uniform distributed data access –Use of GRAPPA portal, possibly over DAGman –Demo ATLAS SW Week – March 2002

2329 Jan 2003 Mike Wilde, University of Chicago Year 2 Plan: ATLAS ATLAS-GriPhyN Challenge Problem II –Virtualization of pipelines to deliver analysis data products: reconstructions and metadata tags –Full chain production and analysis of event data –Prototyping of typical physicist analysis sessions –Graphical monitoring display of event throughput throughout the Grid –Live update display of distributed histogram population from Athena –Virtual data re-materialization from Athena –Grappa job submission and monitoring

2429 Jan 2003 Mike Wilde, University of Chicago Year 2 Plan: SDSS Challenge Problem 1 – Balanced resources –Cluster Galaxy Cataloging –Exercises virtual data derivation tracking Challenge Problem 2 – Compute Intensive –Spatial Correlation Functions and Power Spectra –Provides a research base for scientific knowledge search-engine problems Challenge Problem 3 – Storage Intensive –Weak Lensing –Provides challenging testbed for advanced request planning algorithms

2529 Jan 2003 Mike Wilde, University of Chicago Integration of GriPhyN and iVDGL Tight integration with GriPhyN –Testbeds –VDT support –Outreach –Common External Advisory Committee

2629 Jan 2003 Mike Wilde, University of Chicago iVDGL Management & Coordination Project Coordination Group US External Advisory Committee GLUE Interoperability Team Collaborating Grid Projects TeraGridEDGAsiaDataTAG BTEV LCG? BioALICEGeo? D0PDCCMS HI ? US Project Directors Outreach Team Core Software Team Facilities Team Operations Team Applications Team International Piece US Project Steering Group U.S. Piece GriPhyN Mike Wilde

2729 Jan 2003 Mike Wilde, University of Chicago Global Context: Data Grid Projects U.S. Infrastructure Projects –GriPhyN (NSF) –iVDGL (NSF) –Particle Physics Data Grid (DOE) –TeraGrid (NSF) –DOE Science Grid (DOE) EU, Asia major projects –European Data Grid (EDG) (EU, EC) –EDG related national Projects (UK, Italy, France, …) –CrossGrid (EU, EC) –DataTAG (EU, EC) –LHC Computing Grid (LCG) (CERN) –Japanese Project –Korea project

2829 Jan 2003 Mike Wilde, University of Chicago Coordination with US Efforts Trillium = GriPhyN + iVDGL + PPDG NMI & VDT Networking initiatives –HENP working group within Internet2 –Working closely with National Light Rail New proposals

2929 Jan 2003 Mike Wilde, University of Chicago International Coordination EU DataGrid & DataTAG HICB: HEP Inter-Grid Coordination Board –HICB-JTB: Joint Technical Board –GLUE Participation in LHC Computing Grid (LCG) International networks –Standing Committee on Inter-regional Connectivity –Digital Divide projects, IEEAF