Operated by the Southeastern Universities Research Association for the U.S. Depart. Of Energy Thomas Jefferson National Accelerator Facility Andy Kowalski.

Slides:



Advertisements
Similar presentations
30-31 Jan 2003J G Jensen, RAL/WP5 Storage Elephant Grid Access to Mass Storage.
Advertisements

GridPP July 2003Stefan StonjekSlide 1 SAM middleware components Stefan Stonjek University of Oxford 7 th GridPP Meeting 02 nd July 2003 Oxford.
WP2: Data Management Gavin McCance University of Glasgow November 5, 2001.
Data Management Expert Panel - WP2. WP2 Overview.
Data Management Expert Panel. RLS Globus-EDG Replica Location Service u Joint Design in the form of the Giggle architecture u Reference Implementation.
JLab Lattice Portal – Data Grid Web Service Ying Chen, Chip Watson Thomas Jefferson National Accelerator Facility.
FP7-INFRA Enabling Grids for E-sciencE EGEE Induction Grid training for users, Institute of Physics Belgrade, Serbia Sep. 19, 2008.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
GGF Toronto Spitfire A Relational DB Service for the Grid Peter Z. Kunszt European DataGrid Data Management CERN Database Group.
Office of Science U.S. Department of Energy Grids and Portals at NERSC Presented by Steve Chan.
Data Management for Physics Analysis in PHENIX (BNL, RHIC) Evaluation of Grid architecture components in PHENIX context Barbara Jacak, Roy Lacey, Saskia.
Magda – Manager for grid-based data Wensheng Deng Physics Applications Software group Brookhaven National Laboratory.
Web-based Portal for Discovery, Retrieval and Visualization of Earth Science Datasets in Grid Environment Zhenping (Jane) Liu.
Data Grid Web Services Chip Watson Jie Chen, Ying Chen, Bryan Hess, Walt Akers.
QCDgrid Technology James Perry, George Beckett, Lorna Smith EPCC, The University Of Edinburgh.
The Mass Storage System at JLAB - Today and Tomorrow Andy Kowalski.
Grappa: Grid access portal for physics applications Shava Smallen Extreme! Computing Laboratory Department of Physics Indiana University.
CERN - IT Department CH-1211 Genève 23 Switzerland t Monitoring the ATLAS Distributed Data Management System Ricardo Rocha (CERN) on behalf.
QCDgrid UKQCD Achievements and Future Priorities Who and what Achievements QCDgrid middleware Future priorities Demo of meta-data catalogue browser Alan.
Data Management Kelly Clynes Caitlin Minteer. Agenda Globus Toolkit Basic Data Management Systems Overview of Data Management Data Movement Grid FTP Reliable.
Alexandre A. P. Suaide VI DOSAR workshop, São Paulo, 2005 STAR grid activities and São Paulo experience.
ARGONNE  CHICAGO Ian Foster Discussion Points l Maintaining the right balance between research and development l Maintaining focus vs. accepting broader.
Remote Production and Regional Analysis Centers Iain Bertram 24 May 2002 Draft 1 Lancaster University.
November 7, 2001Dutch Datagrid SARA 1 DØ Monte Carlo Challenge A HEP Application.
Grid Computing Chip Watson Jefferson Lab Hall B Collaboration Meeting 1-Nov-2001.
COMP3019 Coursework: Introduction to GridSAM Steve Crouch School of Electronics and Computer Science.
QCDGrid Progress James Perry, Andrew Jackson, Stephen Booth, Lorna Smith EPCC, The University Of Edinburgh.
Grid Workload Management & Condor Massimo Sgaravatto INFN Padova.
HEPiX Karlsruhe May 9-13, 2005 Operated by the Southeastern Universities Research Association for the U.S. Department of Energy Thomas Jefferson National.
ILDG Middleware Status Chip Watson ILDG-6 Workshop May 12, 2005.
CYBERINFRASTRUCTURE FOR THE GEOSCIENCES Data Replication Service Sandeep Chandra GEON Systems Group San Diego Supercomputer Center.
Web Services BOF This is a proposed new working group coming out of the Grid Computing Environments Research Group, as an outgrowth of their investigations.
10/24/2015OSG at CANS1 Open Science Grid Ruth Pordes Fermilab
D C a c h e Michael Ernst Patrick Fuhrmann Tigran Mkrtchyan d C a c h e M. Ernst, P. Fuhrmann, T. Mkrtchyan Chep 2003 Chep2003 UCSD, California.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
Integrating JASMine and Auger Sandy Philpott Thomas Jefferson National Accelerator Facility Jefferson Ave. Newport News, Virginia USA 23606
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
Author - Title- Date - n° 1 Partner Logo EU DataGrid, Work Package 5 The Storage Element.
Lattice QCD Data Grid Middleware: status report M. Sato, CCS, University of Tsukuba ILDG6, May, 12, 2005.
Grid Architecture William E. Johnston Lawrence Berkeley National Lab and NASA Ames Research Center (These slides are available at grid.lbl.gov/~wej/Grids)
09/02 ID099-1 September 9, 2002Grid Technology Panel Patrick Dreher Technical Panel Discussion: Progress in Developing a Web Services Data Analysis Grid.
1 ILDG Status in Japan  Lattice QCD Archive(LQA) a gateway to ILDG Japan Grid  HEPNet-J/sc an infrastructure for Japan Lattice QCD Grid A. Ukawa Center.
NA-MIC National Alliance for Medical Image Computing UCSD: Engineering Core 2 Portal and Grid Infrastructure.
Replica Management Services in the European DataGrid Project Work Package 2 European DataGrid.
What is SAM-Grid? Job Handling Data Handling Monitoring and Information.
Owen SyngeTitle of TalkSlide 1 Storage Management Owen Synge – Developer, Packager, and first line support to System Administrators. Talks Scope –GridPP.
…building the next IT revolution From Web to Grid…
Overview of grid activities in France in relation to FKPPL FKPPL Workshop Thursday February 26th, 2009 Dominique Boutigny.
Grid User Interface for ATLAS & LHCb A more recent UK mini production used input data stored on RAL’s tape server, the requirements in JDL and the IC Resource.
US LHC OSG Technology Roadmap May 4-5th, 2005 Welcome. Thank you to Deirdre for the arrangements.
PPDGLHC Computing ReviewNovember 15, 2000 PPDG The Particle Physics Data Grid Making today’s Grid software work for HENP experiments, Driving GRID science.
DGC Paris WP2 Summary of Discussions and Plans Peter Z. Kunszt And the WP2 team.
Lattice QCD Data Grid Middleware: Meta Data Catalog (MDC) -- CCS ( tsukuba) proposal -- M. Sato, for ILDG Middleware WG ILDG Workshop, May 2004.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
Jens G Jensen RAL, EDG WP5 Storage Element Overview DataGrid Project Conference Heidelberg, 26 Sep-01 Oct 2003.
AliEn AliEn at OSC The ALICE distributed computing environment by Bjørn S. Nilsen The Ohio State University.
USQCD regional grid Report to ILDG /28/09ILDG14, June 5, US Grid Usage  Growing usage of gauge configurations in ILDG file format.  Fermilab.
Storage and Data Movement at FNAL D. Petravick CHEP 2003.
Data Transfer Service Challenge Infrastructure Ian Bird GDB 12 th January 2005.
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
10 March Andrey Grid Tools Working Prototype of Distributed Computing Infrastructure for Physics Analysis SUNY.
Jefferson Lab Site Report Kelvin Edwards Thomas Jefferson National Accelerator Facility Newport News, Virginia USA
Open Science Grid Consortium Storage on Open Science Grid Placing, Using and Retrieving Data on OSG Resources Abhishek Singh Rana OSG Users Meeting July.
A Web Based Job Submission System for a Physics Computing Cluster David Jones IOP Particle Physics 2004 Birmingham 1.
Hall D Computing Facilities Ian Bird 16 March 2001.
Grid Portal Services IeSE (the Integrated e-Science Environment)
Patrick Dreher Research Scientist & Associate Director
A Web-Based Data Grid Chip Watson, Ian Bird, Jie Chen,
Production Manager Tools (New Architecture)
Presentation transcript:

Operated by the Southeastern Universities Research Association for the U.S. Depart. Of Energy Thomas Jefferson National Accelerator Facility Andy Kowalski SURA Cyberinfrastructure Workshop Georgia State University January 5–7, 2005 Jefferson Lab: Experimental and Theoretical Physics Grids Andy Kowalski

Operated by the Southeastern Universities Research Association for the U.S. Depart. Of Energy Thomas Jefferson National Accelerator Facility Andy Kowalski Jefferson Lab Who are we? Thomas Jefferson National Accelerator Facility Department of Energy Research Laboratory Southeastern Universities Research Association What do we do? High Energy Nuclear Physics quarks and gluons Operate a 6.07 GeV continuous electron beam accelerator Free-Electron Laser (10 kW)

Operated by the Southeastern Universities Research Association for the U.S. Depart. Of Energy Thomas Jefferson National Accelerator Facility Andy Kowalski Jefferson Lab

Operated by the Southeastern Universities Research Association for the U.S. Depart. Of Energy Thomas Jefferson National Accelerator Facility Andy Kowalski Data and Storage Three experimental halls HallA and HallC 100’s GB/day each HallB – CLAS TB/day (currently up to 30MB/sec) Currently store and manage 1 PB of data on tape. Users around the world want access to the data

Operated by the Southeastern Universities Research Association for the U.S. Depart. Of Energy Thomas Jefferson National Accelerator Facility Andy Kowalski Computing Batch Farm 200 dual CPU nodes ~358,060 SPECint2000 Moves 4-7 TB/day Reconstruction Analysis Simulations (CLAS – large) Lattice QCD Machine 3 clusters 128, 256, 384 nodes

Operated by the Southeastern Universities Research Association for the U.S. Depart. Of Energy Thomas Jefferson National Accelerator Facility Andy Kowalski Need for Grids 12 GeV Upgrade HallB – CLAS data rates increase to MB/sec Will export 50% or more of the data Import data from simulations done at Universities This can be a rather large amount HallD – GlueX Same scale as the LHC experiments 100 MB/sec - 3 PB of data per year 1 PB of raw data at JLab 1 PB for analysis (JLAb and offsite) 1 PB for simulations (offsite) Lattice QCD 10 TF machine A significant amount of data Users around the world want access to the data

Operated by the Southeastern Universities Research Association for the U.S. Depart. Of Energy Thomas Jefferson National Accelerator Facility Andy Kowalski JLab: Theory and Experimental Grid Efforts Similarities Focus on Data Grids Desire interfaces definitions for interoperability Chose web services for implementation WSDL defines the interface Theory ILDG and PPDG SRM Replica Catalog Experimenal PPDG and pursuing OSG SRM Job submission interface

Operated by the Southeastern Universities Research Association for the U.S. Depart. Of Energy Thomas Jefferson National Accelerator Facility Andy Kowalski ILDG: Data Grid Services Web Services Architecture File Client Meta Data Catalog Replica Catalog SRM Service Replication Service Storage (disk, silo) File Server(s) Web Services Single Site Meta Data Catalog Replica Catalog SRM Service Replication Service Storage (disk, silo) File Server(s) Replica Catalog Meta Data Catalog Storage (disk, silo) File Server(s) SRM Service Storage (disk, silo) File Server(s) Replication Service Replica Catalog SRM Service Storage (disk, silo) File Server(s) Replication Service Replica Catalog SRM Service (Consistency Agent) Storage (disk, silo) File Server(s) SRM Service Replication Service Replica Catalog Storage (disk, silo) File Server(s) SRM Service * Slide from Chip Watson, ILDG Middleware Project Status

Operated by the Southeastern Universities Research Association for the U.S. Depart. Of Energy Thomas Jefferson National Accelerator Facility Andy Kowalski ILDG: A Three Tier Web Services Architecture Web Browser XML to HTML servlet Web Service Application Web Service Local Backend Services (batch, file, etc.) Web Server (Portal) Authenticated connections Remote Web Server Web Service Storage system Catalogs Web services provide a standard API for clients, and intermediary servlets allow use from a browser (as in a portal) * Slide from Chip Watson, ILDG Middleware Project Status

Operated by the Southeastern Universities Research Association for the U.S. Depart. Of Energy Thomas Jefferson National Accelerator Facility Andy Kowalski Components: Meta Data Catalog Hold metadata for files Hold metadata for a set of files (data set) Process query lookup Queries return (sets of) GFN, (Global File Name = key), and optionally full metadata for each match * Slide from Chip Watson, ILDG Middleware Project Status File Client Meta Data Catalog Replica Catalog SRM Service Replication Service Storage Resource File Server(s) SRM Listener Web Services Single Site

Operated by the Southeastern Universities Research Association for the U.S. Depart. Of Energy Thomas Jefferson National Accelerator Facility Andy Kowalski Components: Replica Catalog Track all copies of a file / data set Get replicas Create replica Remove replica Prototypes exist at Jefferson Lab Fermilab * Slide from Chip Watson, ILDG Middleware Project Status File Client Meta Data Catalog Replica Catalog SRM Service Replication Service Storage Resource File Server(s) SRM Listener Web Services Single Site

Operated by the Southeastern Universities Research Association for the U.S. Depart. Of Energy Thomas Jefferson National Accelerator Facility Andy Kowalski Components: Storage Resource Manager Manage storage system Disk only Disk plus tape 3 party file transfers Negotiate protocols for file retrieval (select a file server) Auto stage a file on get (asynchronous operation) Version 2.1 defined (collaboration) * Slide from Chip Watson, ILDG Middleware Project Status File Client Meta Data Catalog Replica Catalog SRM Service Replication Service Storage Resource File Server(s) SRM Listener Web Services Single Site

Operated by the Southeastern Universities Research Association for the U.S. Depart. Of Energy Thomas Jefferson National Accelerator Facility Andy Kowalski ILDG Components MetaData Catalog (MDC) Each collaboration deploys one A mechanism (not defined yet, under discussion) exists for searching all (a virtual MDC) Replica Catalog (RC) (same comments) Storage Resource Manager (SRM) Each collaboration deploys one or more At each SRM site, there are one or more file servers: http, ftp, gridftp, jparss, bbftp, … * Slide from Chip Watson, ILDG Middleware Project Status

Operated by the Southeastern Universities Research Association for the U.S. Depart. Of Energy Thomas Jefferson National Accelerator Facility Andy Kowalski JLab: Experimental Effort PPDG (Particle Physics Data Grid) Collaboration of computer scientists and physicists Developing and deploying production Grid systems for experiment-specific applications Now supporting OSG (Open Science Grid) SRM (Storage Resource Manager) A common/standard interface to mass storage systems In 2003 FSU used SRM v1 to process monte-carlo for 30 million events In 2004 deployed a v2 implementation for testing Required for production in February 2005 Already working with LBL, Fermi, CERN to define v3 Job Submission PKI Based authentication to Auger (JLab job submission system) Investigated uJDL (a user level job description language) BNL leading this effort

Operated by the Southeastern Universities Research Association for the U.S. Depart. Of Energy Thomas Jefferson National Accelerator Facility Andy Kowalski Envisioned Architecture

Operated by the Southeastern Universities Research Association for the U.S. Depart. Of Energy Thomas Jefferson National Accelerator Facility Andy Kowalski SRM v2 Implemented SRM version Interface to Jasmine via the HPC Disk/Cache Manager. JLab SRM is a Java Web Service. Uses Apache Axis as SOAP Engine Uses Apache Tomcat as Servlet Engine. Uses GridFTP for file movement Testing with CMU Production service required by February Had a hard time using GT3 Cannot just take components that one wants (it is all or nothing)

Operated by the Southeastern Universities Research Association for the U.S. Depart. Of Energy Thomas Jefferson National Accelerator Facility Andy Kowalski SRM v2 Server Deployment Requires Tomcat, MySQL, SRM worker daemon Firewall configuration: SRM port 8443 GRIDFTP ports 2811, Currently only installed at JLab Testing client access with CMU Next step: install an SRM server at CMU

Operated by the Southeastern Universities Research Association for the U.S. Depart. Of Energy Thomas Jefferson National Accelerator Facility Andy Kowalski SRM v2 Client Deployment Installed at JLab and CMU Implements only srmGet and srmPut (permission problem to fix) Requires specific ant and java versions Proper grid certificate request and installation a challenge (?) Use OpenSSL for cert request instead Globus requires a full installation simply to request a cert and run the client Just need grid-proxy-init Note: Curtis' notes are at Currently the only SRM v2 server and client

Operated by the Southeastern Universities Research Association for the U.S. Depart. Of Energy Thomas Jefferson National Accelerator Facility Andy Kowalski Long-Term SRM Work We are considering how the next SRM version could become the primary interface to Jasmine and the primary farm file mover. Use for Local and Remote Access Goal: 25TB/day from tape through SRM. Balancing classes of requests/prioritizing types of data transfers becomes essential. Farm interaction use cases must be modeled: farm input, farm output, scheduling. We are already looking at what SRM v3 will look like. SRM Core Features and Feature Sets (ideas from the last SRM meeting)

Operated by the Southeastern Universities Research Association for the U.S. Depart. Of Energy Thomas Jefferson National Accelerator Facility Andy Kowalski Job Submission uJDL Is this really needed? Is a standard job submission interface what is really needed? Is that Condor-G? Auger interface Uses java web services Uses PKI authentication for authentication Not GSI

Operated by the Southeastern Universities Research Association for the U.S. Depart. Of Energy Thomas Jefferson National Accelerator Facility Andy Kowalski Grid3dev - OSG JLab development effort is limited Grid3 proved successful Atlas and CMS were the major users JLab plans to join Grid3dev as a step toward OSG-INT/OSG We cannot develop everything we need VO management tools, monitoring, etc. Testing and evaluation Integration with facility infrastructure Determine what we need and can use for others

Operated by the Southeastern Universities Research Association for the U.S. Depart. Of Energy Thomas Jefferson National Accelerator Facility Andy Kowalski References