1 Grid Related Activities at Caltech Koen Holtman Caltech/CMS PPDG meeting, Argonne July 13-14, 2000.

Slides:



Advertisements
Similar presentations
WP2: Data Management Gavin McCance University of Glasgow November 5, 2001.
Advertisements

OptorSim: A Replica Optimisation Simulator for the EU DataGrid W. H. Bell, D. G. Cameron, R. Carvajal, A. P. Millar, C.Nicholson, K. Stockinger, F. Zini.
31/03/00 CMS(UK)Glenn Patrick What is the CMS(UK) Data Model? Assume that CMS software is available at every UK institute connected by some infrastructure.
Author - Title- Date - n° 1 GDMP The European DataGrid Project Team
ATLAS Tier-3 in Geneva Szymon Gadomski, Uni GE at CSCS, November 2009 S. Gadomski, ”ATLAS T3 in Geneva", CSCS meeting, Nov 091 the Geneva ATLAS Tier-3.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
GRID DATA MANAGEMENT PILOT (GDMP) Asad Samar (Caltech) ACAT 2000, Fermilab October , 2000.
Workload Management Workpackage Massimo Sgaravatto INFN Padova.
High-Performance Throughput Tuning/Measurements Davide Salomoni & Steffen Luitz Presented at the PPDG Collaboration Meeting, Argonne National Lab, July.
Title US-CMS User Facilities Vivian O’Dell US CMS Physics Meeting May 18, 2001.
The new The new MONARC Simulation Framework Iosif Legrand  California Institute of Technology.
Ian M. Fisk Fermilab February 23, Global Schedule External Items ➨ gLite 3.0 is released for pre-production in mid-April ➨ gLite 3.0 is rolled onto.
Vladimir Litvin, Harvey Newman Caltech CMS Scott Koranda, Bruce Loftis, John Towns NCSA Miron Livny, Peter Couvares, Todd Tannenbaum, Jamie Frey Wisconsin.
Grid Information Systems. Two grid information problems Two problems  Monitoring  Discovery We can use similar techniques for both.
CMS Report – GridPP Collaboration Meeting VI Peter Hobson, Brunel University30/1/2003 CMS Status and Plans Progress towards GridPP milestones Workload.
08/06/00 LHCb(UK) Meeting Glenn Patrick LHCb(UK) Computing/Grid: RAL Perspective Glenn Patrick Central UK Computing (what.
November 15, 2000US CMS Tier2 Plans Matthias Kasemann1 US CMS Software and Computing Tier 2 Center Plans Matthias Kasemann Fermilab DOE/NSF Baseline Review.
ARGONNE  CHICAGO Ian Foster Discussion Points l Maintaining the right balance between research and development l Maintaining focus vs. accepting broader.
3 Sept 2001F HARRIS CHEP, Beijing 1 Moving the LHCb Monte Carlo production system to the GRID D.Galli,U.Marconi,V.Vagnoni INFN Bologna N Brook Bristol.
Alain Romeyer - 15/06/20041 CMS farm Mons Final goal : included in the GRID CMS framework To be involved in the CMS data processing scheme.
D0 SAM – status and needs Plagarized from: D0 Experiment SAM Project Fermilab Computing Division.
LCG and HEPiX Ian Bird LCG Project - CERN HEPiX - FNAL 25-Oct-2002.
1 Object Level Physics Data Replication in the Grid Koen Holtman Caltech/CMS ACAT’2000, Fermilab October 16-20, 2000.
Grid Workload Management & Condor Massimo Sgaravatto INFN Padova.
Jean-Yves Nief CC-IN2P3, Lyon HEPiX-HEPNT, Fermilab October 22nd – 25th, 2002.
Fermilab User Facility US-CMS User Facility and Regional Center at Fermilab Matthias Kasemann FNAL.
DataGrid Applications Federico Carminati WP6 WorkShop December 11, 2000.
Grid Status - PPDG / Magda / pacman Torre Wenaus BNL U.S. ATLAS Physics and Computing Advisory Panel Review Argonne National Laboratory Oct 30, 2001.
Network Tests at CHEP K. Kwon, D. Han, K. Cho, J.S. Suh, D. Son Center for High Energy Physics, KNU, Korea H. Park Supercomputing Center, KISTI, Korea.
Your university or experiment logo here Caitriana Nicholson University of Glasgow Dynamic Data Replication in LCG 2008.
Finnish DataGrid meeting, CSC, Otaniemi, V. Karimäki (HIP) DataGrid meeting, CSC V. Karimäki (HIP) V. Karimäki (HIP) Otaniemi, 28 August, 2000.
11 December 2000 Paolo Capiluppi - DataGrid Testbed Workshop CMS Applications Requirements DataGrid Testbed Workshop Milano, 11 December 2000 Paolo Capiluppi,
DOSAR Workshop, Sao Paulo, Brazil, September 16-17, 2005 LCG Tier 2 and DOSAR Pat Skubic OU.
Ruth Pordes, Fermilab CD, and A PPDG Coordinator Some Aspects of The Particle Physics Data Grid Collaboratory Pilot (PPDG) and The Grid Physics Network.
LHC Computing Review Recommendations John Harvey CERN/EP March 28 th, th LHCb Software Week.
Data Grid projects in HENP R. Pordes, Fermilab Many HENP projects are working on the infrastructure for global distributed simulated data production, data.
Virtual Data Grid Architecture Ewa Deelman, Ian Foster, Carl Kesselman, Miron Livny.
7April 2000F Harris LHCb Software Workshop 1 LHCb planning on EU GRID activities (for discussion) F Harris.
RAL Site Report John Gordon IT Department, CLRC/RAL HEPiX Meeting, JLAB, October 2000.
D0RACE: Testbed Session Lee Lueking D0 Remote Analysis Workshop February 12, 2002.
LCG Phase 2 Planning Meeting - Friday July 30th, 2004 Jean-Yves Nief CC-IN2P3, Lyon An example of a data access model in a Tier 1.
The Internet2 HENP Working Group Internet2 Spring Meeting April 9, 2003.
The UK eScience Grid (and other real Grids) Mark Hayes NIEeS Summer School 2003.
Ian Bird LHC Computing Grid Project Leader LHC Grid Fest 3 rd October 2008 A worldwide collaboration.
PPDG update l We want to join PPDG l They want PHENIX to join NSF also wants this l Issue is to identify our goals/projects Ingredients: What we need/want.
CERN IT Department CH-1211 Genève 23 Switzerland t Frédéric Hemmer IT Department Head - CERN 23 rd August 2010 Status of LHC Computing from.
High Energy Physics and Grids at UF (Dec. 13, 2002)Paul Avery1 University of Florida High Energy Physics.
LIGO-G E LIGO Scientific Collaboration Data Grid Status Albert Lazzarini Caltech LIGO Laboratory Trillium Steering Committee Meeting 20 May 2004.
PPDGLHC Computing ReviewNovember 15, 2000 PPDG The Particle Physics Data Grid Making today’s Grid software work for HENP experiments, Driving GRID science.
UK Grid Meeting Glenn Patrick1 LHCb Grid Activities in UK Grid Prototype and Globus Technical Meeting QMW, 22nd November 2000 Glenn Patrick (RAL)
December 26, 2015 RHIC/USATLAS Grid Computing Facility Overview Dantong Yu Brookhaven National Lab.
Integration of the ATLAS Tag Database with Data Management and Analysis Components Caitriana Nicholson University of Glasgow 3 rd September 2007 CHEP,
A Fully Automated Fault- tolerant System for Distributed Video Processing and Off­site Replication George Kola, Tevfik Kosar and Miron Livny University.
State of LSC Data Analysis and Software LSC Meeting LIGO Hanford Observatory November 11 th, 2003 Kent Blackburn, Stuart Anderson, Albert Lazzarini LIGO.
Scott Koranda, UWM & NCSA 14 January 2016www.griphyn.org Lightweight Data Replicator Scott Koranda University of Wisconsin-Milwaukee & National Center.
US-CMS T2 Centers US-CMS Tier 2 Report Patricia McBride Fermilab GDB Meeting August 31, 2007 Triumf - Vancouver.
July 26, 1999MONARC Meeting CERN MONARC Meeting CERN July 26, 1999.
Super Computing 2000 DOE SCIENCE ON THE GRID Storage Resource Management For the Earth Science Grid Scientific Data Management Research Group NERSC, LBNL.
Doug Benjamin Duke University. 2 ESD/AOD, D 1 PD, D 2 PD - POOL based D 3 PD - flat ntuple Contents defined by physics group(s) - made in official production.
BaBar and the GRID Tim Adye CLRC PP GRID Team Meeting 3rd May 2000.
Tier 1 at Brookhaven (US / ATLAS) Bruce G. Gibbard LCG Workshop CERN March 2004.
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
January 20, 2000K. Sliwa/ Tufts University DOE/NSF ATLAS Review 1 SIMULATION OF DAILY ACTIVITITIES AT REGIONAL CENTERS MONARC Collaboration Alexander Nazarenko.
CLRC Grid Team Glenn Patrick LHCb GRID Plans Glenn Patrick LHCb has formed a GRID technical working group to co-ordinate practical Grid.
Grid Activities in CMS Asad Samar (Caltech) PPDG meeting, Argonne July 13-14, 2000.
Run-time Adaptation of Grid Data Placement Jobs George Kola, Tevfik Kosar and Miron Livny Condor Project, University of Wisconsin.
1 CMS Virtual Data Overview Koen Holtman Caltech/CMS GriPhyN all-hands meeting, Marina del Rey April 9, 2001.
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
PetaCache: Data Access Unleashed Tofigh Azemoon, Jacek Becla, Chuck Boeheim, Andy Hanushevsky, David Leith, Randy Melen, Richard P. Mount, Teela Pulliam,
1 Particle Physics Data Grid (PPDG) project Les Cottrell – SLAC Presented at the NGI workshop, Berkeley, 7/21/99.
Presentation transcript:

1 Grid Related Activities at Caltech Koen Holtman Caltech/CMS PPDG meeting, Argonne July 13-14, 2000

2 Executive Summary Caltech group works on: Developing the (Data) Grid Mostly in collaborative efforts Focus on OODBs, regional centres, WAN Synergy CMS ORCA production To support detector design studies, trigger design studies Real, current needs TBs of data

3 Resources 1 People People who spend most of their time on grid related work: At Caltech: Julian Bunn, Takako Hickey, Koen Holtman At CERN payed by Caltech: Iosif Legrand, Asad Samar, Mehnaz Hafeez (40%), Philippe Galvez, Gregory Denis Many others spend smaller parts of their time Datasets Currently we have ~3 TB of CMS physics data (mostly CMSIM) ~300 GB persistent physics data Software CMS software: CMSIM, ORCA, IGUANA Objectivity (C++&Java), Globus [Condor]

4 Resources 2 Hardware Major CPU: HP X-class (256 CPU) and V-class (128 CPU) systems (Major use for CMS (CMSIM) production planned) Currently using a Condor system at Wisconsin (~50 CPU used during our production) Will build a Linux farm at Caltech - Order in July: CPU, Caltech+SDSC - Rising to 256 CPU if Caltech is one of the Tier2 centres Major Storage: HPSS system at CACR (~350 TB) ~1 TB on disk arrays Networking: Many connections -- of particular interest: NTON link to SLAC (2xOC12, 1244 Mbit/s) US- CERN link (45 Mbit/s, probably 155 Mbit/s in Fall)

5 Grid activities 1 ORCA production Production at Caltech Will generate about 1,000,000 events (~1TB) in summer Serves as testbed/motivation for many activities Participation in production elsewhere Involves a lot of file transfers over network Expect to gain valuable experience in running (distributed) production (e.g. To be fed into MONARC) Currently using Condor system at Winsconsin (Vladimir Litvin) MONARC: see talk by Iosif Legrand

6 Grid activities 2 File based replication system for distributed ORCA production See talk by Asad Samar Mostly done by CERN-based people (EU DataGrid pilot) First prototype August 2000 (to be used in Fall production) Uses Globus components Object based replication See (later) talk by Koen Holtman Object (event subset) based replication important for distributed analysis, Tier 2 regional centres and smaller Not a production effort, initial work targeted at demonstration prototype First demo around Autumn 2000 Uses Globus components Active contacts with Globus team in both efforts

7 Grid activities 3 WAN file replication tests Julian Bunn, Philippe Galvez (network monitoring) Also together with SLAC: see next slide Goals: investigate WAN issues, tune file transport to maximise WAN throughput (TCP window size, routers?) Specific tests (using Globus GSIftp, ORCA files) Caltech-Argonne: tests complete (3.9 Mbytes/s) Caltech-INFN: initial tests complete Will install HPSS enabled version of Globus FTP at CACR as soon as it becomes available

8 Grid activities 4 WAN file replication tests 2 For PPDG 100 MB/s milestone, together with Davide Salomoni and Les Cottrell at SLAC NTON link to SLAC (2xOC12, 2x622 Mbit/s = 1244 Mbit/s) Latest results (Julian Bunn): Achieved 380 Mbits/s (47 MB/s) memory-to-memory between Caltech and SLAC over a single OC12 link With a single TCP/IP stream using the iperf tool

9 Grid activities 5 Distributed resource management Takako Hickey Research agenda: Fault-tolerant service (special focus on network partition) Data-aware scheduling, Scalable software structure Prototype Execution Service: Submit, monitor, kill jobs individually or as a set Service tolerates all types of failures (network partition, fail-stop, message omissions etc.) Mechanism for placing jobs near data Being tested with ORCA production software (with Vladimir Litvin) Works with 30 CPUs, being debugged for 60 CPUs. Future Work: Study of data-aware scheduling policies using MONARC simulation tool (with Iosif Legrand) Extending prototype for scalability

10 Grid activities 6 VRVS Philippe Galvez, Gregory Denis at CERN Tools for videoconferencing and collaborative work Part R&D, part production Also performing a lot of WAN network monitoring

11 Grid activities 7 ALDAP Collaboration between Caltech/CMS and JHU/SDSS Both groups are using Objectivity Goal: share experience, technology, develop new ways of organising (object) data SX: data mining tool used in SDSS Use case is similar to end user physics analysis on tag objects and AOD (10 KB) objects Uses an SQL like language Currently installing SX tool at Caltech Will make replica of SDSS data at Caltech (40 GB now, 200 GB in half a year) Doing comparisons between OODBMS and RDBMS (Objectivity and MS SQL Server) for tag/ntuple type analysis

12 Grid activities 8 Grid related project coordination Lot of resources devoted to coordination/synchronisation Between GryPhyn, PPDG, EU DataGrid, MONARC, RD45, ALDAP,..... Harvey Newman, Julian Bunn, Asad Samar