Data GRID deployment in HEPnet-J Takashi Sasaki Computing Research Center KEK.

Slides:



Advertisements
Similar presentations
1 ALICE Grid Status David Evans The University of Birmingham GridPP 14 th Collaboration Meeting Birmingham 6-7 Sept 2005.
Advertisements

23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
HEP Prospects, J. Yu LEARN Strategy Meeting Prospects on Texas High Energy Physics Network Needs LEARN Strategy Meeting University of Texas at El Paso.
Large scale data flow in local and GRID environment V.Kolosov, I.Korolko, S.Makarychev ITEP Moscow.
LHC’s Second Run Hyunseok Lee 1. 2 ■ Discovery of the Higgs particle.
Grid Computing for High Energy Physics in Japan Hiroyuki Matsunaga International Center for Elementary Particle Physics (ICEPP), The University of Tokyo.
25 February 2000Tim Adye1 Using an Object Oriented Database to Store BaBar's Terabytes Tim Adye Particle Physics Department Rutherford Appleton Laboratory.
High Energy Physics At OSCER A User Perspective OU Supercomputing Symposium 2003 Joel Snow, Langston U.
Cluster currently consists of: 1 Dell PowerEdge Ghz Dual, quad core Xeons (8 cores) and 16G of RAM Original GRIDVM - SL4 VM-Ware host 1 Dell PowerEdge.
Alexandre A. P. Suaide VI DOSAR workshop, São Paulo, 2005 STAR grid activities and São Paulo experience.
Computing for ILC experiment Computing Research Center, KEK Hiroyuki Matsunaga.
Computing Coordination in Japan Takashi Sasaki Computing Research Center KEK, Inter-University Research Institute Corporation High Energy Accelerator Research.
Data GRID Activity in Japan Yoshiyuki WATASE KEK (High energy Accelerator Research Organization) Tsukuba, Japan
Grid Applications for High Energy Physics and Interoperability Dominique Boutigny CC-IN2P3 June 24, 2006 Centre de Calcul de l’IN2P3 et du DAPNIA.
QCDGrid Progress James Perry, Andrew Jackson, Stephen Booth, Lorna Smith EPCC, The University Of Edinburgh.
LHCb computing in Russia Ivan Korolko (ITEP Moscow) Russia-CERN JWGC, October 2005.
8th November 2002Tim Adye1 BaBar Grid Tim Adye Particle Physics Department Rutherford Appleton Laboratory PP Grid Team Coseners House 8 th November 2002.
LHC Computing Plans Scale of the challenge Computing model Resource estimates Financial implications Plans in Canada.
Workshop KEK - CC-IN2P3 KEK new Grid system 27 – 29 Oct. CC-IN2P3, Lyon, France Day2 14: :55 (40min) Koichi Murakami, KEK/CRC.
Instrumentation of the SAM-Grid Gabriele Garzoglio CSC 426 Research Proposal.
International Workshop on HEP Data Grid Nov 9, 2002, KNU Data Storage, Network, Handling, and Clustering in CDF Korea group Intae Yu*, Junghyun Kim, Ilsung.
Computing for ILC experiments Akiya Miyamoto KEK 14 May 2014 AWLC14 Any comments are welcomed.
Site Report from KEK, Japan JP-KEK-CRC-01 and JP-KEK-CRC-02 Go Iwai, KEK/CRC Grid Operations Workshop – 2007 Kungliga Tekniska högskolan, Stockholm, Sweden.
HEP Data Grid in Japan Takashi Sasaki Computing Research Center KEK.
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
Tier-2  Data Analysis  MC simulation  Import data from Tier-1 and export MC data CMS GRID COMPUTING AT THE SPANISH TIER-1 AND TIER-2 SITES P. Garcia-Abia.
Operated by the Southeastern Universities Research Association for the U.S. Depart. Of Energy Thomas Jefferson National Accelerator Facility Andy Kowalski.
1 LCG-France sites contribution to the LHC activities in 2007 A.Tsaregorodtsev, CPPM, Marseille 14 January 2008, LCG-France Direction.
Computing Resources for ILD Akiya Miyamoto, KEK with a help by Vincent, Mark, Junping, Frank 9 September 2014 ILD Oshu City a report on work.
Owen SyngeTitle of TalkSlide 1 Storage Management Owen Synge – Developer, Packager, and first line support to System Administrators. Talks Scope –GridPP.
…building the next IT revolution From Web to Grid…
The LHCb CERN R. Graciani (U. de Barcelona, Spain) for the LHCb Collaboration International ICFA Workshop on Digital Divide Mexico City, October.
Les Les Robertson LCG Project Leader High Energy Physics using a worldwide computing grid Torino December 2005.
GRID Deployment Status and Plan at KEK ISGC2007 Takashi Sasaki KEK Computing Research Center.
ATLAS WAN Requirements at BNL Slides Extracted From Presentation Given By Bruce G. Gibbard 13 December 2004.
David Adams ATLAS DIAL: Distributed Interactive Analysis of Large datasets David Adams BNL August 5, 2002 BNL OMEGA talk.
HIGUCHI Takeo Department of Physics, Faulty of Science, University of Tokyo Representing dBASF Development Team BELLE/CHEP20001 Distributed BELLE Analysis.
Computing Research Center, High Energy Accelerator Organization (KEK) Site Status Report Go Iwai, KEK/CRC, Japan WLCG Tier-2 Workshop Dec. 1 ~ 4, 2006.
International Workshop on HEP Data Grid Aug 23, 2003, KNU Status of Data Storage, Network, Clustering in SKKU CDF group Intae Yu*, Joong Seok Chae Department.
LHC Computing, CERN, & Federated Identities
Computing Resources for ILD Akiya Miyamoto, KEK with a help by Vincent, Mark, Junping, Frank 9 September 2014 ILD Oshu City a report on work.
Computing Issues for the ATLAS SWT2. What is SWT2? SWT2 is the U.S. ATLAS Southwestern Tier 2 Consortium UTA is lead institution, along with University.
KEK GRID for ILC Experiments Akiya Miyamoto, Go Iwai, Katsumasa Ikematsu KEK LCWS March 2010.
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
LHCb 2009-Q4 report Q4 report LHCb 2009-Q4 report, PhC2 Activities in 2009-Q4 m Core Software o Stable versions of Gaudi and LCG-AA m Applications.
1 Particle Physics Data Grid (PPDG) project Les Cottrell – SLAC Presented at the NGI workshop, Berkeley, 7/21/99.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Introduction Salma Saber Electronic.
ATLAS Computing: Experience from first data processing and analysis Workshop TYL’10.
DØ Grid Computing Gavin Davies, Frédéric Villeneuve-Séguier Imperial College London On behalf of the DØ Collaboration and the SAMGrid team The 2007 Europhysics.
GRID & Parallel Processing Koichi Murakami11 th Geant4 Collaboration Workshop / LIP - Lisboa (10-14/Oct./2006) 1 GRID-related activity in Japan Go Iwai,
Particle Physics Sector Young-Kee Kim / Greg Bock Leadership Team Strategic Planning Winter Workshop January 29, 2013.
KEK CC - present and future - Mitsuaki NOZAKi (KEK)
Status and Plans on GRID related activities at KEK
LHC DATA ANALYSIS INFN (LNL – PADOVA)
LHCb Computing Model and Data Handling Angelo Carbone 5° workshop italiano sulla fisica p-p ad LHC 31st January 2008.
Philippe Charpentier CERN – LHCb On behalf of the LHCb Computing Group
NAREGI at KEK and GRID plans
ILD Ichinoseki Meeting
Particle Physics at KISTI
Interoperability of Digital Repositories
High Energy Physics at UTA
Nuclear Physics Data Management Needs Bruce G. Gibbard
 YongPyong-High Jan We appreciate that you give an opportunity to have this talk. Our Belle II computing group would like to report on.
High Energy Physics at UTA
Particle Physics Theory
Using an Object Oriented Database to Store BaBar's Terabytes
Grid related activities at KEK
The LHCb Computing Data Challenge DC06
Presentation transcript:

Data GRID deployment in HEPnet-J Takashi Sasaki Computing Research Center KEK

Who we are? KEK stands for –Kou = High –Enerugi = Energy –Kasokuki Kennkyu Kiko = Accelerator Research Organization We are one of the governmental agencies as like other national universities and national laboratory in Japan since the year 2004 –We are an Inter-University Research Institute Corporation

Major projects at KEK Belle –CP violation K2K, T2K –Neutrino –KEK/Tokai to Kamioka CDF –Hadron collider, top quark –Fermi Lab., US ALTAS –Hadron collider, SUSY –CERN, Switerland J-PARC –Joint project with JAEA –Being built at Tokai ILC –International Linear Collider –Site still note decided International competition Japan has interests to host Lattice QCD –Dedicated IBM blue gene 57.3TFlops Material and life science –Synchrotron radiation Muon and meson science Technology transfer –Medical applications Simulation Accelerator

HENP institutes in Japan KEK is the only central laboratory in Japan Smaller scale centers are exist also –ICEPP(U Tokyo), Riken, Osaka Univ. and a few Majorities are smaller groups in universities –Mostly 1-3 faculties and/or researchers and graduate students No engineers nor no technicians for IT This is not HENP specific, but commonly observed –KEK has a role to offer necessary assistance to them Mostly graduate students in physics are the main human resource to support IT unfortunately

HEPnet-J Originally, KEK organized HEP institutes in Japan to provide the networking among them –We started from 9600bps DECnet in early 1980’s –KEK is one of the first Internet sites and the first web site in Japan (1983? and 1992) This year, Super SInet3 will be introduced with 20Gbps and 10Gbps to main nodes as the final upgrade –Shift to more application oriented rather than the band width –GRID deployment is an issue –Virtual Organization for HEP Japan

History of HEPnet-J 2003 Super SInet (backbone) IP 10Gbps

Belle at KEK

Belle collaboration

Data flow model in Belle At every beam crossing, an interaction between particles happens and final state particles are observed by the detector –Event Different type of interactions may happen at each beam crossings Events are in time sequence Something like one picture in the movie film –Run Something like a role of the movie film Cut at a good file size for later processing (historically a size of a tape, 2GB or 4GB) –Data from the detector (signals) are called as “raw data” Physical properties for each particles are “reconstructed” –Vectorization of images and conversions of units –a signal processing Events are classified into types of interactions (pattern matching) –Data Summary Tape (DST) More condensed events samples are selected from DST –Something like a knowledge discovery in images –Called Mini DST –Detector signals are striped –Sometimes, subset of mini DST, micro DST is produced

Belle data analysis Frequency of reprocessing –Reconstruction from raw data One a year or less –DST production Twice a year or less –Mini DST production Many times –Micro DST production Many times –End users analysis Every day, very many times Monte Carlo production –More than number of real data –More likely CPU intensive jobs Full simulation Fast simulation Event size –40KB in raw data (signal only) Record rate –10MB/sec Accumulated event in total –1 PB

Event processing Reconstruction and DST production is done on site due to large data size Physics analysis jobs are executed locally against miniDST or microDST, and also MC –What they are doing mainly is statistical analysis and visualization of histograms Also software development Official jobs, like MC production, cross the levels –CPU intensive jobs miniDST and microDST production are done by sub- groups and can be localized Most of jobs are integer intensive than floating points –Many branches in the code

Data Distribution Model in Belle Level 0 (a few PB) –Only KEK has raw data and reconstructed data –Whole MC data Level 1 (a few 10TB) –Big institutions may want a replica of DST –Join MC production Level 2 (a few 100GB) –Most of institutions are satisfied with mini DST –Join May join MC production Smaller institutions may satisfied with micro DST even Collaboration wide data set –Raw data –Reconstructed data –DST –MC events (background+ signal) Sub group wide data set –Mini DST –Micro DST –MC events (signals)

GRID deployment at KEK Bare Globus –Up to GT2 and gave up to follow We have our own GRID CA –In production since this January –Accredited by APGRID PMA Two LCG sites and one test bed –KEK-LCG-01 For R&D –KEK-LCG-02 For production Interface to HPSS –Test bed Training and tests NAREGI test bed –Under construction SRB (UCSD) –GSI authentication or password –SRB-DSI became available Works as SRM for the SRB world from LCG side Performance test will be done –Performance tests among RAL, CC-IN2P3 and KEK is on going Gfarm –Collaboration with AIST

GRID deployment ATLAS definitely require LCG/gLite –ICEPP (International Center for Elementary Particle Physics), U of Tokyo will be a tier-2 center of ATLAS They have degraded from tier-1 One professor, one associate professor and a few assistant professors are working on the tier-2 center –No technician, no engineer nor no contractors, but only “physicists” –Can you believe this? –How other ATLAS member institutes, mostly smaller groups, can survive? Belle –Some of the collaborators requested us to support a GRID environment for data distribution and efficient analysis –Some time their collaborators also join either of LHC experiments They want to use the same thing for both

LCG/gLite LCG (LHC Computing GRID) is now based on gLite 3.0. Only middleware available today to satisfy HEP requirements –US people are also developing their own Difficulty –Support Language gaps –Quality assurance –Assumes rich man power

NAREGI What we expect for NAREGI –Better quality –Easier deployment –Better support in the native language What we need but still looks not in NAREGI –File/replica catalogue and data GRID related functionalities Need more assessments Comes a little bit late –Earlier is better for us We need something working today! Require commercial version of PBS for β

First stage plan Ask NAREGI to implement LFC on their middleware –We assume job submission between them will be realized –Share the same file/replica catalogue space between LCG/gLite and NAREGI Move data between them using GridFTP –Try something by ourselves Brute force porting of LFC on NAREGI NAREGI SRB gLite will be tried also Assessments will done for –Command level compatibility (syntax) between NAREGI and gLite –Job description languages –Software in experiments, especially ATLAS How depends on LCG/gLite?

Future strategy ILC, International Linear Collider, will be the target – interoperability among gLite, OSG and NAREGI will be required

Conclusion HE(N)P has a problem to be solved today –GRID seems the solution, however, much human resource consumption is the problem We expect much on NAREGI –Still we cannot escape from gLite –Interoperability is the issue We work on this issue together with NARGI and IN2P3