YongPyong-High1 2015 Jan. 30. 2015 We appreciate that you give an opportunity to have this talk. Our Belle II computing group would like to report on.

Slides:



Advertisements
Similar presentations
Jiri Chudoba for the Pierre Auger Collaboration Institute of Physics of the CAS and CESNET.
Advertisements

December 17th 2008RAL PPD Computing Christmas Lectures 11 ATLAS Distributed Computing Stephen Burke RAL.
The SAMGrid Data Handling System Outline:  What Is SAMGrid?  Use Cases for SAMGrid in Run II Experiments  Current Operational Load  Stress Testing.
3rd June 2004 CDF Grid SAM:Metadata and Middleware Components Mòrag Burgon-Lyon University of Glasgow.
Jean-Yves Nief CC-IN2P3, Lyon HEPiX-HEPNT, Fermilab October 22nd – 25th, 2002.
Marianne BargiottiBK Workshop – CERN - 6/12/ Bookkeeping Meta Data catalogue: present status Marianne Bargiotti CERN.
8th November 2002Tim Adye1 BaBar Grid Tim Adye Particle Physics Department Rutherford Appleton Laboratory PP Grid Team Coseners House 8 th November 2002.
INFSO-RI Enabling Grids for E-sciencE Project Gridification: the UNOSAT experience Patricia Méndez Lorenzo CERN (IT-PSS/ED) CERN,
1 st December 2003 JIM for CDF 1 JIM and SAMGrid for CDF Mòrag Burgon-Lyon University of Glasgow.
Belle II Data Management System Junghyun Kim, Sunil Ahn and Kihyeon Cho * (on behalf of the Belle II Computing Group) *Presenter High Energy Physics Team.
ATLAS and GridPP GridPP Collaboration Meeting, Edinburgh, 5 th November 2001 RWL Jones, Lancaster University.
Bookkeeping Tutorial. Bookkeeping & Monitoring Tutorial2 Bookkeeping content  Contains records of all “jobs” and all “files” that are created by production.
GridPP18 Glasgow Mar 07 DØ – SAMGrid Where’ve we come from, and where are we going? Evolution of a ‘long’ established plan Gavin Davies Imperial College.
F. Fassi, S. Cabrera, R. Vives, S. González de la Hoz, Á. Fernández, J. Sánchez, L. March, J. Salt, A. Lamas IFIC-CSIC-UV, Valencia, Spain Third EELA conference,
BESIII Production with Distributed Computing Xiaomei Zhang, Tian Yan, Xianghu Zhao Institute of High Energy Physics, Chinese Academy of Sciences, Beijing.
DIRAC Review (13 th December 2005)Stuart K. Paterson1 DIRAC Review Exposing DIRAC Functionality.
International Workshop on HEP Data Grid Nov 9, 2002, KNU Data Storage, Network, Handling, and Clustering in CDF Korea group Intae Yu*, Junghyun Kim, Ilsung.
Status of the LHCb MC production system Andrei Tsaregorodtsev, CPPM, Marseille DataGRID France workshop, Marseille, 24 September 2002.
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
22 nd September 2003 JIM for CDF 1 JIM and SAMGrid for CDF Mòrag Burgon-Lyon University of Glasgow.
Dzero MC production on LCG How to live in two worlds (SAM and LCG)
The Advanced Data Searching System The Advanced Data Searching System with 24 February APCTP 2010 J.H Kim & S. I Ahn & K. Cho on behalf of the Belle-II.
Metadata requirements for HEP Paul Millar. Slide 2 12 September 2007 Metadata requirements for HEP Some of the players in this game... WLCG – Umbrella.
Report on Belle II Data Handling Group Kihyeon Cho High Energy Physics Team KISTI (Korea Institute of Science and Technology Information) June 16~18, 2010.
Distributed Computing for CEPC YAN Tian On Behalf of Distributed Computing Group, CC, IHEP for 4 th CEPC Collaboration Meeting, Sep , 2014 Draft.
ATLAS is a general-purpose particle physics experiment which will study topics including the origin of mass, the processes that allowed an excess of matter.
Hussein Suleman University of Cape Town Department of Computer Science Digital Libraries Laboratory February 2008 Data Curation Repositories:
HIGUCHI Takeo Department of Physics, Faulty of Science, University of Tokyo Representing dBASF Development Team BELLE/CHEP20001 Distributed BELLE Analysis.
Managing Data DIRAC Project. Outline  Data management components  Storage Elements  File Catalogs  DIRAC conventions for user data  Data operation.
Integration of the ATLAS Tag Database with Data Management and Analysis Components Caitriana Nicholson University of Glasgow 3 rd September 2007 CHEP,
Performance of The NorduGrid ARC And The Dulcinea Executor in ATLAS Data Challenge 2 Oxana Smirnova (Lund University/CERN) for the NorduGrid collaboration.
UTA MC Production Farm & Grid Computing Activities Jae Yu UT Arlington DØRACE Workshop Feb. 12, 2002 UTA DØMC Farm MCFARM Job control and packaging software.
Jean-Roch Vlimant, CERN Physics Performance and Dataset Project Physics Data & MC Validation Group McM : The Evolution of PREP. The CMS tool for Monte-Carlo.
Bookkeeping Tutorial. 2 Bookkeeping content  Contains records of all “jobs” and all “files” that are produced by production jobs  Job:  In fact technically.
Summary of Software and tracking activities Tokusui Workshop December 2015 Akiya Miyamoto KEK.
The GridPP DIRAC project DIRAC for non-LHC communities.
David Adams ATLAS ATLAS-ARDA strategy and priorities David Adams BNL October 21, 2004 ARDA Workshop.
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
Distributed Analysis Tutorial Dietrich Liko. Overview  Three grid flavors in ATLAS EGEE OSG Nordugrid  Distributed Analysis Activities GANGA/LCG PANDA/OSG.
Finding Data in ATLAS. May 22, 2009Jack Cranshaw (ANL)2 Starting Point Questions What is the latest reprocessing of cosmics? Are there are any AOD produced.
Jiri Chudoba for the Pierre Auger Collaboration Institute of Physics of the CAS and CESNET.
D.Spiga, L.Servoli, L.Faina INFN & University of Perugia CRAB WorkFlow : CRAB: CMS Remote Analysis Builder A CMS specific tool written in python and developed.
The GridPP DIRAC project DIRAC for non-LHC communities.
ATLAS Distributed Analysis DISTRIBUTED ANALYSIS JOBS WITH THE ATLAS PRODUCTION SYSTEM S. González D. Liko
Simulation Production System Science Advisory Committee Meeting UW-Madison March 1 st -2 nd 2007 Juan Carlos Díaz Vélez.
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
Breaking the frontiers of the Grid R. Graciani EGI TF 2012.
Belle II Computing Fabrizio Bianchi INFN and University of Torino Meeting Belle2 Italia 17/12/2014.
CDF SAM Deployment Status Doug Benjamin Duke University (for the CDF Data Handling Group)
Joe Foster 1 Two questions about datasets: –How do you find datasets with the processes, cuts, conditions you need for your analysis? –How do.
HEP Group Activity Kihyeon Cho (KISTI) KISTI-Belle II Open Meeting, KISTI, Daejeon, Korea Monday, February 23, 2014.
CERN IT Department CH-1211 Genève 23 Switzerland t EGEE09 Barcelona ATLAS Distributed Data Management Fernando H. Barreiro Megino on behalf.
on behalf of the Belle II Computing Group
Simulation Production System
Overview of the Belle II computing
Belle II Physics Analysis Center at TIFR
Akiya Miyamoto KEK 1 June 2016
Data Challenge with the Grid in ATLAS
Work report Xianghu Zhao Nov 11, 2014.
Future challenges for the BELLE II experiment
Muon stopping target optimization
LHCb Computing Model and Data Handling Angelo Carbone 5° workshop italiano sulla fisica p-p ad LHC 31st January 2008.
Philippe Charpentier CERN – LHCb On behalf of the LHCb Computing Group
Job workflow Pre production operations:
New developments on the LHCb Bookkeeping
EGEE Middleware: gLite Information Systems (IS)
ATLAS DC2 & Continuous production
Belle II experiment Requirement of data handling system Belle II Metadata service system Data Cache system at Belle II experiment Summary.
The LHCb Computing Data Challenge DC06
Presentation transcript:

 YongPyong-High1 2015 Jan. 30. 2015 We appreciate that you give an opportunity to have this talk. Our Belle II computing group would like to report on Distributed computing and data handling at Belle II experiment in this talk.

we want to introduce Belle II experiment, computing model, distributed computing and data handling and MC production campaign.

Belle II detector KEKB accelerator has 3Km circulation and supply 8.0 GeV electron and 3.5 GeV positron beams for Belle II.

Kobayashi-maskawa won a Nobel Prize from Belle result, which is CP violation measurement, at 2008. Furthermore, Belle will be upgraded to Belle II searching the New Physics from precision measurement. Belle II will reach 50 ab^-1 in 2022. It is corresponding with 50 times of Belle.

Belle II collaboration is joined in 530 members of 94 institutions and they come from 23 countries.

Our Computing group of Belle II designed the computing model as you can see this page. The model is formed of 3 stages. First, KEK site will be produced the raw data from detector. Then the raw data will be saved in tape and storage. This data will be processed to make refined data, mdst. In addition, PNNL, USA site will back raw data up. Second, mdst will be transferred into grid site. Also, grid sites will generate Monte carlo and store it to be processed by end-users. Last, Processing results, ntuples will go to local resource, user interface for analysis. Optionally, we consider Cloud system for generating large MC production.

This is Belle II distributed computing system, This system is constructed with 4 dirac server, 3 AMGA server, LCG sites from Europe and OSG sites from USA as you can see this page.

To realize the computing system, we developed “gbasf2” for distributed computing and data handling. Gbasf2 is the commandline client for submitting grid based basf2 jobs. It is based on existing, well-proven solutions plus extensions for Belle II. Gbasf2 is constructed with DIRAC for job management, AMGA for meta data catalog and CVMFS for software distribution.

Gbasf2 is also based on projects and dataset is software for job distribution. Datset is a collection of basf2 data which are stored in output files. Project is a series of gbasf2 jobs to handle data set. In a project, gbasf2 can read input dataset and generate output dataset. Gbasf2 has some functionalities as job submission, job monitoring, rescheduling and etc. you can see the commands and results in project level at right side picture.

We are still developing the metadata schema based on projects and dataset. Our goal is to book the data information automatically into metadata system and LFC. A dataset registeration is required for Belle II data analysis on the distributed computing.

Our computing group applied AMGA in gbasf2, a part of data handling, to improve the scalability and performance of metadata catalog. Left side picture shows the Data handling system and workflow of gbasf2. Right side picture shows how AMGA is worked in gbasf2.

In addition, Korean group developed AMGA manager that aims at providing an interactive exploration and searching environment for metadata in an user-friendly manner, and hiding complexities for accessing Grid service. It allows users to manipulate metadata schema, entries, attributes, access control, constraints, user and group information through a user-friendly GUI.

Finally, We have the big tests, two MC production Campaign Finally, We have the big tests, two MC production Campaign. First MC production Campaign was taken from February 28th to March 19th this year. We performed event generation and detector simulation in 1st stage. Reconstruction was performed with 240k jobs corresponding with 60M events and 190TB of output data in 2nd stage. We found around 20% failure rate from metadata registeration, input data download and application errors.

Second MC production Campaign was taken in this summer Second MC production Campaign was taken in this summer. We run Simulation and reconstruction with background mixing. In 2nd stage, we performed 630K jobs and obtained 560M events and 8.5TByes of output data. Dramatically, we reduce the failure rate around 10% to 1%.

We want to introduce Korean group activity, AMGA system We want to introduce Korean group activity, AMGA system. It is operated well as a part of gbasf2 in MC production campaign as you can see the picture.

Also, Our Korean site, KISTI resource give a contribution for MC production campaign even if it is small.

Belle II still search for New physics with 50 times more data than current B factories. We designed computing model with projects and datasets. Belle II computing group developed “gbasf2” for distributed computing and data handling based on Dirac, AMGA and CVMFS. We had two MC production campaigns this year. In two tests, Belle II distributed computing system works well. Thanks a lot to technology and resource providers!