Belle computing ACAT'2002, June 24-28, 2002, Moscow Pavel Krokovny BINP, Novosibirsk On behalf of Belle Collaboration.

Slides:



Advertisements
Similar presentations
31/03/00 CMS(UK)Glenn Patrick What is the CMS(UK) Data Model? Assume that CMS software is available at every UK institute connected by some infrastructure.
Advertisements

MEG-Review Feb MEG Software Group MEG Software Status Framework for MC, Schedule, DC reconstruction update and Discussion on a ROOT-based offline.
23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
O. Stézowski IPN Lyon AGATA Week September 2003 Legnaro Data Analysis – Team #3 ROOT as a framework for AGATA.
Belle computing upgrade Ichiro Adachi 22 April 2005 Super B workshop in Hawaii.
ACAT 2002, Moscow June 24-28thJ. Hernández. DESY-Zeuthen1 Offline Mass Data Processing using Online Computing Resources at HERA-B José Hernández DESY-Zeuthen.
16 April 2005 APS 2005 Search for exclusive two body decays of B→D s * h at Belle Luminda Kulasiri University of Cincinnati Outline Motivation Results.
June 2000Status of BELLE1 Beyond e + e - Workshop The Homestead Glen Arbor, MI Daniel Marlow BELLE Princeton University June 17, 2000.
July 2000Results from BELLE1 IV th Rencontres du Vietnam Hanoi, Vietnam Daniel Marlow BELLE Princeton University July 22, 2000.
Large scale data flow in local and GRID environment V.Kolosov, I.Korolko, S.Makarychev ITEP Moscow.
Hall D Online Data Acquisition CEBAF provides us with a tremendous scientific opportunity for understanding one of the fundamental forces of nature. 75.
The D0 Monte Carlo Challenge Gregory E. Graham University of Maryland (for the D0 Collaboration) February 8, 2000 CHEP 2000.
CERN - European Laboratory for Particle Physics HEP Computer Farms Frédéric Hemmer CERN Information Technology Division Physics Data processing Group.
SuperKEKB to search for new sources of flavor mixing and CP violation - Introduction - Introduction - Motivation for L= Motivation for L=
KLOE Offline Status Report Data Reconstruction MonteCarlo updates MonteCarlo production Computing Farm C.Bloise – May 28th 2003.
Online Data Challenges David Lawrence, JLab Feb. 20, /20/14Online Data Challenges.
HEP GRID CHEP, KNU 11/9/2002 Youngjoon Kwon (Yonsei Univ.) 1 Belle Computing / Data Handling  What is Belle and why we need large-scale computing?
W.Smith, U. Wisconsin, ZEUS Computing Board Zeus Executive, March 1, ZEUS Computing Board Report Zeus Executive Meeting Wesley H. Smith, University.
GLAST LAT ProjectDOE/NASA Baseline-Preliminary Design Review, January 8, 2002 K.Young 1 LAT Data Processing Facility Automatically process Level 0 data.
03/27/2003CHEP20031 Remote Operation of a Monte Carlo Production Farm Using Globus Dirk Hufnagel, Teela Pulliam, Thomas Allmendinger, Klaus Honscheid (Ohio.
CDF data production models 1 Data production models for the CDF experiment S. Hou for the CDF data production team.
November 7, 2001Dutch Datagrid SARA 1 DØ Monte Carlo Challenge A HEP Application.
D0 Farms 1 D0 Run II Farms M. Diesburg, B.Alcorn, J.Bakken, T.Dawson, D.Fagan, J.Fromm, K.Genser, L.Giacchetti, D.Holmgren, T.Jones, T.Levshina, L.Lueking,
Farm Management D. Andreotti 1), A. Crescente 2), A. Dorigo 2), F. Galeazzi 2), M. Marzolla 3), M. Morandin 2), F.
An Overview of PHENIX Computing Ju Hwan Kang (Yonsei Univ.) and Jysoo Lee (KISTI) International HEP DataGrid Workshop November 8 ~ 9, 2002 Kyungpook National.
3rd June 2004 CDF Grid SAM:Metadata and Middleware Components Mòrag Burgon-Lyon University of Glasgow.
LHCb computing in Russia Ivan Korolko (ITEP Moscow) Russia-CERN JWGC, October 2005.
Jiawen Zhang, IHEP, 2008, April 10, frascati Status of BEPCII/BESIII and Physics preparation Jiawen Zhang 2008/4/7—10 , PHIPSI08, Frascati.
8th November 2002Tim Adye1 BaBar Grid Tim Adye Particle Physics Department Rutherford Appleton Laboratory PP Grid Team Coseners House 8 th November 2002.
The Computing System for the Belle Experiment Ichiro Adachi KEK representing the Belle DST/MC production group CHEP03, La Jolla, California, USA March.
Ichiro Adachi ACAT03, 2003.Dec.021 Ichiro Adachi KEK representing for computing & DST/MC production group ACAT03, KEK, 2003.Dec.02 Belle Computing System.
ALICE Upgrade for Run3: Computing HL-LHC Trigger, Online and Offline Computing Working Group Topical Workshop Sep 5 th 2014.
David N. Brown Lawrence Berkeley National Lab Representing the BaBar Collaboration The BaBar Mini  BaBar  BaBar’s Data Formats  Design of the Mini 
CDF Offline Production Farms Stephen Wolbers for the CDF Production Farms Group May 30, 2001.
Management System of Event Processing and Data Files Based on XML Software Tools at Belle Ichiro Adachi, Nobu Katayama, Masahiko Yokoyama IPNS, KEK, Tsukuba,
9 February 2000CHEP2000 Paper 3681 CDF Data Handling: Resource Management and Tests E.Buckley-Geer, S.Lammel, F.Ratnikov, T.Watts Hardware and Resources.
KLOE Computing Update Paolo Santangelo INFN LNF KLOE General Meeting University of Rome 2, Tor Vergata 2002, December
EGEE is a project funded by the European Union under contract IST HEP Use Cases for Grid Computing J. A. Templon Undecided (NIKHEF) Grid Tutorial,
The LHCb CERN R. Graciani (U. de Barcelona, Spain) for the LHCb Collaboration International ICFA Workshop on Digital Divide Mexico City, October.
Hall-D/GlueX Software Status 12 GeV Software Review III February 11[?], 2015 Mark Ito.
The KLOE computing environment Nuclear Science Symposium Portland, Oregon, USA 20 October 2003 M. Moulson – INFN/Frascati for the KLOE Collaboration.
KISTI & Belle experiment Eunil Won Korea University On behalf of the Belle Collaboration.
Status of OPAL data Matthias Schröder
PC clusters in KEK A.Manabe KEK(Japan). 22 May '01LSCC WS '012 PC clusters in KEK s Belle (in KEKB) PC clusters s Neutron Shielding Simulation cluster.
HIGUCHI Takeo Department of Physics, Faulty of Science, University of Tokyo Representing dBASF Development Team BELLE/CHEP20001 Distributed BELLE Analysis.
Particle Identification at BESIII Kanglin He April 23, 2007, Amsterdam.
Offline Status Report A. Antonelli Summary presentation for KLOE General Meeting Outline: Reprocessing status DST production Data Quality MC production.
Status report of the KLOE offline G. Venanzoni – LNF LNF Scientific Committee Frascati, 9 November 2004.
5 June 2003Alan Norton / Focus / EP Topics1 Other EP Topics Some 2003 Running Experiments - NA48/2 (Flavio Marchetto) - Compass (Benigno Gobbo) - NA60.
Computing Resources for ILD Akiya Miyamoto, KEK with a help by Vincent, Mark, Junping, Frank 9 September 2014 ILD Oshu City a report on work.
Large scale data flow in local and GRID environment Viktor Kolosov (ITEP Moscow) Ivan Korolko (ITEP Moscow)
Computing Issues for the ATLAS SWT2. What is SWT2? SWT2 is the U.S. ATLAS Southwestern Tier 2 Consortium UTA is lead institution, along with University.
Computing for Belle CHEP2004 September 27, 2004 Nobu Katayama KEK.
A High Statistics Study of the Decay M. Fujikawa for the Belle Collaboration Outline 1.Introduction 2.Experiment Belle detector 3.Analysis Event selection.
D0 Farms 1 D0 Run II Farms M. Diesburg, B.Alcorn, J.Bakken, R. Brock,T.Dawson, D.Fagan, J.Fromm, K.Genser, L.Giacchetti, D.Holmgren, T.Jones, T.Levshina,
First results from CMD-3 detector at VEPP-2000 collider Budker Institute of Nuclear Physics, SB RAS, Novosibirsk, Russia September 2011 E.Solodov (for.
Preservation of LEP Data There is still hope Is there? Marcello Maggi, Ulrich Schwickerath, Matthias Schröder, , DPHEP7 1.
Markus Frank (CERN) & Albert Puig (UB).  An opportunity (Motivation)  Adopted approach  Implementation specifics  Status  Conclusions 2.
Jianming Qian, UM/DØ Software & Computing Where we are now Where we want to go Overview Director’s Review, June 5, 2002.
Status Report on Data Reconstruction May 2002 C.Bloise Results of the study of the reconstructed events in year 2001 Data Reprocessing in Y2002 DST production.
MAUS Status A. Dobbs CM43 29 th October Contents MAUS Overview Infrastructure Geometry and CDB Detector Updates CKOV EMR KL TOF Tracker Global Tracking.
The Software Library of the Belle II Experiment
BESIII data processing
Overview of the Belle II computing
SuperB and its computing requirements
The COMPASS event store in 2002
Special edition: Farewell for Stephen Bailey
HEC Beam Test Software schematic view T D S MC events ASCII-TDS
Development of LHCb Computing Model F Harris
Presentation transcript:

Belle computing ACAT'2002, June 24-28, 2002, Moscow Pavel Krokovny BINP, Novosibirsk On behalf of Belle Collaboration

Goal of B-factory CPV due to complex phases in CKM matrix 11 22 33 V td V tb V cd V cb V ud V ub * * * 1) Establish CPV 2) Precise/Redundant Measurement of CKM : angles and length 3) Beyond SM Last summer ! Next Step CKM matrix Unitarity triangle

The Belle Collaboration A World-Wide Activity Involving 50 Institutions ~300 members BINP

KEKB asymmetric e + e  collider Two separate rings e  (HER) : 8.0 GeV e + (LER) : 3.5 GeV E CM : GeV at  (4S) Design: Luminosity: cm -2 s -1 Current: 2.6 / 1.1A ±11 mrad crossing angle (LER HER) Beam size:  y  3  m  x  100  m  = 0.425

Integrated Luminosity & Data 400pb -1 /day Integrated luminosity/day Total accumulated luminosity 88 fb -1

Reconstructed Two-B  event

KEKB computer system

Sparc CPUs Belle’s reference platform –Solaris workgroup servers (500Hz, 4CPU) 38 compute servers (500Hz, 4CPU) –LSF batch system –40 tape drives (2 each on 20 servers) Fast access to disk servers

Intel CPUs Compute servers Linux RH 6.2/7.2) –4 CPU (Pentium Xeon MHz) servers~96 units –2 CPU (Pentium III 0.8~1.26GHz) servers~167 units User terminals to log onto the group servers) –106 PCs (~50Win2000+X window, ~60 Linux) Compute/file servers at universities –A few to a few each institution –Used in generic MC production as well as physics analyses at each institution –Novosibirsk: one group server used for analyses and callibration + user terminals

Belle jargon, data sizes Raw: 30KB average DST: 120KB/ hadronic event mDST: 10(21)KB/ hadronic(BBbar MC) event –zlib compressed, four vector + physics information only –(i.e. tracks, photons, etc) production/reprocess –Rerun all reconstruction code –reprocess: process ALL events using a new version of software generic MC –QQ (jetset c,u,d,s pairs/ generic B decays) –used for background study

Data storage requirements Raw data 1GB/pb  1 (100TB for 100 fb  1 ) DST:1.5GB/pb  1 /copy (150TB for 100 fb  1 ) Skims for calibration:1.3GB/pb  1 mDST:45GB/fb  1 (4.5TB for 100 fb  1 ) Other physics skims:~30GB/fb  1 Generic MC:MDST: ~10TB/year

Disk 8TB NFS file servers 120TB HSM (4.5TB staging disk) –DST skims –User data files 500TB tape library (direct access) –40 tape drives on 20 sparc servers –DTF2: 200GB/tape, 24MB/s IO speed –Raw, DST files –generic MC files are stored and read by users(batch jobs) ~12TB local data disks on PCs –Not used efficiently at this point

Software C++ –gcc3 (compiles with SunCC) No commercial software –QQ, (EvtGen), GEANT3, CERNLIB, CLHEP, Postgres Legacy FORTRAN code –GSIM/GEANT3/ and old calibration/reconstruction code) I/O: home-grown serial I/O package+zlib –The only data format for all stages (from DAQ to final user analysis skim files) Framework: Basf

Framework (BASF) Event parallelism on SMP (1995~) –Using fork (for legacy Fortran common blocks) Event parallelism on multi-compute servers (dbasf, 2001~) Users code/reconstruction code are dynamically loaded The only framework for all processing stages (from DAQ to final analysis)

DST production cluster I/O server is Sparc –Input rate: 2.5MB/s 15 compute servers –4 Pentium III Xeon 0.7GHz 200 pb  1 /day Several such clusters may be used to process DST Using perl and postgres to manage production Overhead at the startup time –Wait for comunication –Database access –Need optimization Single output stream

Belle Software Library CVS (no remote check in/out) –Check-ins are done by authorized persons A few releases (two major releases last year) –Usually it takes a few weeks to settle down after a release. It has been left to the developers to check the “new” version” of the code. We are now trying to establish a procedure to compare against old versions –All data are reprocessed/All generic MC are regenerated with a new major release of the software (at most once per year, though)

DST production 300GHz Pentium III~1fb  1 /day Need ~40 4CPU servers to keep up with data taking at this moment Reprocessing strategy –Goal: 3 months to reprocess all data using all KEK computing servers –Often limited by determination of calibration constants

Skims Calibration skims (DST level) –QED: (Radiative) Bhabha, (Radiative) Mupair –Tau, Cosmic, Low multiplicity, Random Physics skims (mDST level) –Hadron A, B, C (from loose to very tight cut), –J/ , Low multiplicity, ,  c etc Users skims (mDST level) –For the physics analysis

Data quality monitor DQM (online data quality monitor) –run by run histograms for sub detectors –viewed by shifters and detector experts QAM (offline quality assurance monitor) –data quality monitor using DST outputs –WEB based –Viewed by detector experts and monitoring group –histograms, run dependence

MC production 400GHz Pentium III~1fb  /day 240GB/fb  data in the compressed format No intermediate (GEANT3 hits/raw) hits are kept. –When a new release of the library comes, we have to produce new generic MC sample For every real data taking run, we try to generate 3 times as many events as in the real run, taking into account: –Run dependence –Detector background are taken from random trigger events of the run being simulated

Postgres database system The only database system –Other than simple UNIX files and directories –Recently moved from version 6 to 7 –A few years ago, we were afraid that nobody use Postgres but it seems Postgres is the only database on Linux and is well maintained One master, one copy at KEK, many copies at institutions/on personal PCs –~20 thousand records –IP profile is the largest/most popular

Reconstruction software 30~40 people have contributed in the last few years For most reconstruction software, we only have one package, except for muon identification software. Very little competition –Good and bad Identify weak points and ask someone to improve it –Mostly organized within the sub detector groups –Physics motivated, though Systematic effort to improve tracking software but very slow progress

Analysis software Several people have contributed –Kinematical and vertex fitter –Flavor tagging –Vertexing –Particle ID (Likelihood) –Event shape –Likelihood/Fisher analysis People tend to use standard packages

Human resources KEKB computer system + Network –Supported by the computer center (1 researcher, 6~7 system engineers+1 hardware eng., 2~3 operators) PC farms and Tape handling –2 Belle support staffs (they help productions as well) DST/MC production management –2 KEK/Belle researchers, 1 postdoc or student at a time from collaborating institutions Library/Constants database –2 KEK/Belle researchers + sub detector groups

Networks KEKB computer system –internal NFS network –user network –inter compute server network –firewall KEK LAN, WAN, Firewall, Web servers Special network to a few remote institutions –Hope to share KEKB comp. disk servers with remote institutions via NFS TV conference, moving to the H323 IP conference Now possible to participate form Novosibirsk!

Data transfer to universities A firewall and login servers make the data transfer miserable (100Mbps max.) DAT tapes to copy compressed hadron files and MC generated by outside institutions Dedicated GbE network to a few institutions are now being added Total 10Gbit to/from KEK being added Slow network to most of collaborators (Novosibirsk: 0.5Mbps)

Plans More CPU for DST/MC production Distributed analysis (with local data disks) Better constants management More man power on reconstruction software and everything else –Reduce systematic errors, better efficiencies

Summary

1 Day Accelerator Performance Snapshot 1 Day Accelerator Performance Snapshot

The Belle Detector SVD: 3 DSSD lyr  ~  55  m CDC: 50 layers TOF:  ~  95ps Aerogel (n = 1.01 ~ 1.03) CsI :  E /E  ~ 1.8%   (dE/dx)  ~  7   p /p ~ 0.35% 1.5T B-field K/  ~3.5 Gev/c KLM: RPC 14 lyr

Three phases of KEKB L=5x > 10 35