1 March 8-9, 2011 Soonwook Hwang KISTI FKPPL VO: Status and Perspectives of Grid Infrastructure The 3 rd FKPPL KIAS.

Slides:



Advertisements
Similar presentations
FP7-INFRA Enabling Grids for E-sciencE EGEE Induction Grid training for users, Institute of Physics Belgrade, Serbia Sep. 19, 2008.
Advertisements

Polish Infrastructure for Supporting Computational Science in the European Research Space EUROPEAN UNION Services and Operations in Polish NGI M. Radecki,
GRID INTEROPERABILITY USING GANGA Soonwook Hwang (KISTI) YoonKee Lee and EunSung Kim (Seoul National Uniersity) KISTI-CCIN2P3 FKPPL Workshop December 1,
Oct 24 MOST MeetingDenis Perret-Gallix Asia-Pacific Cooperation.
Overview of LCG-France Tier-2s and Tier-3s Frédérique Chollet (IN2P3-LAPP) on behalf of the LCG-France project and Tiers representatives CMS visit to Tier-1.
Korea CDF Workshop, Kyungpook National University April KIhyeon Cho (KISTI) 한 프랑스 CDF 공동연구 1.
KISTI’s Activities on the NA4 Biomed Cluster Soonwook Hwang, Sunil Ahn, Jincheol Kim, Namgyu Kim and Sehoon Lee KISTI e-Science Division.
FKPPL workshop May 2012 BUI The Quang Prof. Vincent Breton Prof. Doman Kim Prof. NGUYEN Hong Quang Prof. PHAM Quoc Long Grid enabled in silico drug discovery.
Computing for ILC experiment Computing Research Center, KEK Hiroyuki Matsunaga.
Nicholas LoulloudesMarch 3 rd, 2009 g-Eclipse Testing and Benchmarking Grid Infrastructures using the g-Eclipse Framework Nicholas Loulloudes On behalf.
INFSO-RI Enabling Grids for E-sciencE The US Federation Miron Livny Computer Sciences Department University of Wisconsin – Madison.
11 ALICE Computing Activities in Korea Beob Kyun Kim e-Science Division, KISTI
:: ::::: ::::: ::::: ::::: ::::: ::::: ::::: ::::: ::::: ::::: ::::: :: GridKA School 2009 MPI on Grids 1 MPI On Grids September 3 rd, GridKA School 2009.
Enabling Grids for E-sciencE ENEA and the EGEE project gLite and interoperability Andrea Santoro, Carlo Sciò Enea Frascati, 22 November.
Grid Technologies  Slide text. What is Grid?  The World Wide Web provides seamless access to information that is stored in many millions of different.
F. Fassi, S. Cabrera, R. Vives, S. González de la Hoz, Á. Fernández, J. Sánchez, L. March, J. Salt, A. Lamas IFIC-CSIC-UV, Valencia, Spain Third EELA conference,
BESIII Production with Distributed Computing Xiaomei Zhang, Tian Yan, Xianghu Zhao Institute of High Energy Physics, Chinese Academy of Sciences, Beijing.
Responsibilities of ROC and CIC in EGEE infrastructure A.Kryukov, SINP MSU, CIC Manager Yu.Lazin, IHEP, ROC Manager
1 Soonwook Hwang, Hangi Kim KISTI Seyong Kim Sejong University XQCD Workshop 2009 August 3-5, 2009 Large-scale Deployment of two-color QCD on the FKPPL.
EGEE-II INFSO-RI Enabling Grids for E-sciencE An Introduction to the EGEE Project Presented by Min Tsai ISGC 2007, Taipei With thanks.
09/02 ID099-1 September 9, 2002Grid Technology Panel Patrick Dreher Technical Panel Discussion: Progress in Developing a Web Services Data Analysis Grid.
June 24-25, 2008 Regional Grid Training, University of Belgrade, Serbia Introduction to gLite gLite Basic Services Antun Balaž SCL, Institute of Physics.
T3 analysis Facility V. Bucard, F.Furano, A.Maier, R.Santana, R. Santinelli T3 Analysis Facility The LHCb Computing Model divides collaboration affiliated.
Overview of grid activities in France in relation to FKPPL FKPPL Workshop Thursday February 26th, 2009 Dominique Boutigny.
Grid User Interface for ATLAS & LHCb A more recent UK mini production used input data stored on RAL’s tape server, the requirements in JDL and the IC Resource.
US LHC OSG Technology Roadmap May 4-5th, 2005 Welcome. Thank you to Deirdre for the arrangements.
11 November 2010 Natascha Hörmann Computing at HEPHY Evaluation 2010.
Glite. Architecture Applications have access both to Higher-level Grid Services and to Foundation Grid Middleware Higher-Level Grid Services are supposed.
E-Science paradigm for CDF experiment Kihyeon Cho High Energy Physics Team KISTI (Korea Institute of Science and Technology Information) October 8,2010.
Development of e-Science Application Portal on GAP WeiLong Ueng Academia Sinica Grid Computing
ALICE-Korea Activities and Plans Yong-Wook Baek Kangnung National University* and L.P.C. Clermont-Ferrand* *France-Korea Particle Physics Laboratory (International.
DataGrid is a project funded by the European Commission under contract IST rd EU Review – 19-20/02/2004 The EU DataGrid Project Three years.
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI VO auger experience with large scale simulations on the grid Jiří Chudoba.
Breaking the frontiers of the Grid R. Graciani EGI TF 2012.
EGEE-II INFSO-RI Enabling Grids for E-sciencE Overview of gLite, the EGEE middleware Mike Mineter Training Outreach Education National.
Geant4 Simulation Study for Proton Therapy Proton Therapy Center Se Byeong Lee Medical Physicist, Ph.D. The 3 rd FKPPL workshop, March 8-9, 2011, Seoul,
Geant4 GRID production Sangwan Kim, Vu Trong Hieu, AD At KISTI.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Introduction Salma Saber Electronic.
Multi-community e-Science service connecting grids & clouds R. Graciani 1, V. Méndez 2, T. Fifield 3, A. Tsaregordtsev 4 1 University of Barcelona 2 University.
Enabling Grids for E-sciencE Claudio Cherubino INFN DGAS (Distributed Grid Accounting System)
GRID & Parallel Processing Koichi Murakami11 th Geant4 Collaboration Workshop / LIP - Lisboa (10-14/Oct./2006) 1 GRID-related activity in Japan Go Iwai,
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI EGI Overview for ENVRI Gergely Sipos, Malgorzata Krakowian EGI.eu
1 The Life-Science Grid Community Tristan Glatard 1 1 Creatis, CNRS, INSERM, Université de Lyon, France The Spanish Network for e-Science 2/12/2010.
GridMaGrid Users & Applications Conclusions 16/ Grid activities in Morocco Abderrahman El Kharrim CNRST - MaGrid Team Morocco Grid Workshop - Rabat,
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI EGI solution for high throughput data analysis Peter Solagna EGI.eu Operations.
Accessing the VI-SEEM infrastructure
Grid2Win Porting of gLite middleware to Windows XP platform
Simulation of secondary beam production at RISP
Open Science Grid Progress and Status
U.S. ATLAS Grid Production Experience
Belle II Physics Analysis Center at TIFR
Data Challenge with the Grid in ATLAS
Roberto Barbera (a nome di Livia Torterolo)
Christos Markou Institute of Nuclear Physics NCSR ‘Demokritos’
Long-term Grid Sustainability
Update on Plan for KISTI-GSDC
Grid2Win: Porting of gLite middleware to Windows XP platform
Introduction to Grid Technology
OpenGATE meeting/Grid tutorial, mars 9nd 2005
Short update on the latest gLite status
Welcome to the third FKPPL workshop
Particle Physics at KISTI
Patrick Dreher Research Scientist & Associate Director
e-Science paradigm for CDF experiment
e-Science for High Energy Physics
Pierre Girard ATLAS Visit
Overview of the EGEE project and the gLite middleware
EGEE Middleware: gLite Information Systems (IS)
Data Processing for CDF Computing
GRIF : an EGEE site in Paris Region
Presentation transcript:

1 March 8-9, 2011 Soonwook Hwang KISTI FKPPL VO: Status and Perspectives of Grid Infrastructure The 3 rd FKPPL KIAS

Introduction to FKPPL  FKPPL (France-Korea Particle Physics Laboratory)  International Associated Laboratory between French and Korean laboratories  Promote joint cooperative activities (research projects) under a scientific research program in the area of  Particle Physics  LHC  ILC  e-Science  Bioinformatics  Grid Computing  Geant4

 FKPPL focuses on particle physics and e-science Both require international collaboration Project nameCoordinatorsPartners ILC calorimeter (particle physics) Jongseo Chai, SKKU C de LA TAILLE, OMEGA-LAL Sung Kyun Kwan Univ., Korea Institute of Radiological and medical Sciencces, Pohang Accel. Lab. OMEGA-LAL, LLR ILC electronics (particle physics) Jongseo Chai, SKKU Christoph de la Taille, LAL Sung Kyun Kwan Univ., Korea Institute of Radiological and medical Sciencces, Pohang Accel. Lab. LAL, LLR Grid computingS. Hwang, KISTI D. Boutigny, CC-IN2P3 KISTI, CC-IN2P3 WISDOM (insilico drug discovery) Doman Kim, CNU V. Breton, LPC Chonnam Nat. Univ., KISTI, Kangnung Nat. Univ., LPC ALICE ( heavy ions physics) Yongwook Baek, KNU Pascal Dupieux, LPC Kangnung Nat. Univ. LPC CDF (particle physics) Kihyeon Cho, KISTI Aurore Savoy-Navarro, LPNHE KISTI, LPNHE CREAM (Astroparticle Physics) Yong-Yeon Keum, Ewha Pierre SALATI, LAPTH Ewha Womans Univ. LAPTH Geant4 (Medical Physics) Se Byeong Lee, NCC Sebastien Incert, IN2P3-CENBG NCC, IN2P3-CENBG, IN2P3-LPC FKPPL scientific projects FKPPL Scientific Projects

Grid FKPPL  Participating Organizations  CC-IN2P3 and KISTI  Group Leaders  Dominique Boutigny, Director of CC-IN2P3, France  Soonwook Hwang, KISTI, Korea  ’10 Budget  France: ~6000 Euro  Mainly traveling cost funded by CNRS  Korea: 20,000,000 Won  traveling cost and organizing grid workshop and training  funded by KRCF under the framework of the CNRS-KRCFST Joint Programme  Common Interest  Joint in Grid computing  Collaboration on ALICE computing: CC-IN2P3 (Tier1) and KISTI (Tier2)  Joint operation and maintenance of production grid infrastructure

Objective  Background  Collaborative work between KISTI in Korea and CC-IN2P3 in France in the area of Grid computing  Objective  Provide computing facilities and user support needed to foster the scientific applications established under the framework of FKPPL collaboration and beyond  Promote the adoption of grid technology and grid awareness in Korea and France by providing scientists and researchers with production Grid infrastructure and technical support necessary for them

FKPPL VO Grid  Built based on the gLite middleware services  Has been up and running since October 2008, providing ~10,000 CPU cores and ~30 TBytes of disk storage  Since last December, KEK has joined FKPPL VO, contributing ~1,600 CPU cores and 27 Tbytes of Disks VOMS WMS CESEUICESE FKPPL VO LFC WIKI CE WMS SE

FKPPL VO Grid Testbed ServiceHostSite UIkenobi.kisti.re.krKISTI ncc-ui.ncc.re.krNCC VOMSpalpatine.kisti.re.krKISTI WMS/LBsnow.kisti.re.kr, snow02, snow03(LB)KISTI kek2-wms.cc.kek.jpKEK SEccsrm02.in2p3.fr (0.5TB)CC-IN2P3 kek2-se.cc.kek.jp (26.6TB)KEK hansolo.kisti.re.kr(3TB)KISTI CECclcgceli01[3].in2p3.fr (8000 CPU cores)CC-IN2P3 Kek2-ce01[5|6].cc.kek.jp (1600 CPU cores)KEK darthvader.kisti.re.kr (128 CPU cores)KISTI VOMS WMS CESEUICESE FKPPL VO KISTI IN2P3 LFC WIKI WMS SE KEK CE

gLite Grid Services on FKPPL VO User Interface (UI) User Interface (UI): The place where users logon to access the Grid Workload Management System (WMS) Workload Management System (WMS): Matches the user requirements with the available resources on the Grid File and replica catalog File and replica catalog: Location of grid files and grid file replicas Computing Element (CE) Computing Element (CE): A batch queue on a site’s computers where the user’s job is executed Storage Element (SE) Storage Element (SE): provides (large-scale) storage for files

9 Computing Element Storage Element IN2P3 Information System Submit job (executable + small inputs) Submit job query Retrieve output WMS User Interface publish state File and Replica Catalog VO Management Service (DB of VO users) query create proxy process Retrieve status & (small) output files Logging and bookkeeping Job status Logging Input file(s) Output file(s) Register file Job Submission Example

Resource and Service CC-IN2P3  CPU Used  ~5.1 millions of hours HS06  Number of Jobs executed  193,434  72 years for 1 processor intel Xenon 2.5 GHz

Number of Jobs

Cumulative CPU usage

User Support  FKPPL VO Wiki site   User Accounts on UI  104 User accounts has been created  FKPPL VO Membership Registration  70 Users have been registered at FKPPL VO membership

Grid Training (1/2)  In February 2010, we organized Geant4 and Grid tutorial 2010 for Korean medical physics communities  Co-hosted by KISTI and NCC  About 34 participants from major hospitals in Korea  About 20 new users joined the FKPPL VO membership

Grid Training (2/2)  “2010 Summer Training Course on Geant4, GATE and Grid computing” held in Seoul in July  Co-hosted by KISTI and NCC  About 50 participants from about 20 institutes in Korea

Application porting Support on FKPPL VO  Deployment of Geant4 applications  Used extensively by the National Cancel Center in Korea to carry out compute-intensive simulations relevant to cancer treatment planning  In collaboration with National Cancer Center in Korea  Deployment of two-color QCD (Quantum ChromoDynamics) simulations in theoretical Physics  Several hundreds or thousands of QCD jobs are required to be run on the Grid, with each jobs taking about 10 days.  In collaboration with Prof. Seyong Kim of Sejong University

17 User Community Support User Community Support  Sejong Univerisity  Porting of two-color QCD (Quantum chromo-dynamics) simulations on the Grid and large-scale execution on it  National Cancer Center  Porting of Geant4 simulations on the Grid for planning  KAIST  Used as a testbed for a grid and distributed computing course in computer science department  East-West Neo Medical Center in Kyung Hee University  Porting of Geant4 simulations on the Grid  Ewha Womans University  Porting of Gate applications on the grid

Deployment of QCD Simulations on the FKPPL VO

Our Two-color QCD Applications (1/2)  Large-scale  A large number of simulation jobs to be run with a wide range of different parameters  In our case, we have planned to run a total of 360 different QCD jobs with a different parameter set  beta = [1.50, 1.51, 1.52, …, 2.09, 2.10] (61)  J = [0.04, 0.05, 0.06] (3)  mu = [0.0, 0.57, 0.65, 0.7, 0.75, 0.8, 0.85, 0.9, 1] (9)  Independent  Each job runs independently  Long-duration  Each QCD job goes through 400 steps to complete, each step taking an average of 1 Hour, so each QCD job takes an average of 400 hours

 Need a computing facility to run a large number of jobs  FKPPL VO provides computing resources sufficient to run the 360 QCD jobs all together concurrently  Need some grid tool to effectively maintain such a large-scale jobs running on the Grid without having to know the details of the underlying Grid  Ganga seems to be appropriate as a tool for managing such large number of jobs on the Grid Our Two-color QCD Applications (2/2)

Issues relating to long-running jobs on the Grid  Long-running jobs often fail to complete on the Grid  It is not straightforward to successfully get done a long-duration job like our two-color QCD simulation on the Grid  A Grid proxy certificate expires before the job’s completion  By default, the proxy has a lifetime of 12 hours  Each Grid site has its own site operational policy such as the maximum CPU time for a job to be allowed to run at a time SitesCE NodesMaxCPUTime (Min) CC-IN2P3 Jobmanager-bqs-short5 Jobmanager-bqs-medium227 Jobmanager-bqs-long1429 KISTIJobmanager-lcgpbs-fkppl2880

Application-level Checkpointing/Restarting  We have modified the original two-color QCD simulation code to support an application- level checkpointing scheme  The two color QCD code takes 400 steps to complete  Once a QCD job is launched successfully on the Grid, a intermediate result is generated at each step and saved into the checkpoint server  When a QCD job is detected to be stopped for some reason,  Ganga restarts it from where it has left by resubmitting it along with the latest intermediate result

23 Computing Element Storage Element IN2P3 (Re)Submit QCD job (executable + small inputs) Submit QCD job Retrieve output WMS Su2.x Retrieve status & (small) output files Input file(s) Output file(s) Overview of QCD Simulation Runs on the Grid Heartbeat Monitor Check status & Intermediate Result send intermediate result CheckpointServer Retrieve the latest Intermediate result

Preliminary Results published in CPC

Two-color QCD in Production  ~ about 75 CPU years  1,647 runs * 400 step * 1 hours = 658,800 hours  ~ 1,647 concurrent QCD runs on the FKPPL VO  The simulation started at the end of August and completed on December 8 th for ~ 3.5 months  Now, it’s under analysis by Prof. Seyong Kim of Sejong University

Number of QCD Jobs on the FKPPL VO

Towards France-Asia VO

FJKPPL Workshop on Grid Computing  FJPPL/FKPPL Joint workshop on Grid computing was held in KEK on December  Hosted by KEK  CC-IN2P3, KEK and KISTI agreed to move forward towards France-Asia VO

Perspectives for France-Asia VO  Computing Infrastructure  gLite Middleware based  Computing centers offering resources  CC-IN2P3, KEK, KISTI  IHEP in China ?  IOIT in Vietnam?  Data Infrastructure  IRODS (Integrated Rule-Oriented Data System)  Both KEK and CC-IN2P3 have some expertise on the operation and management the IRODS service and might be able to provide IRODS service in the future  Applications  It is important to have some applications with scientific results  As of now, we have applications such as:  In-silico docking applications  QCD simulations  Geant4/GATE applications

 User Communities  As of now, we have user communities mainly from Korea  We might be able to have communities from Japan and Vietnam  Geant4 communities (Japan)  In silico drug discovery communities (Vietnam)  ?? (China)  High-level Tools/Services  It is important to provide users with an easy-to-use high level tools  Some of tools that we have some expertise on  WISDOM  Ganga  JSAGA  DIRAC  ??  Training  In order to promote the awareness of the France-Aisa VO, it is important to organize some tutorials on glite middleware, high-level tools and applications Perspectives for France-Asia VO

Thank you for your attention