Status of Storm+Lustre and Multi-VO Support

Slides:



Advertisements
Similar presentations
Status of BESIII Distributed Computing BESIII Workshop, Mar 2015 Xianghu Zhao On Behalf of the BESIII Distributed Computing Group.
Advertisements

Copyright GeneGo CONFIDENTIAL »« MetaCore TM (System requirements and installation) Systems Biology for Drug Discovery.
K.Harrison CERN, 23rd October 2002 HOW TO COMMISSION A NEW CENTRE FOR LHCb PRODUCTION - Overview of LHCb distributed production system - Configuration.
Reproducible Environment for Scientific Applications (Lab session) Tak-Lon (Stephen) Wu.
DIRAC API DIRAC Project. Overview  DIRAC API  Why APIs are important?  Why advanced users prefer APIs?  How it is done?  What is local mode what.
How Computers Work. A computer is a machine f or the storage and processing of information. Computers consist of hardware (what you can touch) and software.
Distributed Computing for CEPC YAN Tian On Behalf of Distributed Computing Group, CC, IHEP for 4 th CEPC Collaboration Meeting, Sep ,
Customized cloud platform for computing on your terms !
1 port BOSS on Wenjing Wu (IHEP-CC)
Test Of Distributed Data Quality Monitoring Of CMS Tracker Dataset H->ZZ->2e2mu with PileUp - 10,000 events ( ~ 50,000 hits for events) The monitoring.
YAN, Tian On behalf of distributed computing group Institute of High Energy Physics (IHEP), CAS, China CHEP-2015, Apr th, OIST, Okinawa.
Wenjing Wu Andrej Filipčič David Cameron Eric Lancon Claire Adam Bourdarios & others.
Status of StoRM+Lustre and Multi-VO Support YAN Tian Distributed Computing Group Meeting Oct. 14, 2014.
BESIII Production with Distributed Computing Xiaomei Zhang, Tian Yan, Xianghu Zhao Institute of High Energy Physics, Chinese Academy of Sciences, Beijing.
DIRAC Review (13 th December 2005)Stuart K. Paterson1 DIRAC Review Exposing DIRAC Functionality.
Enabling Grids for E-sciencE EGEE-III INFSO-RI Using DIANE for astrophysics applications Ladislav Hluchy, Viet Tran Institute of Informatics Slovak.
Distributed Computing for CEPC YAN Tian On Behalf of Distributed Computing Group, CC, IHEP for 4 th CEPC Collaboration Meeting, Sep , 2014 Draft.
Foundation API The Fast Version. Follow my examples Example: building App for GMAP on Stampede Materials are in the Data Store: – Community Data/iplantcollaborative/example_data/gmap.
Getting started DIRAC Project. Outline  DIRAC information system  Documentation sources  DIRAC users and groups  Registration with DIRAC  Getting.
Running Kuali: A Technical Perspective Ailish Byrne (Indiana University) Jonathan Keller (University of California, Davis)
OPTIMIZATION OF DIESEL INJECTION USING GRID COMPUTING Miguel Caballer Universidad Politécnica de Valencia.
Testing Infrastructure Wahid Bhimji Sam Skipsey Intro: what to test Existing testing frameworks A proposal.
Breaking the frontiers of the Grid R. Graciani EGI TF 2012.
StoRM + Lustre Proposal YAN Tian On behalf of Distributed Computing Group
Geant4 GRID production Sangwan Kim, Vu Trong Hieu, AD At KISTI.
StoRM+Lustre Performance Test with 10Gbps Network YAN Tian for Distributed Computing Group Meeting Nov. 4th, 2014.
Progress of Work on SE and DMS YAN Tian April. 16, 2014.
Fermilab Scientific Computing Division Fermi National Accelerator Laboratory, Batavia, Illinois, USA. Off-the-Shelf Hardware and Software DAQ Performance.
Status of BESIII Distributed Computing BESIII Collaboration Meeting, Nov 2014 Xiaomei Zhang On Behalf of the BESIII Distributed Computing Group.
Using the Grid for the ILC Mokka and Marlin on the Grid ILC Software Meeting, Cambridge 2006.
The CMS Beijing Tier 2: Status and Application Xiaomei Zhang CMS IHEP Group Meeting December 28, 2007.
Progress on Design and Implement of Job Management System Suo Bing, Yan Tian, Zhao Xianghu
ANL T3g infrastructure S.Chekanov (HEP Division, ANL) ANL ASC Jamboree September 2009.
Advanced Computing Facility Introduction
Compute and Storage For the Farm at Jlab
Accessing the VI-SEEM infrastructure
Outline Installing Gem5 SPEC2006 for Gem5 Configuring Gem5.
ILD MCProduction with ILCDirac
Status of BESIII Distributed Computing
SuperB – INFN-Bari Giacinto DONVITO.
Large Output and Shared File Systems
HPC usage and software packages
The Beijing Tier 2: status and plans
By Chris immanuel, Heym Kumar, Sai janani, Susmitha
Chapter 1: A Tour of Computer Systems
Report of Dubna discussion
Heterogeneous Computation Team HybriLIT
Practical: The Information Systems
Akiya Miyamoto KEK 1 June 2016
Added value of new features of the ATLAS computing model and a shared Tier-2 and Tier-3 facilities from the community point of view Gabriel Amorós on behalf.
Work report Xianghu Zhao Nov 11, 2014.
PROOF – Parallel ROOT Facility
Usecase Dynamo Moderate Requirements:
Spark Presentation.
Grid Application Support Group Case study Schrodinger equations on the Grid Status report 16. January, Created by Akos Balasko
LHCb Computing Model and Data Handling Angelo Carbone 5° workshop italiano sulla fisica p-p ad LHC 31st January 2008.
Generator Services planning meeting
Computing at CEPC Xiaomei Zhang Xianghu Zhao
Virtualization in the gLite Grid Middleware software process
Simulation use cases for T2 in ALICE
Discussions on group meeting
Xiaomei Zhang On behalf of CEPC software & computing group Nov 6, 2017
CCR Advanced Seminar: Running CPLEX Computations on the ISE Cluster
Support for ”interactive batch”
Alice Software Demonstration
Getting Started With Solr
The CMS Beijing Site: Status and Application
Status and plans for bookkeeping system and production tools
Production Manager Tools (New Architecture)
Thursday AM, Lecture 2 Brian Lin OSG
Presentation transcript:

Status of Storm+Lustre and Multi-VO Support YAN Tian for Distributed Computing Group Meeting Oct. 23, 2014

StoRM + Lustre: Test Bed SE server configuration: This test machine is originally prepared for dCache+Lustre frontend, thus with good network performance Model Dell PowerEdger R620 CPU Xeon E5-2609 v2 @ 2.50 GHz CPU Cores 8 cores Memory 64 GB HDD scsi 300 GB Network 1 eth0: 1 Gbps Network 2 eth4: 10 Gbps A symblic link to Lustre directory , User can access files in this directory through StoRM webDAV portal

StoRM + Lustre Test 1: single thread download Test time: Oct 15, 17:50--18:40 Lustre is not busy (load 7%, out 80 MB/s) 20 files of size 1 GB Average download speed: 10.6 MB/s with eth0: 1 Gbps load of SE: 0.8~1.1 load, 11~13% wa v.s. When Lustre is busy: out 500~1400 MB/s

StoRM + Lustre Test 2: multi threads/processes download Multi-tread download tool: mytget, can’t start multi-thread mode for Lustre Multi-process wget download, do not improve much, 22~33 M/s 4 processes 8 processes

StoRM + Lustre Test 3: Symbolic Link Problem Modify namespace.xml is under trying.

Storm+Lustre Test: To Do Solve symblik link problem Dataset transfer test between IHEPD-USER Open ports 50000:55000 Dataset transfer test between WHU/USTC-USER

ILC-DIRAC Study: User Interface Python code, which can be directly execute A job script example: from DIRAC.Core.Base import Script Script.parseCommandLine() from ILCDIRAC.Interfaces.API.DiracILC import DiracILC dirac = DiracILC(True,“my_job_repository.rep") from ILCDIRAC.Interfaces.API.NewInterface.UserJob import UserJob job = UserJob() job.setName("MyJobName") job.setJobGroup("Agroup") job.setCPUTime(86400) from ILCDIRAC.Interface.API.NewInterface.Application import Mokka, Marlin mo = Mokka() mo.setLogFile(“sim-job.log”) mo.setInputFile(“init.macro”) mo.setOutputFile(“E250-CDR_wo_Pnnh.eL.eR.001.slcio”) mo.setNumberOfEvents(1000) job.append(mo) mar = Marlin() mar.setParameters(“value”) mar.getInputFromApp(mo) job.append(mar) job.submit(dirac) dirac instance is job reciever define and set para. for app. applications stack

ILC-DIRAC Study: Job Repository Repo. Contains all necessary information of jobs for Job Monitoring $ dirac-repo-monitor repo.cfg for Retrive all the output sandbox and output data $ dirac-repo-retrieve-jobs-output –r –O repo.cfg Repository is a functionality provided by DIRAC call 3 methods

ILC-DIRAC Study: Applications Many applications Generation: Whizard, Pythia, StdHepCut Simulation: Mokka, SLIC Reconstruction: Marlin, LCSIM, SLICPandora Analysis: Marlin, ROOT, Druid, etc… A command for user to querry avaliable app. and it’s version $ dirac-ilc-show-software Applications are all defined in module ILCDIRAC.Interfaces.API.NewInterface.Application (base class) ILCDIRAC.Interfaces.API.NewInterface.Applications In job script: from ILCDIRAC.Interface.API.NewInterface.Application import Mokka mo = Mokka() mo.setParameters1(“value1”) mo.setParameters2(“value2”) job.append(mo) A Generic application for executable outside ILCsoft, e.g. ga = GenericApplication() ga.setScript(“boss.exe”) ga.setArguments(“jobOptions.txt”)

ILC-DIRAC Study: User Input Data For ILC analysis jobs, users always need his own lib. file *.so ILC solution: upload to SE, download to WN $ tar czf lib.tar.gz lib/ $ dirac-dms-add-files /ilc/user/i/initial/some/path/lib.tar.gz lib.tar.gz CERN-SRM $ dirac-dms-remove-files /ilc/user/i/initial/some/path/lib.tar.gz In Job Script: job.setInputSandbox("LFN:/ilc/user/i/initial/some/path/lib.tar.gz") ILC allow user to use $ dirac-dms-filecatalog

ILC-DIRAC Study: Class Inheritance DIRAC classes ILC-DIRAC classes Dirac DiracILC Spliter? UserJob Job ProductionJob Application Applications MokkaAnalysis ModuleBase MarlinAnalysis PhthiaAnalysis etc…

ILC-DIRAC: Module Example (MokkaAnalysis) called by Job Agent from ILCDIRAC.Workflow.Modules import MokkaAnalysis ma = MokkaAnalysis() ma.execute() In this module: retrieve job parameters write a shell script to set environment run application return status code