1 Building application portals with DIRAC A.Tsaregorodtsev, CPPM-IN2P3-CNRS, Marseille 27 April 2010, Journée LuminyGrid, Marseille.

Slides:



Advertisements
Similar presentations
High Performance Computing Course Notes Grid Computing.
Advertisements

1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
1 Grid services based architectures Growing consensus that Grid services is the right concept for building the computing grids; Recent ARDA work has provoked.
Network Management Overview IACT 918 July 2004 Gene Awyzio SITACS University of Wollongong.
Workload Management Workpackage Massimo Sgaravatto INFN Padova.
Nikolay Tomitov Technical Trainer SoftAcad.bg.  What are Amazon Web services (AWS) ?  What’s cool when developing with AWS ?  Architecture of AWS 
1 Bridging Clouds with CernVM: ATLAS/PanDA example Wenjing Wu
Stuart K. PatersonCHEP 2006 (13 th –17 th February 2006) Mumbai, India 1 from DIRAC.Client.Dirac import * dirac = Dirac() job = Job() job.setApplication('DaVinci',
Web-based Portal for Discovery, Retrieval and Visualization of Earth Science Datasets in Grid Environment Zhenping (Jane) Liu.
DIRAC API DIRAC Project. Overview  DIRAC API  Why APIs are important?  Why advanced users prefer APIs?  How it is done?  What is local mode what.
LHC Experiment Dashboard Main areas covered by the Experiment Dashboard: Data processing monitoring (job monitoring) Data transfer monitoring Site/service.
1 Managing distributed computing resources with DIRAC A.Tsaregorodtsev, CPPM-IN2P3-CNRS, Marseille September 2011, NEC’11, Varna.
DIRAC Web User Interface A.Casajus (Universitat de Barcelona) M.Sapunov (CPPM Marseille) On behalf of the LHCb DIRAC Team.
CGW 2003 Institute of Computer Science AGH Proposal of Adaptation of Legacy C/C++ Software to Grid Services Bartosz Baliś, Marian Bubak, Michał Węgiel,
Computing for ILC experiment Computing Research Center, KEK Hiroyuki Matsunaga.
Grid Initiatives for e-Science virtual communities in Europe and Latin America DIRAC TEAM CPPM – CNRS DIRAC Grid Middleware.
CERN IT Department CH-1211 Genève 23 Switzerland t EIS section review of recent activities Harry Renshall Andrea Sciabà IT-GS group meeting.
YAN, Tian On behalf of distributed computing group Institute of High Energy Physics (IHEP), CAS, China CHEP-2015, Apr th, OIST, Okinawa.
INFSO-RI Enabling Grids for E-sciencE SA1: Cookbook (DSA1.7) Ian Bird CERN 18 January 2006.
INFSO-RI Enabling Grids for E-sciencE Logging and Bookkeeping and Job Provenance Services Ludek Matyska (CESNET) on behalf of the.
Grid Workload Management & Condor Massimo Sgaravatto INFN Padova.
EGEE-III INFSO-RI Enabling Grids for E-sciencE Julia Andreeva CERN (IT/GS) CHEP 2009, March 2009, Prague New job monitoring strategy.
CERN IT Department CH-1211 Genève 23 Switzerland t Internet Services Job Monitoring for the LHC experiments Irina Sidorova (CERN, JINR) on.
PanDA Multi-User Pilot Jobs Maxim Potekhin Brookhaven National Laboratory Open Science Grid WLCG GDB Meeting CERN March 11, 2009.
1 DIRAC – LHCb MC production system A.Tsaregorodtsev, CPPM, Marseille For the LHCb Data Management team CHEP, La Jolla 25 March 2003.
The Grid System Design Liu Xiangrui Beijing Institute of Technology.
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
Cracow Grid Workshop October 2009 Dipl.-Ing. (M.Sc.) Marcus Hilbrich Center for Information Services and High Performance.
Tool Integration with Data and Computation Grid GWE - “Grid Wizard Enterprise”
EGEE-III INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks WMSMonitor: a tool to monitor gLite WMS/LB.
Enabling Grids for E-sciencE System Analysis Working Group and Experiment Dashboard Julia Andreeva CERN Grid Operations Workshop – June, Stockholm.
June 24-25, 2008 Regional Grid Training, University of Belgrade, Serbia Introduction to gLite gLite Basic Services Antun Balaž SCL, Institute of Physics.
NA-MIC National Alliance for Medical Image Computing UCSD: Engineering Core 2 Portal and Grid Infrastructure.
Getting started DIRAC Project. Outline  DIRAC information system  Documentation sources  DIRAC users and groups  Registration with DIRAC  Getting.
Overview of Privilege Project at Fermilab (compilation of multiple talks and documents written by various authors) Tanya Levshina.
Conference name Company name INFSOM-RI Speaker name The ETICS Job management architecture EGEE ‘08 Istanbul, September 25 th 2008 Valerio Venturi.
1 Andrea Sciabà CERN Critical Services and Monitoring - CMS Andrea Sciabà WLCG Service Reliability Workshop 26 – 30 November, 2007.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
CHEP 2006, February 2006, Mumbai 1 LHCb use of batch systems A.Tsaregorodtsev, CPPM, Marseille HEPiX 2006, 4 April 2006, Rome.
System Center Lesson 4: Overview of System Center 2012 Components System Center 2012 Private Cloud Components VMM Overview App Controller Overview.
Managing Data DIRAC Project. Outline  Data management components  Storage Elements  File Catalogs  DIRAC conventions for user data  Data operation.
INFSO-RI Enabling Grids for E-sciencE ARDA Experiment Dashboard Ricardo Rocha (ARDA – CERN) on behalf of the Dashboard Team.
Development of e-Science Application Portal on GAP WeiLong Ueng Academia Sinica Grid Computing
Transformation System report Luisa Arrabito 1, Federico Stagni 2 1) LUPM CNRS/IN2P3, France 2) CERN 5 th DIRAC User Workshop 27 th – 29 th May 2015, Ferrara.
Testing and integrating the WLCG/EGEE middleware in the LHC computing Simone Campana, Alessandro Di Girolamo, Elisa Lanciotti, Nicolò Magini, Patricia.
GraDS MacroGrid Carl Kesselman USC/Information Sciences Institute.
GRID ANATOMY Advanced Computing Concepts – Dr. Emmanuel Pilli.
Tool Integration with Data and Computation Grid “Grid Wizard 2”
DIRAC Pilot Jobs A. Casajus, R. Graciani, A. Tsaregorodtsev for the LHCb DIRAC team Pilot Framework and the DIRAC WMS DIRAC Workload Management System.
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI How to integrate portals with the EGI monitoring system Dusan Vudragovic.
DIRAC Project A.Tsaregorodtsev (CPPM) on behalf of the LHCb DIRAC team A Community Grid Solution The DIRAC (Distributed Infrastructure with Remote Agent.
EGI Technical Forum Amsterdam, 16 September 2010 Sylvain Reynaud.
EGEE-II INFSO-RI Enabling Grids for E-sciencE Practical using WMProxy advanced job submission.
The GridPP DIRAC project DIRAC for non-LHC communities.
Grid Execution Management for Legacy Code Architecture Exposing legacy applications as Grid services: the GEMLCA approach Centre.
Breaking the frontiers of the Grid R. Graciani EGI TF 2012.
EGEE-II INFSO-RI Enabling Grids for E-sciencE Overview of gLite, the EGEE middleware Mike Mineter Training Outreach Education National.
LHCb/DIRAC week A.Tsaregorodtsev, CPPM 7 April 2011.
SAM architecture EGEE 07 Service Availability Monitor for the LHC experiments Simone Campana, Alessandro Di Girolamo, Nicolò Magini, Patricia Mendez Lorenzo,
DIRAC for Grid and Cloud Dr. Víctor Méndez Muñoz (for DIRAC Project) LHCb Tier 1 Liaison at PIC EGI User Community Board, October 31st, 2013.
DIRAC in LHCb and beyond Philippe Charpentier (LHCb-CERN) Slides courtesy of A.Tsaregorodtsev BigPanda Workshop, October Click to edit Master title.
DIRAC Distributed Computing Services A. Tsaregorodtsev, CPPM-IN2P3-CNRS FCPPL Meeting, 29 March 2013, Nanjing.
1 DIRAC Project Status A.Tsaregorodtsev, CPPM-IN2P3-CNRS, Marseille 10 March, DIRAC Developer meeting.
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI EGI Services for Distributed e-Infrastructure Access Tiziana Ferrari on behalf.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Introduction Salma Saber Electronic.
Enabling Grids for E-sciencE Claudio Cherubino INFN DGAS (Distributed Grid Accounting System)
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI EGI solution for high throughput data analysis Peter Solagna EGI.eu Operations.
DIRAC services.
Management of Virtual Execution Environments 3 June 2008
Monitoring of the infrastructure from the VO perspective
Presentation transcript:

1 Building application portals with DIRAC A.Tsaregorodtsev, CPPM-IN2P3-CNRS, Marseille 27 April 2010, Journée LuminyGrid, Marseille

2 Outline  Scope of the DIRAC project  DIRAC Overview  User Interfaces  Web Portals  Conclusion

Large VO issues  HEP experiments collect unprecedented volumes of data to be processed on large amount of geographically distributed computing resources  10s of PBytes of data per year  10s of thousands CPUs in 100s of centers  1000s of users from 100s of institutions  However, other application domains are quickly approaching these scales  Large user communities (Virtual Organizations) have specific problems  Dealing with heterogeneous resources Various computing clusters, grids, etc  Dealing with the intracommunity workload management User group quotas and priorities Priorities of different activities  Dealing with a variety of applications Massive data productions Individual user applications, etc 3

General problems and solutions  Overcome deficiencies of the standard grid middleware  Inefficiencies, failures Production managers can afford that, users can not  Lacking specific functionality  Alleviate the excessive burden from sites – resource providers – in supporting multiple VOs  Avoid complex VO specific configuration on sites  Avoid VO specific services on sites  The complexity of managing the VO workload resulted in specific software layer on top of the standard grid middleware. Among the LHC experiments  AliEn in Alice  PanDA in Atlas  GlideIn WMS, Phedex in CMS  DIRAC in LHCb 4

DIRAC and LHCb  LHCb experiment is dedicated to the study of CP-violation in b-systems  Smallest of the 4 big LHC experiments  ~500 physicists  ~60 institutes from 15 countries  Nevertheless, computing is also a challenge….  DIRAC was originally developed as a distributed data production and analysis system used by the LHCb experiment  Includes workload and data management components  Started with the MC data production tasks  Extended to data processing and user analysis  The goal was: ●integrate all the heterogeneous computing resources available to LHCb ●Minimize human intervention at LHCb sites 5

DIRAC Project scope  Large volumes of Computing and Storage resources are supplied by various institutes  Grids define common rules of how to access these resources  Users organized in virtual communities are providing payloads to be executes on the Grid resources  Middleware is a set of software components enabling users to exploit the Grid resources  DIRAC provides a complete set of middleware components for workload and data management tasks  User community perspective 6

Brief DIRAC Overview 7

DIRAC Framework  Services oriented architecture  DIRAC systems consist of services, light distributed agents and client tools  All the communications between the distributed components are secure  DISET custom client/service protocol Control and data communications  X509, GSI security standards  Fine grained authorization rules Per individual user FQAN Per service interface method Per job 8

DIRAC base services  Redundant Configuration Service  Provides service discovery and setup parameters for all the DIRAC components  Full featured proxy management system  Proxy storage and renewal mechanism  Support for multiuser pilot jobs  System Logging service  Collect essential error messages from all the components  Monitoring service  Monitor the service and agents behavior 9

Workload Management  Workload Management System with Pilot Jobs  Originally introduced by DIRAC  Increase the visible user job efficiency  Allow efficient and precise application of the community policies  Allow to aggregate heterogeneous resources 10

Physicist User EGEE Pilot Director EGEE Grid NDG Pilot Director NDG Grid EELA Pilot Director EELA Grid CREAM Pilot Director CREAM CE Matcher Service Production Manager

WMS performance  DIRAC performance measured in the recent Data Challenges and Production runs  Up to 25K concurrent jobs in ~120 distinct sites  One mid-range central server hosting DIRAC services  Further optimizations to increase capacity are possible ●Hardware, database optimizations, service load balancing, etc 12

Support for MPI Jobs  MPI Service developed for applications in the EELA Grid  Non-HEP applications: Astrophysics, BioMed, Seismology  No special MPI support on sites is required MPI software installed by Pilot Jobs Use site MPI support if exists  MPI ring usage optimization Ring reuse for multiple jobs  Lower load on the gLite WMS Variable ring sizes for different jobs 13

Other DIRAC components  Request Management System  Collect and execute asynchronously any kind of operation that can fail Data upload and registration Job status and parameter reports, etc  Essential in the ever unstable Grid environment  Production Management System  Automatic data processing jobs creation and submission according to predefined scenarios Complex Workflow management  Organization and handling of O(100K) of jobs  Data Management System  Full featured File Replica and Metadata Catalogs  Automatic data replication  Storage resources monitoring, data integrity checking 14

User Interfaces 15

DIRAC user interfaces  Easy client installation for various platforms (Linux, MacOS)  Includes security components  JDL notation for job description  Simplified with respect to the « standard » JDL  Command line tools  à la gLite UI commands  e.g. dirac-wms-job-submit  Extensive Python API for all the tasks  Job creation and manipulation, results retrieval Possibility to use complex workflow templates  Data operations, catalog inspection  Used by the GANGA user front-end 16

17 Example job submission from DIRAC.Interfaces.API.Dirac import Dirac from Extensions.LHCb.API.LHCbJob import LHCbJob … myJob = LHCbJob() myJob.setCPUTime(50000) myJob.setSystemConfig('slc4_ia32_gcc34') myJob.setApplication('Brunel','v32r3p1','RealDataDst200Evts.opts','LogFileName.log') myJob.setName('DIRAC3-Job') myJob.setInputData(['/lhcb/data/CCRC08/RAW/LHCb/CCRC/420157/420157_ raw']) #myJob.setDestination('LCG.CERN.ch') dirac = Dirac() jobID = dirac.submit(myJob) … dirac.status( ) dirac.parameters( ) dirac.loggingInfo( ) … dirac.getOutputSandbox( )

DIRAC: Secure Web Portal  Web portal with intuitive desktop application like interface  Ajax, Pylons, ExtJS Javascript library  Monitoring and control of all activities  User job monitoring and manipulation  Data production controls  DIRAC Systems configuration  Secure access  Standard grid certificates  Fine grained authorization rules  Web pages for standard DIRAC tasks  System Configuration, Services administration  Job Monitoring and Controls  Resources Accounting and Monitoring 18

Web Portal: example interfaces 19

LHCb Web: Bookkeeping page 20  Interface to the LHCb Metadata Catalog  Part of the LHCb DIRAC Web Portal

LHCb Web: Production Requests 21  Comprehensive forms to define Data Production requests  Multiple input parameters with the help for the parameter choices  Support for the complex request verification and approval procedure

Web Portal: user tasks  Job submission through the Web Portal  Full GSI security  Sandboxes uploading and downloading  Generic Job Launchpad panel exists in the basic DIRAC Web Portal Can be useful for newcomers and occasional users  Specific application Web Portals can be derived  Community Application Servers All the grid computational tasks steered on the web  VO “formation” DIRAC instance to be deployed at CC/IN2P3 22

DIRAC: Getting started  Get your Grid certificate  usercert.p12  dirac-cert-convert.sh to convert it to PEM format  Register in a Grid VO  To have access to the Grid resources  Delegate user proxy  proxy-init –g dirac_user  Start using the DIRAC Web Portal 23

DIRAC Installations  Latin American EELA Grid  Part of the production infrastructure of the GISELA Grid  Astrophysics, BioMed, Seismology applications  HEP Experiments  ILC Collaboration  Belle Collaboration at KEK, Japan Using Amazon EC2 Cloud Computing  Installation at CC/IN2P3, Lyon for the VO vo.formation.idgrilles.fr (training program)  dirac.in2p3.fr  Documentation is in preparation 24

DIRAC Installations  An installation can be prepared and maintained for the LuminyGrid users  This is a proposal to be discussed  Ensure user access to the LuminyGrid, France NGI and EGEE resources  Multiple VO support with the single DIRAC instance No need to have VO experts in the DIRAC services administration  Support for specific VO applications Porting the applications to the grid Help in developing specific Web portals Interfacing existing Web Portals to the DIRAC backend 25

26 Conclusions  DIRAC project provides a secure framework for building distributed computing (grid) system  DIRAC is providing a complete middleware stack  Can integrate standard (gLite) services as well  The DIRAC Framework can be used to build application specific services and Web Portals  Based on our many years experience, we are looking for ways how to help the other users in porting their applications to the Grid and making the Grid usage a fun

Backup slides 27

28 DIRAC development environment  Python is the main development language  Fast prototyping/development cycle  Platform independence  MySQL database for the main services  ORACLE database backend for the LHCb Metadata Catalog  Modular architecture allowing an easy customization for the needs of a particular community  Simple framework for building custom services and agents