Thomas Jefferson National Accelerator Facility Page 1 Clas12 Reconstruction and Analysis Framework V. Gyurjyan S. Mancilla.

Slides:



Advertisements
Similar presentations
Building Portals to access Grid Middleware National Technical University of Athens Konstantinos Dolkas, On behalf of Andreas Menychtas.
Advertisements

BARNALI CHAKRABARTY. What is an Operating System ?
Distributed Systems Major Design Issues Presented by: Christopher Hector CS8320 – Advanced Operating Systems Spring 2007 – Section 2.6 Presentation Dr.
Apache Struts Technology
©Ian Sommerville 2004Software Engineering, 7th edition. Chapter 9 Distributed Systems Architectures Slide 1 1 Chapter 9 Distributed Systems Architectures.
Distributed Systems Architectures
CS 501: Software Engineering Fall 2000 Lecture 16 System Architecture III Distributed Objects.
Ch 12 Distributed Systems Architectures
The Architecture of Transaction Processing Systems
Software Engineering Module 1 -Components Teaching unit 3 – Advanced development Ernesto Damiani Free University of Bozen - Bolzano Lesson 2 – Components.
DISTRIBUTED PROCESS IMPLEMENTAION BHAVIN KANSARA.
Overview SAP Basis Functions. SAP Technical Overview Learning Objectives What the Basis system is How does SAP handle a transaction request Differentiating.
UNIT-V The MVC architecture and Struts Framework.
Client/Server Architectures
Thomas Jefferson National Accelerator Facility Page 1 Clas12 Reconstruction and Analysis Framework V. Gyurjyan.
 Cloud computing  Workflow  Workflow lifecycle  Workflow design  Workflow tools : xcp, eucalyptus, open nebula.
©Ian Sommerville 2006Software Engineering, 8th edition. Chapter 12 Slide 1 Distributed Systems Architectures.
CLAS12 CalCom Activity CLAS Collaboration Meeting, March 6 th 2014.
Thomas Jefferson National Accelerator Facility Page 1 Clas12 Reconstruction and Analysis Framework V. Gyurjyan S. Mancilla.
Introduction and Overview Questions answered in this lecture: What is an operating system? How have operating systems evolved? Why study operating systems?
Thomas Jefferson National Accelerator Facility Page 1 12 GeV Upgrade Software Review Jefferson Lab November 25-26, 2013 Software Project for Hall B 12.
Robert Fourer, Jun Ma, Kipp Martin Copyright 2006 An Enterprise Computational System Built on the Optimization Services (OS) Framework and Standards Jun.
The Pipeline Processing Framework LSST Applications Meeting IPAC Feb. 19, 2008 Raymond Plante National Center for Supercomputing Applications.
Module 7: Fundamentals of Administering Windows Server 2008.
SOFTWARE DESIGN AND ARCHITECTURE LECTURE 07. Review Architectural Representation – Using UML – Using ADL.
CS 390- Unix Programming Environment CS 390 Unix Programming Environment Topics to be covered: Distributed Computing Fundamentals.
Architecture styles Pipes and filters Object-oriented design Implicit invocation Layering Repositories.
SOFTWARE DESIGN AND ARCHITECTURE LECTURE 09. Review Introduction to architectural styles Distributed architectures – Client Server Architecture – Multi-tier.
1 of of 25 3 of 25 ORBs (Object Request Broker) – A distributed software bus for communication among middleware services and applications – To.
SUMA: A Scientific Metacomputer Cardinale, Yudith Figueira, Carlos Hernández, Emilio Baquero, Eduardo Berbín, Luis Bouza, Roberto Gamess, Eric García,
N ATIONAL E NERGY R ESEARCH S CIENTIFIC C OMPUTING C ENTER Charles Leggett The Athena Control Framework in Production, New Developments and Lessons Learned.
 Apache Airavata Architecture Overview Shameera Rathnayaka Graduate Assistant Science Gateways Group Indiana University 07/27/2015.
CSC 480 Software Engineering Lecture 18 Nov 6, 2002.
Distributed System Concepts and Architectures 2.3 Services Fall 2011 Student: Fan Bai
Framework for MDO Studies Amitay Isaacs Center for Aerospace System Design and Engineering IIT Bombay.
Computing Division Requests The following is a list of tasks about to be officially submitted to the Computing Division for requested support. D0 personnel.
Distribution and components. 2 What is the problem? Enterprise computing is Large scale & complex: It supports large scale and complex organisations Spanning.
LOGO Development of the distributed computing system for the MPD at the NICA collider, analytical estimations Mathematical Modeling and Computational Physics.
GRID Overview Internet2 Member Meeting Spring 2003 Sandra Redman Information Technology and Systems Center and Information Technology Research Center National.
Faucets Queuing System Presented by, Sameer Kumar.
David Adams ATLAS DIAL: Distributed Interactive Analysis of Large datasets David Adams BNL August 5, 2002 BNL OMEGA talk.
HIGUCHI Takeo Department of Physics, Faulty of Science, University of Tokyo Representing dBASF Development Team BELLE/CHEP20001 Distributed BELLE Analysis.
© FPT SOFTWARE – TRAINING MATERIAL – Internal use 04e-BM/NS/HDCV/FSOFT v2/3 JSP Application Models.
CSI 3125, Preliminaries, page 1 SERVLET. CSI 3125, Preliminaries, page 2 SERVLET A servlet is a server-side software program, written in Java code, that.
Preface IIntroduction Objectives I-2 Course Overview I-3 1Oracle Application Development Framework Objectives 1-2 J2EE Platform 1-3 Benefits of the J2EE.
V Gyurjyan, D Abbott, J Carbonneau, G Gilfoyle, D Heddle, G Heyes, S Paul, C Timmer, D Weygand V. Gyurjyan JLAB data acquisition and analysis group.
Slide 1 Service-centric Software Engineering. Slide 2 Objectives To explain the notion of a reusable service, based on web service standards, that provides.
Data Communications and Networks Chapter 9 – Distributed Systems ICT-BVF8.1- Data Communications and Network Trainer: Dr. Abbes Sebihi.
Dual Target Design for CLAS12 Omair Alam and Gerard Gilfoyle Department of Physics, University of Richmond Introduction One of the fundamental goals of.
Thomas Jefferson National Accelerator Facility Page 1 Overview Talk Content Break-out Sessions Planning 12 GeV Upgrade Software Review Jefferson Lab November.
General requirements for BES III offline & EF selection software Weidong Li.
Thomas Jefferson National Accelerator Facility Page 1 ClaRA Stress Test V. Gyurjyan S. Mancilla.
Cloud Computing from a Developer’s Perspective Shlomo Swidler CTO & Founder mydrifts.com 25 January 2009.
VI/ CERN Dec 4 CMS Software Architecture vs Hybrid Store Vincenzo Innocente CMS Week CERN, Dec
Two New UML Diagram Types Component Diagram Deployment Diagram.
Fermilab Scientific Computing Division Fermi National Accelerator Laboratory, Batavia, Illinois, USA. Off-the-Shelf Hardware and Software DAQ Performance.
Distributed Systems Architectures. Topics covered l Client-server architectures l Distributed object architectures l Inter-organisational computing.
CLAS12 Commissioning with Cosmincs
Controlling a large CPU farm using industrial tools
The Client/Server Database Environment
Introduction to Operating System (OS)
CLARA Based Application Vertical Elasticity
Vincenzo Innocente CERN/EP/CMC
Service-centric Software Engineering
Lecture 1: Multi-tier Architecture Overview
Clas Reconstruction and Analyses framework
Physics data processing with SOA
CLARA . What’s new? CLAS Collaboration Meeting. March 6, 2019
Presentation transcript:

Thomas Jefferson National Accelerator Facility Page 1 Clas12 Reconstruction and Analysis Framework V. Gyurjyan S. Mancilla

Thomas Jefferson National Accelerator Facility Page 2 JLab Thomas Jefferson National Accelerator Facility (TJNAF), commonly called Jefferson Lab or JLab, is a U.S. national laboratory located in Newport News, Virginia Superconducting RF technology based accelerator will provide a 12 GeV continuous electron beam with a bunch length of less than 1 picosecond. Nuclear physics experiments in 4 end- stations (A,B,C,D) CLAS is a large acceptance spectrometer installed in Hall B to study Quark-gluon interactions with nuclei Nucleon-nucleon correlations Nucleon quark structure imaging, etc.

Thomas Jefferson National Accelerator Facility Page 3 CLAS12 Computing Requirements Enhance utilization, accessibility, contribution and collaboration Reusability of components On-demand data processing. Location independent resource pooling. Software agility Utilization of multicore processor systems Multi-threading Ability to expand computing power with minimal capital expenditure Dynamic elasticity. Utilization of IT resources of collaborating Universities. Take advantage of available commercial computing resources.

Thomas Jefferson National Accelerator Facility Page 4 Ideas Adopted From GAUDI Clear separation between data and algorithms Services encapsulate algorithms Services communicate data Three basic types of data: event, detector, statistics Clear separation between persistent and transient data No code or algorithmic solution was borrowed.

Thomas Jefferson National Accelerator Facility Page 5 Computing Model and Architecture Choice ClaRA is an implementation of the SOA Data processing major components as services Multilingual support –Services can be written in C++, Java and Python Physics application design/composition based on services Supports both traditional and cloud computing models Single process as well as distributed application design modes Centralized batch processing Distributed cloud processing Multi-Threaded

Thomas Jefferson National Accelerator Facility Page 6 Attributes of ClaRA Services Communicate data through shared memory and /or pub-sub messaging system Well defined, easy-to-use, data-centric interface Self-contained with no dependencies on other services Loose coupling. Coupled through communicating data. Always available but idle until requests arrival Location transparent Services are defined by unique names, and are discovered through discovery services Combine existing services into composite services or applications Services can also be combined in a single process (run-time environment), communicating data through shared memory (traditional computing model).

Thomas Jefferson National Accelerator Facility Page 7 ClaRA Design Architecture PDP Service Bus (pub-sub and/or shared-memory) Service layer Orchestration Layer Rule invocation Identification Filtration Control Layer Data flow control Load balancing Error recovery Managing errors and exceptions Security, validation and auditing Administration SaaS, IaaS, DaaS Registration Service Service Inventory

Thomas Jefferson National Accelerator Facility Page 8 ClaRA Components DPE Platform (cloud controller) C S DPE C S C S Orchestrator

Thomas Jefferson National Accelerator Facility Page 9 Platform (Cloud Controller) ClaRA administration Service registration and discovery. Keeps an inventory of all running DPEs and all deployed services. Used by orchestrators to discover and check services availability and distribution. Cloud Controller Service Bus (pub-sub server) Registration Discovery Administration Governing Platform Cloud Control Node

Thomas Jefferson National Accelerator Facility Page 10 Data Processing Environment (DPE) Main ClaRA processes. Each node acts as a DPE. All services are deployed and executed by threads inside the DPE process. Global memory to share data between services. Service Bus Pub-sub server Administration Service deployment Service removal Service recovery Monitoring DPE Computing Node 1

Thomas Jefferson National Accelerator Facility Page 11 Service Container Group and manage services in a DPE Can be used as namespaces to separate services. o The same service engine can be deployed in different containers in the same DPE. Handle service execution and its output. Service container presents a user engine as an SOA service (SaaS implementation). Engine interface Message processing Service Engine

Thomas Jefferson National Accelerator Facility Page 12 Transient Data Envelope

Thomas Jefferson National Accelerator Facility Page 13 Transient Data Object EVIO 4.1 is the default event format. o Data is just a byte buffer (avoid serialization). o Complete API (Java, C++,Python) to get data from the buffer. o A set of wrappers to work with the common CLAS12 bank format.

Thomas Jefferson National Accelerator Facility Page 14 Service Communication Transient Data Storage Service Bus Service 1 Service 2 Service N Service 1 Service 2 Service N Java DPE C++ DPE Computing Node 1 Service Bus Computing Node 2 Service Bus Computing Node 1 Service Bus Computing Node N

Thomas Jefferson National Accelerator Facility Page 15 The fundamental unit of ClaRA based application. Receives an input data in an envelope, and generates an output data. o The data envelope is the same for all services. Implements ClaRA standard interface o A configure method o An execute method. o Several description/identification methods. Must be thread-safe. o The same service engine can be executed in parallel multiple times. Service Engine

Thomas Jefferson National Accelerator Facility Page 16 Design and control ClaRA applications Coordinate services execution and data flow. Usually run outside of the DPE. Deploy services to DPEs. o Each deployed service is identified by the following canonical name:dpe_name/container_name/service_engine_name Link services together. o The output of a service is sent as the input to its linked service. Orchestrator

Thomas Jefferson National Accelerator Facility Page 17 Request services execution. o Async requests: service output is sent to all its linked services. o Sync request: service output is returned to the requester. Monitor services execution. o Data, done, warning and error monitoring. o Run custom callback code when a notification is received. Orchestrator

Thomas Jefferson National Accelerator Facility Page 18 Application Graphical Designer

Thomas Jefferson National Accelerator Facility Page 19 Read EVIO events from input file. Events pass from service to service in the chain. o Services add more banks to the event. Write events to output file. RS1S2SNW Single Event Reconstruction

Thomas Jefferson National Accelerator Facility Page 20 Multi-Core Reconstruction R S1S2SN W S1S2SN S1S2SN O DPE

Thomas Jefferson National Accelerator Facility Page 21 Multi-Core Reconstruction

Thomas Jefferson National Accelerator Facility Page 22 Multi-Core Reconstruction

Thomas Jefferson National Accelerator Facility Page 23 Multi-Core Reconstruction

Thomas Jefferson National Accelerator Facility Page 24 Multi-Node Reconstruction R S1SN DO S2 S1SNS2 S1SNS2 DPEn S1SNS2 S1SNS2 S1SNS2 DPE2 S1SNS2 S1SNS2 S1SNS2 DPE1 W DPEio MO

Thomas Jefferson National Accelerator Facility Page 25 Multi-Node Reconstruction

Thomas Jefferson National Accelerator Facility Page 26 Batch Deployment

Thomas Jefferson National Accelerator Facility Page 27 Clas12 Detector Electronics Clas12 Detector Electronics Trigger Slow Controls ET Online Transient Data Storage ET Online Transient Data Storage Permanent Data Storage Permanent Data Storage Online EB Services Online EB Services Online Monitoring Services Online Monitoring Services Event Visualization Services Event Visualization Services Online Calibration Services Online Calibration Services Calibration Database Calibration Database Conditions Database Conditions Database Geometry Calibration Services Geometry Calibration Services Run Conditions Services Run Conditions Services Cloud Control Service Registration Service Control Online Farm Online Application Orchestrator Online Application Orchestrator Cloud Control Service Registration Service Control Physics Data Processing Application Orchestrator Physics Data Processing Application Orchestrator Geant 4 GEMC Simulation GEMC Simulation EB Services EB Services DST Histogram Visualization Services DST Histogram Visualization Services Analysis Services Analysis Services Calibration Services Calibration Services Run Conditions Services Run Conditions Services Geometry Calibration Services Permanent Data Storage Permanent Data Storage Permanent Data Storage Permanent Data Storage Service Control Service Registration Cloud Control Permanent Data Storage Permanent Data Storage Calibration Database Calibration Database Conditions Database Conditions Database Service Control Service Registration Cloud Control Permanent Data Storage Permanent Data Storage Calibration Database Calibration Database Conditions Database Conditions Database Offline University Cloud 1 Offline University Cloud n Cloud Scheduler Offline JLAB Farm

Thomas Jefferson National Accelerator Facility Page 28 Challenges Increases author and user pools. Management and administration. Strict service canonization rules Workloads of different clients may overwhelm a single service. Service and Cloud governance Network security Client authentication and message encryption

Thomas Jefferson National Accelerator Facility Page 29 Summary and Conclusion A multi-treaded analyses framework, based on SOA PDP application based on specialized services Small, independent Easy to test, update and maintain Building and running PDP application does not require CS skills. List of applications has been developed using the framework Charge particle tracking (central, forward) EC reconstruction FTOF reconstruction PID HTCC PCAL Detector calibration Event Building Histogram services Database application –Geometry, calibration constants and run conditions services ClaRA supports both traditional and cloud computing models and if need be we are ready for cloud deployment.

Thomas Jefferson National Accelerator Facility Page 30 e e Links

Thomas Jefferson National Accelerator Facility Page 31 Service Bus Performance measurements Control messages only <0.03ms/message

Thomas Jefferson National Accelerator Facility Page 32 Service Bus Performance measurements