Remote Site C Pilot Scheduler Pilots and Pilot Schedulers Jobs Statistics Production Dashboard Dynamic Data Movement Monitor Panda Server (Apache) Development.

Slides:



Advertisements
Similar presentations
Module 13: Performance Tuning. Overview Performance tuning methodologies Instance level Database level Application level Overview of tools and techniques.
Advertisements

CASSANDRA-A Decentralized Structured Storage System Presented By Sadhana Kuthuru.
Based on the text by Jimmy Lin and Chris Dryer; and on the yahoo tutorial on mapreduce at index.html
Serverless Network File Systems. Network File Systems Allow sharing among independent file systems in a transparent manner Mounting a remote directory.
Chapter 9 Designing Systems for Diverse Environments.
Database Systems: Design, Implementation, and Management Eighth Edition Chapter 11 Database Performance Tuning and Query Optimization.
McGraw-Hill/Irwin Copyright © 2007 by The McGraw-Hill Companies, Inc. All rights reserved. Chapter 17 Client-Server Processing, Parallel Database Processing,
©Silberschatz, Korth and Sudarshan18.1Database System Concepts Centralized Systems Run on a single computer system and do not interact with other computer.
Data Warehousing: Defined and Its Applications Pete Johnson April 2002.
MultiJob PanDA Pilot Oleynik Danila 28/05/2015. Overview Initial PanDA pilot concept & HPC Motivation PanDA Pilot workflow at nutshell MultiJob Pilot.
Microsoft ® Application Virtualization 4.5 Infrastructure Planning and Design Series.
Distributed Databases
Dynamics AX Technical Overview Application Architecture Dynamics AX Technical Overview.
Simplify your Job – Automatic Storage Management Angelo Session id:
Russ Houberg Senior Technical Architect, MCM KnowledgeLake, Inc.
Microsoft ® Application Virtualization 4.6 Infrastructure Planning and Design Published: September 2008 Updated: February 2010.
How WebMD Maintains Operational Flexibility with NoSQL Rajeev Borborah, Sr. Director, Engineering Matt Wilson – Director, Production Engineering – Consumer.
Presented by: Alvaro Llanos E.  Motivation and Overview  Frangipani Architecture overview  Similar DFS  PETAL: Distributed virtual disks ◦ Overview.
Distributed Data Stores – Facebook Presented by Ben Gooding University of Arkansas – April 21, 2015.
 Cloud computing  Workflow  Workflow lifecycle  Workflow design  Workflow tools : xcp, eucalyptus, open nebula.
CERN - IT Department CH-1211 Genève 23 Switzerland t Monitoring the ATLAS Distributed Data Management System Ricardo Rocha (CERN) on behalf.
Database Systems: Design, Implementation, and Management Tenth Edition Chapter 11 Database Performance Tuning and Query Optimization.
Database Systems Design, Implementation, and Management Coronel | Morris 11e ©2015 Cengage Learning. All Rights Reserved. May not be scanned, copied or.
Database Systems: Design, Implementation, and Management Eighth Edition Chapter 10 Database Performance Tuning and Query Optimization.
IT The Relational DBMS Section 06. Relational Database Theory Physical Database Design.
Ch 5. The Evolution of Analytic Processes
Performance Concepts Mark A. Magumba. Introduction Research done on 1058 correspondents in 2006 found that 75% OF them would not return to a website that.
OSG Area Coordinator’s Report: Workload Management February 9 th, 2011 Maxim Potekhin BNL
OSG Area Coordinator’s Report: Workload Management April 20 th, 2011 Maxim Potekhin BNL
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
Designing a Scalable Enterprise Project Management Architecture Ken Toole Platform Test Manager MS Project Microsoft Corporation.
NOVA Networked Object-based EnVironment for Analysis P. Nevski, A. Vaniachine, T. Wenaus NOVA is a project to develop distributed object oriented physics.
Development of a noSQL storage solution for the Panda Monitoring System “Database Futures” Workshop at CERN June 7 th, 2011 Maxim Potekhin Brookhaven National.
OSG Area Coordinator’s Report: Workload Management Maxim Potekhin BNL
Tarball server (for Condor installation) Site Headnode Worker Nodes Schedd glidein - special purpose Condor pool master DB Panda Server Pilot Factory -
Plethora: A Wide-Area Read-Write Storage Repository Design Goals, Objectives, and Applications Suresh Jagannathan, Christoph Hoffmann, Ananth Grama Computer.
What is SAM-Grid? Job Handling Data Handling Monitoring and Information.
11 CLUSTERING AND AVAILABILITY Chapter 11. Chapter 11: CLUSTERING AND AVAILABILITY2 OVERVIEW  Describe the clustering capabilities of Microsoft Windows.
VMware vSphere Configuration and Management v6
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
CS525: Big Data Analytics MapReduce Computing Paradigm & Apache Hadoop Open Source Fall 2013 Elke A. Rundensteiner 1.
Copyright 2007, Information Builders. Slide 1 Machine Sizing and Scalability Mark Nesson, Vashti Ragoonath June 2008.
Creating SmartArt 1.Create a slide and select Insert > SmartArt. 2.Choose a SmartArt design and type your text. (Choose any format to start. You can change.
PanDA Status Report Kaushik De Univ. of Texas at Arlington ANSE Meeting, Nashville May 13, 2014.
INFSO-RI Enabling Grids for E-sciencE ARDA Experiment Dashboard Ricardo Rocha (ARDA – CERN) on behalf of the Dashboard Team.
Global ADC Job Monitoring Laura Sargsyan (YerPhI).
Database Systems, 8 th Edition SQL Performance Tuning Evaluated from client perspective –Most current relational DBMSs perform automatic query optimization.
OSG Area Coordinator’s Report: Workload Management February 9 th, 2011 Maxim Potekhin BNL
Cloud Computing: Pay-per-Use for On-Demand Scalability Developing Cloud Computing Applications with Open Source Technologies Shlomo Swidler.
1 Copyright © 2008, Oracle. All rights reserved. Repository Basics.
BIG DATA/ Hadoop Interview Questions.
Cofax Scalability Document Version Scaling Cofax in General The scalability of Cofax is directly related to the system software, hardware and network.
A Scalable and Resilient PanDA Service for Open Science Grid Dantong Yu Grid Group RHIC and US ATLAS Computing Facility.
Table General Guidelines for Better System Performance
Flash Storage 101 Revolutionizing Databases
Open Source distributed document DB for an enterprise
Spark Presentation.
Maximum Availability Architecture Enterprise Technology Centre.
Database Performance Tuning &
CHAPTER 3 Architectures for Distributed Systems
Database Performance Tuning and Query Optimization
Copyright © 2011 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Chapter 2 Database System Concepts and Architecture.
1 Demand of your DB is changing Presented By: Ashwani Kumar
Oracle Architecture Overview
Introduction to Apache
Table General Guidelines for Better System Performance
Cloud computing mechanisms
Chapter 11 Database Performance Tuning and Query Optimization
Performance And Scalability In Oracle9i And SQL Server 2000
Database System Architectures
Presentation transcript:

Remote Site C Pilot Scheduler Pilots and Pilot Schedulers Jobs Statistics Production Dashboard Dynamic Data Movement Monitor Panda Server (Apache) Development of noSQL data storage for the ATLAS PanDA Monitoring System M.Potekhin Brookhaven National Laboratory for the ATLAS Collaboration DATABASE Task Buffer (job queue) Brokerage Policy, Priority Job Dispatcher CE Pilot Job Job Request and Dispatch Main components of Panda Architecture and Current Monitor Implementation Job Submission Interfaces Job Resource Manager Pilot Launch (Condor, PBS, LSF, etc) Overview of Panda Workload Management System Panda is built on the concept of Pilot Job Framework. Workload is assigned to successfully activated and validated Pilot Jobs, which are lightweight processes that probe the environment and act as a ‘smart wrapper’ for the payload. This 'late binding' of workload jobs to processing slots prevents latencies and failure modes in slot acquisition from impacting the jobs, and maximizes the flexibility of job allocation to globally distributed resources. A central and crucial component of the Panda architecture is its database, which at any given time reflects the state of both pilot and payload jobs, as well as stores a variety of vital configuration information. It is driven by the Panda Server, which is implemented as an Apache-based Python application and performs a variety of brokerage and workload management tasks. Guaranteed consistency of the database is crucial for correct operation of the Panda server. By accessing this database, another Panda component – the Monitoring System – offers to its users and operators a comprehensive and coherent view of the system and job execution, from high level summaries to detailed drill-down job diagnostics. Regional Production Distributed Analysis Remote Site A CE Remote Site B CE Characteristics of the currently deployed Panda Monitoring System and its database backend Panda Monitoring System is a Web application that performs complex tasks of database queries and data aggregation, in order to provide the end user with a variety of interconnected views of the state of the system and its configuration. The state of the system includes the status of pilot jobs, payload jobs, data that’s being transmitted in support of payload execution, and other parameters. It is reflected in the database, which in its current implementation relies on Oracle RDBM. The data is indexed based on most popular queries. Other optimization tools are employed such as data partitioning, hints, bind variables. Data is largely de-normalized. Also, the data is segregated into a “live” segment (table) which is frequently updated by the server, and the archive table, which remains largely static. Motivations For System Evolution With more than 500,000 new entries generated daily, and multi-terabyte data load, the Oracle database used in Panda is approaching the scale of what is called “Big Data”. Despite ongoing query optimization effort, there are capability and performance limits imposed by available resources of the current Oracle installation. We observe that for purposes of monitoring applications which access archived data we do not use any of RDBMS functionality (such as capability to perform joins, or hard consistency). In situations like this, a variety of noSQL database solutions are often employed by major Web services that require extreme degrees of availability, resilience and scalability, such as Google, Amazon, Facebook, Twitter, Digg and many others, as well as “brick and mortar” businesses. Our goal is to make our systems future-proof and gain experience with technology we may use going forward. Panda Monitor (Apache) Operations Dashboard Database Access Layer and other utilities File-based cache Caching Logic Views: SQL Users and Operators HTTP/HTML Role of noSQL DB in Panda Monitoring We plan to use noSQL back-end for finalized, reference portion of the data where consistency is not essential but sheer performance and scalability are. Access to the data is implemented as a Web service, which furnishes data to clients in JSON format. Choice of the noSQL platform Based on current usage pattern, the criteria for choosing a noSQL platform include: Horizontally scalable read performance High write performance in order to efficiently absorb data coming from Oracle Indexing capability (either built-in or not difficult to implement by hand) roughly matching that in Oracle There are a number of noSQL solutions that we could use. Based on experience already existing in ATLAS, we chose Cassandra as our platform for monitoring data migration, as it satisfies all of the above requirements. Architecture of the ATLAS PanDA Monitoring System using noSQL ORACLE DB Containers and staging Data Extractor (finalized job, file and other data) Cassandra cluster External Clients and Systems Features of Cassandra Schema-less (“column families” are collections of dictionaries, searchable ) Ring-like architecture of the cluster; key space divided according to md5 hash; transparent load balancing Incrementally and linearly scalable – capacity can be added with no downtime Fault-tolerant, with data replication within the cluster as well as cross-data center Durability of write operation (once completed, data will survive hardware failure) Cache tuning options Local Disk Cache Web Service (Apache/Django) csv files in Amazon S3 HTTP/JSON Pycassa client Indexing Service Loader Implementation of indexes Two types of indexing solution were tested: custom indexes implemented in the application, as distinct column families,(trade ease of maintenance for disk space and performance) “native” indexing available in Cassandra since version 0.7 in 2010 (trade disk space for ease of maintenance) Based on experience, we chose native indexing scheme. To make it work, we augment the data with composite elements, i.e. add columns like PRODUSERNAME+JOBSETID, COMPUTINGSITE+PRODSOURCELABEL+DATE etc Hardware considerations Write operation is CPU bound Read operation is disk bound RAM will help efficient caching of keys and data Network speed is important Multiple data centers can be used for increased redundancy and fault tolerance VM is not optimal for that application since Cassandra is resource-hungry Panda Server Hardware configuration Three different configurations were tested in progression: 1.Cluster of 4 VMs at CERN(each with 2 cores, 4GB RAM) 2.Dedicated cluster of 3 servers at BNL (each with 24 cores, 48GB RAM, 6 spindles in RAID0) 3.Dedicated cluster of 3 servers at BNL (each with 24 cores, 48GB RAM, SSD) Data was loaded either from a disk cache residing on a separate machine, or via an application accessing our repository in Amazon S3 Scalability testing and comparison to Oracle We have analyzed the pattern of queries done against the Oracle database and used most popular queries for our scalability testing and comparison of our Cassandra cluster to the existing Oracle installation. We find that with configuration utilizing SSD storage, the Cassandra cluster was capable of scaling to at least 3 times the load typically experienced by the Oracle installation. The load was controlled by number of clients hitting the database in parallel and the test was bound by the number of client threads available. In primary key queries, in terms of plain speed Cassandra consistently outperformed Oracle with 3 to 5ms per query vs 50ms with Oracle. For queries relying in indexes, the Cassandra cluster had a response time of about 4ms vs a spread of 0.5 to 15 ms in comparable Oracle queries. Conclusions Current setup of the BNL Cassandra cluster provides acceptable performance and scaling characteristics for use in monitoring We demonstrated flexibility and functionality of composite indexes in Cassandra Focus will now shift to building a Web service that would serve JSON data to clients