An agile big data analytic engine. The BigObject: What and Why? to analyze big data on real-time basis Mission an agile analytic engine which stores and.

Slides:



Advertisements
Similar presentations
Indexing Large Data COMP # 22
Advertisements

Database Scalability, Elasticity, and Autonomy in the Cloud Agrawal et al. Oct 24, 2011.
10 REASONS Why it makes a good option for your DB IN-MEMORY DATABASES Presenter #10: Robert Vitolo.
CS 153 Design of Operating Systems Spring 2015
OS Fall ’ 02 Introduction Operating Systems Fall 2002.
Chapter 8 Operating System Support
Chapter 9 Virtual Memory Produced by Lemlem Kebede Monday, July 16, 2001.
Virtual Memory Deung young, Moon ELEC 5200/6200 Computer Architecture and Design Lectured by Dr. V. Agrawal Lectured by Dr. V.
Memory Management April 28, 2000 Instructor: Gary Kimura.
Databases Chapter 11.
Data Warehousing: Defined and Its Applications Pete Johnson April 2002.
Virtual Memory By: Dinouje Fahih. Definition of Virtual Memory Virtual memory is a concept that, allows a computer and its operating system, to use a.
Tutorial 6 Memory Management
Database Systems – Data Warehousing
Bioinformatics Sean Langford, Larry Hale. What is it?  Bioinformatics is a scientific field involving many disciplines that focuses on the development.
Overview of Computing. Computer Science What is computer science? The systematic study of computing systems and computation. Contains theories for understanding.
1 INTRODUCTION TO DATABASE MANAGEMENT SYSTEM L E C T U R E
Hadoop/MapReduce Computing Paradigm 1 Shirish Agale.
Introduction to Apache OODT Yang Li Mar 9, What is OODT Object Oriented Data Technology Science data management Archiving Systems that span scientific.
Chapter 1 Introduction. Objectives To explain the definition of computer architecture To discuss the history of computers To describe the von-neumann.
(C) 2008 Clusterpoint(C) 2008 ClusterPoint Ltd. Empowering You to Manage and Drive Down Database Costs April 17, 2009 Gints Ernestsons, CEO © 2009 Clusterpoint.
Copyright © 2015 McGraw-Hill Education. All rights reserved. No reproduction or distribution without the prior written consent of McGraw-Hill Education.
What is a Database? SECTION 1. Database Technology and its Evolution Decades long evolution Early data processing systems Today's systems New technology.
Introduction to Computer Architecture. What is binary? We use the decimal (base 10) number system Binary is the base 2 number system Ten different numbers.
OOP with Java, David J. Barnes Bits, Bytes, and Java1 The Challenge Writing programs well is a challenge. –From coder to software engineer. We want programs.
Operating Systems. Overview What is an Operating System (OS) What is an Operating System (OS) What Operating Systems do. What Operating Systems do. Operating.
What is Big Data and Why Do We Need it?
1. 2 Preface In the time since the 1986 edition of this book, the world of compiler design has changed significantly 3.
+ Big Data IST210 Class Lecture. + Big Data Summary by EMC Corporation ( More videos that.
Caching Principles & Paging Performance CS-502 (EMC) Fall Caching Principles and Paging Performance CS-502, Operating Systems Fall 2009 (EMC) (Slides.
CS 127 Introduction to Computer Science. What is a computer?  “A machine that stores and manipulates information under the control of a changeable program”
Introduction: Memory Management 2 Ideally programmers want memory that is large fast non volatile Memory hierarchy small amount of fast, expensive memory.
Copyright © 2006, GemStone Systems Inc. All Rights Reserved. Increasing computation throughput with Grid Data Caching Jags Ramnarayan Chief Architect GemStone.
 Data Sets too Large & Complex for Companies to Manage Within Traditional IT Systems  3 “Vs” › Volume › Velocity › Variety  Opportunity.
Chapter 8: Installing Linux The Complete Guide To Linux System Administration.
Enterprise Solutions Chapter 11 – In-memory Technology.
Chapter 1 Introduction. Components of a Computer CPU (central processing unit) Executing instructions –Carrying out arithmetic and logical operations.
What we know or see What’s actually there Wikipedia : In information technology, big data is a collection of data sets so large and complex that it.
IMS 4212: Database Implementation 1 Dr. Lawrence West, Management Dept., University of Central Florida Physical Database Implementation—Topics.
CS4315A. Berrached:CMS:UHD1 Introduction to Operating Systems Chapter 1.
Copyright © 2016 Pearson Education, Inc. Modern Database Management 12 th Edition Jeff Hoffer, Ramesh Venkataraman, Heikki Topi CHAPTER 11: BIG DATA AND.
Cloud Computing: Pay-per-Use for On-Demand Scalability Developing Cloud Computing Applications with Open Source Technologies Shlomo Swidler.
Virtual Memory By CS147 Maheshpriya Venkata. Agenda Review Cache Memory Virtual Memory Paging Segmentation Configuration Of Virtual Memory Cache Memory.
What is the Big Data Challenge? Organizations are seeking solutions that combine the real-time analytics capabilities of SAP HANA and accessibility to.
Operating Systems, Winter Semester 2011 Practical Session 9, Memory 1.
Computer Architecture Organization and Architecture
Managing Data Resources File Organization and databases for business information systems.
CT101: Computing Systems Introduction to Operating Systems.
© 2017 by McGraw-Hill Education. This proprietary material solely for authorized instructor use. Not authorized for sale or distribution in any manner.
Virtual Memory (Section 9.3). The Need For Virtual Memory Many computers don’t have enough memory in RAM to accommodate all the programs a user wants.
W HAT IS SAP HANA? HANA - High-Performance Analytic Appliance What is SAP HANA ? Is SAP HANA An another database …. ? A modern column store database ….?
Chapter 3 Data Representation
Aga Private computer Institute Prepared by: Srwa Mohammad
Big Data Enterprise Patterns
CS422 Principles of Database Systems Course Overview
Hierarchy of Data in a Database
ICT Database Lesson 1 What is a Database?.
Ministry of Higher Education
به نام خدا Big Data and a New Look at Communication Networks Babak Khalaj Sharif University of Technology Department of Electrical Engineering.
CSCI206 - Computer Organization & Programming
O.S Lecture 13 Virtual Memory.
Data Structures and Algorithms
GEOMATIKA UNIVERSITY COLLEGE CHAPTER 2 OPERATING SYSTEM PRINCIPLES
Simulation of computer system
Caching Principles and Paging Performance
Practical Session 9, Memory
Caching Principles and Paging Performance
Big DATA.
IBM Tivoli Storage Manager
Presentation transcript:

An agile big data analytic engine

The BigObject: What and Why? to analyze big data on real-time basis Mission an agile analytic engine which stores and computes big data on commodity machines. Position

The BigObject: Target 10 6 Computations/sec Mega Operations 10 9 Computation/sec Giga Operations 10 6 – 10 9 per sec ► Exploit the power of hardware ► Maximize the performance a single machine can achieve Our Focus

Application Computing Engine for BI Tool –Multi-dimensional Infographics –Pre-built Components: Association for Recommendation System Significance for Market Segmentation Log Analysis –Exception Detection Computing Engine in Scientific Fields –Climate Simulation –DNA Sequencing

Performance Compute 1M Data Records Compute 100M Data Records 0.07 BigObject MySQL PostgresSQL Commercial DB

Why the BigObject Outperforms Others? Technical Principles of the BigObject

Dealing With Big Data The Portraits of Big Data –3Vs of big data: Volume, Velocity, Variety –We have dealt with Volume by Scale-out and Velocity by Scale-up Another V shouldn’t be neglected – Valence, the degree of interdependency among data components, which normally causes data shuffling. In-place Computing Model is designed to solve high-valence big data problems.

Scale-Out, Scale-Up, and Spread out the data components over a flat, virtually infinite address space, where data are ready for the CPU to compute. Scale-In Scale-Out Scale-Up

2 64 : Nearly Infinite The introduction of 64-bit CPU evokes revisit to the software discipline. When address space is nearly infinite, why don’t we place the data components over it and make them ready for computations instead of moving the data to program?

Time to Revisit the Computing Model Traditional Approach In-place Approach CPU File Virtual Memory Space In-place Object CPU DISK Database Table RAM 1. Data Retrievals 2. Juggling 3. Swapping Paging (On-demand)

The BigObject Data live and work in a flat and infinite address space. Big data is organized in the way able to trade time with space. In-place Computing ApproachAn in-memory database (no ACID) Can be implemented on every data node. Complementary to Distributed Approach

Website: Contact: