Download presentation
Presentation is loading. Please wait.
Published byShanon Leonard Modified over 9 years ago
1
SEMINAR ON Guided by: Prof. D.V.Chaudhari Seminar by: Namrata Sakhare Roll No: 65 B.E.Comp
2
Large businesses needed to go through terabytes and petabytes of data. This data was initially control by a single powerful computer. But due to its limitation, it can handle data up to certain limits. To solve this problem, Google publicized MapReduce. MapReduce : A system which supports distributed computing on large data sets on clusters. Many other businesses were facing the same problem of scaling. Therefore, Doug Cutting developed an open source version of MapReduce system called HADOOP.
3
Hadoop is framework of tools. The objective of hadoop is,it supports running application on big data. It is an open source set of tools and distributed under Apache License. It is powerful tool designed for deep analysis and transaction of very large data.
4
The keyword behind hadoop is BIG DATA. Big data facing challenges Velocity Variety Volume Big Data
5
BIG DATA Powerful Computer Processed by BIG DATA Powerful Computer Processing Limits
6
BIG DATA Broken Into Pieces
7
BIG DATA Computation Combined Result
8
MapReduce HDFS Task tracker Name Node Date Node Job Tracker
9
Task tracker Data node Task tracker Data node Name node Task trackerJob tracker Master Slave
10
Task tracker Data node Task tracker Data node Name node Task trackerJob tracker
11
Task tracker Data node Task tracker Data node Name node Task trackerJob tracker Master Slave
12
Task tracker Data node Task tracker Data node
13
Task tracker Data node Task tracker Data node Name node Task trackerJob tracker Master Slave HDFS
14
Task tracker Data node Task tracker Data node Name node Task trackerJob tracker Master Slave MAPREDUCE
15
Task tracker Data node Task tracker Data node Name node Task trackerJob tracker Master Slave Tables are backed up
16
Where the file is located How to manage failures How to break computations into pieces How to program for scaling Don’t have to worry about Programme r
17
Main Features Of Hadoop : Works on distributed model.. :It Works on numerous low cost computer instead of single powerful computer. Linux based set of tools. : It Works On Linux Operating System.
18
Tools Of HADOOP Scoop Flume Oozie Pig Mahout Hbase Hive
19
Yahoo IBM FACEBOOK AMAZON AMERICAN AIRLINES THE NEWYORK TIMES EBAY
Similar presentations
© 2024 SlidePlayer.com. Inc.
All rights reserved.