2. 2
The Three V’s of Big Data
Varity
Unstructured and semi-structured data is becoming as
strategic as traditional structured data. (Text, Machine
logs, clickstream, Social blog..etc)
Volume
Data coming in from new source as well as increased
regulation in multiple areas means storing more data for
longer period of time.
Velocity
Machine data as well as data coming for new source is
being ingested at speeds not even imagined a few
years ago.
3. 3
6 Key Hadoop Data Types
• How your Customer FeelsSentiment
• Website Visitors DataClickstream
• Data from Remote Sensors and MachinesSensor/Machine
• Location Based dataGeographic
• Log files automatically created by serversServer Logs
• Millions of web pages, emails and documentsText
4. 4
Changes in Analyzing Data
Big Data is fundamentally changing the way we analyze information.
Ability to analyze vast amounts of data rather than evaluating sample
sets.
Historically we have had to look at causes. Now we can look at patterns
and correlate in data that give us much better prospective.
5. 5
The Need of Hadoop
Store and use all types of data
Process all the data
Scalability
Commodity hardware
Scale (Storage and Processing)
Traditional
DBMS
EDW
MPP
Analytics
No SQL Hadoop
Platform
6. 6
Integrating Hadoop
RDBMS
Streams
Social Media
Data Marts
Machine Logs
Sqoop/Flume
O
p
e
n
S
o
u
r
c
e
E
T
L
S
t
r
e
a
m
i
n
g
E
T
L
Direct Access
No SQL
O
p
e
n
S
o
u
r
c
e
E
T
L
Data Warehouse Access
EDW
D
a
t
a
M
a
r
t
A
c
c
e
s
s
Big Data Access
B
u
si
n
e
s
s
In
te
lli
g
e
n
c
e
P
la
tf
o
r
m
Cloud
Performance
& Adhoc
reporting
OLAP
Dashboards
Exploratory
Visualization
Statistical
Analytics
Machine
Learning
7. 7
What is Hadoop ?
o Framework for solving data intensive processes
o Designed to scale massively
o Processes all the contents of a file (instead of attempting to read
portion of a file)
o Hadoop is very fast for very large jobs
o Hadoop is not fast for small jobs
o It does not provide caching or indexing (tools like HBase can
provide these features if needed)
o Designed for hardware and software failures
8. 8
What is Hadoop 2.0?
The Apache Hadoop 2.0 project consists of the following modules
o Hadoop Common: the utilities that provide support for the other
Hadoop modules
o HDFS: the Hadoop Distributed File System
o YARN: a framework for job scheduling and cluster resource
management.
o MapReduce: for processing large data sets in a scalable and
parallel fashion
9. 9
What is Yarn?
YARN is a sub-project of Hadoop at the Apache Software Foundation that takes
Hadoop beyond batch processing to enable broader data-processing
It extends the Hadoop platform by supporting non-MapReduce workloads
associated with other programming models
The core concept of YARN was born out of a need to have Hadoop work for more
real-time and streaming capabilities
As more and more data landed in Hadoop, enterprises have demanded that
Hadoop extend its capabilities
As part of Hadoop 2.0, YARN takes the resource management capabilities that
were in MapReduce and packages them so they can be used by new engines
Streamlines MapReduce to do what it does best - process data
Run multiple applications in Hadoop, all sharing a common resource
management
Many organization are already building application on YARN in order to bring
them IN to Hadoop
With Hadoop 2.0 and YARN, organizations can use Hadoop for streaming,
interactive and a world of other Hadoop-based applications
10. 10
Yarn taking Hadoop Beyond Batch
With YARN, Applications run natively in Hadoop
Applications Run Natively IN Hadoop
HDFS2 (Redundant, Reliable Storage)
YARN (Cluster Resource Management)
BATCH
(MapReduce)
INTERACTIVE
(Tez)
STREAMING
(Storm, S4,…)
GRAPH
(Giraph)
IN-MEMORY
(Spark)
HPC MPI
(OpenMPI)
ONLINE
(HBase)
OTHER
(Search)
(Weave…)
11. 11
How it Works
Raw
Data
1. Put the data into HDFS in Raw
Format 2. Use Pig to explore and Transform
3. Data Analytics use Hive to query the
data
4. Data Scientist use MapReduce, R
and Mahout to mine the data
Hadoop Distributed File
System
Structured
Data
Answer to
Question = $$
Predictive KPI =
##
12. 12
Getting Relational Data & Raw Data in to Hadoop
Raw Data
Hadoop Distributed File
System
Table in RDBMS
Sqoop Job Sqoop is a tool to transfer data between
RDBMS to Hadoop
Flume Agent
Flume is a tool to
streaming data in to
Hadoop
13. 13
What is Pig?
Pig is an extension of Hadoop that simplifies the ability to query
large HDFS datasets
Pig is made up of two main components:
• A data processing language called Pig Latin
• A compiler that compiles and runs Pig Latin scripts
Pig was created at Yahoo! to make it easier to analyze the data in
HDFS without the complexities of writing a traditional MapReduce
program
With Pig, you can develop MapReduce jobs with a few lines of Pig
Latin
14. 14
What is Hive?
Hive is a subproject of the Apache Hadoop project that provides a
data warehousing layer built on top of Hadoop
Hive allows you to define a structure for your unstructured big data,
simplifying the process of performing analysis and queries by
introducing a familiar, SQL-like language called HiveQL
Hive is for data analysts familiar with SQL who need to do ad-hoc
queries, summarization and data analysis on their HDFS data
Hive is not a relational database
Hive uses a database to store metadata, but the data that Hive
processes is stored in HDFS
Hive is not designed for on-line transaction processing and does not
offer real-time queries and row level updates
15. 15
1. Client sends a request to the NameNode
and a file to HDFS
2. NameNode tells client how and where to
distribute the blocks
Big Data
3. Client breaks the data in to blocks and
distributes the blocks to the DataNode
4. DataNode replicates the blocks (as instructed
by NameNode
How HDFS Works?
19. 19
Hortonworks HDP: Enterprise Hadoop 1.x Distribution
OS Cloud VM Appliance
PLATFORM
SERVICES
HADOOP
CORE
Enterprise Readiness
High Availability, Disaster Recovery,
Security and Snapshots
HORTONWORKS
DATA PLATFORM (HDP)
OPERATIONAL
SERVICES
DATA
SERVICES
HIVE
(HCATALOG)
PIG HBASE
OOZIE
AMBARI
HDFS
MAP REDUCE
Hortonworks
Data Platform (HDP)
Enterprise Hadoop
• The ONLY 100% open source and
complete distribution
• Enterprise grade, proven and
tested at scale
• Ecosystem endorsed to ensure
interoperability
SQOOP
FLUME
NFS
LOAD &
EXTRACT
WebHDFS
20. 20
Hadoop 2.0… The Enterprise Generation
Business Value
Big Data
Transactions,
Interactions,
Observations
Single Platform
Multiple Use
BATCH
INTERACTIVE
ONLINE
1.0 Architected for the Large Web Properties
2.0 Architected for the Broad Enterprise
Enterprise Requirements Hadoop 2.0 Features
Mixed workloads YARN
Interactive Query Hive on Tez
Reliability Full Stack HA
Point in time Recovery Snapshots
Multi Data Center Disaster Recovery
ZERO downtime Rolling Upgrades
Security Knox Gateway
21. 21
HDP: Enterprise Hadoop 2.0 Distribution
OS/VM Cloud Appliance
PLATFORM
SERVICES
HADOOP
CORE
Enterprise Readiness
High Availability, Disaster
Recovery, Rolling Upgrades,
Security and Snapshots
HORTONWORKS
DATA PLATFORM (HDP)
OPERATIONAL
SERVICES
DATA
SERVICES
HIVE &
HCATALOG
PIG HBASE
HDFS
MAP
Hortonworks
Data Platform (HDP)
Enterprise Hadoop
• The ONLY 100% open source and
complete distribution
• Enterprise grade, proven and
tested at scale
• Ecosystem endorsed to ensure
interoperability
SQOOP
FLUME
NFS
LOAD &
EXTRACT
WebHDFS
KNOX*
OOZIE
AMBARI
FALCON*
YARN*
TEZ* OTHERREDUCE
22. 22
Current Architecture in my Project
Source System Integration Layer Enterprise Data Warehouse
Layer
Semantic
Layer
Presentation Layer
Oracle
DB2
SQL
Server
Data
Profiling
Data
Extraction
Data
Quality
Data
Transformation
Data
Loading
MetadataManagement
Scheduling
Semantic/Mart
Enterprise Data
Warehouse
Staging
Flat-
file/.cvs
XML
Metadata Management
Data Governence
Data Quality
SAP BO
Universe
SAP BO Reports
Landing
Other
Systems
Get the Data using Sqoop
Use Hive External & Managed
Table
23. 23
Lambda Architecture
The Lambda-Architecture aims to satisfy the needs for a robust system that is fault-tolerant, both against hardware failures
and human mistakes, being able to serve a wide range of workloads and use cases, and in which low-latency reads and
updates are required. The resulting system should be linearly scalable, and it should scale out rather than up.
1. All data entering the system is
dispatched to both the batch layer
and the speed layer for processing.
2. The batch layer has two functions:
(i) managing the master dataset (an
immutable, append-only set of raw
data), and (ii) to pre-compute the
batch views.
3. The serving layer indexes the batch
views so that they can be queried in
low-latency, ad-hoc way.
4. The speed layer compensates for
the high latency of updates to the
serving layer and deals with recent
data only.
5. Any incoming query can be
answered by merging results from
batch views and real-time views.