• Share
  • Email
  • Embed
  • Like
  • Save
  • Private Content
Hadoop Ecosystem Architecture Overview
 

Hadoop Ecosystem Architecture Overview

on

  • 1,225 views

Hadoop Ecosystems overview and diagrams - helps to understand list of subprojects in hadoop in diagramatic way.

Hadoop Ecosystems overview and diagrams - helps to understand list of subprojects in hadoop in diagramatic way.

Statistics

Views

Total Views
1,225
Views on SlideShare
1,223
Embed Views
2

Actions

Likes
6
Downloads
2
Comments
0

1 Embed 2

https://twitter.com 2

Accessibility

Upload Details

Uploaded via as Adobe PDF

Usage Rights

© All Rights Reserved

Report content

Flagged as inappropriate Flag as inappropriate
Flag as inappropriate

Select your reason for flagging this presentation as inappropriate.

Cancel
  • Full Name Full Name Comment goes here.
    Are you sure you want to
    Your message goes here
    Processing…
Post Comment
Edit your comment

    Hadoop Ecosystem Architecture Overview Hadoop Ecosystem Architecture Overview Presentation Transcript

    • Hadoop Technologies Architecture Overview @senthil245 Mail - senthil245@gmail.com
    • DISTRIBUTED CLUSTER ARCHITECTURE: MASTER/SLAVE
    • HADOOP CORE
    • MAPREDUCE PATTERNS
    • WHEN MAPREDUCE Since the MapReduce is running within a cluster of computing nodes, the architecture is very scalable. • In other words, if the data size is increased by the factor of x, the performance should be still constant if we are adding a predictable/fixed factor of y. The graph on the right is illustrating the relationship between the size of the data (xaxis) and processing time (y-axis). •The blue color curve is the process using traditional programming. On the other hand, the black color curve is the process using Hadoop. When the data size is small, traditional programming is better performance because the bootstrap of Hadoop is expensive (Copy the data within the cluster, inter-nodes communication, etc.). Once the data size is big enough, the penalty of the Hadoop bootstrap becomes invisible. •Hence Hadoop is best suited for Big Data crunching ideally in terms of petaBytes and is not suited for implementing common data integration patterns
    • APACHE SQOOP
    • APACHE FLUME
    • APACHE CHUKWA
    • HDFS
    • APACHE OOZIE – WORKFLOW SCHEDULER (CHECK AZKABAN & LINKEDIN OPENSOURCE)
    • PIG AND HQL (DO NOT USE HQL)
    • APACHE S4 (STREAM PROCESSING)(ALSO CHECK KAFKA AND STORM)
    • APACHE ZOOKEEPER SERVICE (ALSO CHECK APACHE HUE)
    • APACHE HIVE
    • APACHE HCATALOG, HIVE AND HBASE