Hug India Jul10 Hadoop Map

3,525 views

Published on

Hadoop Ecosystem Map

0 Comments
0 Likes
Statistics
Notes
  • Be the first to comment

  • Be the first to like this

No Downloads
Views
Total views
3,525
On SlideShare
0
From Embeds
0
Number of Embeds
2
Actions
Shares
0
Downloads
11
Comments
0
Likes
0
Embeds 0
No embeds

No notes for slide
  • How did it start- huge data on the web Nutch built to crawl this web data Huge data had to saved- HDFS was born How to use this data Map reduce framework built for coding and running analytics – java, any language-streaming, pipes How to get in unstructured data – Web logs, Click streams, Apache logs, Server logs- fuse,webdav, chukwa, flume, Scribe RDBMS – the next data sources Hiho and sqoop for loading data into HDFS High level interfaces required over low level map reduce programming– Pig, Hive, Jaql Hive -SQL like interface- BI tools with advanced UI reporting- drilldown etc- Intellicus Workflow tools over Map reduce processes and High level languages Monitor and manage hadoop, run jobs/hive, view HDFS – high level view- Hue, karmasphere, eclipse plugin Support frameworks- Avro (Serialization), Zookeeper (Coordination) More High level interfaces/uses- Mahout, Elastic map Reduce OLTP- also possible - Hbase
  • Hug India Jul10 Hadoop Map

    1. 1. HUG - July 2010 Hadoop Today! Presenter- Sanjay Sharma
    2. 2. Hadoop Ecosystem Map Java Applications Web Data RDBMS Scribe OLTP Structured Data hiho Sqoop File system Engine + Logic Unstructured Data High Level Interfaces Workflow Cascading Monitor/manage Hadoop ecosystem Support Cascading More High Level Interfaces
    3. 3. Q & A At the end of sessions Thank You <ul><li>http://code.google.com/p/hadoop-toolkit/ </li></ul><ul><li>Hadoop performance monitoring tool </li></ul><ul><li>Hadoop cluster setup tool </li></ul>

    ×