• Share
  • Email
  • Embed
  • Like
  • Save
  • Private Content
Hadoop at datasift
 

Hadoop at datasift

on

  • 515 views

Presentation given at Edinburgh University Student Tech-Meetup on 6th Feb, 2013.

Presentation given at Edinburgh University Student Tech-Meetup on 6th Feb, 2013.

Statistics

Views

Total Views
515
Views on SlideShare
515
Embed Views
0

Actions

Likes
1
Downloads
7
Comments
0

0 Embeds 0

No embeds

Accessibility

Categories

Upload Details

Uploaded via as Adobe PDF

Usage Rights

© All Rights Reserved

Report content

Flagged as inappropriate Flag as inappropriate
Flag as inappropriate

Select your reason for flagging this presentation as inappropriate.

Cancel
  • Full Name Full Name Comment goes here.
    Are you sure you want to
    Your message goes here
    Processing…
Post Comment
Edit your comment

    Hadoop at datasift Hadoop at datasift Presentation Transcript

    • HADOOP ATDATASIFT
    • ABOUT ME Jairam Chandar Big Data Engineer Datasift @jairamc http://about.me/jairam http://jairam.meAnd I’m a Formula 1 Fan!
    • OUTLINE• What is Datasift ?• Where do we use Hadoop ? • The Numbers • The Use-cases • The Lessons
    • !! SALES PITCH ALERT !!
    • WHAT IS DATASIFT?
    • WHAT IS DATASIFT?
    • WHAT IS DATASIFT?
    • WHAT IS DATASIFT?
    • WHAT IS DATASIFT?
    • WHAT IS DATASIFT?
    • WHAT IS DATASIFT?
    • WHAT IS DATASIFT?
    • WHAT IS DATASIFT?
    • THE NUMBERS• Machines • HBase • 60 Machines as RegionServers •1 HMaster •3 Zookeeper nodes
    • THE NUMBERS• Machines • Hadoop • 135 Machines divided into 2 clusters • Datanodes/Tasktrakers • Namenodes with High-Availability Failover •1 Jobtracker each
    • THE NUMBERS• Machines • DL380 Gen8 • 2 * Intel Xeon E5646 @ 2.40GHz (24 core total) • 48GB RAM • 6 * 2 TB disks in JBOD (small partition on first disk for OS, rest is storage) • 1 Gigabit network links
    • THE NUMBERS• Data • Average load of 7500 interactions per second • Peak loads of 15000 interactions per second sustained over a min • Peak of 21000 interactions per second during superbowl • Total current capacity ~ 1.6 PB; Total current usage ~ 800 TB • Avg size of interaction 2 KB – thats ~ 1GB a min or ~ 2 TB a day with replication (RF = 3) • And that’s not it!
    • THE USE CASES• HBase • Recordings • Archive• Map/Reduce • Exports • Historics • Migration
    • THE USE CASES• Recordings • User defined streams • Stored in HBase for later retrieval • Export to multiple output formats and stores • <recording-id><interaction-uuid> • Recording-id is a SHA-1 hash • Allows recordings to be distributed by their key without generating hot- spots.
    • THE RECORDER
    • THE USE CASES• Exporter • Export data from HBase for customer • Export files ~ 5 – 10 GB or ~ 3-6 million records • MR over HBase using TableInputFormat • But the data needs to be sorted • TotalOrderPartioner
    • EXPORTER
    • HISTORICS
    • THE USE CASES• Twitter Import •2 years of Tweets • About 95,000,000,000 tweets • Over 300 TB with added augmentation • Import was not as simple as you would imagine
    • THE USE CASES• Archive • Not just the Firehose but the Ultrahose • Stored in HBase as well • HBase architecture (BigTable) creates Hotspots with Time Series data • Leading randomizing bit (see HBaseWD) • Pre-split regions • Concurrent writes
    • THE USE CASES• Historics • Export archive data • Slightly different from Exporter • Much larger time lines (1 – 3 months) • Controlled access to Hadoop cluster with efficient job scheduling • Unfiltered Input Data • Therefore longer processing time • Hence more optimizations required
    • HISTORICS
    • THE LESSONS• Tune Tune Tune (Default == BAD)• Based on use case tune - • Heap • Block Size • Memstore size• Keep number of column families low• Be aware of hot-spotting issue when writing time-series data
    • THE LESSONS• Use compression (eg. Snappy)• Ops need intimate understanding of system• Monitor system metrics (GC, CPU, Compaction, I/O) and application metrics (writes/sec etc)• Dont be afraid to fiddle with HBase code• Using a distribution is advisable
    • QUESTIONS? We are hiringhttp://datasift.com/about-us/careers