Build Your Web Analytics with node.js, Amazon DynamoDB and Amazon EMR (BDT203) | AWS re:Invent 2013
Upcoming SlideShare
Loading in...5
×
 

Build Your Web Analytics with node.js, Amazon DynamoDB and Amazon EMR (BDT203) | AWS re:Invent 2013

on

  • 3,109 views

Want to learn how to build your own Google Analytics? Learn how to build a scalable architecture using node.js, Amazon DynamoDB, and Amazon EMR. This architecture is used by ScribbleLive to track ...

Want to learn how to build your own Google Analytics? Learn how to build a scalable architecture using node.js, Amazon DynamoDB, and Amazon EMR. This architecture is used by ScribbleLive to track billions of engagement minutes per month. In this session, we go over the code in node.js, how to store the data in Amazon DynamoDB, and how to roll-up the data using Hadoop and Hive. Attend this session to learn how to move data quickly at any scale as well as how to use genomic analysis tools and pipelines for next generation sequencers using Globus on AWS.

Statistics

Views

Total Views
3,109
Views on SlideShare
3,109
Embed Views
0

Actions

Likes
11
Downloads
43
Comments
0

0 Embeds 0

No embeds

Accessibility

Categories

Upload Details

Uploaded via as Adobe PDF

Usage Rights

© All Rights Reserved

Report content

Flagged as inappropriate Flag as inappropriate
Flag as inappropriate

Select your reason for flagging this presentation as inappropriate.

Cancel
  • Full Name Full Name Comment goes here.
    Are you sure you want to
    Your message goes here
    Processing…
Post Comment
Edit your comment

Build Your Web Analytics with node.js, Amazon DynamoDB and Amazon EMR (BDT203) | AWS re:Invent 2013 Build Your Web Analytics with node.js, Amazon DynamoDB and Amazon EMR (BDT203) | AWS re:Invent 2013 Presentation Transcript

  • Building Your Own Web Analytics Service with node.js, Amazon DynamoDB, and Amazon Elastic MapReduce Jonathan Keebler - Founder, CTO - ScribbleLive November 13, 2013 © 2013 Amazon.com, Inc. and its affiliates. All rights reserved. May not be copied, modified, or distributed in whole or in part without the express consent of Amazon.com, Inc.
  • Who Am I? •Jonathan Keebler @keebler •Built video player for all CTV properties –Worked on news sites like CTV, TSN, CP24 •CTO, Founder of ScribbleLive •Bootstrapped a high scalability startup –Credit card limit wasn’t that high, had to find cheap ways to handle the load of top tier news sites
  • What is ScribbleLive? •Leading provider of real-time engagement management solutions •We enable real-time publication and syndication of digital content •Our platform is transforming the way the world’s largest brands and media approach communication and content creation, creating true real-time engagement
  • Some of our customers
  • Today •Learn to build your own analytics service – Seriously, we’re going to do it •node.js on Amazon EC2: web servers •Amazon DynamoDB: database •Hadoop/Hive on Amazon Elastic MapReduce (EMR): roll-up data
  • Why would we do this? •ScribbleLive tracks “engagement minutes” (EMs) across all customer sites – e.g., ESPN.com, CNN.com, Reuters.com – EM = 1 minute of a user watching a webpage – 2.5B per month, 120M+ per hour •Big analytics providers couldn’t do it – Didn’t have the features – Too inaccurate
  • How are we going to do this? Visitors Elastic Load Balancing node.js node.js node.js DynamoDB node.js
  • DynamoDB: data structure •Separate tables by timeframe – Minute (written by node.js directly) – Hour (EMR from minute data) – Day (EMR from hour data) – Month (EMR from day data) •Structure – Hash: Item (page id) – Range: Time (rounded to min, hour, day) – { Hits: 1 }
  • Elastic Load Balancing: AMI setup •Custom AMI – Loads source from SVN – Launches node.js
  • Elastic Load Balancing: Load balancing •1 load balancer •Cookies keep unique user on same instance •Auto-scaling – CPU >50% or network-in 50M bytes, triggers new servers coming online and added to Elastic Load Balancing
  • node.js: Overview of code •Accepts GET /?item={ID}&uid={UserID} •Dictionary/Array of how many GETs per item in this minute – Hits[Minute][“{ID}”]++ – Example: Hits[“1/1/2014 1:23:00”][“abcd”]++ •Dictionary/Array of Users already counted in Item:Minute (prevent double-counting) •At end of minute, write data back to DynamoDB
  • node.js: Bulk writing to DynamoDB •Writing all data back immediately in a loop = BAD! – Throughput would spike in that ~second – Would have to use higher throughput limit – More $$$$ •Instead, figure out how many writes need to happen / 60 seconds = how many writes per second you should do
  • node.js: Bulk writing to DynamoDB •Call to DynamoDB per item: – update: (atomic) add X to {ID}:{Minute}
  • Hadoop: What we map and reduce •To go from minute to hourly data – Round every minute down to the nearest hour (floor( Minute / 3600 ) * 3600) – Sum the # of “Hits” from each data point •Just look at the past 24 hours to save time •Do the same for hourly to daily, daily to monthly
  • Hadoop: Hive scripts INSERT OVERWRITE TABLE MetricsHourly SELECT Item, (floor( Time / 3600 ) * 3600) AS Time, SUM(Hits) AS Hits, from_unixtime(floor( Time / 3600 ) * 3600 ) AS TimeFriendly FROM Metrics WHERE Time >= floor( unix_timestamp() / 86400 ) * 86400 - ( 86400 * 1 ) GROUP BY Item, floor( Time / 3600 ) * 3600;
  • Hadoop: Setting Up EMR
  • Hadoop: Setting Up EMR • “Start an Interactive Hive Session” • Run a cron job every 15 minutes to check if the Hive job is complete • If complete, downloads newest Hive script and restarts the job • Amazon CloudWatch alarms if jobs taking longer than 12 hours
  • Hadoop: Cron Job #!/bin/sh JOBID=$(hadoop job -list | grep job_ | cut -f1) if [ -n "$JOBID" ];then echo "Another job already running"; else echo "Starting Hive job..." echo `date` starting >> /var/log/metricsdaily_starting wget -qO- http://DEPLOY/metrics/rollups.sql > /tmp/rollups.sql && hive -f /tmp/rollups.sql fi
  • Application API •RESTful API in the language of your choice •Calls to DynamoDB: –query: Hash:{ID} w/ Range:{Time A}-{Time B} •Since M-R could take a day to run, need to reconstruct hourly data from minutes for most recent 24 hours –e.g. if you want hourly data for last 2 days, take 24 hourly data pts from yesterday, and 24*60 minute data pts from today (convert to hourly data pts in code)
  • Performance
  • Performance
  • Please give us your feedback on this presentation BDT203 As a thank you, we will select prize winners daily for completed surveys!