SlideShare a Scribd company logo
1 of 57
Introduction to Big data
A deep introduction about big data topic along with real advices of how to start a career in this hot topic.
Be ready to digest a concentrated big data tablet that will put you on the right way.
Agenda:
Data nowadays:
- Data types
- Fun facts about data nowadays.
- From where we generate data.
- Lake of data effect on business decisions.
- Future of data size.
Unlocking Big data solutions:
- Hadoop.
- Hadoop ecosystem Zoo
- Big data landscape
- Top Big data companies
- How to start a career in Big data
- Questions
Big Data:
- What’s big data?
- How big is the big data?
- The famous Vs about big data.
- Challenges of dealing with such data amount.
- Why to consider a career in big data?
Data unit measures:
2
Data Types:
information with a degree of
organization that is readily
searchable and quickly
consolidate into facts.
Examples: RDMBS, spreadsheet
information with a lack of
structure that is time and energy
consuming to search and find and
consolidate into facts
Exemples: email, documents, images,
reports
Semi Structured data : XML data
Challenges for Unstructured data:
How do you store
Billions of Files?
How long does it take to
migrate 100’s of TB’s or
data every 3-5 years
Data has no
structure
Resources Limitation
Data Redundancy Data Backup
Sources of data generation:
Social Media Sensors Cell Phones GPS Purchase
WWW E-mails Media streaming Healthcare IOT
Facts about data:
2
Facts about data:
70% of data is created by Individuals – but enterprises are
responsible for storing and managing 80% of it.
52% of travelers use social media to plan for their vacations.
35% of purchases on Amazon are though recommendations
75% of what people watch on Netflix are recommendations.
Facts about data:
Lake of data and business decisions :
2
Cost:
Even if RDBMS is used to handle and store “big data,” it will turn out to be very expensive.
Velocity:
Also, “big data” is generated at a very high velocity. RDBMS lacks in high velocity because it’s
designed for steady data retention rather than rapid growth.
Can traditional DBMS solve this ?
Data types:
Second, the majority of the data comes in a semi-structured or unstructured format from social
media, audio, video, texts, and emails. However, the second problem related to unstructured data is
outside the purview of RDBMS because relational databases just can’t categorize unstructured data.
They’re designed and structured to accommodate structured data such as weblog sensor and
financial data.
Size:
First, the data size has increased tremendously to the range of petabytes—one petabyte = 1,024
terabytes. RDBMS finds it challenging to handle such huge data volumes. To address this, RDBMS
added more central processing units (or CPUs) or more memory to the database management
system to scale up vertically.
What is Big data:
Big data is a term that describes the large volume of data – both structured and
unstructured – that generates on a day-to-day basis. But it’s not the amount of data
that’s important. It’s what organizations do with the data that matters. Big data can
be analyzed for insights that lead to better decisions and strategic business moves.
Big data is high-volume, high-velocity and/or high-variety information assets that
demand cost-effective, innovative forms of information processing that enable
enhanced insight, decision making, and process automation
Big data is a term for data sets that are so large or complex that traditional data
processing applications are inadequate to deal with them. Challenges include
analysis, capture, data curation, search, sharing, storage, transfer, visualization,
querying, updating and information privacy.
The Vs:
Velocity Volume Variety
Veracity Variability Visualization Value
Big data in action:
UPS stores a large amount of data – much of which comes
from sensors in its vehicles - GPS
ORION (On-Road Integration Optimization and Navigation)
Data Analytics
and Data Science
the world's
largest
operations
research project
savings of more
than 8.4 million
85 million miles
off of daily routes
Saved
$30 million/Day
Big data in action:
Walmart collects 2.5 petabytes of information from 1 million customers.
from 6000 store
Big data System (Kosmix)
Pricing
strategies
Advertising
campaigns
30% on their
Online sales
Revenue got
increased by 40%
“We want to know what every product in
the world is. We want to know who every
person in the world is. And we want to
have the ability to connect them together
in a transaction.”
-Neil Ashe, CEO of Global E-commerce at
Walmart
Big data in quotes:
“Without big data analytics, companies are blind and deaf, wandering out onto the web like
deer on a freeway.” – Geoffrey Moore
“Data is the new science. Big Data holds the answers.” – Pat Gelsinger
“With too little data, you won’t be able to make any conclusions that you trust. With loads of
data you will find relationships that aren’t real… Big data isn’t about bits, it’s about talent.” –
Douglas Merrill
“The world is one big data problem.” – Andrew McAfee
Big data market forecast:
The “big data” market is expected to cross $50 billion by 2017.
Big data jobs trend:
The median advertised salary for professionals with big data expertise
is $124,000 a year.
IBM , Cisco and Oracle together advertised 26,488 open positions that
required big data expertise in the last twelve months.
124000 usd to sar =
465012
/12
= 38751 SAR/Month
How to solve big data
Hadoop: is a big data analysis engine
What is Hadoop
The Apache Hadoop software library is a framework that allows for the distributed
processing of large data sets across clusters of computers using simple programming
models.
Hadoop is an open-source software framework for storing data and running applications
on clusters of commodity hardware. It provides massive storage for any kind of data,
enormous processing power and the ability to handle virtually limitless concurrent tasks
or jobs.
Hadoop history
Nutch is a well matured, production ready Web crawler. that enables fine
grained configuration, relying on Apache Hadoop™ data structures, which
are great for batch processing.
Why Hadoop is important ?
Ability to store and process huge amounts of
any kind of data, quickly.
With data volumes and varieties constantly
increasing, especially from social media and the
Internet of Things (IoT), that's a key
consideration.
Computing power. Hadoop's distributed computing model processes big data fast. The
more computing nodes you use, the more processing power you have.
Fault tolerance. Data and application processing are protected against hardware failure.
If a node goes down, jobs are automatically redirected to other nodes to make sure the
distributed computing does not fail. Multiple copies of all data are stored automatically.
Why Hadoop is important ?
Flexibility. Unlike traditional relational databases, you
don’t have to preprocess data before storing it. You
can store as much data as you want and decide how
to use it later. That includes unstructured data like
text, images and videos.
Low cost. The open-source framework is free and uses commodity hardware
to store large quantities of data.
Scalability. You can easily grow your system to handle more data simply by
adding nodes. Little administration is required.
Scalability
Horizontal scaling means that you scale by adding more
machines into your pool of resources
Vertical scaling means that you scale by adding more
power (CPU, RAM) to an existing machine
How is Hadoop being used?
Going beyond its original goal of searching millions (or billions) of web pages and returning
relevant results, many organizations are looking to Hadoop as their next big data platform.
Popular uses today include:
How is Hadoop being used?
How is Hadoop being used?
Hadoop ecosystem
Data
Ingestion
Sqoop
Flume
Storm
NiFi
Kafka
Querying Layer
(Hive,Impala)
Processing Layer
(MapReduce,Spark,Pig)
Storage Layer
(HDFS,Hbase,Hcatalog)
Cluster monitoring, provisioning and management
Hadoop | Data Ingestion
Apache Sqoop is a tool designed for efficiently transferring bulk data between
Apache Hadoop and structured data stores such as relational databases.
Hadoop | Data Ingestion
Flume is a distributed, reliable, and available service for efficiently collecting,
aggregating, and moving large amounts of log data.
It is robust and fault tolerant with tunable reliability mechanisms and many
failover and recovery mechanisms.
Hadoop | Data Ingestion
Storm is real-time computation system. Storm makes it easy to reliably process
unbounded streams of data, doing for real-time processing.
A Storm topology consumes streams of data and processes those streams in
arbitrarily complex ways, repartitioning the streams between each stage of
the computation however needed.
Hadoop | Data Ingestion
An easy to use, powerful, and reliable system to process and distribute data.
Apache NiFi supports powerful and scalable directed graphs of data routing,
transformation, and system mediation logic in a Web-based user interface
Hadoop | Data Ingestion
Kafka™ is used for building real-time data pipelines and streaming apps. It is
horizontally scalable, fault-tolerant, wicked fast, and runs in production in
thousands of companies.
Hadoop | Data Ingestion
Large scale log aggregator, and analytics.
Fluentd is an open source data collector for unified logging
layer.
Fluentd allows you to unify data collection and consumption
for a better use and understanding of data.
Apache Samza is a distributed stream processing framework.
It uses Apache Kafka for messaging, and Apache Hadoop YARN
to provide fault tolerance, processor isolation, security, and
resource management.
Hadoop | Data Storage Layer
Hadoop Distributed File System (HDFS) offers a way to store large files across
multiple machines. Hadoop and HDFS was derived from Google File System
(GFS) paper.
Hadoop | Data Storage Layer
Hadoop | Data Storage Layer
A distributed, column-oriented database. HBase uses HDFS for its underlying
storage, and supports both batch-style computations using MapReduce and
point queries (random reads)
Hadoop | Data Storage Layer
A metadata and table management system for Hadoop. It shares the metadata
with other tools like map reduce, Pig and Hive.
It provides one constant data model for all Hadoop tools along with a shared
schema.
Hadoop | Data Processing Layer
MapReduce is the heart of Hadoop. It is this programming paradigm that
allows for massive scalability across hundreds or thousands of servers in a
Hadoop cluster with a parallel, distributed algorithm.
Hadoop | Data Processing Layer
Hadoop | Data Processing Layer
A scripting SQL based language and execution environment for creating complex
MapReduce transformations. Functions are written in Pig Latin (the language)
and translated into executable MapReduce jobs. Pig also allows the user to
create extended functions (UDFs) using Java.
Hadoop | Data Processing Layer
In memory data analytics cluster computing framework originally developed in the
AMPLab at UC Berkeley.
Run programs up to 100x faster than Hadoop MapReduce in memory, or 10x faster on
disk.
Hadoop | Data Querying Layer
A distributed data warehouse built on top of HDFS to manage and organize
large amounts of data. Hive provides a query language based on SQL semantic
(HiveQL) which is translated by the runtime engine to MapReduce jobs for
querying the data.
Hadoop | Data Querying Layer
pen source massively parallel processing (MPP) SQL query engine for data
stored in a computer cluster running Apache Hadoop.
Hadoop | Management Layer
intuitive, easy-to-use Hadoop management web UI. Apache Ambari was
donated by Hortonworks team. It's a powerful and nice interface for Hadoop
and other typical applications from the Hadoop ecosystem.
Big data existing solutions:
Big data existing solutions:
Data Sources
YARN: A framework for job scheduling and cluster resource management.
A platform for manipulating data
stored in HDFS via a high-level
language called Pig Latin. It does
data extractions,
transformations and loading, and
basic analysis in patch mode
A data warehousing and SQL-like
query language that presents
data in the form of tables. Hive
programming is similar to
database programming.
Ambari: A web interface for managing, configuring and testing Hadoop services and components.
An open-source
cluster computing
framework with
in-memory
analytics.
HDFS :A platform for manipulating data
stored in HDFS via a high-level
language called Pig Latin. It does
data extractions,
transformations and loading, and
basic analysis in patch mode
HBase:A distributed, column-oriented
database. HBase uses HDFS for
its underlying storage, and
supports both batch-style
computations using MapReduce
and point queries.
HCatalog
A table and storage management
layer for Hadoop that enables Hadoop
applications (Pig, MapReduce, and
Hive) to read and write data to a
tabular form as opposed to files.
A distributed data processing
model and execution
environment that runs on large
clusters of commodity
machines.
MapReduce
A Scalable machine learning and data mining library
A high-performance
coordination service for
distributed applications.
is a Java Web application
used to schedule Apache
Hadoop jobs
data collection system for monitoring large
distributed systems.
A web-based tool for
provisioning, managing, and
monitoring Apache Hadoop
clusters. Ambari also provides a
dashboard for viewing cluster
health such as heatmaps and
ability to view MapReduce, Pig
and Hive applications visually
along with features to diagnose
their performance.
is a data serialization system
Other apache projects:
Apache Flink
is an open source
platform for
distributed stream
and batch data
processing.
Apache Falcon
Feed management
and data processing
platform
Apache Ranger
Ranger is a framework
to enable, monitor and
manage comprehensive
data security across the
Hadoop platform.
Apache Tez
to develop a generic
application which
can be used to
process complex
data-processing task
Apache Tika
toolkit detects and
extracts metadata
and text from over
a thousand
different file types
Apache Parquet
columnar storage
format available to
any project in the
Hadoop ecosystem
Apache Zeppelin
A web-based notebook
that enables interactive
data analytics.
Apache Drill
Schema-free SQL
Query Engine for
Hadoop, NoSQL
and Cloud Storage
Top Leading Big data companies
The Apache Software Foundation (ASF) is an American non-
profit corporation to support Apache projects
How to start
1. Identify business use cases tied to business
outcomes, metrics and your big data roadmap
2. Identify big data champions from both the
business and IT sides of your organization
3. Select infrastructure, tools and architecture
for your big data POC/implementation
4. Staff the project with the right big data skills
or a strategic big data implementation partner
5. Run your project/POC in sprints or short
projects with tangible and measurable
outcomes
6. Try to scale your success POC up to test your
Logic implementation against the big dataset.
What can I do now ?
Certification Path | Administration
Cloudera HortonWorks
Certification Path | Development
Cloudera HortonWorks
Certification Path | Data Science
Cloudera HortonWorks
Questions
Thanks For your time

More Related Content

Similar to Lecture 5 - Big Data and Hadoop Intro.ppt

Hadoop and Big Data Analytics | Sysfore
Hadoop and Big Data Analytics | SysforeHadoop and Big Data Analytics | Sysfore
Hadoop and Big Data Analytics | SysforeSysfore Technologies
 
Big Data using NoSQL Technologies
Big Data using NoSQL TechnologiesBig Data using NoSQL Technologies
Big Data using NoSQL TechnologiesAmit Singh
 
Exploring the Wider World of Big Data
Exploring the Wider World of Big DataExploring the Wider World of Big Data
Exploring the Wider World of Big DataNetApp
 
Big data data lake and beyond
Big data data lake and beyond Big data data lake and beyond
Big data data lake and beyond Rajesh Kumar
 
Big Data: Its Characteristics And Architecture Capabilities
Big Data: Its Characteristics And Architecture CapabilitiesBig Data: Its Characteristics And Architecture Capabilities
Big Data: Its Characteristics And Architecture CapabilitiesAshraf Uddin
 
Big data and data mining
Big data and data miningBig data and data mining
Big data and data miningPolash Halder
 
Big Data Session 1.pptx
Big Data Session 1.pptxBig Data Session 1.pptx
Big Data Session 1.pptxElsonPaul2
 
big data and hadoop
 big data and hadoop big data and hadoop
big data and hadoopahmed alshikh
 
Enterprise Data Lake
Enterprise Data LakeEnterprise Data Lake
Enterprise Data Lakesambiswal
 
Enterprise Data Lake - Scalable Digital
Enterprise Data Lake - Scalable DigitalEnterprise Data Lake - Scalable Digital
Enterprise Data Lake - Scalable Digitalsambiswal
 
Big data with hadoop
Big data with hadoopBig data with hadoop
Big data with hadoopAnusha sweety
 
The book of elephant tattoo
The book of elephant tattooThe book of elephant tattoo
The book of elephant tattooMohamed Magdy
 
Big data4businessusers
Big data4businessusersBig data4businessusers
Big data4businessusersBob Hardaway
 

Similar to Lecture 5 - Big Data and Hadoop Intro.ppt (20)

Data mining with big data
Data mining with big dataData mining with big data
Data mining with big data
 
Big Data
Big DataBig Data
Big Data
 
Hadoop and Big Data Analytics | Sysfore
Hadoop and Big Data Analytics | SysforeHadoop and Big Data Analytics | Sysfore
Hadoop and Big Data Analytics | Sysfore
 
Hadoop(Term Paper)
Hadoop(Term Paper)Hadoop(Term Paper)
Hadoop(Term Paper)
 
Big Data using NoSQL Technologies
Big Data using NoSQL TechnologiesBig Data using NoSQL Technologies
Big Data using NoSQL Technologies
 
Exploring the Wider World of Big Data
Exploring the Wider World of Big DataExploring the Wider World of Big Data
Exploring the Wider World of Big Data
 
Big data data lake and beyond
Big data data lake and beyond Big data data lake and beyond
Big data data lake and beyond
 
TSE_Pres12.pptx
TSE_Pres12.pptxTSE_Pres12.pptx
TSE_Pres12.pptx
 
Big Data: Its Characteristics And Architecture Capabilities
Big Data: Its Characteristics And Architecture CapabilitiesBig Data: Its Characteristics And Architecture Capabilities
Big Data: Its Characteristics And Architecture Capabilities
 
Big data and data mining
Big data and data miningBig data and data mining
Big data and data mining
 
Big Data Session 1.pptx
Big Data Session 1.pptxBig Data Session 1.pptx
Big Data Session 1.pptx
 
big data and hadoop
 big data and hadoop big data and hadoop
big data and hadoop
 
Unit 1
Unit 1Unit 1
Unit 1
 
Enterprise Data Lake
Enterprise Data LakeEnterprise Data Lake
Enterprise Data Lake
 
Enterprise Data Lake - Scalable Digital
Enterprise Data Lake - Scalable DigitalEnterprise Data Lake - Scalable Digital
Enterprise Data Lake - Scalable Digital
 
Hadoop
HadoopHadoop
Hadoop
 
Big data with hadoop
Big data with hadoopBig data with hadoop
Big data with hadoop
 
The book of elephant tattoo
The book of elephant tattooThe book of elephant tattoo
The book of elephant tattoo
 
Big Data
Big DataBig Data
Big Data
 
Big data4businessusers
Big data4businessusersBig data4businessusers
Big data4businessusers
 

Recently uploaded

Interdisciplinary_Insights_Data_Collection_Methods.pptx
Interdisciplinary_Insights_Data_Collection_Methods.pptxInterdisciplinary_Insights_Data_Collection_Methods.pptx
Interdisciplinary_Insights_Data_Collection_Methods.pptxPooja Bhuva
 
dusjagr & nano talk on open tools for agriculture research and learning
dusjagr & nano talk on open tools for agriculture research and learningdusjagr & nano talk on open tools for agriculture research and learning
dusjagr & nano talk on open tools for agriculture research and learningMarc Dusseiller Dusjagr
 
Beyond_Borders_Understanding_Anime_and_Manga_Fandom_A_Comprehensive_Audience_...
Beyond_Borders_Understanding_Anime_and_Manga_Fandom_A_Comprehensive_Audience_...Beyond_Borders_Understanding_Anime_and_Manga_Fandom_A_Comprehensive_Audience_...
Beyond_Borders_Understanding_Anime_and_Manga_Fandom_A_Comprehensive_Audience_...Pooja Bhuva
 
Details on CBSE Compartment Exam.pptx1111
Details on CBSE Compartment Exam.pptx1111Details on CBSE Compartment Exam.pptx1111
Details on CBSE Compartment Exam.pptx1111GangaMaiya1
 
Basic Intentional Injuries Health Education
Basic Intentional Injuries Health EducationBasic Intentional Injuries Health Education
Basic Intentional Injuries Health EducationNeilDeclaro1
 
Simple, Complex, and Compound Sentences Exercises.pdf
Simple, Complex, and Compound Sentences Exercises.pdfSimple, Complex, and Compound Sentences Exercises.pdf
Simple, Complex, and Compound Sentences Exercises.pdfstareducators107
 
latest AZ-104 Exam Questions and Answers
latest AZ-104 Exam Questions and Answerslatest AZ-104 Exam Questions and Answers
latest AZ-104 Exam Questions and Answersdalebeck957
 
AIM of Education-Teachers Training-2024.ppt
AIM of Education-Teachers Training-2024.pptAIM of Education-Teachers Training-2024.ppt
AIM of Education-Teachers Training-2024.pptNishitharanjan Rout
 
Towards a code of practice for AI in AT.pptx
Towards a code of practice for AI in AT.pptxTowards a code of practice for AI in AT.pptx
Towards a code of practice for AI in AT.pptxJisc
 
Sensory_Experience_and_Emotional_Resonance_in_Gabriel_Okaras_The_Piano_and_Th...
Sensory_Experience_and_Emotional_Resonance_in_Gabriel_Okaras_The_Piano_and_Th...Sensory_Experience_and_Emotional_Resonance_in_Gabriel_Okaras_The_Piano_and_Th...
Sensory_Experience_and_Emotional_Resonance_in_Gabriel_Okaras_The_Piano_and_Th...Pooja Bhuva
 
Basic Civil Engineering first year Notes- Chapter 4 Building.pptx
Basic Civil Engineering first year Notes- Chapter 4 Building.pptxBasic Civil Engineering first year Notes- Chapter 4 Building.pptx
Basic Civil Engineering first year Notes- Chapter 4 Building.pptxDenish Jangid
 
HMCS Max Bernays Pre-Deployment Brief (May 2024).pptx
HMCS Max Bernays Pre-Deployment Brief (May 2024).pptxHMCS Max Bernays Pre-Deployment Brief (May 2024).pptx
HMCS Max Bernays Pre-Deployment Brief (May 2024).pptxEsquimalt MFRC
 
How to Manage Global Discount in Odoo 17 POS
How to Manage Global Discount in Odoo 17 POSHow to Manage Global Discount in Odoo 17 POS
How to Manage Global Discount in Odoo 17 POSCeline George
 
How to Add New Custom Addons Path in Odoo 17
How to Add New Custom Addons Path in Odoo 17How to Add New Custom Addons Path in Odoo 17
How to Add New Custom Addons Path in Odoo 17Celine George
 
On_Translating_a_Tamil_Poem_by_A_K_Ramanujan.pptx
On_Translating_a_Tamil_Poem_by_A_K_Ramanujan.pptxOn_Translating_a_Tamil_Poem_by_A_K_Ramanujan.pptx
On_Translating_a_Tamil_Poem_by_A_K_Ramanujan.pptxPooja Bhuva
 
How to setup Pycharm environment for Odoo 17.pptx
How to setup Pycharm environment for Odoo 17.pptxHow to setup Pycharm environment for Odoo 17.pptx
How to setup Pycharm environment for Odoo 17.pptxCeline George
 
REMIFENTANIL: An Ultra short acting opioid.pptx
REMIFENTANIL: An Ultra short acting opioid.pptxREMIFENTANIL: An Ultra short acting opioid.pptx
REMIFENTANIL: An Ultra short acting opioid.pptxDr. Ravikiran H M Gowda
 
TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...
TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...
TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...Nguyen Thanh Tu Collection
 
SOC 101 Demonstration of Learning Presentation
SOC 101 Demonstration of Learning PresentationSOC 101 Demonstration of Learning Presentation
SOC 101 Demonstration of Learning Presentationcamerronhm
 

Recently uploaded (20)

Interdisciplinary_Insights_Data_Collection_Methods.pptx
Interdisciplinary_Insights_Data_Collection_Methods.pptxInterdisciplinary_Insights_Data_Collection_Methods.pptx
Interdisciplinary_Insights_Data_Collection_Methods.pptx
 
dusjagr & nano talk on open tools for agriculture research and learning
dusjagr & nano talk on open tools for agriculture research and learningdusjagr & nano talk on open tools for agriculture research and learning
dusjagr & nano talk on open tools for agriculture research and learning
 
Beyond_Borders_Understanding_Anime_and_Manga_Fandom_A_Comprehensive_Audience_...
Beyond_Borders_Understanding_Anime_and_Manga_Fandom_A_Comprehensive_Audience_...Beyond_Borders_Understanding_Anime_and_Manga_Fandom_A_Comprehensive_Audience_...
Beyond_Borders_Understanding_Anime_and_Manga_Fandom_A_Comprehensive_Audience_...
 
Details on CBSE Compartment Exam.pptx1111
Details on CBSE Compartment Exam.pptx1111Details on CBSE Compartment Exam.pptx1111
Details on CBSE Compartment Exam.pptx1111
 
Basic Intentional Injuries Health Education
Basic Intentional Injuries Health EducationBasic Intentional Injuries Health Education
Basic Intentional Injuries Health Education
 
Simple, Complex, and Compound Sentences Exercises.pdf
Simple, Complex, and Compound Sentences Exercises.pdfSimple, Complex, and Compound Sentences Exercises.pdf
Simple, Complex, and Compound Sentences Exercises.pdf
 
latest AZ-104 Exam Questions and Answers
latest AZ-104 Exam Questions and Answerslatest AZ-104 Exam Questions and Answers
latest AZ-104 Exam Questions and Answers
 
AIM of Education-Teachers Training-2024.ppt
AIM of Education-Teachers Training-2024.pptAIM of Education-Teachers Training-2024.ppt
AIM of Education-Teachers Training-2024.ppt
 
Towards a code of practice for AI in AT.pptx
Towards a code of practice for AI in AT.pptxTowards a code of practice for AI in AT.pptx
Towards a code of practice for AI in AT.pptx
 
Sensory_Experience_and_Emotional_Resonance_in_Gabriel_Okaras_The_Piano_and_Th...
Sensory_Experience_and_Emotional_Resonance_in_Gabriel_Okaras_The_Piano_and_Th...Sensory_Experience_and_Emotional_Resonance_in_Gabriel_Okaras_The_Piano_and_Th...
Sensory_Experience_and_Emotional_Resonance_in_Gabriel_Okaras_The_Piano_and_Th...
 
Basic Civil Engineering first year Notes- Chapter 4 Building.pptx
Basic Civil Engineering first year Notes- Chapter 4 Building.pptxBasic Civil Engineering first year Notes- Chapter 4 Building.pptx
Basic Civil Engineering first year Notes- Chapter 4 Building.pptx
 
HMCS Max Bernays Pre-Deployment Brief (May 2024).pptx
HMCS Max Bernays Pre-Deployment Brief (May 2024).pptxHMCS Max Bernays Pre-Deployment Brief (May 2024).pptx
HMCS Max Bernays Pre-Deployment Brief (May 2024).pptx
 
How to Manage Global Discount in Odoo 17 POS
How to Manage Global Discount in Odoo 17 POSHow to Manage Global Discount in Odoo 17 POS
How to Manage Global Discount in Odoo 17 POS
 
How to Add New Custom Addons Path in Odoo 17
How to Add New Custom Addons Path in Odoo 17How to Add New Custom Addons Path in Odoo 17
How to Add New Custom Addons Path in Odoo 17
 
Call Girls in Uttam Nagar (delhi) call me [🔝9953056974🔝] escort service 24X7
Call Girls in  Uttam Nagar (delhi) call me [🔝9953056974🔝] escort service 24X7Call Girls in  Uttam Nagar (delhi) call me [🔝9953056974🔝] escort service 24X7
Call Girls in Uttam Nagar (delhi) call me [🔝9953056974🔝] escort service 24X7
 
On_Translating_a_Tamil_Poem_by_A_K_Ramanujan.pptx
On_Translating_a_Tamil_Poem_by_A_K_Ramanujan.pptxOn_Translating_a_Tamil_Poem_by_A_K_Ramanujan.pptx
On_Translating_a_Tamil_Poem_by_A_K_Ramanujan.pptx
 
How to setup Pycharm environment for Odoo 17.pptx
How to setup Pycharm environment for Odoo 17.pptxHow to setup Pycharm environment for Odoo 17.pptx
How to setup Pycharm environment for Odoo 17.pptx
 
REMIFENTANIL: An Ultra short acting opioid.pptx
REMIFENTANIL: An Ultra short acting opioid.pptxREMIFENTANIL: An Ultra short acting opioid.pptx
REMIFENTANIL: An Ultra short acting opioid.pptx
 
TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...
TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...
TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...
 
SOC 101 Demonstration of Learning Presentation
SOC 101 Demonstration of Learning PresentationSOC 101 Demonstration of Learning Presentation
SOC 101 Demonstration of Learning Presentation
 

Lecture 5 - Big Data and Hadoop Intro.ppt

  • 1. Introduction to Big data A deep introduction about big data topic along with real advices of how to start a career in this hot topic. Be ready to digest a concentrated big data tablet that will put you on the right way.
  • 2.
  • 3. Agenda: Data nowadays: - Data types - Fun facts about data nowadays. - From where we generate data. - Lake of data effect on business decisions. - Future of data size. Unlocking Big data solutions: - Hadoop. - Hadoop ecosystem Zoo - Big data landscape - Top Big data companies - How to start a career in Big data - Questions Big Data: - What’s big data? - How big is the big data? - The famous Vs about big data. - Challenges of dealing with such data amount. - Why to consider a career in big data?
  • 5. Data Types: information with a degree of organization that is readily searchable and quickly consolidate into facts. Examples: RDMBS, spreadsheet information with a lack of structure that is time and energy consuming to search and find and consolidate into facts Exemples: email, documents, images, reports Semi Structured data : XML data
  • 6. Challenges for Unstructured data: How do you store Billions of Files? How long does it take to migrate 100’s of TB’s or data every 3-5 years Data has no structure Resources Limitation Data Redundancy Data Backup
  • 7. Sources of data generation: Social Media Sensors Cell Phones GPS Purchase WWW E-mails Media streaming Healthcare IOT
  • 9. Facts about data: 70% of data is created by Individuals – but enterprises are responsible for storing and managing 80% of it. 52% of travelers use social media to plan for their vacations. 35% of purchases on Amazon are though recommendations 75% of what people watch on Netflix are recommendations.
  • 11. Lake of data and business decisions : 2
  • 12. Cost: Even if RDBMS is used to handle and store “big data,” it will turn out to be very expensive. Velocity: Also, “big data” is generated at a very high velocity. RDBMS lacks in high velocity because it’s designed for steady data retention rather than rapid growth. Can traditional DBMS solve this ? Data types: Second, the majority of the data comes in a semi-structured or unstructured format from social media, audio, video, texts, and emails. However, the second problem related to unstructured data is outside the purview of RDBMS because relational databases just can’t categorize unstructured data. They’re designed and structured to accommodate structured data such as weblog sensor and financial data. Size: First, the data size has increased tremendously to the range of petabytes—one petabyte = 1,024 terabytes. RDBMS finds it challenging to handle such huge data volumes. To address this, RDBMS added more central processing units (or CPUs) or more memory to the database management system to scale up vertically.
  • 13.
  • 14. What is Big data: Big data is a term that describes the large volume of data – both structured and unstructured – that generates on a day-to-day basis. But it’s not the amount of data that’s important. It’s what organizations do with the data that matters. Big data can be analyzed for insights that lead to better decisions and strategic business moves. Big data is high-volume, high-velocity and/or high-variety information assets that demand cost-effective, innovative forms of information processing that enable enhanced insight, decision making, and process automation Big data is a term for data sets that are so large or complex that traditional data processing applications are inadequate to deal with them. Challenges include analysis, capture, data curation, search, sharing, storage, transfer, visualization, querying, updating and information privacy.
  • 15. The Vs: Velocity Volume Variety Veracity Variability Visualization Value
  • 16. Big data in action: UPS stores a large amount of data – much of which comes from sensors in its vehicles - GPS ORION (On-Road Integration Optimization and Navigation) Data Analytics and Data Science the world's largest operations research project savings of more than 8.4 million 85 million miles off of daily routes Saved $30 million/Day
  • 17. Big data in action: Walmart collects 2.5 petabytes of information from 1 million customers. from 6000 store Big data System (Kosmix) Pricing strategies Advertising campaigns 30% on their Online sales Revenue got increased by 40% “We want to know what every product in the world is. We want to know who every person in the world is. And we want to have the ability to connect them together in a transaction.” -Neil Ashe, CEO of Global E-commerce at Walmart
  • 18. Big data in quotes: “Without big data analytics, companies are blind and deaf, wandering out onto the web like deer on a freeway.” – Geoffrey Moore “Data is the new science. Big Data holds the answers.” – Pat Gelsinger “With too little data, you won’t be able to make any conclusions that you trust. With loads of data you will find relationships that aren’t real… Big data isn’t about bits, it’s about talent.” – Douglas Merrill “The world is one big data problem.” – Andrew McAfee
  • 19. Big data market forecast: The “big data” market is expected to cross $50 billion by 2017.
  • 20. Big data jobs trend: The median advertised salary for professionals with big data expertise is $124,000 a year. IBM , Cisco and Oracle together advertised 26,488 open positions that required big data expertise in the last twelve months. 124000 usd to sar = 465012 /12 = 38751 SAR/Month
  • 21. How to solve big data Hadoop: is a big data analysis engine
  • 22. What is Hadoop The Apache Hadoop software library is a framework that allows for the distributed processing of large data sets across clusters of computers using simple programming models. Hadoop is an open-source software framework for storing data and running applications on clusters of commodity hardware. It provides massive storage for any kind of data, enormous processing power and the ability to handle virtually limitless concurrent tasks or jobs.
  • 23. Hadoop history Nutch is a well matured, production ready Web crawler. that enables fine grained configuration, relying on Apache Hadoop™ data structures, which are great for batch processing.
  • 24. Why Hadoop is important ? Ability to store and process huge amounts of any kind of data, quickly. With data volumes and varieties constantly increasing, especially from social media and the Internet of Things (IoT), that's a key consideration. Computing power. Hadoop's distributed computing model processes big data fast. The more computing nodes you use, the more processing power you have. Fault tolerance. Data and application processing are protected against hardware failure. If a node goes down, jobs are automatically redirected to other nodes to make sure the distributed computing does not fail. Multiple copies of all data are stored automatically.
  • 25. Why Hadoop is important ? Flexibility. Unlike traditional relational databases, you don’t have to preprocess data before storing it. You can store as much data as you want and decide how to use it later. That includes unstructured data like text, images and videos. Low cost. The open-source framework is free and uses commodity hardware to store large quantities of data. Scalability. You can easily grow your system to handle more data simply by adding nodes. Little administration is required. Scalability Horizontal scaling means that you scale by adding more machines into your pool of resources Vertical scaling means that you scale by adding more power (CPU, RAM) to an existing machine
  • 26. How is Hadoop being used? Going beyond its original goal of searching millions (or billions) of web pages and returning relevant results, many organizations are looking to Hadoop as their next big data platform. Popular uses today include:
  • 27. How is Hadoop being used?
  • 28. How is Hadoop being used?
  • 29. Hadoop ecosystem Data Ingestion Sqoop Flume Storm NiFi Kafka Querying Layer (Hive,Impala) Processing Layer (MapReduce,Spark,Pig) Storage Layer (HDFS,Hbase,Hcatalog) Cluster monitoring, provisioning and management
  • 30. Hadoop | Data Ingestion Apache Sqoop is a tool designed for efficiently transferring bulk data between Apache Hadoop and structured data stores such as relational databases.
  • 31. Hadoop | Data Ingestion Flume is a distributed, reliable, and available service for efficiently collecting, aggregating, and moving large amounts of log data. It is robust and fault tolerant with tunable reliability mechanisms and many failover and recovery mechanisms.
  • 32. Hadoop | Data Ingestion Storm is real-time computation system. Storm makes it easy to reliably process unbounded streams of data, doing for real-time processing. A Storm topology consumes streams of data and processes those streams in arbitrarily complex ways, repartitioning the streams between each stage of the computation however needed.
  • 33. Hadoop | Data Ingestion An easy to use, powerful, and reliable system to process and distribute data. Apache NiFi supports powerful and scalable directed graphs of data routing, transformation, and system mediation logic in a Web-based user interface
  • 34. Hadoop | Data Ingestion Kafka™ is used for building real-time data pipelines and streaming apps. It is horizontally scalable, fault-tolerant, wicked fast, and runs in production in thousands of companies.
  • 35. Hadoop | Data Ingestion Large scale log aggregator, and analytics. Fluentd is an open source data collector for unified logging layer. Fluentd allows you to unify data collection and consumption for a better use and understanding of data. Apache Samza is a distributed stream processing framework. It uses Apache Kafka for messaging, and Apache Hadoop YARN to provide fault tolerance, processor isolation, security, and resource management.
  • 36. Hadoop | Data Storage Layer Hadoop Distributed File System (HDFS) offers a way to store large files across multiple machines. Hadoop and HDFS was derived from Google File System (GFS) paper.
  • 37. Hadoop | Data Storage Layer
  • 38. Hadoop | Data Storage Layer A distributed, column-oriented database. HBase uses HDFS for its underlying storage, and supports both batch-style computations using MapReduce and point queries (random reads)
  • 39. Hadoop | Data Storage Layer A metadata and table management system for Hadoop. It shares the metadata with other tools like map reduce, Pig and Hive. It provides one constant data model for all Hadoop tools along with a shared schema.
  • 40. Hadoop | Data Processing Layer MapReduce is the heart of Hadoop. It is this programming paradigm that allows for massive scalability across hundreds or thousands of servers in a Hadoop cluster with a parallel, distributed algorithm.
  • 41. Hadoop | Data Processing Layer
  • 42. Hadoop | Data Processing Layer A scripting SQL based language and execution environment for creating complex MapReduce transformations. Functions are written in Pig Latin (the language) and translated into executable MapReduce jobs. Pig also allows the user to create extended functions (UDFs) using Java.
  • 43. Hadoop | Data Processing Layer In memory data analytics cluster computing framework originally developed in the AMPLab at UC Berkeley. Run programs up to 100x faster than Hadoop MapReduce in memory, or 10x faster on disk.
  • 44. Hadoop | Data Querying Layer A distributed data warehouse built on top of HDFS to manage and organize large amounts of data. Hive provides a query language based on SQL semantic (HiveQL) which is translated by the runtime engine to MapReduce jobs for querying the data.
  • 45. Hadoop | Data Querying Layer pen source massively parallel processing (MPP) SQL query engine for data stored in a computer cluster running Apache Hadoop.
  • 46. Hadoop | Management Layer intuitive, easy-to-use Hadoop management web UI. Apache Ambari was donated by Hortonworks team. It's a powerful and nice interface for Hadoop and other typical applications from the Hadoop ecosystem.
  • 47. Big data existing solutions:
  • 48. Big data existing solutions: Data Sources YARN: A framework for job scheduling and cluster resource management. A platform for manipulating data stored in HDFS via a high-level language called Pig Latin. It does data extractions, transformations and loading, and basic analysis in patch mode A data warehousing and SQL-like query language that presents data in the form of tables. Hive programming is similar to database programming. Ambari: A web interface for managing, configuring and testing Hadoop services and components. An open-source cluster computing framework with in-memory analytics. HDFS :A platform for manipulating data stored in HDFS via a high-level language called Pig Latin. It does data extractions, transformations and loading, and basic analysis in patch mode HBase:A distributed, column-oriented database. HBase uses HDFS for its underlying storage, and supports both batch-style computations using MapReduce and point queries. HCatalog A table and storage management layer for Hadoop that enables Hadoop applications (Pig, MapReduce, and Hive) to read and write data to a tabular form as opposed to files. A distributed data processing model and execution environment that runs on large clusters of commodity machines. MapReduce A Scalable machine learning and data mining library A high-performance coordination service for distributed applications. is a Java Web application used to schedule Apache Hadoop jobs data collection system for monitoring large distributed systems. A web-based tool for provisioning, managing, and monitoring Apache Hadoop clusters. Ambari also provides a dashboard for viewing cluster health such as heatmaps and ability to view MapReduce, Pig and Hive applications visually along with features to diagnose their performance. is a data serialization system
  • 49. Other apache projects: Apache Flink is an open source platform for distributed stream and batch data processing. Apache Falcon Feed management and data processing platform Apache Ranger Ranger is a framework to enable, monitor and manage comprehensive data security across the Hadoop platform. Apache Tez to develop a generic application which can be used to process complex data-processing task Apache Tika toolkit detects and extracts metadata and text from over a thousand different file types Apache Parquet columnar storage format available to any project in the Hadoop ecosystem Apache Zeppelin A web-based notebook that enables interactive data analytics. Apache Drill Schema-free SQL Query Engine for Hadoop, NoSQL and Cloud Storage
  • 50. Top Leading Big data companies The Apache Software Foundation (ASF) is an American non- profit corporation to support Apache projects
  • 51. How to start 1. Identify business use cases tied to business outcomes, metrics and your big data roadmap 2. Identify big data champions from both the business and IT sides of your organization 3. Select infrastructure, tools and architecture for your big data POC/implementation 4. Staff the project with the right big data skills or a strategic big data implementation partner 5. Run your project/POC in sprints or short projects with tangible and measurable outcomes 6. Try to scale your success POC up to test your Logic implementation against the big dataset.
  • 52. What can I do now ?
  • 53. Certification Path | Administration Cloudera HortonWorks
  • 54. Certification Path | Development Cloudera HortonWorks
  • 55. Certification Path | Data Science Cloudera HortonWorks

Editor's Notes

  1. https://datafloq.com/read/3vs-sufficient-describe-big-data/166 Velocity The Velocity is the speed at which the data is created, stored, analyzed and visualized. In the past, when batch processing was common practice, it was normal to receive an update from the database every night or even every week. Computers and servers required substantial time to process the data and update the databases. In the big data era, data is created in real-time or near real-time. With the availability of Internet connected devices, wireless or wired, machines and devices can pass-on their data the moment it is created. The speed at which data is created currently is almost unimaginable: Every minute we upload 100 hours of video on YouTube. In addition, every minute over 200 million emails are sent, around 20 million photos are viewed and 30.000 uploaded on Flickr, almost 300.000 tweets are sent and almost 2,5 million queries on Google are performed. The challenge organizations have is to cope with the enormous speed the data is created and used in real-time. Volume 90% of all data ever created, was created in the past 2 years. From now on, the amount of data in the world will double every two years. By 2020, we will have 50 times the amount of data as that we had in 2011. The sheer volume of the data is enormous and a very large contributor to the ever expanding digital universe is the Internet of Things with sensors all over the world in all devices creating data every second. The era of a trillion sensors is upon us. If we look at airplanes they generate approximately 2,5 billion Terabyte of data each year from the sensors installed in the engines. Self-driving cars will generate 2 Petabyte of data every year. Also the agricultural industry generates massive amounts of data with sensors installed in tractors. Shell uses super-sensitive sensors to find additional oil in wells and if they install these sensors at all 10.000 wells they will collect approximately 10 Exabyte of data annually. That again is absolutely nothing if we compare it to the Square Kilometer Array Telescope that will generate 1 Exabyte of data per day. In the past, the creation of so much data would have caused serious problems. Nowadays, with decreasing storage costs, better storage solutions like Hadoop and the algorithms to create meaning from all that data this is not a problem at all. Variety In the past, all data that was created was structured data, it neatly fitted in columns and rows but those days are over. Nowadays, 90% of the data that is generated by organization is unstructured data. Data today comes in many different formats: structured data, semi-structured data, unstructured data and even complex structured data. The wide variety of data requires a different approach as well as different techniques to store all raw data. There are many different types of data and each of those types of data require different types of analyses or different tools to use. Social media like Facebook posts or Tweets can give different insights, such as sentiment analysis on your brand, while sensory data will give you information about how a product is used and what the mistakes are. Are you looking for Big Data Jobs or Candidates? Please go to our WORK section The Four Additional V’s Now that the context is set regarding the traditional V’s, let’s see which other V’s are important for organizations to keep in mind when they develop a big data strategy. Veracity Having a lot of data in different volumes coming in at high speed is worthless if that data is incorrect. Incorrect data can cause a lot of problems for organizations as well as for consumers. Therefore, organizations need to ensure that the data is correct as well as the analyses performed on the data are correct. Especially in automated decision-making, where no human is involved anymore, you need to be sure that both the data and the analyses are correct. If you want your organization to become information-centric, you should be able to trust that data as well as the analyses. accountability. Variability Big data is extremely variable. Brian Hopkins, a Forrester principal analyst, defines variability as the “variance in meaning, in lexicon”. He refers to the supercomputer Watson who won Jeopardy. The supercomputer had to “dissect an answer into its meaning and […] to figure out what the right question was”. That is extremely difficult because words have different meanings an all depends on the context. For the right answer, Watson had to understand the context. Variability is often confused with variety. Say you have bakery that sells 10 different breads. That is variety. Now imagine you go to that bakery three days in a row and every day you buy the same type of bread but each day it tastes and smells different. That is variability. Variability is thus very relevant in performing sentiment analyses. Variability means that the meaning is changing(rapidly). In (almost) the same tweets a word can have a totally different meaning. In order to perform a proper sentiment analyses, algorithms need to be able to understand the context and be able to decipher the exact meaning of a word in that context. This is still very difficult. Visualization This is the hard part of big data. Making all that vast amount of data comprehensible in a manner that is easy to understand and read. With the right analyses and visualizations, raw data can be put to use otherwise raw data remains essentially useless. Visualizations of course do not mean ordinary graphs or pie charts. They mean complex graphs that can include many variables of data while still remaining understandable and readable. Visualizing might not be the most technological difficult part; it sure is the most challenging part. Telling a complex story in a graph is very difficult but also extremely crucial. Luckily there are more and more big data startups appearing that focus on this aspect and in the end, visualizations will make the difference. One of them is future this will be the direction to go, where visualizations help organizations answer questions they did not know to ask. Value All that available data will create a lot of value for organizations, societies and consumers. Big data means big business and every industry will reap the benefits from big data. McKinsey states that potential annual value of big data to the US Health Care is $ 300 billion, more than double the total annual health care spending of Spain. They also mention that big data has a potential annual value of € 250 billion to the Europe’s public sector administration. Even more, in their well-regarded report from 2011, they state that the potential annual consumer surplus from using personal location data globally can be up to $ 600 billion in 2020. That is a lot of value. Of course, data in itself is not valuable at all. The value is in the analyses done on that data and how the data is turned into information and eventually turning it into knowledge. The value is in how organisations will use that data and turn their organisation into an information-centric company that relies  on insights derived from data analyses for their decision-making.
  2. http://bridg.com/blog/walmart-big-data/
  3. One such project was an open-source web search engine called Nutch – the brainchild of Doug Cutting and Mike Cafarella. They wanted to return web search results faster by distributing data and calculations across different computers so multiple tasks could be accomplished simultaneously. During this time, another search engine project called Google was in progress. It was based on the same concept – storing and processing data in a distributed, automated way so that relevant web search results could be returned faster. In 2006, Cutting joined Yahoo and took with him the Nutch project as well as ideas based on Google’s early work with automating distributed data storage and processing. The Nutch project was divided – the web crawler portion remained as Nutch and the distributed computing and processing portion became Hadoop (named after Cutting’s son’s toy elephant). In 2008, Yahoo released Hadoop as an open-source project. Today, Hadoop’s framework and ecosystem of technologies are managed and maintained by the non-profit Apache Software Foundation (ASF), a global community of software developers and contributors.  
  4. http://www.predictiveanalyticstoday.com/data-ingestion-tools/
  5. http://static.googleusercontent.com/media/research.google.com/en//archive/gfs-sosp2003.pdf
  6. http://static.googleusercontent.com/media/research.google.com/en//archive/gfs-sosp2003.pdf
  7. http://thebigdatablog.weebly.com/blog/the-hadoop-ecosystem-overview
  8. http://thebigdatablog.weebly.com/blog/the-hadoop-ecosystem-overview
  9. http://thebigdatablog.weebly.com/blog/the-hadoop-ecosystem-overview