Apache Hadoop is an open-source software framework for distributed storage and processing of large datasets across clusters of computers. It allows for processing of structured, semi-structured, and unstructured data using simple programming models. Hadoop can handle failures at the application layer and provides highly available services across large clusters of computers that may be prone to failures. Case studies show Hadoop has been used for fraud detection in banking, analyzing customer spending, predicting weather patterns and climate changes, processing astronomical images, and more wherever high volumes of unstructured data are growing rapidly.
We are living in the world of “Big Data”. “Big Data” is mainly expressed with three Vs – Volume, Velocity and Variety. The presentation will discuss how Big Data impacts us and how SAS programmers can use SAS skills in Big Data environment
The presentation will introduce Big Data Storage solution – Hadoop and NoSQL. In Hadoop, the presentation will discuss two major Hadoop capabilities - Hadoop Distributed File System (HDFS) and Map/Reduce (parallel computing in Hadoop). The presentation will show how SAS can work with Hadoop using HDFS LIBNAME, FILENAME, SAS/ACCESS to Hadoop HIVE and SAS GRID Managers to Hadoop YARN. The presentation will also introduce the concepts of NoSQL database for a big data solution.
The presentation will also introduce how SAS can work with the variety of data format, especially XML and JSON. The presentation will show the use case of converting XML documents to SAS datasets using LIBNAME XMLV2 XMLMAP statement. The presentation will also introduce REST API to extract data through internets and will demonstrate how SAS PROC HTTP can move the data through REST API.
This is a power point presentation on Hadoop and Big Data. This covers the essential knowledge one should have when stepping into the world of Big Data.
This course is available on hadoop-skills.com for free!
This course builds a basic fundamental understanding of Big Data problems and Hadoop as a solution. This course takes you through:
• This course builds Understanding of Big Data problems with easy to understand examples and illustrations.
• History and advent of Hadoop right from when Hadoop wasn’t even named Hadoop and was called Nutch
• What is Hadoop Magic which makes it so unique and powerful.
• Understanding the difference between Data science and data engineering, which is one of the big confusions in selecting a carrier or understanding a job role.
• And most importantly, demystifying Hadoop vendors like Cloudera, MapR and Hortonworks by understanding about them.
This course is available for free on hadoop-skills.com
Technological geeks Hindi Video 1 -
https://youtu.be/LSvAoo4pYjs
Contents :-
What is Big Data ?
Big Data characteristics
Big Data sources
Use cases of Big Data
Hadoop Daemons
Hadoop Master slave architecture
Hadoop cluster
Secondary namenode
We are living in the world of “Big Data”. “Big Data” is mainly expressed with three Vs – Volume, Velocity and Variety. The presentation will discuss how Big Data impacts us and how SAS programmers can use SAS skills in Big Data environment
The presentation will introduce Big Data Storage solution – Hadoop and NoSQL. In Hadoop, the presentation will discuss two major Hadoop capabilities - Hadoop Distributed File System (HDFS) and Map/Reduce (parallel computing in Hadoop). The presentation will show how SAS can work with Hadoop using HDFS LIBNAME, FILENAME, SAS/ACCESS to Hadoop HIVE and SAS GRID Managers to Hadoop YARN. The presentation will also introduce the concepts of NoSQL database for a big data solution.
The presentation will also introduce how SAS can work with the variety of data format, especially XML and JSON. The presentation will show the use case of converting XML documents to SAS datasets using LIBNAME XMLV2 XMLMAP statement. The presentation will also introduce REST API to extract data through internets and will demonstrate how SAS PROC HTTP can move the data through REST API.
This is a power point presentation on Hadoop and Big Data. This covers the essential knowledge one should have when stepping into the world of Big Data.
This course is available on hadoop-skills.com for free!
This course builds a basic fundamental understanding of Big Data problems and Hadoop as a solution. This course takes you through:
• This course builds Understanding of Big Data problems with easy to understand examples and illustrations.
• History and advent of Hadoop right from when Hadoop wasn’t even named Hadoop and was called Nutch
• What is Hadoop Magic which makes it so unique and powerful.
• Understanding the difference between Data science and data engineering, which is one of the big confusions in selecting a carrier or understanding a job role.
• And most importantly, demystifying Hadoop vendors like Cloudera, MapR and Hortonworks by understanding about them.
This course is available for free on hadoop-skills.com
Technological geeks Hindi Video 1 -
https://youtu.be/LSvAoo4pYjs
Contents :-
What is Big Data ?
Big Data characteristics
Big Data sources
Use cases of Big Data
Hadoop Daemons
Hadoop Master slave architecture
Hadoop cluster
Secondary namenode
Significance Of Hadoop For Data ScienceRobert Smith
Hadoop is an important tool for data science when the volume of data exceeds the system memory or when the business case requires data to be distributed across multiple servers.
10 Popular Hadoop Technical Interview QuestionsZaranTech LLC
Big Data has been attested as one of the fastest growing technologies of this decade and thus potent enough to produce a large number of jobs. While enterprises across industrial stretch have started building teams, Hadoop technical interview questions could vary from simple definitions to critical case studies. Let’s take quick glimpse at the most obvious ones.
Hadoop is a cluster computing framework.
Hadoop tools empower more developers and more organizations to leverage Hadoop for big data management. There’s been a growing demand for Hadoop tools that can make Hadoop's vast processing power more accessible. I’m going to present a Brief explanation of the various applications and tools that are associated with Hadoop. Also, I would be presenting a project how on how some of these tools where used to analyze the percentage of brain injured person in New England in the month of December 2010 survey to determine if brain transplant was an option to solve brain problem in the Nation.
Content presented at a talk on Aug. 29th. Purpose is to inform a fairly technical audience on the primary tenets of Big Data and the hadoop stack. Also, did a walk-thru' of hadoop and some of the hadoop stack i.e. Pig, Hive, Hbase.
Big Data Hadoop Training in Pune-Course Content Advanto SoftwareAdvanto Software
Learn Big Data Hadoop Training from Real time Industry Experts with more than 10 years of Industry experience. Advanto Software is the best big data Hadoop training institute in pune
The presentation covers following topics: 1) Hadoop Introduction 2) Hadoop nodes and daemons 3) Architecture 4) Hadoop best features 5) Hadoop characteristics. For more further knowledge of Hadoop refer the link: http://data-flair.training/blogs/hadoop-tutorial-for-beginners/
Introduction to Hadoop.
What are Hadoop, MapReeduce, and Hadoop Distributed File System.
Who uses Hadoop?
How to run Hadoop?
What are Pig, Hive, Mahout?
Improving performance of apriori algorithm using hadoopeSAT Journals
Abstract Spatial data is a data having a geological information. This paper explores the use of Hadoop framework to improve the performance of Apriori algorithm for spatial data mining. FP growth algorithm is better than Apriori but it fails in certain situations. By applying the Apriori algorithm parallely using Hadoop framework to spatial data, we can perform well as compare to FP growth. This paper includes clustering based on geological location, classification based on mineral resource type and spatial coherence between mineral resources. Spatial data mining find out the different association rules by observing the spatial data by using Apriori algorithm. The result of the paper will indicate the accurate prediction of occurrence of commodity with respect to other commodity of mineral resources. Keywords: Hadoop, data mining, association rules, clustering, spatial coherence
.NET Usergroup Oldenburg 28. Mai 2015 - von Dr. Yvette Teiken
Big Data ist in aller Munde. Auch Microsoft ist mit HDInsight auf den Zug aufgesprungen. Aber wie passt das zusammen, Open Source, Hadoop und Microsoft? Wo sind die Anknüpfungspunkte zu klassischem BI? Wie werden Daten gespeichert und analysiert? Was ändert sich mit Big Data und was nicht? Unter anderem soll es gehen um.
Erstellung, Anfragen und Export von Hive Tabellen
Umsetzung von ETL-Prozessen mit Hilfe von PIG
Entwicklung nativer Map Reduce-Jobs mit C#
Interaktion mit traditionellen RDBMS und Streaming-Technologien
Datenspeicherung mit DocumentDB
Skalierung von Analysen
Introduction to SARA's Hadoop Hackathon - dec 7th 2010Evert Lammerts
This was the first of two introduction presentations to the first Hadoop Hackathon at SARA, the Dutch center for High Performance Computing and Networking.
This video is a recording of a tech talk where we explain the basics of Big Data. It has certainly been the buzzword in the IT industry and this is an effort towards a level 100 talk where people would learn about the history, basics, current market needs and in and out of Big Data.
Significance Of Hadoop For Data ScienceRobert Smith
Hadoop is an important tool for data science when the volume of data exceeds the system memory or when the business case requires data to be distributed across multiple servers.
10 Popular Hadoop Technical Interview QuestionsZaranTech LLC
Big Data has been attested as one of the fastest growing technologies of this decade and thus potent enough to produce a large number of jobs. While enterprises across industrial stretch have started building teams, Hadoop technical interview questions could vary from simple definitions to critical case studies. Let’s take quick glimpse at the most obvious ones.
Hadoop is a cluster computing framework.
Hadoop tools empower more developers and more organizations to leverage Hadoop for big data management. There’s been a growing demand for Hadoop tools that can make Hadoop's vast processing power more accessible. I’m going to present a Brief explanation of the various applications and tools that are associated with Hadoop. Also, I would be presenting a project how on how some of these tools where used to analyze the percentage of brain injured person in New England in the month of December 2010 survey to determine if brain transplant was an option to solve brain problem in the Nation.
Content presented at a talk on Aug. 29th. Purpose is to inform a fairly technical audience on the primary tenets of Big Data and the hadoop stack. Also, did a walk-thru' of hadoop and some of the hadoop stack i.e. Pig, Hive, Hbase.
Big Data Hadoop Training in Pune-Course Content Advanto SoftwareAdvanto Software
Learn Big Data Hadoop Training from Real time Industry Experts with more than 10 years of Industry experience. Advanto Software is the best big data Hadoop training institute in pune
The presentation covers following topics: 1) Hadoop Introduction 2) Hadoop nodes and daemons 3) Architecture 4) Hadoop best features 5) Hadoop characteristics. For more further knowledge of Hadoop refer the link: http://data-flair.training/blogs/hadoop-tutorial-for-beginners/
Introduction to Hadoop.
What are Hadoop, MapReeduce, and Hadoop Distributed File System.
Who uses Hadoop?
How to run Hadoop?
What are Pig, Hive, Mahout?
Improving performance of apriori algorithm using hadoopeSAT Journals
Abstract Spatial data is a data having a geological information. This paper explores the use of Hadoop framework to improve the performance of Apriori algorithm for spatial data mining. FP growth algorithm is better than Apriori but it fails in certain situations. By applying the Apriori algorithm parallely using Hadoop framework to spatial data, we can perform well as compare to FP growth. This paper includes clustering based on geological location, classification based on mineral resource type and spatial coherence between mineral resources. Spatial data mining find out the different association rules by observing the spatial data by using Apriori algorithm. The result of the paper will indicate the accurate prediction of occurrence of commodity with respect to other commodity of mineral resources. Keywords: Hadoop, data mining, association rules, clustering, spatial coherence
.NET Usergroup Oldenburg 28. Mai 2015 - von Dr. Yvette Teiken
Big Data ist in aller Munde. Auch Microsoft ist mit HDInsight auf den Zug aufgesprungen. Aber wie passt das zusammen, Open Source, Hadoop und Microsoft? Wo sind die Anknüpfungspunkte zu klassischem BI? Wie werden Daten gespeichert und analysiert? Was ändert sich mit Big Data und was nicht? Unter anderem soll es gehen um.
Erstellung, Anfragen und Export von Hive Tabellen
Umsetzung von ETL-Prozessen mit Hilfe von PIG
Entwicklung nativer Map Reduce-Jobs mit C#
Interaktion mit traditionellen RDBMS und Streaming-Technologien
Datenspeicherung mit DocumentDB
Skalierung von Analysen
Introduction to SARA's Hadoop Hackathon - dec 7th 2010Evert Lammerts
This was the first of two introduction presentations to the first Hadoop Hackathon at SARA, the Dutch center for High Performance Computing and Networking.
This video is a recording of a tech talk where we explain the basics of Big Data. It has certainly been the buzzword in the IT industry and this is an effort towards a level 100 talk where people would learn about the history, basics, current market needs and in and out of Big Data.
Tools and Methods for Big Data Analytics by Dahl WintersMelinda Thielbar
Research Triangle Analysts October presentation on Big Data by Dahl Winters (formerly of Research Triangle Institute). Dahl takes her viewers on a whirlwind tour of big data tools such as Hadoop and big data algorithms such as MapReduce, clustering, and deep learning. These slides document the many resources available on the internet, as well as guidelines of when and where to use each.
Tools and Methods for Big Data Analytics by Dahl WintersMelinda Thielbar
Research Triangle Analysts October presentation on Big Data by Dahl Winters (formerly of Research Triangle Institute). Dahl takes her viewers on a whirlwind tour of big data tools such as Hadoop and big data algorithms such as MapReduce, clustering, and deep learning. These slides document the many resources available on the internet, as well as guidelines of when and where to use each.
Predictive Analytics: Context and Use Cases
Historical context for successful implementation of predictive analytic techniques and examples of implementation of successful use cases.
What is popular in the manufacturing industry today? I think it’s going to be digital conversion, Industry 4.0, artificial intelligence...
Let’s take a look at how AI is changing manufacturing.
seminar on Big Data Technology
report on big data technology
webinar on big data technology
topic on big data technology
ppt presentation on big data technology
March Towards Big Data - Big Data Implementation, Migration, Ingestion, Manag...Experfy
Gartner, IBM, Accenture and many others have asserted that 80% or more of the world’s information is unstructured – and inherently hard to analyze. What does that mean? And what is required to extract insight from unstructured data?
Unstructured data is infinitely variable in quality and format, because it is produced by humans who can be fastidious, unpredictable, ill-informed, or even cynical, but always unique, not standard in any way. Recent advances in natural language processing provides the notion that unstructured content can be included in data analysis.
Serious growth and value companies are committed to data. The exponential growth of Big Data has posed major challenges in data governance and data analysis. Good data governance is pivotal for business growth.
Therefore, it is of paramount importance to slice and dice Big Data that addresses data governance and data analysis issues. In order to support high quality business decision making, it is important to fully harness the potential of Big Data by implementing proper Data Migration, Data Ingestion, Data Management, Data Analysis, Data Visualization and Data Virtualization tools.
Check it out: https://www.experfy.com/training/courses/march-towards-big-data-big-data-implementation-migration-ingestion-management-visualization
To view recording of this webinar please use below URL:
http://wso2.com/library/webinars/2016/06/analytics-in-your-enterprise/
Big data spans many fields and brings together technologies like distributed systems, machine learning, statistics and Internet of Things (IoT). It has now become a multi-billion dollar industry with use cases ranging from targeted advertising and fraud detection to product recommendations and market surveys.
Some use cases such as urban planning can be slower (done in batch mode), while others such as the stock market needs results in milliseconds (done is a streaming fashion). Different technologies are used for each case; MapReduce for batch analytics, complex event processing for real-time analytics and machine learning for predictive analytics. Furthermore, the type of analysis ranges from basic statistics to complicated prediction models.
This webinar will discuss the big data landscape including
Concepts, use cases and technologies
Capabilities and applications of the WSO2 analytics platform
WSO2 Data Analytics Server
WSO2 Complex Event Processor
WSO2 Machine Learner
Memory Management in BigData: A Perpective Viewijtsrd
The requirement to perform complicated statistic analysis of big data by institutions of engineering, scientific research, health care, commerce, banking and computer research is immense. However, the limitations of the widely used current desktop software like R, excel, minitab and spss gives a researcher limitation to deal with big data and big data analytic tools like IBM BigInsight, HP Vertica, SAP HANA & Pentaho come at an overpriced license. Apache Hadoop is an open source distributed computing framework that uses commodity hardware. With this project, I intend to collaborate Apache Hadoop and R software to develop an analytic platform that stores big data (using open source Apache Hadoop) and perform statistical analysis (using open source R software).Due to the limitations of vertical scaling of computer unit, data storage is handled by several machines and so analysis becomes distributed over all these machines. Apache Hadoop is what comes handy in this environment. To store massive quantities of data as required by researchers, we could use commodity hardware and perform analysis in distributed environment. Bhavna Bharti | Prof. Avinash Sharma"Memory Management in BigData: A Perpective View" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-2 | Issue-4 , June 2018, URL: http://www.ijtsrd.com/papers/ijtsrd14436.pdf http://www.ijtsrd.com/engineering/computer-engineering/14436/memory-management-in-bigdata-a-perpective-view/bhavna-bharti
Role of Big Data in Start-ups - PersianMobin Ranjbar
This presentation is about role of big data in businesses and start-ups and how companies can take advantage of it.It's in Persian.It's provided for Startup Weekend Sari event.
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024Albert Hoitingh
In this session I delve into the encryption technology used in Microsoft 365 and Microsoft Purview. Including the concepts of Customer Key and Double Key Encryption.
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...DanBrown980551
Do you want to learn how to model and simulate an electrical network from scratch in under an hour?
Then welcome to this PowSyBl workshop, hosted by Rte, the French Transmission System Operator (TSO)!
During the webinar, you will discover the PowSyBl ecosystem as well as handle and study an electrical network through an interactive Python notebook.
PowSyBl is an open source project hosted by LF Energy, which offers a comprehensive set of features for electrical grid modelling and simulation. Among other advanced features, PowSyBl provides:
- A fully editable and extendable library for grid component modelling;
- Visualization tools to display your network;
- Grid simulation tools, such as power flows, security analyses (with or without remedial actions) and sensitivity analyses;
The framework is mostly written in Java, with a Python binding so that Python developers can access PowSyBl functionalities as well.
What you will learn during the webinar:
- For beginners: discover PowSyBl's functionalities through a quick general presentation and the notebook, without needing any expert coding skills;
- For advanced developers: master the skills to efficiently apply PowSyBl functionalities to your real-world scenarios.
Accelerate your Kubernetes clusters with Varnish CachingThijs Feryn
A presentation about the usage and availability of Varnish on Kubernetes. This talk explores the capabilities of Varnish caching and shows how to use the Varnish Helm chart to deploy it to Kubernetes.
This presentation was delivered at K8SUG Singapore. See https://feryn.eu/presentations/accelerate-your-kubernetes-clusters-with-varnish-caching-k8sug-singapore-28-2024 for more details.
JMeter webinar - integration with InfluxDB and GrafanaRTTS
Watch this recorded webinar about real-time monitoring of application performance. See how to integrate Apache JMeter, the open-source leader in performance testing, with InfluxDB, the open-source time-series database, and Grafana, the open-source analytics and visualization application.
In this webinar, we will review the benefits of leveraging InfluxDB and Grafana when executing load tests and demonstrate how these tools are used to visualize performance metrics.
Length: 30 minutes
Session Overview
-------------------------------------------
During this webinar, we will cover the following topics while demonstrating the integrations of JMeter, InfluxDB and Grafana:
- What out-of-the-box solutions are available for real-time monitoring JMeter tests?
- What are the benefits of integrating InfluxDB and Grafana into the load testing stack?
- Which features are provided by Grafana?
- Demonstration of InfluxDB and Grafana using a practice web application
To view the webinar recording, go to:
https://www.rttsweb.com/jmeter-integration-webinar
Securing your Kubernetes cluster_ a step-by-step guide to success !KatiaHIMEUR1
Today, after several years of existence, an extremely active community and an ultra-dynamic ecosystem, Kubernetes has established itself as the de facto standard in container orchestration. Thanks to a wide range of managed services, it has never been so easy to set up a ready-to-use Kubernetes cluster.
However, this ease of use means that the subject of security in Kubernetes is often left for later, or even neglected. This exposes companies to significant risks.
In this talk, I'll show you step-by-step how to secure your Kubernetes cluster for greater peace of mind and reliability.
Generating a custom Ruby SDK for your web service or Rails API using Smithyg2nightmarescribd
Have you ever wanted a Ruby client API to communicate with your web service? Smithy is a protocol-agnostic language for defining services and SDKs. Smithy Ruby is an implementation of Smithy that generates a Ruby SDK using a Smithy model. In this talk, we will explore Smithy and Smithy Ruby to learn how to generate custom feature-rich SDKs that can communicate with any web service, such as a Rails JSON API.
UiPath Test Automation using UiPath Test Suite series, part 3DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 3. In this session, we will cover desktop automation along with UI automation.
Topics covered:
UI automation Introduction,
UI automation Sample
Desktop automation flow
Pradeep Chinnala, Senior Consultant Automation Developer @WonderBotz and UiPath MVP
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdf91mobiles
91mobiles recently conducted a Smart TV Buyer Insights Survey in which we asked over 3,000 respondents about the TV they own, aspects they look at on a new TV, and their TV buying preferences.
Key Trends Shaping the Future of Infrastructure.pdfCheryl Hung
Keynote at DIGIT West Expo, Glasgow on 29 May 2024.
Cheryl Hung, ochery.com
Sr Director, Infrastructure Ecosystem, Arm.
The key trends across hardware, cloud and open-source; exploring how these areas are likely to mature and develop over the short and long-term, and then considering how organisations can position themselves to adapt and thrive.
Software Delivery At the Speed of AI: Inflectra Invests In AI-Powered QualityInflectra
In this insightful webinar, Inflectra explores how artificial intelligence (AI) is transforming software development and testing. Discover how AI-powered tools are revolutionizing every stage of the software development lifecycle (SDLC), from design and prototyping to testing, deployment, and monitoring.
Learn about:
• The Future of Testing: How AI is shifting testing towards verification, analysis, and higher-level skills, while reducing repetitive tasks.
• Test Automation: How AI-powered test case generation, optimization, and self-healing tests are making testing more efficient and effective.
• Visual Testing: Explore the emerging capabilities of AI in visual testing and how it's set to revolutionize UI verification.
• Inflectra's AI Solutions: See demonstrations of Inflectra's cutting-edge AI tools like the ChatGPT plugin and Azure Open AI platform, designed to streamline your testing process.
Whether you're a developer, tester, or QA professional, this webinar will give you valuable insights into how AI is shaping the future of software delivery.
Slack (or Teams) Automation for Bonterra Impact Management (fka Social Soluti...Jeffrey Haguewood
Sidekick Solutions uses Bonterra Impact Management (fka Social Solutions Apricot) and automation solutions to integrate data for business workflows.
We believe integration and automation are essential to user experience and the promise of efficient work through technology. Automation is the critical ingredient to realizing that full vision. We develop integration products and services for Bonterra Case Management software to support the deployment of automations for a variety of use cases.
This video focuses on the notifications, alerts, and approval requests using Slack for Bonterra Impact Management. The solutions covered in this webinar can also be deployed for Microsoft Teams.
Interested in deploying notification automations for Bonterra Impact Management? Contact us at sales@sidekicksolutionsllc.com to discuss next steps.
2. What is Apache Hadoop?
The Apache Hadoop software library is a framework that allows for
the distributed processing of large data sets across clusters of
.computers using simple programming models It is designed to scale
,up from single servers to thousands of machines each ofering local
.computation and storage
- ,Rather than rely on hardware to deliver high availability the library
itself is designed to detect and handle failures at the application
, -layer so delivering a highly available service on top of a cluster of
, .computers each of which may be prone to failures
Source: http://hadoop.apache.com
3. Why Apache Hadoop?
● .It is distributed
● / - /It can handle structured semi structured unstructured
.data
● .It does not need expensive hardwares
● .It is open source
● .It has a variety of tools for machine learning and etc
● .And etc
5. Economy and Financial Systems
● .Fraud Detection in bank accounts
● Analysing customers data to know how are they spending
.money
● .Who is more eligible to get a loan based on their history
● Modern currencies like Bitcoin can take advantage of it in
.mining process
6. Communications
● - /Content centric Networking Named Data Networking
● .Analysing internet trafc in miliseconds
7. News and Media
● .News classifcation and recommendation systems
● What do people like to watch and read without asking them
( ).what they like based on their behaviour
8. Medical and Pharmacy
● ,Accurate medical prescription based on patients' lifestyle
.nutrition and genetic disorders
● .Analyzing and decreasing prescription drug side efects
● .Discovering cancers
● .Brain genome mapping and sequencing
9. Sales and Marketing
● .Analyzing customers behaviour Where do they look the
.most in a market store
● Finding the related products based on their previous
.purchases
10. Oil and Gas Exploration and Discovery
● , ,Weather soil and equipment data can be analyzed to predict the
success of drilling operations and make more intelligent
.decisions about drilling sites
● , ,Oilfeld managers need to analyze well data seismic data
industry news and potentially social media to evaluate potential
.oilfelds
● Processing seismic data that can be used to discover seismic
.trace signatures that were previously overlooked
11. Weather Forecasting
● .More accurate weather forecasting
● .Analysing and forecasting the risk of food
● .Discovering climate changes in the future
12. Geology
● .Forecasting the risk of earthquick based on aftershocks data
● .Analysing volcanoes' behaviour
13. Astronomy and Space
● - .Astronomical Hi Res Image Processing
● .Exploring new galaxies and planets
● .Analysing Meteorite impact models
14. Green and Renewable Energy
● : -GreenHadoop Leveraging Green Energy in Data Processing
.Frameworks
● Vestas utilized big data to maximize power generation and
.reduce energy costs
● Predicting the amount of solar energy that will be available in
.the near future
15. Crime and Urban Management
● .Accurate election result prediction
● .Preventing crime before it happens
16. And anywhere you can fnd high
volume of unstructured data with high
,rate of growth Apache Hadoop will
.change the game
17. THANK YOU
Presented by Mobin Ranjbar
:Twitter @MobinRanjbar
: . / /Linkedin linkedin com in mobinranjbar