This document discusses cloud computing, big data, Hadoop, and data analytics. It begins with an introduction to cloud computing, explaining its benefits like scalability, reliability, and low costs. It then covers big data concepts like the 3 Vs (volume, variety, velocity), Hadoop for processing large datasets, and MapReduce as a programming model. The document also discusses data analytics, describing different types like descriptive, diagnostic, predictive, and prescriptive analytics. It emphasizes that insights from analyzing big data are more valuable than raw data. Finally, it concludes that cloud computing can enhance business efficiency by enabling flexible access to computing resources for tasks like big data analytics.
The rise of “Big Data” on cloud computing: Review and open research issues
Paper Link: https://www.researchgate.net/publication/264624667_The_rise_of_Big_Data_on_cloud_computing_Review_and_open_research_issues
Cloud computing introduction and concept as per the RGPV, BE syllabus. PPt contains the material from various cloud Draft (NIST) and other research material to fulfill the Syllabus requirement.
The practice of using a network of remote servers hosted on the Internet to store, manage, and process data, rather than a local server or a personal computer.
The rise of “Big Data” on cloud computing: Review and open research issues
Paper Link: https://www.researchgate.net/publication/264624667_The_rise_of_Big_Data_on_cloud_computing_Review_and_open_research_issues
Cloud computing introduction and concept as per the RGPV, BE syllabus. PPt contains the material from various cloud Draft (NIST) and other research material to fulfill the Syllabus requirement.
The practice of using a network of remote servers hosted on the Internet to store, manage, and process data, rather than a local server or a personal computer.
Edge computing is a distributed computing paradigm that brings computation and data storage closer to the sources of data. This ppt contains everything about Edge Computing Starting from its Definition, needs, terms involved to its merits, demerits and application use cases
This deck is from Interpol Conference 2017, these slides shows the holistic view of machine learning in cyber security for better organization readiness
Cloud computing :
Accessibility: Cloud computing facilitates the access of applications and data from any location worldwide and from any device with an internet connection.
Cost savings: Cloud computing offers businesses scalable computing resources hence saving them on the cost of acquiring and maintaining them.
Security: Cloud providers especially those offering private cloud services, have strived to implement the best security standards and procedures in order to protect client’s data saved in the cloud.
Disaster recovery: Cloud computing offers the most efficient means for small, medium, and even large enterprises to backup and restore their data and applications in a fast and reliable way.
Big data Analytics is a process to extract meaningful insight from big such as hidden patterns, unknown correlations, market trends and customer preferences
Big data is a term that describes the large volume of data – both structured and unstructured – that inundates a business on a day-to-day basis. But it’s not the amount of data that’s important. It’s what organizations do with the data that matters. Big data can be analyzed for insights that lead to better decisions and strategic business moves.
Introduction to Cloud computing and Big Data-HadoopNagarjuna D.N
Cloud Computing Evolution
Why Cloud Computing needed?
Cloud Computing Models
Cloud Solutions
Cloud Jobs opportunities
Criteria for Big Data
Big Data challenges
Technologies to process Big Data- Hadoop
Hadoop History and Architecture
Hadoop Eco-System
Hadoop Real-time Use cases
Hadoop Job opportunities
Hadoop and SAP HANA integration
Summary
Edge computing is a distributed computing paradigm that brings computation and data storage closer to the sources of data. This ppt contains everything about Edge Computing Starting from its Definition, needs, terms involved to its merits, demerits and application use cases
This deck is from Interpol Conference 2017, these slides shows the holistic view of machine learning in cyber security for better organization readiness
Cloud computing :
Accessibility: Cloud computing facilitates the access of applications and data from any location worldwide and from any device with an internet connection.
Cost savings: Cloud computing offers businesses scalable computing resources hence saving them on the cost of acquiring and maintaining them.
Security: Cloud providers especially those offering private cloud services, have strived to implement the best security standards and procedures in order to protect client’s data saved in the cloud.
Disaster recovery: Cloud computing offers the most efficient means for small, medium, and even large enterprises to backup and restore their data and applications in a fast and reliable way.
Big data Analytics is a process to extract meaningful insight from big such as hidden patterns, unknown correlations, market trends and customer preferences
Big data is a term that describes the large volume of data – both structured and unstructured – that inundates a business on a day-to-day basis. But it’s not the amount of data that’s important. It’s what organizations do with the data that matters. Big data can be analyzed for insights that lead to better decisions and strategic business moves.
Introduction to Cloud computing and Big Data-HadoopNagarjuna D.N
Cloud Computing Evolution
Why Cloud Computing needed?
Cloud Computing Models
Cloud Solutions
Cloud Jobs opportunities
Criteria for Big Data
Big Data challenges
Technologies to process Big Data- Hadoop
Hadoop History and Architecture
Hadoop Eco-System
Hadoop Real-time Use cases
Hadoop Job opportunities
Hadoop and SAP HANA integration
Summary
In computer networking, cloud computing is a phrase used to describe a variety of computing concepts that involve a large number of computers connected through a communication network such as the Internet. It is very similar to the concept of utility computing. In science, cloud computing is a synonym for distributed computing over a network, and means the ability to run a program or application on many connected computers at the same time.
The phrase is often used in reference to network-based services, which appear to be provided by real server hardware, and are in fact served up by virtual hardware, simulated by software running on one or more real machines. Such virtual servers do not physically exist and can therefore be moved around and scaled up or down on the fly without affecting the end user, somewhat like a cloud becoming larger or smaller without being a physical object.
All Things Open 2014 - Day 1
Wednesday, October 22nd, 2014
Mark Hinkle
Senior Director & Citrix Open Source Business Office for Citrix
Cloud
Crash Course in Cloud Computing
Find more of Mark's talks here: http://www.slideshare.net/socializedsoftware
Big Data is all about extracting 4Vs
Cloud focuses on On-Demand, Elastic, Scalable, Pay-Per use Self Service models
“Big data is not a stand-alone technology; rather, it is a combination of the last 50 years of technology evolution”
Wave 1: Creating manageable data structures
Wave 2: Web and content management
Wave 3: Managing big data
Big Data and virtualization are two of the most exciting trends in the industry today. In this session you will learn about the components of Big Data systems, and how real-time, interactive and distributed processing systems like Hadoop integrate with existing applications and databases. The combination of Big Data systems with virtualization gives Hadoop and other Big Data technologies the key benefits of cloud computing: elasticity, multi-tenancy and high availability. A new open source project that VMware will announce at the Hadoop Summit will make it easy to deploy, configure and manage Hadoop on a virtualized infrastructure. We will discuss reference architectures for key Hadoop distributions anddiscuss future directions of this new open source project.
Overview: Big Data Use Cases in Telecom, Retail, Insurance, Automotive, Media & Banking & Finances Industry Segments. How can we map these business challenges to Solutions on AWS Cloud? Let's Find Out!
Big Data is Growing Bigger & Bigger with a prediction of 40 Zeta Bytes of Data by 2020.
> What are the 4 Vs of Big Data?
> Big Data Industry Use Cases:
- Telecommunications
- Retail
- Insurance
- Automotive
- Media
- Banking
Which AWS Components can be mapped to each stage of the Big Data Life Cycle:
AWS S3, AWS EC2, AWS EMR, AWS Redshift, Data Pipelines & many more.
Cloud Migration, Application Modernization and Security for PartnersAmazon Web Services
As AWS continues to expand, enterprise customers are increasingly looking to our partner ecosystem to assist in migrating their workloads to the cloud. This session describes the challenges, lessons learned, and best practices for large-scale application migrations. We will use real examples from our consulting partners and AWS Professional Services to illustrate how to move workloads to the cloud while modernizing the associated applications to take advantage of the unique benefits of AWS. We will also dive into how to use an array of AWS services and features to improve customers' security posture as they migrate and once they are up and running in the cloud.
A brief intro on the idea of what is Big Data and it's potential. This is primarily a basic study & I have quoted the source of infographics, stats & text at the end. If I have missed any reference due to human error & you recognize another source, please mention.
The Power of your Data Achieved - Next Gen ModernizationHortonworks
Fueled by ever-changing customer behaviors and an increasing number of industry disruptions, the modern enterprise requires analytics to stay ahead of the game. Today’s data warehouse needs continuous enhancements to address new requirements for advanced analytics, real-time streaming data, Big Data, and unstructured data. The focus should be on developing a forward-looking, future-proof view and holistically addressing the combination of forces that are impacting the existing operational model.
Adopting Hadoop to manage your Big Data is an important step, but not the end-solution to your Big Data challenges. Here are some of the additional considerations you must face:
Choosing the right cloud for the job: The massive computing and storage resources that are needed to support Big Data applications make cloud environments an ideal fit, and more than ever, there is a growing number of choices of cloud infrastructure types and providers. Given the diverse options, and the dynamic environments involved, it becomes ever more important to maintain the flexibility for all your IT needs.
Big Data is a complex beast: It involves many and different moving parts, in large clusters, and is continually growing and evolving. Managing such an environment manually is not a viable option. The question is, how can you achieve automation of all this complexity?
The world beyond Hadoop: Big Data is not just Hadoop – there is a whole rapidly growing ecosystem to contend with, including NoSQL, data processing, analytics tools… As well as your own application services. How can you manage deployment, configuration, scaling and failover of all the different pieces, in a consistent way?
In this session, you’ll learn how to deploy and manage your Hadoop cluster on any Cloud, as well as manage the rest of your big data application stack using a new open source framework called Cloudify.
SMAC - Social, Mobile, Analytics and Cloud - An overview Rajesh Menon
In this presentation, all the aspects of SMAC are covered in as much detail as possible. You will find some ideas worth sharing and also get attuned to Social, Mobile, Analytics and Cloud
Most common technology which is used to store meta data and large databases.we can find numerous applications in the real world.It is the very useful for creating new database oriented apps
In this paper, we discuss security issues for cloud computing, Map Reduce and Hadoop
environment. We also discuss various possible solutions for the issues in cloud computing
security and Hadoop. Today, Cloud computing security is developing at a rapid pace which
includes computer security, network security and information security. Cloud computing plays a
very vital role in protecting data, applications and the related infrastructure with the help of
policies, technologies and controls.
HIGH LEVEL VIEW OF CLOUD SECURITY: ISSUES AND SOLUTIONScscpconf
In this paper, we discuss security issues for cloud computing, Map Reduce and Hadoop
environment. We also discuss various possible solutions for the issues in cloud computing
security and Hadoop. Today, Cloud computing security is developing at a rapid pace which
includes computer security, network security and information security. Cloud computing plays a
very vital role in protecting data, applications and the related infrastructure with the help of
policies, technologies and controls.
Big data is data that, by virtue of its velocity, volume, or variety (the three Vs), cannot be easily stored or analyzed with traditional methods. Hadoop is an open-source software framework for storing data and running applications on clusters of commodity hardware.
Simplifying Your Cloud Architecture with a Logical Data Fabric (APAC)Denodo
Watch full webinar here: https://bit.ly/3dudL6u
It's not if you move to the cloud, but when. Most organisations are well underway with migrating applications and data to the cloud. In fact, most organisations - whether they realise it or not - have a multi-cloud strategy. Single, hybrid, or multi-cloud…the potential benefits are huge - flexibility, agility, cost savings, scaling on-demand, etc. However, the challenges can be just as large and daunting. A poorly managed migration to the cloud can leave users frustrated at their inability to get to the data that they need and IT scrambling to cobble together a solution.
In this session, we will look at the challenges facing data management teams as they migrate to cloud and multi-cloud architectures. We will show how the Denodo Platform can:
- Reduce the risk and minimise the disruption of migrating to the cloud.
- Make it easier and quicker for users to find the data that they need - wherever it is located.
- Provide a uniform security layer that spans hybrid and multi-cloud environments.
Neuro-symbolic is not enough, we need neuro-*semantic*Frank van Harmelen
Neuro-symbolic (NeSy) AI is on the rise. However, simply machine learning on just any symbolic structure is not sufficient to really harvest the gains of NeSy. These will only be gained when the symbolic structures have an actual semantics. I give an operational definition of semantics as “predictable inference”.
All of this illustrated with link prediction over knowledge graphs, but the argument is general.
Search and Society: Reimagining Information Access for Radical FuturesBhaskar Mitra
The field of Information retrieval (IR) is currently undergoing a transformative shift, at least partly due to the emerging applications of generative AI to information access. In this talk, we will deliberate on the sociotechnical implications of generative AI for information access. We will argue that there is both a critical necessity and an exciting opportunity for the IR community to re-center our research agendas on societal needs while dismantling the artificial separation between the work on fairness, accountability, transparency, and ethics in IR and the rest of IR research. Instead of adopting a reactionary strategy of trying to mitigate potential social harms from emerging technologies, the community should aim to proactively set the research agenda for the kinds of systems we should build inspired by diverse explicitly stated sociotechnical imaginaries. The sociotechnical imaginaries that underpin the design and development of information access technologies needs to be explicitly articulated, and we need to develop theories of change in context of these diverse perspectives. Our guiding future imaginaries must be informed by other academic fields, such as democratic theory and critical theory, and should be co-developed with social science scholars, legal scholars, civil rights and social justice activists, and artists, among others.
Let's dive deeper into the world of ODC! Ricardo Alves (OutSystems) will join us to tell all about the new Data Fabric. After that, Sezen de Bruijn (OutSystems) will get into the details on how to best design a sturdy architecture within ODC.
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
Epistemic Interaction - tuning interfaces to provide information for AI supportAlan Dix
Paper presented at SYNERGY workshop at AVI 2024, Genoa, Italy. 3rd June 2024
https://alandix.com/academic/papers/synergy2024-epistemic/
As machine learning integrates deeper into human-computer interactions, the concept of epistemic interaction emerges, aiming to refine these interactions to enhance system adaptability. This approach encourages minor, intentional adjustments in user behaviour to enrich the data available for system learning. This paper introduces epistemic interaction within the context of human-system communication, illustrating how deliberate interaction design can improve system understanding and adaptation. Through concrete examples, we demonstrate the potential of epistemic interaction to significantly advance human-computer interaction by leveraging intuitive human communication strategies to inform system design and functionality, offering a novel pathway for enriching user-system engagements.
UiPath Test Automation using UiPath Test Suite series, part 4DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 4. In this session, we will cover Test Manager overview along with SAP heatmap.
The UiPath Test Manager overview with SAP heatmap webinar offers a concise yet comprehensive exploration of the role of a Test Manager within SAP environments, coupled with the utilization of heatmaps for effective testing strategies.
Participants will gain insights into the responsibilities, challenges, and best practices associated with test management in SAP projects. Additionally, the webinar delves into the significance of heatmaps as a visual aid for identifying testing priorities, areas of risk, and resource allocation within SAP landscapes. Through this session, attendees can expect to enhance their understanding of test management principles while learning practical approaches to optimize testing processes in SAP environments using heatmap visualization techniques
What will you get from this session?
1. Insights into SAP testing best practices
2. Heatmap utilization for testing
3. Optimization of testing processes
4. Demo
Topics covered:
Execution from the test manager
Orchestrator execution result
Defect reporting
SAP heatmap example with demo
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Key Trends Shaping the Future of Infrastructure.pdfCheryl Hung
Keynote at DIGIT West Expo, Glasgow on 29 May 2024.
Cheryl Hung, ochery.com
Sr Director, Infrastructure Ecosystem, Arm.
The key trends across hardware, cloud and open-source; exploring how these areas are likely to mature and develop over the short and long-term, and then considering how organisations can position themselves to adapt and thrive.
Software Delivery At the Speed of AI: Inflectra Invests In AI-Powered QualityInflectra
In this insightful webinar, Inflectra explores how artificial intelligence (AI) is transforming software development and testing. Discover how AI-powered tools are revolutionizing every stage of the software development lifecycle (SDLC), from design and prototyping to testing, deployment, and monitoring.
Learn about:
• The Future of Testing: How AI is shifting testing towards verification, analysis, and higher-level skills, while reducing repetitive tasks.
• Test Automation: How AI-powered test case generation, optimization, and self-healing tests are making testing more efficient and effective.
• Visual Testing: Explore the emerging capabilities of AI in visual testing and how it's set to revolutionize UI verification.
• Inflectra's AI Solutions: See demonstrations of Inflectra's cutting-edge AI tools like the ChatGPT plugin and Azure Open AI platform, designed to streamline your testing process.
Whether you're a developer, tester, or QA professional, this webinar will give you valuable insights into how AI is shaping the future of software delivery.
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...DanBrown980551
Do you want to learn how to model and simulate an electrical network from scratch in under an hour?
Then welcome to this PowSyBl workshop, hosted by Rte, the French Transmission System Operator (TSO)!
During the webinar, you will discover the PowSyBl ecosystem as well as handle and study an electrical network through an interactive Python notebook.
PowSyBl is an open source project hosted by LF Energy, which offers a comprehensive set of features for electrical grid modelling and simulation. Among other advanced features, PowSyBl provides:
- A fully editable and extendable library for grid component modelling;
- Visualization tools to display your network;
- Grid simulation tools, such as power flows, security analyses (with or without remedial actions) and sensitivity analyses;
The framework is mostly written in Java, with a Python binding so that Python developers can access PowSyBl functionalities as well.
What you will learn during the webinar:
- For beginners: discover PowSyBl's functionalities through a quick general presentation and the notebook, without needing any expert coding skills;
- For advanced developers: master the skills to efficiently apply PowSyBl functionalities to your real-world scenarios.
PHP Frameworks: I want to break free (IPC Berlin 2024)Ralf Eggert
In this presentation, we examine the challenges and limitations of relying too heavily on PHP frameworks in web development. We discuss the history of PHP and its frameworks to understand how this dependence has evolved. The focus will be on providing concrete tips and strategies to reduce reliance on these frameworks, based on real-world examples and practical considerations. The goal is to equip developers with the skills and knowledge to create more flexible and future-proof web applications. We'll explore the importance of maintaining autonomy in a rapidly changing tech landscape and how to make informed decisions in PHP development.
This talk is aimed at encouraging a more independent approach to using PHP frameworks, moving towards a more flexible and future-proof approach to PHP development.
Kubernetes & AI - Beauty and the Beast !?! @KCD Istanbul 2024Tobias Schneck
As AI technology is pushing into IT I was wondering myself, as an “infrastructure container kubernetes guy”, how get this fancy AI technology get managed from an infrastructure operational view? Is it possible to apply our lovely cloud native principals as well? What benefit’s both technologies could bring to each other?
Let me take this questions and provide you a short journey through existing deployment models and use cases for AI software. On practical examples, we discuss what cloud/on-premise strategy we may need for applying it to our own infrastructure to get it to work from an enterprise perspective. I want to give an overview about infrastructure requirements and technologies, what could be beneficial or limiting your AI use cases in an enterprise environment. An interactive Demo will give you some insides, what approaches I got already working for real.
DevOps and Testing slides at DASA ConnectKari Kakkonen
My and Rik Marselis slides at 30.5.2024 DASA Connect conference. We discuss about what is testing, then what is agile testing and finally what is Testing in DevOps. Finally we had lovely workshop with the participants trying to find out different ways to think about quality and testing in different parts of the DevOps infinity loop.
Essentials of Automations: Optimizing FME Workflows with ParametersSafe Software
Are you looking to streamline your workflows and boost your projects’ efficiency? Do you find yourself searching for ways to add flexibility and control over your FME workflows? If so, you’re in the right place.
Join us for an insightful dive into the world of FME parameters, a critical element in optimizing workflow efficiency. This webinar marks the beginning of our three-part “Essentials of Automation” series. This first webinar is designed to equip you with the knowledge and skills to utilize parameters effectively: enhancing the flexibility, maintainability, and user control of your FME projects.
Here’s what you’ll gain:
- Essentials of FME Parameters: Understand the pivotal role of parameters, including Reader/Writer, Transformer, User, and FME Flow categories. Discover how they are the key to unlocking automation and optimization within your workflows.
- Practical Applications in FME Form: Delve into key user parameter types including choice, connections, and file URLs. Allow users to control how a workflow runs, making your workflows more reusable. Learn to import values and deliver the best user experience for your workflows while enhancing accuracy.
- Optimization Strategies in FME Flow: Explore the creation and strategic deployment of parameters in FME Flow, including the use of deployment and geometry parameters, to maximize workflow efficiency.
- Pro Tips for Success: Gain insights on parameterizing connections and leveraging new features like Conditional Visibility for clarity and simplicity.
We’ll wrap up with a glimpse into future webinars, followed by a Q&A session to address your specific questions surrounding this topic.
Don’t miss this opportunity to elevate your FME expertise and drive your projects to new heights of efficiency.
2. Introduction
Why Cloud Computing
Benefits of Cloud Computing
Characteristics
Advantages of Cloud Computing
Disadvantages of Cloud
Computing
How Cloud Computing Works
Challenges of Cloud Computing
Layers of Cloud Computing
Components of Cloud Computing
Big Data
3 Vs of Big Data
Importance of Big Data
What Comes Under Big Data
Hadoop
Hadoop Architecture
Hadoop With Big Data
Map Reduce
Why Data Analytics
Types of Analysis
Types of Data Analytics
Big Data Analytics
Conclusion
References
Thanking You
2
3. Cloud computing is an internet based computer
technology. It is the next stage technology that
uses the clouds to provide the services
whenever and wherever the user need it. It
provides a method to access several servers
world wide.
What is Cloud?
A cloud is a combination of networks,
hardware, services, storage, and interfaces
that helps in delivering computing as a
service.
What is Cloud Computing ?
3
5. Benefits of Cloud Computing
Cloud computing enables companies and
applications, which are system
infrastructure dependent, to be
infrastructure-less.
By using the Cloud infrastructure on “pay
as used and on demand”, all of us can save
in capital and operational investment!
Clients can:-
Put their data on the platform instead of on their
own desktop PCs and/or on their own servers.
They can put their applications on the cloud and
use the servers within the cloud to do processing
and data manipulations etc.
5
8. Disadvantages of Cloud Computing
Requires a constant Internet
connection
Stored data might not be secured
Limited control and flexibility
More risk on information leakage
Users cannot be aware of the
network
Dependencies on service suppliers for
implementing data management
8
9.
10. Use of cloud computing means dependence on
others and that could possibly limit flexibility
and innovation
Security could prove to be a big issue:
It is still unclear how safe out-sourced data is and when using these services
Ownership of data is not always clear.
Data Centre can become environmental
hazards: Green Cloud
Cloud Interoperability is still an issue.
11. Layers of Cloud Computing
Infrastructure as a service (IaaS):-It provides cloud infrastructure
in terms of hardware as like memory, processor, speed etc.
Platform as a service (PaaS):It provides cloud application
platform for the developer.
Software as a service (SaaS)::It provides the cloud applications
to users directly without installing anything on the system.
These applications remains on cloud.
13. Big Data
Big Data refers to a collection of data sets so large
and complex. It is impossible to process them with
the usual databases and tools because of its size and
associated numbers. Big data is hard to capture, store,
search, share, analyze and visualize.
14. 3 Vs of Big Data
The “BIG” in big data isn’t just about volume
Volume
Variety
Velocity
15. Importance of Big Data
The importance of big data does not revolve around how much data you have ,
but what you do with it.
You can take data from any source and analyze it to find answer that enables,
Cost reductions.
Time reductions.
New product development and optimized offerings .
Smart decision making.
16. Black Box Data
Social Media Data
Stock Exchange Data
Power Grid Data
Transport Data
Search Engine Data
Structured data
Semi Structured data
Unstructured data
17. What is Hadoop ?
Hadoop is an open-source software framework for storing
data and running applications on clusters of commodity
hardware. It provides massive storage for any kind of
data, enormous processing power and the ability to handle
virtually limitless concurrent tasks or jobs.
The software framework that supports HDFS,
MapReduce and other related entities is called the project
Hadoop or simply Hadoop.
This is open source and distributed by Apache.
18. Hadoop Ecosystem
Apache Oozie (Workflow)
Pig Latin
Data Analysis
Mahout
Machine Learning
HDFS (Hadoop Distributed File System)
Map Reduce Framework
Flume Sqoop
Unstructured or
Semi-Structured data
Structured data
Pig Latin
Data Analysis
Mahout
Machine Learning
H Base
Hive
DW System
19. With Big Data
Hadoop is the core platform for
structuring Big Data, and solves the
problem of formatting it for
subsequent analytics
purposes. Hadoop uses a distributed
computing architecture consisting of
multiple servers using commodity
hardware, making it relatively
20. Cost Effective System
Large Cluster of Notes
Parallel Processing
Distributive Data
Automatic failover management
Data Locality optimization
Heterogeneous Cluster
Scalability
21. Map Reduce
MapReduce is a programming model that Google has used
successfully in processing its “big-data” sets (~ 20000 peta bytes
per day)
A map function extracts some intelligence from
raw data.
A reduce function aggregates according to some
guides the data output by the map.
Users specify the computation in terms of a map
and a reduce function,
Underlying runtime system automatically
parallelizes the computation across large-scale
clusters of machines, and
Underlying system also handles machine failures,
efficient communications, and performance issues.
22. Broken into pieces
[ MAP ]
Computation
Computation
Computation
Computation
Computation
Computation
Shuffle and Sort
23. Why Data Analysis?
It is important to remember that the primary
value from big data does not come from the
data in its raw form but from the processing
and analysis of it and the insights, products
and services that emerge from analysis.
24. For unstructured data to be useful it must be analysed to extract and
expose the information it contains
Different types of analysis are possible, such as:-
Entity analysis – people, organisations, objects and events, and the relationships
between them
Topic analysis – topics or themes, and their relative importance
Sentiment analysis – subjective view of a person to a particular topic
Feature analysis – Inherent characteristics that are significant for a particular analytical
perspective (e.g. land coverage in satellite imagery)
Types Of Analysis
25. Types Of Data Analytics
Analytic Excellence leads to better decisions:-
Descriptive Analytics : What is happening?
Diagnostic Analytics : Why did it happen?
Predictive Analytics : What is likely going to
happen?
Prescriptive Analytics : What should we do about it?
26. Analytics
Focus On :-
Predictive Analysis
Data Science
Data Sets:-
Large Scale Data Sets
More type of Data
Raw Data
Complex Data Models
Supports:-
Correlations – new insight more accurate answer
27. Two IT initiatives are currently top of mind for organizations across the globe i.e.
Big Data Analytics
Cloud Computing
As a delivery model for IT services , cloud computing has the potential to enhance
business agility and productivity while enabling greater efficiencies and reducing
costs.
In the current scenario , Big Data is a big challenge for the organizations .
To store and process such large volume of data , variety of data and velocity of data
Hadoop came into existence.
Our presentation is all about Cloud Computing , Big Data & Big Data Analytics.