Big data solutions on the cloud are the future according to the author. Big data analytics and cloud computing both offer organizations valuable insights, innovations, and revenue opportunities while also enhancing agility, efficiency, and reducing costs. Hadoop is commonly used for big data but Spark is emerging as more flexible. The cloud is becoming a popular option for big data due to elastic infrastructure and ability to process data close to its source. Common big data use cases include optimizing sales funnels, behavioral analytics, customer segmentation, predictive maintenance, and market analysis.
Real time trade surveillance in financial marketsHortonworks
Who’s winning the deep forensic analysis ‘arms race’ for compliance? Real-time trade surveillance in global financial markets has created a data tsunami. With greater volumes of data comes greater compliance risk. CNBC reports U.S. Banks have been fined over $200B since the financial crisis. How are compliance teams fighting back to make more of the data and stay out of regulatory hot water? Rapid response to suspect trades means compliance teams need to access and visualize trade patterns, real time and historic data, to navigate the data in depth and flag possible violations. Join Hortonworks and Arcadia for this live webinar: we’ll cover the use case at a top 50 Global Bank who now has deep forensic analysis of trade activity. The result: interactive, ad hoc data visualization and access across multiple platforms – without limits on historic data – to detect irregularities as they happen. In-depth expert presentations by:
Shailesh Ambike, Executive Co-Chair of Compliance & Legal Section (CLS) Education Sub-Committee of the Investment Industry Regulatory Organization of Canada (IIROC)
Vamsi K Chemitiganti, GM – Financial Services at Hortonworks
Informatica Becomes Part of the Business Data Lake EcosystemCapgemini
Informatica is now part of the Business Data Lake ecosystem developed by Capgemini and Pivotal. Customers worldwide will now be able to leverage Informatica’s data integration software in addition to Pivotal’s advanced big data, analytics and application software, and Capgemini’s industry and implementation expertise. Informatica will deliver certified technologies for Data Integration, Data Quality and Master Data Management (MDM) to help enterprises distill raw data into actionable insights.
http://www.capgemini.com/resources/the-business-data-lake-delivering-the-speed-and-accuracy-to-solve-your-big-data-problems
Real time trade surveillance in financial marketsHortonworks
Who’s winning the deep forensic analysis ‘arms race’ for compliance? Real-time trade surveillance in global financial markets has created a data tsunami. With greater volumes of data comes greater compliance risk. CNBC reports U.S. Banks have been fined over $200B since the financial crisis. How are compliance teams fighting back to make more of the data and stay out of regulatory hot water? Rapid response to suspect trades means compliance teams need to access and visualize trade patterns, real time and historic data, to navigate the data in depth and flag possible violations. Join Hortonworks and Arcadia for this live webinar: we’ll cover the use case at a top 50 Global Bank who now has deep forensic analysis of trade activity. The result: interactive, ad hoc data visualization and access across multiple platforms – without limits on historic data – to detect irregularities as they happen. In-depth expert presentations by:
Shailesh Ambike, Executive Co-Chair of Compliance & Legal Section (CLS) Education Sub-Committee of the Investment Industry Regulatory Organization of Canada (IIROC)
Vamsi K Chemitiganti, GM – Financial Services at Hortonworks
Informatica Becomes Part of the Business Data Lake EcosystemCapgemini
Informatica is now part of the Business Data Lake ecosystem developed by Capgemini and Pivotal. Customers worldwide will now be able to leverage Informatica’s data integration software in addition to Pivotal’s advanced big data, analytics and application software, and Capgemini’s industry and implementation expertise. Informatica will deliver certified technologies for Data Integration, Data Quality and Master Data Management (MDM) to help enterprises distill raw data into actionable insights.
http://www.capgemini.com/resources/the-business-data-lake-delivering-the-speed-and-accuracy-to-solve-your-big-data-problems
The Modern Data Architecture for Predictive Analytics with Hortonworks and Re...Revolution Analytics
Hortonworks and Revolution Analytics have teamed up to bring the predictive analytics power of R to Hortonworks Data Platform.
Hadoop, being a disruptive data processing framework, has made a large impact in the data ecosystems of today. Enabling business users to translate existing skills to Hadoop is necessary to encourage the adoption and allow businesses to get value out of their Hadoop investment quickly. R, being a prolific and rapidly growing data analysis language, now has a place in the Hadoop ecosystem.
This presentation covers:
- Trends and business drivers for Hadoop
- How Hortonworks and Revolution Analytics play a role in the modern data architecture
- How you can run R natively in Hortonworks Data Platform to simply move your R-powered analytics to Hadoop
Presentation replay at:
http://www.revolutionanalytics.com/news-events/free-webinars/2013/modern-data-architecture-revolution-hortonworks/
Enterprise Search: Addressing the First Problem of Big Data & Analytics - Sta...StampedeCon
This session addresses the first problems of Big Data & Analytics–Identifying, Indexing, Connecting and Gaining Insight of Existing Data to Drive Value. HPE’s Chief Field Technologist will give her perspectives on Enterprise Search as a Fundamental Cornerstone of Building a Data Driven Enterprise.
Nov 2014 talk to SW Data Meetup by Mike Olson, co-founder and chairman of Cloudera.
In business, we often deal with hype around trends in society, politics, economy and technology. We know we need to take claims of the next big thing with a grain of salt and that we should be careful not to set expectations too high. However, with Big Data analytics, the opposite is true. The hype that accompanies it actually conceals the enormity of its impact on the way we do business. In this talk I’ll discuss how new 'Data Driven' economies are emerging through relentless innovation across the public and private sectors.
Mike (co-founded Cloudera in 2008 and served as its CEO until 2013 when he took on his current role of chief strategy officer (CSO.) As CSO, Mike is responsible for Cloudera’s product strategy, open source leadership, engineering alignment and direct engagement with customers. Prior to Cloudera Mike was CEO of Sleepycat Software, makers of Berkeley DB, the open source embedded database engine. Mike spent two years at Oracle Corporation as vice president for Embedded Technologies after Oracle’s acquisition of Sleepycat in 2006. Prior to joining Sleepycat, Mike held technical and business positions at database vendors Britton Lee, Illustra Information Technologies and Informix Software. Mike has a Bachelor’s and a Master’s Degree in Computer Science from the University of California, Berkeley.
The Big Data Journey – How Companies Adopt Hadoop - StampedeCon 2016StampedeCon
Hadoop adoption is a journey. Depending on the business the process can take weeks, months, or even years. Hadoop is a transformative technology so the challenges have less to do with the technology and more to do with how a company adapts itself to a new way of thinking about data. There are challenges for companies who have lived with an application driven business for the last two decades to suddenly become data driven. Companies need to begin thinking less in terms of single, silo’d servers and more about “the cluster”.
The concept of the cluster becomes the center of data gravity drawing all the applications to it. Companies, especially the IT organizations, embark on a process of understanding how to maintain and operationalize this environment and provide the data lake as a service to the businesses. They must empower the business by providing the resources for the use cases which drive both renovation and innovation. IT needs to adopt new technologies and new methodologies which enable the solutions. This is not technology for technology sake. Hadoop is a data platform servicing and enabling all facets of an organization. Building out and expanding this platform is the ongoing journey as word gets out to businesses that they can have any data they want and any time. Success is what drives the journey.
The length of the journey varies from company to company. Sometimes the challenges are based on the size of the company but many times the challenges are based on the difficulty of unseating established IT processes companies have adopted without forethought for the past two decades. Companies must navigate through the noise. Sifting through the noise to find those solutions which bring real value takes time. As the platform matures and becomes mainstream, more and more companies are finding it easier to adopt Hadoop. Hundreds of companies have already taken many steps; hundreds more have already taken the first step. As the wave of successful Hadoop adoption continues, more and more companies will see the value in starting the journey and paving the way for others.
GITEX Big Data Conference 2014 – SAP PresentationPedro Pereira
Big, Fast and Predictive Data: How to Extract Real Business Value – in real time.
90% of the world’s data was created in the last two years. If you can harness it, it will revolutionize the way you do business. Big Data solutions can help extract real business value – in real time.
Revolution in Business Analytics-Zika Virus ExampleBardess Group
Even from the “man in the street” perspective, there is a sense that we are living in an increasingly algorithmic world. Self-driving cars, pizza delivery by drone, and smart houses are commonplace. The technologies enabling this revolution are both simultaneously mature and evolving rapidly.
In this session, we’ll took a look at a real world problem, the recent global outbreak of the ZIka virus, and used data analytics technologies to gain valuable insights that can assist authorities and the general public to understand and potentially prevent the spread of this disease.
Bardess Group, a sponsor of the event and business analytics consulting firm, will demonstrate how huge, extremely jagged data from a variety of sources can be collected and prepared and rapidly made available for analysis. Advanced machine learning and predictive analysis further enhance the value of those insights.
Finally, Bardess will make the case that using a systematic approach to conceptually visualize the strategic journey to insightful business analytics, the analytics value chain, can assist any organization prepare for this revolution in analytics.
Also see http://cloudera.qlik.com for the demos.
Big Data Hadoop Briefing Hosted by Cisco, WWT and MapR: MapR Overview Present...ervogler
Learn more about how MapR gives you the most technologically advanced distribution for Hadoop, with the product, services, and partner network to ensure production success and continued success.
Transforming GE Healthcare with Data Platform StrategyDatabricks
Data and Analytics is foundational to the success of GE Healthcare’s digital transformation and market competitiveness. This use case focuses on a heavy platform transformation that GE Healthcare drove in the last year to move from an On prem legacy data platforming strategy to a cloud native and completely services oriented strategy. This was a huge effort for an 18Bn company and executed in the middle of the pandemic. It enables GE Healthcare to leap frog in the enterprise data analytics strategy.
"Big Data Use Cases" was presented to Lansing BigData and Hadoop Users Group Kickoff meeting on 2/24/2015 by Vijay Mandava and Lan Jiang. The demo was built on top of CDH 5.3, HDP 2.2 and AWS cloud
A global survey of more than 300 data management professionals conducted by independent research firm Dimensional Research® showed that enterprises of all sizes face challenges on a range of key data performance management issues from stopping bad data to keeping data flows operating effectively. In particular, 87 percent of respondents report flowing bad data into their data stores while just 12 percent consider themselves good at the key aspects of data flow performance management.
How to Become an Analytics Ready Insurer - with Informatica and HortonworksHortonworks
Whether you are an insurer, reinsurer, broker or insurance service provider; everything you do is based on analytics. From underwriting to claims to agency and marketing, the smartest and most streamlined business operations at insurance companies are driven by advanced and intelligent analytics. But is your data ready? Are you an “Analytics Ready” insurer? Great analytics starts with great data management. Join us as industry experts from Informatica and Hortonworks share industry trends and best practices to show you how to become an “Analytics Ready” insurer.
Fighting Financial Crime with Artificial IntelligenceDataWorks Summit
How can we take the state of the art in deep learning and AI research, and transplant it into a large bank to deliver useful results which impact the general public? To answer this broad-reaching question, we take the viewer through a solution Think Big Analytics recently deployed at a major European bank for fraud detection, using state of the art AI techniques and a near-real time open-source architecture. We show how financial transactions can be transposed into a form where the latest AI techniques in image recognition can be leveraged, in surprisingly novel ways. We have been able to more accurately detect fraud and reduce financial crime, cutting losses and improving customer experience. We describe some architectures which can be used to do this in production, at scale, in global financial institutions.
Speaker:
Tim Seears, Director of Data Science, Think Big Analytics, a Teradata Company
Top 5 Strategies for Retail Data AnalyticsHortonworks
It’s an exciting time for retailers as technology is driving a major disruption in the market. Whether you are just beginning to build a retail data analytics program or you have been gaining advanced insights from your data for quite some time, join Eric and Shish as we explore the trends, drivers and hurdles in retail data analytics
There has been an explosion of data digitising our physical world – from cameras, environmental sensors and embedded devices, right down to the phones in our pockets. Which means that, now, companies have new ways to transform their businesses – both operationally, and through their products and services – by leveraging this data and applying fresh analytical techniques to make sense of it. But are they ready? The answer is “no” in most cases.
In this session, we’ll be discussing the challenges facing companies trying to embrace the Analytics of Things, and how Teradata has helped customers work through and turn those challenges to their advantage.
Hear how Manulife Asia has built an environment that enables the company to solve business-critical problems across many countries. What began in 2017 as an update to their enterprise architecture now spans everything from infrastructure to applications, powering their entire digital backbone. It includes fraud identification, real-time investment dashboards, advanced analytics and machine learning, and digital connection apps that talk to customers for claims, support, and more. Learn the importance hard work, coordination, discipline, and an agile methodology play in deciding which use cases they will focus on to deliver new services in an environment where everything is time sensitive and business requirements shift regularly.
Speaker: Ellen Wu, Head of Asia Data Office, Global Data Enablement and Governance, Manulife
Explore how data integration (or “mashups”) can maximize analytic value and help business teams create streamlined data pipelines that enables ad-hoc analytic inquiries. You’ll learn why businesses increasingly focused on blending data on demand and at the source, the concrete analytic advantages that this approach delivers, and the type of architectures required for delivering trusted, blended data. We provide a checklist to assess your data integration needs and capabilities, and review some real-world examples of how blending various data types has created significant analytic value and concrete business impact.
Hadoop 2.0: YARN to Further Optimize Data ProcessingHortonworks
Data is exponentially increasing in both types and volumes, creating opportunities for businesses. Watch this video and learn from three Big Data experts: John Kreisa, VP Strategic Marketing at Hortonworks, Imad Birouty, Director of Technical Product Marketing at Teradata and John Haddad, Senior Director of Product Marketing at Informatica.
Multiple systems are needed to exploit the variety and volume of data sources, including a flexible data repository. Learn more about:
- Apache Hadoop 2 and YARN
- Data Lakes
- Intelligent data management layers needed to manage metadata and usage patterns as well as track consumption across these data platforms.
Getting started with Hadoop on the Cloud with BluemixNicolas Morales
Silicon Valley Code Camp -- October 11, 2014.
Session: Getting started with Hadoop on the Cloud.
Hadoop and Cloud is an almost perfect marriage. Hadoop is a distributed computing framework that leverages a cluster built on commodity hardware. The Cloud simplifies provisioning of machines and software. Getting started with Hadoop on the Cloud makes it simple to provision your environment quickly and actually get started using Hadoop. IBM Bluemix has democratized Hadoop for the masses! This session will provide a brief introduction to what Hadoop is, how does cloud work and will then focus on how to get started via a series of demos. We will conclude with a discussion around the tutorials and public datasets - all of the tools needed to get you started quickly.
Learn more about BigInsights for Hadoop: https://developer.ibm.com/hadoop/
The Modern Data Architecture for Predictive Analytics with Hortonworks and Re...Revolution Analytics
Hortonworks and Revolution Analytics have teamed up to bring the predictive analytics power of R to Hortonworks Data Platform.
Hadoop, being a disruptive data processing framework, has made a large impact in the data ecosystems of today. Enabling business users to translate existing skills to Hadoop is necessary to encourage the adoption and allow businesses to get value out of their Hadoop investment quickly. R, being a prolific and rapidly growing data analysis language, now has a place in the Hadoop ecosystem.
This presentation covers:
- Trends and business drivers for Hadoop
- How Hortonworks and Revolution Analytics play a role in the modern data architecture
- How you can run R natively in Hortonworks Data Platform to simply move your R-powered analytics to Hadoop
Presentation replay at:
http://www.revolutionanalytics.com/news-events/free-webinars/2013/modern-data-architecture-revolution-hortonworks/
Enterprise Search: Addressing the First Problem of Big Data & Analytics - Sta...StampedeCon
This session addresses the first problems of Big Data & Analytics–Identifying, Indexing, Connecting and Gaining Insight of Existing Data to Drive Value. HPE’s Chief Field Technologist will give her perspectives on Enterprise Search as a Fundamental Cornerstone of Building a Data Driven Enterprise.
Nov 2014 talk to SW Data Meetup by Mike Olson, co-founder and chairman of Cloudera.
In business, we often deal with hype around trends in society, politics, economy and technology. We know we need to take claims of the next big thing with a grain of salt and that we should be careful not to set expectations too high. However, with Big Data analytics, the opposite is true. The hype that accompanies it actually conceals the enormity of its impact on the way we do business. In this talk I’ll discuss how new 'Data Driven' economies are emerging through relentless innovation across the public and private sectors.
Mike (co-founded Cloudera in 2008 and served as its CEO until 2013 when he took on his current role of chief strategy officer (CSO.) As CSO, Mike is responsible for Cloudera’s product strategy, open source leadership, engineering alignment and direct engagement with customers. Prior to Cloudera Mike was CEO of Sleepycat Software, makers of Berkeley DB, the open source embedded database engine. Mike spent two years at Oracle Corporation as vice president for Embedded Technologies after Oracle’s acquisition of Sleepycat in 2006. Prior to joining Sleepycat, Mike held technical and business positions at database vendors Britton Lee, Illustra Information Technologies and Informix Software. Mike has a Bachelor’s and a Master’s Degree in Computer Science from the University of California, Berkeley.
The Big Data Journey – How Companies Adopt Hadoop - StampedeCon 2016StampedeCon
Hadoop adoption is a journey. Depending on the business the process can take weeks, months, or even years. Hadoop is a transformative technology so the challenges have less to do with the technology and more to do with how a company adapts itself to a new way of thinking about data. There are challenges for companies who have lived with an application driven business for the last two decades to suddenly become data driven. Companies need to begin thinking less in terms of single, silo’d servers and more about “the cluster”.
The concept of the cluster becomes the center of data gravity drawing all the applications to it. Companies, especially the IT organizations, embark on a process of understanding how to maintain and operationalize this environment and provide the data lake as a service to the businesses. They must empower the business by providing the resources for the use cases which drive both renovation and innovation. IT needs to adopt new technologies and new methodologies which enable the solutions. This is not technology for technology sake. Hadoop is a data platform servicing and enabling all facets of an organization. Building out and expanding this platform is the ongoing journey as word gets out to businesses that they can have any data they want and any time. Success is what drives the journey.
The length of the journey varies from company to company. Sometimes the challenges are based on the size of the company but many times the challenges are based on the difficulty of unseating established IT processes companies have adopted without forethought for the past two decades. Companies must navigate through the noise. Sifting through the noise to find those solutions which bring real value takes time. As the platform matures and becomes mainstream, more and more companies are finding it easier to adopt Hadoop. Hundreds of companies have already taken many steps; hundreds more have already taken the first step. As the wave of successful Hadoop adoption continues, more and more companies will see the value in starting the journey and paving the way for others.
GITEX Big Data Conference 2014 – SAP PresentationPedro Pereira
Big, Fast and Predictive Data: How to Extract Real Business Value – in real time.
90% of the world’s data was created in the last two years. If you can harness it, it will revolutionize the way you do business. Big Data solutions can help extract real business value – in real time.
Revolution in Business Analytics-Zika Virus ExampleBardess Group
Even from the “man in the street” perspective, there is a sense that we are living in an increasingly algorithmic world. Self-driving cars, pizza delivery by drone, and smart houses are commonplace. The technologies enabling this revolution are both simultaneously mature and evolving rapidly.
In this session, we’ll took a look at a real world problem, the recent global outbreak of the ZIka virus, and used data analytics technologies to gain valuable insights that can assist authorities and the general public to understand and potentially prevent the spread of this disease.
Bardess Group, a sponsor of the event and business analytics consulting firm, will demonstrate how huge, extremely jagged data from a variety of sources can be collected and prepared and rapidly made available for analysis. Advanced machine learning and predictive analysis further enhance the value of those insights.
Finally, Bardess will make the case that using a systematic approach to conceptually visualize the strategic journey to insightful business analytics, the analytics value chain, can assist any organization prepare for this revolution in analytics.
Also see http://cloudera.qlik.com for the demos.
Big Data Hadoop Briefing Hosted by Cisco, WWT and MapR: MapR Overview Present...ervogler
Learn more about how MapR gives you the most technologically advanced distribution for Hadoop, with the product, services, and partner network to ensure production success and continued success.
Transforming GE Healthcare with Data Platform StrategyDatabricks
Data and Analytics is foundational to the success of GE Healthcare’s digital transformation and market competitiveness. This use case focuses on a heavy platform transformation that GE Healthcare drove in the last year to move from an On prem legacy data platforming strategy to a cloud native and completely services oriented strategy. This was a huge effort for an 18Bn company and executed in the middle of the pandemic. It enables GE Healthcare to leap frog in the enterprise data analytics strategy.
"Big Data Use Cases" was presented to Lansing BigData and Hadoop Users Group Kickoff meeting on 2/24/2015 by Vijay Mandava and Lan Jiang. The demo was built on top of CDH 5.3, HDP 2.2 and AWS cloud
A global survey of more than 300 data management professionals conducted by independent research firm Dimensional Research® showed that enterprises of all sizes face challenges on a range of key data performance management issues from stopping bad data to keeping data flows operating effectively. In particular, 87 percent of respondents report flowing bad data into their data stores while just 12 percent consider themselves good at the key aspects of data flow performance management.
How to Become an Analytics Ready Insurer - with Informatica and HortonworksHortonworks
Whether you are an insurer, reinsurer, broker or insurance service provider; everything you do is based on analytics. From underwriting to claims to agency and marketing, the smartest and most streamlined business operations at insurance companies are driven by advanced and intelligent analytics. But is your data ready? Are you an “Analytics Ready” insurer? Great analytics starts with great data management. Join us as industry experts from Informatica and Hortonworks share industry trends and best practices to show you how to become an “Analytics Ready” insurer.
Fighting Financial Crime with Artificial IntelligenceDataWorks Summit
How can we take the state of the art in deep learning and AI research, and transplant it into a large bank to deliver useful results which impact the general public? To answer this broad-reaching question, we take the viewer through a solution Think Big Analytics recently deployed at a major European bank for fraud detection, using state of the art AI techniques and a near-real time open-source architecture. We show how financial transactions can be transposed into a form where the latest AI techniques in image recognition can be leveraged, in surprisingly novel ways. We have been able to more accurately detect fraud and reduce financial crime, cutting losses and improving customer experience. We describe some architectures which can be used to do this in production, at scale, in global financial institutions.
Speaker:
Tim Seears, Director of Data Science, Think Big Analytics, a Teradata Company
Top 5 Strategies for Retail Data AnalyticsHortonworks
It’s an exciting time for retailers as technology is driving a major disruption in the market. Whether you are just beginning to build a retail data analytics program or you have been gaining advanced insights from your data for quite some time, join Eric and Shish as we explore the trends, drivers and hurdles in retail data analytics
There has been an explosion of data digitising our physical world – from cameras, environmental sensors and embedded devices, right down to the phones in our pockets. Which means that, now, companies have new ways to transform their businesses – both operationally, and through their products and services – by leveraging this data and applying fresh analytical techniques to make sense of it. But are they ready? The answer is “no” in most cases.
In this session, we’ll be discussing the challenges facing companies trying to embrace the Analytics of Things, and how Teradata has helped customers work through and turn those challenges to their advantage.
Hear how Manulife Asia has built an environment that enables the company to solve business-critical problems across many countries. What began in 2017 as an update to their enterprise architecture now spans everything from infrastructure to applications, powering their entire digital backbone. It includes fraud identification, real-time investment dashboards, advanced analytics and machine learning, and digital connection apps that talk to customers for claims, support, and more. Learn the importance hard work, coordination, discipline, and an agile methodology play in deciding which use cases they will focus on to deliver new services in an environment where everything is time sensitive and business requirements shift regularly.
Speaker: Ellen Wu, Head of Asia Data Office, Global Data Enablement and Governance, Manulife
Explore how data integration (or “mashups”) can maximize analytic value and help business teams create streamlined data pipelines that enables ad-hoc analytic inquiries. You’ll learn why businesses increasingly focused on blending data on demand and at the source, the concrete analytic advantages that this approach delivers, and the type of architectures required for delivering trusted, blended data. We provide a checklist to assess your data integration needs and capabilities, and review some real-world examples of how blending various data types has created significant analytic value and concrete business impact.
Hadoop 2.0: YARN to Further Optimize Data ProcessingHortonworks
Data is exponentially increasing in both types and volumes, creating opportunities for businesses. Watch this video and learn from three Big Data experts: John Kreisa, VP Strategic Marketing at Hortonworks, Imad Birouty, Director of Technical Product Marketing at Teradata and John Haddad, Senior Director of Product Marketing at Informatica.
Multiple systems are needed to exploit the variety and volume of data sources, including a flexible data repository. Learn more about:
- Apache Hadoop 2 and YARN
- Data Lakes
- Intelligent data management layers needed to manage metadata and usage patterns as well as track consumption across these data platforms.
Getting started with Hadoop on the Cloud with BluemixNicolas Morales
Silicon Valley Code Camp -- October 11, 2014.
Session: Getting started with Hadoop on the Cloud.
Hadoop and Cloud is an almost perfect marriage. Hadoop is a distributed computing framework that leverages a cluster built on commodity hardware. The Cloud simplifies provisioning of machines and software. Getting started with Hadoop on the Cloud makes it simple to provision your environment quickly and actually get started using Hadoop. IBM Bluemix has democratized Hadoop for the masses! This session will provide a brief introduction to what Hadoop is, how does cloud work and will then focus on how to get started via a series of demos. We will conclude with a discussion around the tutorials and public datasets - all of the tools needed to get you started quickly.
Learn more about BigInsights for Hadoop: https://developer.ibm.com/hadoop/
Is your big data journey stalling? Take the Leap with Capgemini and ClouderaCloudera, Inc.
Transitioning to a Big Data architecture is a big step; and the complexity of moving existing analytical services onto modern platforms like Cloudera, can seem overwhelming.
Apache Hadoop and its role in Big Data architecture - Himanshu Barijaxconf
In today’s world of exponentially growing big data, enterprises are becoming increasingly more aware of the business utility and necessity of harnessing, storing and analyzing this information. Apache Hadoop has rapidly evolved to become a leading platform for managing and processing big data, with the vital management, monitoring, metadata and integration services required by organizations to glean maximum business value and intelligence from their burgeoning amounts of information on customers, web trends, products and competitive markets. In this session, Hortonworks' Himanshu Bari will discuss the opportunities for deriving business value from big data by looking at how organizations utilize Hadoop to store, transform and refine large volumes of this multi-structured information. Connolly will also discuss the evolution of Apache Hadoop and where it is headed, the component requirements of a Hadoop-powered platform, as well as solution architectures that allow for Hadoop integration with existing data discovery and data warehouse platforms. In addition, he will look at real-world use cases where Hadoop has helped to produce more business value, augment productivity or identify new and potentially lucrative opportunities.
How telecommunication companies can leverage power Hadoop and Big Data to derive use cases.
Based on Cloudera Whitepaper - Big Data Use Cases for Telcos
Modern apps and services are leveraging data to change the way we engage with users in a more personalized way. Skyla Loomis talks big data, analytics, NoSQL, SQL and how IBM Cloud is open for data.
Learn more by visiting our Bluemix Hybrid page: http://ibm.co/1PKN23h
MapR on Azure: Getting Value from Big Data in the Cloud -MapR Technologies
Public cloud adoption is exploding and big data technologies are rapidly becoming an important driver of this growth. According to Wikibon, big data public cloud revenue will grow from 4.4% in 2016 to 24% of all big data spend by 2026. Digital transformation initiatives are now a priority for most organizations, with data and advanced analytics at the heart of enabling this change. This is key to driving competitive advantage in every industry.
There is nothing better than a real-world customer use case to help you understand how to get value from big data in the cloud and apply the learnings to your business. Join Microsoft, MapR, and Sullexis on November 10th to:
Hear from Sullexis on the business use case and technical implementation details of one of their oil & gas customers
Understand the integration points of the MapR Platform with other Azure services and why they matter
Know how to deploy the MapR Platform on the Azure cloud and get started easily
You will also get to hear about customer use cases of the MapR Converged Data Platform on Azure in other verticals such as real estate and retail.
Speakers
Rafael Godinho
Technical Evangelist
Microsoft Azure
Tim Morgan
Managing Director
Sullexis
Big Data brings big promise and also big challenges, the primary and most important one being the ability to deliver Value to business stakeholders who are not data scientists!
When and How Data Lakes Fit into a Modern Data ArchitectureDATAVERSITY
Whether to take data ingestion cycles off the ETL tool and the data warehouse or to facilitate competitive Data Science and building algorithms in the organization, the data lake – a place for unmodeled and vast data – will be provisioned widely in 2020.
Though it doesn’t have to be complicated, the data lake has a few key design points that are critical, and it does need to follow some principles for success. Avoid building the data swamp, but not the data lake! The tool ecosystem is building up around the data lake and soon many will have a robust lake and data warehouse. We will discuss policy to keep them straight, send data to its best platform, and keep users’ confidence up in their data platforms.
Data lakes will be built in cloud object storage. We’ll discuss the options there as well.
Get this data point for your data lake journey.
Big Data Everywhere Chicago: Platfora - Practices for Customer Analytics on H...BigDataEverywhere
Hadoop use cases have historically trended towards cost reduction through data warehouse offload. More recently, an uptick around customer-centric use cases have proven the ability for Hadoop to drive top-line revenue. In this session, Platfora solution architect Rob Rosen will discuss how the ability to coreelate multi-structured data in Hadoop leads to greater customer adoption, expanded cross-selling and reduced customer churn for enterprises deploying Hadoop-centric data lakes.
Using real time big data analytics for competitive advantageAmazon Web Services
Many organisations find it challenging to successfully perform real-time data analytics using their own on premise IT infrastructure. Building a system that can adapt and scale rapidly to handle dramatic increases in transaction loads can potentially be quite a costly and time consuming exercise.
Most of the time, infrastructure is under-utilised and it’s near impossible for organisations to forecast the amount of computing power they will need in the future to serve their customers and suppliers.
To overcome these challenges, organisations can instead utilise the cloud to support their real-time data analytics activities. Scalable, agile and secure, cloud-based infrastructure enables organisations to quickly spin up infrastructure to support their data analytics projects exactly when it is needed. Importantly, they can ‘switch off’ infrastructure when it is not.
BluePi Consulting and Amazon Web Services (AWS) are giving you the opportunity to discover how organisations are using real time data analytics to gain new insights from their information to improve the customer experience and drive competitive advantage.
Using AWS to design and build your data architecture has never been easier to gain insights and uncover new opportunities to scale and grow your business. Join this workshop to learn how you can gain insights at scale with the right big data applications.
Similar to Big data solutions on cloud – the way forward (20)
Explore our comprehensive data analysis project presentation on predicting product ad campaign performance. Learn how data-driven insights can optimize your marketing strategies and enhance campaign effectiveness. Perfect for professionals and students looking to understand the power of data analysis in advertising. for more details visit: https://bostoninstituteofanalytics.org/data-science-and-artificial-intelligence/
Techniques to optimize the pagerank algorithm usually fall in two categories. One is to try reducing the work per iteration, and the other is to try reducing the number of iterations. These goals are often at odds with one another. Skipping computation on vertices which have already converged has the potential to save iteration time. Skipping in-identical vertices, with the same in-links, helps reduce duplicate computations and thus could help reduce iteration time. Road networks often have chains which can be short-circuited before pagerank computation to improve performance. Final ranks of chain nodes can be easily calculated. This could reduce both the iteration time, and the number of iterations. If a graph has no dangling nodes, pagerank of each strongly connected component can be computed in topological order. This could help reduce the iteration time, no. of iterations, and also enable multi-iteration concurrency in pagerank computation. The combination of all of the above methods is the STICD algorithm. [sticd] For dynamic graphs, unchanged components whose ranks are unaffected can be skipped altogether.
1. Big Data Solutions on Cloud – the way
forward
By: K. A. Kiththi Perera
Chief Enterprise and Wholesale Officer
Sri Lanka Telecom
ITU-TRCSL Symposium on Cloud Computing 2015
Colombo
Session 04: Big Data Strategy in the Cloud and Applications
2. Big Data Analytics and
Cloud Computing
• Two ICT initiatives are currently top of mind for organizations;
– Big Data Analytics and
– Cloud Computing
• Big Data Analytics offer;
– Valuable insights to create competitive advantage
– Spark new innovations and
– Drive Revenue
• Cloud Computing offer;
– Enhance Business Agility and Productivity
– Enable greater efficiencies and
– Reduce Costs
Both Technologies continue to evolve
6. What’s driving Big Data
- Ad-hoc querying and reporting
- Data mining techniques
- Structured data, typical sources
- Small to mid-size datasets
- Optimizations and predictive analytics
- Complex statistical analysis
- All types of data, and many sources
- Very large datasets
- More of a real-time
7. Value of Big Data Analytics
• Big Data is more real-
time in nature than
traditional DW
applications
• Traditional DW
Architectures (e.g.
Exadata, Teradata) are
not well-suited for big
data apps
• Shared, massively
parallel processing, scale
out architectures are
well-suited for big data
apps
8. “Without big data, you are blind
and deaf in the middle of a
freeway”
Geoffrey Moore, management consultant and theorist
Need to have a high-performance and easy-to-use data
transformation and analytic solution for Big Data
10. Hadoop Functional Blocks
Hive - A high-level language built on top of MapReduce for analyzing large data sets .
Pig - Enables the analysis of large data sets using Pig Latin.
Sqoop - ("SQL to Hadoop") is a Java-based application designed for transferring bulk data between
Apache Hadoop and non-Hadoop data stores
11. Hadoop Core Components
• HDFS – Hadoop Distributed File System (Distributed Storage);
– Distributed across multiple “nodes”
– Natively redundant
– “NameNode” tracks locations
• Map Reduce (Distributed Processing);
– Split a task across processors
– Self-Healing, High Bandwidth
– Clustered Storage
– JobTracker manages TaskTrackers
14. Alternatives to Hadoop
• Many believe that Big Data and Hadoop is the only option
• Hadoop's historic focus on Batch Processing of data was well
supported by ‘MapReduce’
• But there is a need for more flexible developer tool to support;
– The larger market of 'mid-size data sets’ and
– Use cases that call for ‘real-time processing’
• Apache Spark: Preparing for the Next Wave of Reactive Big Data
18. Economics of Cloud Users
Unused resources
• Pay by use instead of provisioning for peak
Static data center Data center in the cloud
Demand
Capacity
Time
Resources
Demand
Capacity
TimeResources
19. Cloud Computing Modalities
• Hosted Applications and services
• Pay-as-you-go model
• Scalability, fault-tolerance,
elasticity, and self-manageability
• Very large data repositories
• Complex analysis
• Distributed and parallel data
processing
“Can we outsource our IT software and
hardware infrastructure?”
“We have terabytes of click-stream data –
what can we do with it?”
EDBT 2011 Tutorial
20. Big Data - Cloud Option
and Challenges
• Key to big data success;
– Elastic Infrastructure and
– Data gravity
• Cloud is emerging as increasingly popular option for new
analytics applications and processing big data
• Challenge - movement of hundreds of terabytes or petabytes
of data across the network
– Traditional data is largely located in Enterprise Data Warehouse
– Limited speed in the WAN
• New data sets – weather data, census data, machine and
sensor data originate from outside the enterprise
– Cloud becomes the ideal place to capture and data processing
Cloud Service Providers to offer “Hadoop/Spark as a service”
bundled with “High Speed Connectivity”
21. SLT “akaza” cloud services
IAAS
Infrastructure
as a Service
SAAS
Software as
a Service
DAAS
Desktop as a
Service
CAAS
Communicati
on as a
Service
PAAS
Platform as a
Service
22. Big Data Use Cases
Optimize Funnel Conversion01
Behavioral Analytics02
Customer Segmentation03
Predictive Support04
Market Analysis and pricing optimization05
Predict Security Threats06
23. Big data analytics allows companies to track
leads through the entire sales conversion
process, from a click on an adword ad to the
final transaction, in order to uncover insights
on how the conversion process can be
improved.
Optimize Funnel Conversion
25. With access to data on consumer behavior,
companies can learn what prompts a customer
to stick around longer as well as learn more
about their customer’s characteristics and
purchasing habits in order to improve
marketing efforts and boost profits.
Behavioral Analytics
26. PURPOSE:
McDonalds tracks vast amounts of data in order to improve operations and
boost the customer experience. The company looks at factors such as the
design of the drive-thru, information provided on the menu, wait times,
size of orders and ordering patterns in order to optimize each restaurant
to its particular market.
Company
McDonald’s
Industry
Food and Beverage
Employees
750,000
Type
Behavioral Analytics
Behavioral Analytics
27. By accessing data about the consumer from
multiple sources, such as social media data
and transaction history, companies can better
segment and target their customers and start
to make personalized offers to those
customers.
Customer Segmentation
29. Through sensors and other machine-generated
data, companies can identify when a
malfunction is likely to occur. The company can
then proactively order parts and make repairs
in order to avoid downtime and lost profits.
Predictive Support