Slides from a presentation I gave at the 5th SOA, Cloud + Service Technology Symposium (September 2012, Imperial College, London). The goal of this presentation was to explore with the audience use cases at the intersection of SOA, Big Data and Fast Data. If you are working with both SOA and Big Data I would would be very interested to hear about your projects.
Oracle's BigData solutions consist of a number of new products and solutions to support customers looking to gain maximum business value from data sets such as weblogs, social media feeds, smart meters, sensors and other devices that generate massive volumes of data (commonly defined as ‘Big Data’) that isn’t readily accessible in enterprise data warehouses and business intelligence applications today.
Strata 2015 presentation from Oracle for Big Data - we are announcing several new big data products including GoldenGate for Big Data, Big Data Discovery, Oracle Big Data SQL and Oracle NoSQL
Expand a Data warehouse with Hadoop and Big Datajdijcks
After investing years in the data warehouse, are you now supposed to start over? Nope. This session discusses how to leverage Hadoop and big data technologies to augment the data warehouse with new data, new capabilities and new business models.
A modern approach to streaming data integration, event processing with a big data (kappa style) data architecture. Key patterns are discussed with pros/cons of newer approaches and open source technologies. Focus on Oracle and GoldenGate technology. OpenWorld 2018 presentation.
Oracle Big Data Discovery working together with Cloudera Hadoop is the fastest way to ingest and understand data. Powerful data transformation capabilities mean that data can quickly be prepared for consumption by the extended organisation.
Oracle's BigData solutions consist of a number of new products and solutions to support customers looking to gain maximum business value from data sets such as weblogs, social media feeds, smart meters, sensors and other devices that generate massive volumes of data (commonly defined as ‘Big Data’) that isn’t readily accessible in enterprise data warehouses and business intelligence applications today.
Strata 2015 presentation from Oracle for Big Data - we are announcing several new big data products including GoldenGate for Big Data, Big Data Discovery, Oracle Big Data SQL and Oracle NoSQL
Expand a Data warehouse with Hadoop and Big Datajdijcks
After investing years in the data warehouse, are you now supposed to start over? Nope. This session discusses how to leverage Hadoop and big data technologies to augment the data warehouse with new data, new capabilities and new business models.
A modern approach to streaming data integration, event processing with a big data (kappa style) data architecture. Key patterns are discussed with pros/cons of newer approaches and open source technologies. Focus on Oracle and GoldenGate technology. OpenWorld 2018 presentation.
Oracle Big Data Discovery working together with Cloudera Hadoop is the fastest way to ingest and understand data. Powerful data transformation capabilities mean that data can quickly be prepared for consumption by the extended organisation.
Hortonworks Oracle Big Data Integration Hortonworks
Slides from joint Hortonworks and Oracle webinar on November 11, 2014. Covers the Modern Data Architecture with Apache Hadoop and Oracle Data Integration products.
The Value of the Modern Data Architecture with Apache Hadoop and Teradata Hortonworks
This webinar discusses why Apache Hadoop most typically the technology underpinning "Big Data". How it fits in a modern data architecture and the current landscape of databases and data warehouses that are already in use.
CON6619 - OpenWorld Presentation. Oracle data integration, big data, data governance, and cloud integration. Replication, ETL, Data Quality, Streaming Big Data, and Data Preparation
Contexti / Oracle - Big Data : From Pilot to ProductionContexti
Big Data is moving from hype to reality for many organisations. The value proposition is clear and sponsorship is high, but how do organisations execute?
Join Oracle and Contexti to discuss the typical journey of a big data project from concept to pilot to production.
• Discuss our experience with a regional Telco
• Common Use Cases across key verticals
• Defining and prioritising use cases
• The challenge of moving from Pilot to Production
• Common Operating Models for Big Data
• Funding a Big Data Capability going forward
• Pilots - common mistakes; challenges; success criteria
Beyond a Big Data Pilot: Building a Production Data Infrastructure - Stampede...StampedeCon
At StampedeCon 2014, Stephen O’Sullivan (Silicon Valley Data Science) presented "Beyond a Big Data Pilot: Building a Production Data Infrastructure."
Creating a data architecture involves many moving parts. By examining the data value chain, from ingestion through to analytics, we will explain how the various parts of the Hadoop and big data ecosystem fit together to support batch, interactive and realtime analytical workloads.
By tracing the flow of data from source to output, we’ll explore the options and considerations for components, including data acquisition, ingestion, storage, data services, analytics and data management. Most importantly, we’ll leave you with a framework for understanding these options and making choices.
Teradata specializes in storing and analyzing structured, relational data. It has recently purchased Aster Data Systems, Inc. in order to extend its platform to include the capability of handling what is often called ‘big’, ‘semi-structured’ or multi-structured (see below) data.
Insurance companies of all sizes are challenged to keep up with emerging technologies that deliver a competitive advantage. Recording: https://www.brighttalk.com/webcast/9573/192877
Big data holds the key to greater customer insight and stronger customer relationships. But risk of sensitive data exposure — and compliance violations — keeps many insurers from pursuing big data initiatives and reaping the rewards of business-driven analytics. Join Dataguise and Hortonworks for this live webinar to learn how you can free your organization from traditional information security constraints and unlock the power of your most valuable business assets.
• What do you need to know about PII/PHI privacy before embarking on big data initiatives?
• Why do so many big data initiatives fail before they’ve even begun—and what can you do about it?
• How can IT security organizations help data scientists extract more business value from their data?
• How are leading insurance companies leveraging big data to gain competitive advantage?
GoldenGate and Stream Processing with Special Guest RakutenJeffrey T. Pollock
Oracle OpenWorld roadmap presentation for GoldenGate, stream processing, analytics and big data use cases with special guest presenters from Rakuten Travel.
The Next Generation of Big Data AnalyticsHortonworks
Apache Hadoop has evolved rapidly to become a leading platform for managing and processing big data. If your organization is examining how you can use Hadoop to store, transform, and refine large volumes of multi-structured data, please join us for this session where we will discuss, the emergence of "big data" and opportunities for deriving business value, the evolution of Apache Hadoop and future directions, essential components required in a Hadoop-powered platform, and solution architectures that integrate Hadoop with existing data discovery and data warehouse platforms.
Talend Open Studio and Hortonworks Data PlatformHortonworks
Data Integration is a key step in a Hadoop solution architecture. It is the first obstacle encountered once your cluster is up and running. OK, I have a cluster…now what? Complex scripts? For wide scale adoption of Apache Hadoop, an intuitive set of tools that abstract away the complexity of integration is necessary.
Accelerating the Value of Big Data Analytics for P&C Insurers with Hortonwork...Hortonworks
As the Big Data Analytics and the Apache Hadoop ecosystem has matured and gained increasing traction in established industries with faster adoption in the insurance market than originally anticipated, it is clear that the potential benefits for data management and business intelligence are staggering. At the same time, many big data programs have stalled or failed to deliver on their aspirational value proposition, resulting in a substantial gap between expectations of analytics consumers and the ability of big data analytics programs to deliver. Join Hortonworks and Clarity as we review the common needs of Property and Casualty (P&C) Insurers and how to unlock the true value of big data analytics:
Information agility – Centralization of data and decentralization of analysis
Expanded capability – Conventional analysis combined with real-time analytics demands
Reduced expense – Lower costs through cheaper storage while maintaining scalability
We will discuss a modern data architecture that constitutes a mature, enterprise strength Hadoop framework for P&C Insurers that answers the need for governance processes across the enterprise stack. We will cover how a modern data architecture allows organizations to collect, store, analyze and manipulate massive quantities of data on their own terms—regardless of the source of that data - accelerating the real lifetime value of big data and Hadoop analytics for claims, customer sentiment and telematics.
Data Governance, Compliance and Security in Hadoop with ClouderaCaserta
In our recent Big Data Warehousing Meetup, we discussed Data Governance, Compliance and Security in Hadoop.
As the Big Data paradigm becomes more commonplace, we must apply enterprise-grade governance capabilities for critical data that is highly regulated and adhere to stringent compliance requirements. Caserta and Cloudera shared techniques and tools that enables data governance, compliance and security on Big Data.
For more information, visit www.casertaconcepts.com
Oracle has been working hard for several years in building Oracle Fusion Applications which are slated to be released sometime during 2010. In this session, you will learn basic concepts of Fusion Applications, User Experience/UI Shell, features and functionality of Applciations. Present information about the Oracle Fusion Applications Present key concepts and ideas behind Fusion Applications Discuss the key technologies used by Oracle for Fusion Applications
Hortonworks Oracle Big Data Integration Hortonworks
Slides from joint Hortonworks and Oracle webinar on November 11, 2014. Covers the Modern Data Architecture with Apache Hadoop and Oracle Data Integration products.
The Value of the Modern Data Architecture with Apache Hadoop and Teradata Hortonworks
This webinar discusses why Apache Hadoop most typically the technology underpinning "Big Data". How it fits in a modern data architecture and the current landscape of databases and data warehouses that are already in use.
CON6619 - OpenWorld Presentation. Oracle data integration, big data, data governance, and cloud integration. Replication, ETL, Data Quality, Streaming Big Data, and Data Preparation
Contexti / Oracle - Big Data : From Pilot to ProductionContexti
Big Data is moving from hype to reality for many organisations. The value proposition is clear and sponsorship is high, but how do organisations execute?
Join Oracle and Contexti to discuss the typical journey of a big data project from concept to pilot to production.
• Discuss our experience with a regional Telco
• Common Use Cases across key verticals
• Defining and prioritising use cases
• The challenge of moving from Pilot to Production
• Common Operating Models for Big Data
• Funding a Big Data Capability going forward
• Pilots - common mistakes; challenges; success criteria
Beyond a Big Data Pilot: Building a Production Data Infrastructure - Stampede...StampedeCon
At StampedeCon 2014, Stephen O’Sullivan (Silicon Valley Data Science) presented "Beyond a Big Data Pilot: Building a Production Data Infrastructure."
Creating a data architecture involves many moving parts. By examining the data value chain, from ingestion through to analytics, we will explain how the various parts of the Hadoop and big data ecosystem fit together to support batch, interactive and realtime analytical workloads.
By tracing the flow of data from source to output, we’ll explore the options and considerations for components, including data acquisition, ingestion, storage, data services, analytics and data management. Most importantly, we’ll leave you with a framework for understanding these options and making choices.
Teradata specializes in storing and analyzing structured, relational data. It has recently purchased Aster Data Systems, Inc. in order to extend its platform to include the capability of handling what is often called ‘big’, ‘semi-structured’ or multi-structured (see below) data.
Insurance companies of all sizes are challenged to keep up with emerging technologies that deliver a competitive advantage. Recording: https://www.brighttalk.com/webcast/9573/192877
Big data holds the key to greater customer insight and stronger customer relationships. But risk of sensitive data exposure — and compliance violations — keeps many insurers from pursuing big data initiatives and reaping the rewards of business-driven analytics. Join Dataguise and Hortonworks for this live webinar to learn how you can free your organization from traditional information security constraints and unlock the power of your most valuable business assets.
• What do you need to know about PII/PHI privacy before embarking on big data initiatives?
• Why do so many big data initiatives fail before they’ve even begun—and what can you do about it?
• How can IT security organizations help data scientists extract more business value from their data?
• How are leading insurance companies leveraging big data to gain competitive advantage?
GoldenGate and Stream Processing with Special Guest RakutenJeffrey T. Pollock
Oracle OpenWorld roadmap presentation for GoldenGate, stream processing, analytics and big data use cases with special guest presenters from Rakuten Travel.
The Next Generation of Big Data AnalyticsHortonworks
Apache Hadoop has evolved rapidly to become a leading platform for managing and processing big data. If your organization is examining how you can use Hadoop to store, transform, and refine large volumes of multi-structured data, please join us for this session where we will discuss, the emergence of "big data" and opportunities for deriving business value, the evolution of Apache Hadoop and future directions, essential components required in a Hadoop-powered platform, and solution architectures that integrate Hadoop with existing data discovery and data warehouse platforms.
Talend Open Studio and Hortonworks Data PlatformHortonworks
Data Integration is a key step in a Hadoop solution architecture. It is the first obstacle encountered once your cluster is up and running. OK, I have a cluster…now what? Complex scripts? For wide scale adoption of Apache Hadoop, an intuitive set of tools that abstract away the complexity of integration is necessary.
Accelerating the Value of Big Data Analytics for P&C Insurers with Hortonwork...Hortonworks
As the Big Data Analytics and the Apache Hadoop ecosystem has matured and gained increasing traction in established industries with faster adoption in the insurance market than originally anticipated, it is clear that the potential benefits for data management and business intelligence are staggering. At the same time, many big data programs have stalled or failed to deliver on their aspirational value proposition, resulting in a substantial gap between expectations of analytics consumers and the ability of big data analytics programs to deliver. Join Hortonworks and Clarity as we review the common needs of Property and Casualty (P&C) Insurers and how to unlock the true value of big data analytics:
Information agility – Centralization of data and decentralization of analysis
Expanded capability – Conventional analysis combined with real-time analytics demands
Reduced expense – Lower costs through cheaper storage while maintaining scalability
We will discuss a modern data architecture that constitutes a mature, enterprise strength Hadoop framework for P&C Insurers that answers the need for governance processes across the enterprise stack. We will cover how a modern data architecture allows organizations to collect, store, analyze and manipulate massive quantities of data on their own terms—regardless of the source of that data - accelerating the real lifetime value of big data and Hadoop analytics for claims, customer sentiment and telematics.
Data Governance, Compliance and Security in Hadoop with ClouderaCaserta
In our recent Big Data Warehousing Meetup, we discussed Data Governance, Compliance and Security in Hadoop.
As the Big Data paradigm becomes more commonplace, we must apply enterprise-grade governance capabilities for critical data that is highly regulated and adhere to stringent compliance requirements. Caserta and Cloudera shared techniques and tools that enables data governance, compliance and security on Big Data.
For more information, visit www.casertaconcepts.com
Oracle has been working hard for several years in building Oracle Fusion Applications which are slated to be released sometime during 2010. In this session, you will learn basic concepts of Fusion Applications, User Experience/UI Shell, features and functionality of Applciations. Present information about the Oracle Fusion Applications Present key concepts and ideas behind Fusion Applications Discuss the key technologies used by Oracle for Fusion Applications
Governance as Sustainability in the Enterprise Architecture Discipline Eric Stephens
In this talk today I will speak about the fundamental need for EA governance in order to effectively manage ICT investments. I will then talk about recent advances in ICT that require a fresh look from a governance perspective. I will then describe some of the key characteristics of effective EA governance. Finally, I’ll talk about how to instill a governance-driven Technology Adoption process.
Executive Summary Posted with TOG
Corporations are being bombarded with a number of new technologies promising seemingly magical fixes to traditional computing and managerial problems. Technologies such as SOA, cloud computing, and mobility typically be aligned with corporate objectives and realize new enterprise capabilities. However, they are also subject to architectural abuse - especially as they become easier to use (e.g., Ten minutes to spin up an Amazon AWS instance).
With each new technological innovation, it is critical to re-evaluate an EA governance program to ensure the necessary processes, controls, and standards are in place. Otherwise the architectural fragmentation that exists with legacy technology will manifest itself by an order of magnitude with newer generations of technology. And corporate objectives will be missed and enterprise capabilities unrealized.
All Grown Up: Maturation of Analytics in the CloudInside Analysis
The Briefing Room with Wayne Eckerson and Birst
Live Webcast on Nov. 6, 2012
The desire for analytics today extends far beyond the traditional domain of Business Intelligence. The challenge is that operational systems come in countless shapes and sizes. Furthermore, each application treats data somewhat differently. But there are patterns of data flow and transformation that pervade all such systems. And there's one big place where all these data types and use cases have come together architecturally: the Cloud.
Watch this episode of the Briefing Room to hear veteran Analyst Wayne Eckerson explain how Cloud computing is ushering in a new era of analytics and intelligence. He'll be briefed by Brad Peters of Birst who will tout his company's purpose-built analytics platform. He'll discuss how the Birst engine processes and delivers raw data from disparate systems, offering the deployment flexibility of Software-as-a-Service, together with the capabilities of enterprise-class BI.
Watch full webinar here: https://bit.ly/2SaBj5l
You will often hear that "data is the new gold". In this context, data management is one of the areas that has received more attention by the software community in recent years. From Artificial Intelligence and Machine Learning to new ways to store and process data, the landscape for data management is in constant evolution. From the privileged perspective of an enterprise middleware platform, we at Denodo have the advantage of seeing many of these changes happen.
In this webinar we will discuss the technology trends that will drive the enterprise data strategies in the years to come. Don't miss it if you want to keep yourself informed about how to convert your data to strategic assets in order to complete the data-driven transformation in your company.
Join us for an exciting session that will cover:
- The most interesting trends in data management
- How to build a logical data fabric architecture?
- How to manage your data integration strategy in the new hybrid world?
- Our predictions on how those trends will change the data management world
- How can companies monetize the data through data-as-a-service infrastructure?
- What is the role of the voice computing in the future of data analytics?
Self-Service Access and Exploration of Big DataInside Analysis
The Briefing Room with Robin Bloor and Cirro
Live Webcast on Dec. 11, 2012
As the information landscape expands with all kinds of Big Data, businesses are searching for ways to unite their traditional analytics with this new source of insight. One ambitious approach involves federating access to multiple data sources, even across various operating systems. The idea is to take analytic processing to the data, then intelligently assemble the results for a business user. Could this be the long-awaited alternative to data virtualization?
Check out this episode of The Briefing Room to hear veteran Analyst Robin Bloor explain how federated access to data sources can pave the way for a truly integrated data fabric. Bloor will be briefed by Mark Theissen of Cirro, who will tout his company's patent-pending Data Hub, which simplifies data access by federating queries across multiple sources of structured, semi-structured, and unstructured data. He'll discuss Cirro's cost based optimizer, smart caching, dynamic query plan re-optimization, normalization of cost estimates and a metadata repository for unstructured data sources.
Visit: http://www.insideanalysis.com
All kinds of data Large volumes Valuable insight, but difficult to extract (structured and unstructured data) Often extremely time sensitive Most of the vast data types portrayed here is consumer data and while the business will want to leverage Oracle Event Processing for business and application data, they are also impacted by this consumer data and information from the vast array or sensors where stream events showing temperatures in a container mid-pacific may destroy high cost food goods unless immediate action is taken or ….. For Starbucks immediately analyzing tweats after launching a new coffee, seeing spikes of negative comments, and very quickly figuring out that the negative reactions came from stores that were serving a particular warmed cheese sandwich, whose aroma did not go with the new coffee smell….. Huge ROI due to quick analysis and specific targeted response. And as you can see from the Spanish (La Caxia) bank solution, a customers Tweets are also being analzed by Oracle Event Processing and stored in Big Data to augment his preferences and influence his/her real time targetted campaigns
Scripting languages supported via Hadoop Streaming, equivalent to Unix streaming
Facebook, Google, Netflix, etc.Hadron Collider, NSF, etc.
Being able to preserve info over long term (without copy/filtering) could be very interesting for historical analysis, shipping & process optimization
SmartMeter example: want all data to do in-depth energy usage analysis but also want real-time analysis for things like leak detection.