The blessing and curse of today's database market? So many choices! While relational databases still dominate the day-to-day business, a host of alternatives has evolved around very specific use cases: graph, document, NoSQL, hybrid (HTAP), column store, the list goes on. And the database tools market is teeming with activity as well. Register for this special Research Webcast to hear Dr. Robin Bloor share his early findings about the evolving database market. He'll be joined by Steve Sarsfield of HPE Vertica, and Robert Reeves of Datical in a roundtable discussion with Bloor Group CEO Eric Kavanagh. Send any questions to info@insideanalysis.com, or tweet with #DBSurvival.
The Model Enterprise: A Blueprint for Enterprise Data GovernanceEric Kavanagh
What gets measured, gets managed; but what gets governed, generates real value. That's one major reason why data governance has risen to a top priority for most organizations. Another reason is the rapid onboarding of big data, which often comes from beyond the traditional firewall. And then there are the authorities: issues like privacy, security and fiduciary responsibility are combining to make data governance a must-have. Register for this episode of The Briefing Room to hear veteran Analyst Dr. Robin Bloor explain why governance should be viewed as a positive change agent for the modern enterprise. He'll be briefed by Ron Huizenga of IDERA, who will discuss a practical, model-based approach to enterprise data governance, with a focus on Master Data Management.
Caserta Concepts, Datameer and Microsoft shared their combined knowledge and a use case on big data, the cloud and deep analytics. Attendes learned how a global leader in the test, measurement and control systems market reduced their big data implementations from 18 months to just a few.
Speakers shared how to provide a business user-friendly, self-service environment for data discovery and analytics, and focus on how to extend and optimize Hadoop based analytics, highlighting the advantages and practical applications of deploying on the cloud for enhanced performance, scalability and lower TCO.
Agenda included:
- Pizza and Networking
- Joe Caserta, President, Caserta Concepts - Why are we here?
- Nikhil Kumar, Sr. Solutions Engineer, Datameer - Solution use cases and technical demonstration
- Stefan Groschupf, CEO & Chairman, Datameer - The evolving Hadoop-based analytics trends and the role of cloud computing
- James Serra, Data Platform Solution Architect, Microsoft, Benefits of the Azure Cloud Service
- Q&A, Networking
For more information on Caserta Concepts, visit our website: http://casertaconcepts.com/
The Future of Data Warehousing and Data IntegrationEric Kavanagh
The rise of big data, data lakes and the cloud, coupled with increasingly stringent enterprise requirements, are reinventing the role of data warehousing in modern analytics ecosystems. The emerging generation of data warehouses is more flexible, agile and cloud-based than their predecessors, with a strong need for automation and real-time data integration.
Join this live webinar to learn:
-Typical requirements for data integration
-Common use cases and architectural patterns
-Guidelines and best practices to address data requirements
-Guidelines and best practices to apply architectural patterns
The Data Lake - Balancing Data Governance and Innovation Caserta
Joe Caserta gave the presentation "The Data Lake - Balancing Data Governance and Innovation" at DAMA NY's one day mini-conference on May 19th. Speakers covered emerging trends in Data Governance, especially around Big Data.
For more information on Caserta Concepts, visit our website at http://casertaconcepts.com/.
Better Together: The New Data Management OrchestraCloudera, Inc.
To ingest, store, process and leverage big data for maximum business impact requires integrating systems, processing frameworks, and analytic deployment options. Learn how Cloudera’s enterprise data hub framework, MongoDB, and Teradata Data Warehouse working in concert can enable companies to explore data in new ways and solve problems that not long ago might have seemed impossible.
Gone are the days of NoSQL and SQL competing for center stage. Visionary companies are driving data subsystems to operate in harmony. So what’s changed?
In this webinar, you will hear from executives at Cloudera, Teradata and MongoDB about the following:
How to deploy the right mix of tools and technology to become a data-driven organization
Examples of three major data management systems working together
Real world examples of how business and IT are benefiting from the sum of the parts
Join industry leaders Charles Zedlewski, Chris Twogood and Kelly Stirman for this unique panel discussion, moderated by BI Research analyst, Colin White.
The Model Enterprise: A Blueprint for Enterprise Data GovernanceEric Kavanagh
What gets measured, gets managed; but what gets governed, generates real value. That's one major reason why data governance has risen to a top priority for most organizations. Another reason is the rapid onboarding of big data, which often comes from beyond the traditional firewall. And then there are the authorities: issues like privacy, security and fiduciary responsibility are combining to make data governance a must-have. Register for this episode of The Briefing Room to hear veteran Analyst Dr. Robin Bloor explain why governance should be viewed as a positive change agent for the modern enterprise. He'll be briefed by Ron Huizenga of IDERA, who will discuss a practical, model-based approach to enterprise data governance, with a focus on Master Data Management.
Caserta Concepts, Datameer and Microsoft shared their combined knowledge and a use case on big data, the cloud and deep analytics. Attendes learned how a global leader in the test, measurement and control systems market reduced their big data implementations from 18 months to just a few.
Speakers shared how to provide a business user-friendly, self-service environment for data discovery and analytics, and focus on how to extend and optimize Hadoop based analytics, highlighting the advantages and practical applications of deploying on the cloud for enhanced performance, scalability and lower TCO.
Agenda included:
- Pizza and Networking
- Joe Caserta, President, Caserta Concepts - Why are we here?
- Nikhil Kumar, Sr. Solutions Engineer, Datameer - Solution use cases and technical demonstration
- Stefan Groschupf, CEO & Chairman, Datameer - The evolving Hadoop-based analytics trends and the role of cloud computing
- James Serra, Data Platform Solution Architect, Microsoft, Benefits of the Azure Cloud Service
- Q&A, Networking
For more information on Caserta Concepts, visit our website: http://casertaconcepts.com/
The Future of Data Warehousing and Data IntegrationEric Kavanagh
The rise of big data, data lakes and the cloud, coupled with increasingly stringent enterprise requirements, are reinventing the role of data warehousing in modern analytics ecosystems. The emerging generation of data warehouses is more flexible, agile and cloud-based than their predecessors, with a strong need for automation and real-time data integration.
Join this live webinar to learn:
-Typical requirements for data integration
-Common use cases and architectural patterns
-Guidelines and best practices to address data requirements
-Guidelines and best practices to apply architectural patterns
The Data Lake - Balancing Data Governance and Innovation Caserta
Joe Caserta gave the presentation "The Data Lake - Balancing Data Governance and Innovation" at DAMA NY's one day mini-conference on May 19th. Speakers covered emerging trends in Data Governance, especially around Big Data.
For more information on Caserta Concepts, visit our website at http://casertaconcepts.com/.
Better Together: The New Data Management OrchestraCloudera, Inc.
To ingest, store, process and leverage big data for maximum business impact requires integrating systems, processing frameworks, and analytic deployment options. Learn how Cloudera’s enterprise data hub framework, MongoDB, and Teradata Data Warehouse working in concert can enable companies to explore data in new ways and solve problems that not long ago might have seemed impossible.
Gone are the days of NoSQL and SQL competing for center stage. Visionary companies are driving data subsystems to operate in harmony. So what’s changed?
In this webinar, you will hear from executives at Cloudera, Teradata and MongoDB about the following:
How to deploy the right mix of tools and technology to become a data-driven organization
Examples of three major data management systems working together
Real world examples of how business and IT are benefiting from the sum of the parts
Join industry leaders Charles Zedlewski, Chris Twogood and Kelly Stirman for this unique panel discussion, moderated by BI Research analyst, Colin White.
Joe Caserta was a featured speaker, along with MIT Sloan School faculty and other industry thought-leaders. His session 'You're the New CDO, Now What?' discussed how new CDOs can accomplish their strategic objectives and overcome tactical challenges in this emerging executive leadership role.
In its tenth year, the MIT CDOIQ Symposium 2016 continues to explore the developing role of the Chief Data Officer.
For more information, visit http://casertaconcepts.com/
If you also got the Big Data itch, here is something to ease the pain :-)
Answers to this questions will be available soon (more info in the attached link)
Which Big Data Appliance should YOU use?
(click on the attached link for Poll results)
Appliances are Small and Quick, Right?
Revealing the 6 Types of Big Data Appliances
Uncovering the Main Players
Challenges, Pitfalls, and Winning the Big Data Game
Where is all this leading YOU to?
In this presentation at DAMA New York, Joe started by asking a key question: why are we doing this? Why analyze and share all these massive amounts of data? Basically, it comes down to the belief that in any organization, in any situation, if we can get the data and make it correct and timely, insights from it will become instantly actionable for companies to function more nimbly and successfully. Enabling the use of data can be a world-changing, world-improving activity and this session presents the steps necessary to get you there. Joe explained the concept of the "data lake" and also emphasizes the role of a strong data governance strategy that incorporates seven components needed for a successful program.
For more information on this presentation or Caserta Concepts, visit our website at http://casertaconcepts.com/.
Balance agility and governance with #TrueDataOps and The Data CloudKent Graziano
DataOps is the application of DevOps concepts to data. The DataOps Manifesto outlines WHAT that means, similar to how the Agile Manifesto outlines the goals of the Agile Software movement. But, as the demand for data governance has increased, and the demand to do “more with less” and be more agile has put more pressure on data teams, we all need more guidance on HOW to manage all this. Seeing that need, a small group of industry thought leaders and practitioners got together and created the #TrueDataOps philosophy to describe the best way to deliver DataOps by defining the core pillars that must underpin a successful approach. Combining this approach with an agile and governed platform like Snowflake’s Data Cloud allows organizations to indeed balance these seemingly competing goals while still delivering value at scale.
Given in Montreal on 14-Dec-2021
Data Governance, Compliance and Security in Hadoop with ClouderaCaserta
In our recent Big Data Warehousing Meetup, we discussed Data Governance, Compliance and Security in Hadoop.
As the Big Data paradigm becomes more commonplace, we must apply enterprise-grade governance capabilities for critical data that is highly regulated and adhere to stringent compliance requirements. Caserta and Cloudera shared techniques and tools that enables data governance, compliance and security on Big Data.
For more information, visit www.casertaconcepts.com
Modern Data Management for Federal ModernizationDenodo
Watch full webinar here: https://bit.ly/2QaVfE7
Faster, more agile data management is at the heart of government modernization. However, Traditional data delivery systems are limited in realizing a modernized and future-proof data architecture.
This webinar will address how data virtualization can modernize existing systems and enable new data strategies. Join this session to learn how government agencies can use data virtualization to:
- Enable governed, inter-agency data sharing
- Simplify data acquisition, search and tagging
- Streamline data delivery for transition to cloud, data science initiatives, and more
How to Optimize Sales Analytics Using 10x the Data at 1/10th the CostAtScale
Being able to analyze sales at the most granular level with up-to-date data, provides a competitive advantage for unlocking additional revenue -- especially for e-commerce and retail companies heading into the holiday season.
It Takes a Village: Organizational Alignment to Deliver Big Data Value in Hea...DataWorks Summit
The business and technology teams within a health insurer must align the company’s central data platform with its data strategy. That requires substantial organizational alignment. Hear the firsthand perspective from Health Care Service Corporation (HCSC), the largest customer-owned health insurance company in the United States. The speaker will cover how they integrated membership information, regulatory compliance, and the general ledger, to improve overall healthcare management. At HCSC, the strong alignment between executive leadership, business portfolio direction, architectural strategy, technology delivery, and program management have helped create leading-edge capabilities which help the company respond nimbly to a quickly evolving healthcare industry.
In this document, we will present a very brief introduction to BigData (what is BigData?), Hadoop (how does Hadoop fits the picture?) and Cloudera Hadoop (what is the difference between Cloudera Hadoop and regular Hadoop?).
Please note that this document is for Hadoop beginners looking for a place to start.
Big Data International Keynote Speaker Mark van Rijmenam shared his vision on Hadoop Data Lakes during a Zaloni Webinar. What are the Hadoop Data Lake trends for 2016, what are the data lake challenges and how can organizations benefit from data lakes.
Cloudera Federal Forum 2014: The Building Blocks of the Enterprise Data HubCloudera, Inc.
Chief Technologist, Office of the CTO at Cloudera Eli Collins, shares the story of the enterprise data hub and how it relates to the enterprise data warehouse.
Couchbase and Apache Kafka - Bridging the gap between RDBMS and NoSQLDATAVERSITY
Thousands of companies, from Uber and Netflix to Goldman Sachs and Cisco, use Apache Kafka to transform and reshape their data architectures. Kafka is frequently used as the bridge between legacy RDBMS and new NoSQL database systems, effectively transforming SQL table data into JSON documents and vice versa. Many companies also use Kafka for business-critical applications that drive real-time stream processing and analytics, intersystem messaging, high-volume data ingestion, and operational metrics collection.
Couchbase and Kafka can be used together to address high throughput, distributed data management, and transformation challenges.
In this webinar we’ll explore:
Where Kafka fits into the big data ecosystem
How companies are using Kafka for both real-time processing and as a bus for data exchange
An example of how Kafka can bridge legacy RDBMS and new NoSQL database systems
Several real-world use case architectures
Data volumes have experienced explosive growth in recent years, and that data is being generated from sources that are increasingly complex and varied. Harnessing and refining value from this data requires a new approach as data extraction, transformation, and loading (ETL) becoming increasingly more costly and difficult to scale.
Organizations are looking to leverage Hadoop as an enterprise data hub—also called a “data lake” or “data reservoir”—as a key component of their data architecture to augment their data warehouse, ETL and analytical systems in order to maximize their existing investments, reduce costs, and unlock new business value from their data.
In this webinar, you will learn:
Real-world examples that illustrate why Hadoop is the best low-cost data hub, data lake, or data landing zone (staging area) option for ETL processing
Proof points that demonstrate advantages of Hadoop and its ability to scale to manage increasing data volumes and support exploratory big data analytics
Proven best practices for a cost-effective, reliable way to implement a data management platform for your entire big data analytical ecosystem
Hidden issues to be aware of in deploying your data hub/data lake
Webinar - Delivering Enhanced Message Processing at Scale With an Always-on D...DataStax
Managing 3.8 million e-prescriptions daily for more than 1 million healthcare professionals is no small feat. And, with rapid growth in the number of digital transactions and expansion of its network, Surescripts needed to replace its legacy relational database system to address a new set of data management challenges while meeting their customers’ demanding SLAs. Join us for this on-demand webinar to hear from Keith Willard, Chief Architect at Surescripts, to learn how and why Surescripts leverages DataStax Enterprise to deliver enhanced message processing at scale.
View recording: https://youtu.be/1T6V1XAoaJQ
Explore all DataStax webinars: https://www.datastax.com/resources/webinars
Simplifying Real-Time Architectures for IoT with Apache KuduCloudera, Inc.
3 Things to Learn About:
*Building scalable real time architectures for managing data from IoT
*Processing data in real time with components such as Kudu & Spark
*Customer case studies highlighting real-time IoT use cases
Using real time big data analytics for competitive advantageAmazon Web Services
Many organisations find it challenging to successfully perform real-time data analytics using their own on premise IT infrastructure. Building a system that can adapt and scale rapidly to handle dramatic increases in transaction loads can potentially be quite a costly and time consuming exercise.
Most of the time, infrastructure is under-utilised and it’s near impossible for organisations to forecast the amount of computing power they will need in the future to serve their customers and suppliers.
To overcome these challenges, organisations can instead utilise the cloud to support their real-time data analytics activities. Scalable, agile and secure, cloud-based infrastructure enables organisations to quickly spin up infrastructure to support their data analytics projects exactly when it is needed. Importantly, they can ‘switch off’ infrastructure when it is not.
BluePi Consulting and Amazon Web Services (AWS) are giving you the opportunity to discover how organisations are using real time data analytics to gain new insights from their information to improve the customer experience and drive competitive advantage.
Joe Caserta was a featured speaker, along with MIT Sloan School faculty and other industry thought-leaders. His session 'You're the New CDO, Now What?' discussed how new CDOs can accomplish their strategic objectives and overcome tactical challenges in this emerging executive leadership role.
In its tenth year, the MIT CDOIQ Symposium 2016 continues to explore the developing role of the Chief Data Officer.
For more information, visit http://casertaconcepts.com/
If you also got the Big Data itch, here is something to ease the pain :-)
Answers to this questions will be available soon (more info in the attached link)
Which Big Data Appliance should YOU use?
(click on the attached link for Poll results)
Appliances are Small and Quick, Right?
Revealing the 6 Types of Big Data Appliances
Uncovering the Main Players
Challenges, Pitfalls, and Winning the Big Data Game
Where is all this leading YOU to?
In this presentation at DAMA New York, Joe started by asking a key question: why are we doing this? Why analyze and share all these massive amounts of data? Basically, it comes down to the belief that in any organization, in any situation, if we can get the data and make it correct and timely, insights from it will become instantly actionable for companies to function more nimbly and successfully. Enabling the use of data can be a world-changing, world-improving activity and this session presents the steps necessary to get you there. Joe explained the concept of the "data lake" and also emphasizes the role of a strong data governance strategy that incorporates seven components needed for a successful program.
For more information on this presentation or Caserta Concepts, visit our website at http://casertaconcepts.com/.
Balance agility and governance with #TrueDataOps and The Data CloudKent Graziano
DataOps is the application of DevOps concepts to data. The DataOps Manifesto outlines WHAT that means, similar to how the Agile Manifesto outlines the goals of the Agile Software movement. But, as the demand for data governance has increased, and the demand to do “more with less” and be more agile has put more pressure on data teams, we all need more guidance on HOW to manage all this. Seeing that need, a small group of industry thought leaders and practitioners got together and created the #TrueDataOps philosophy to describe the best way to deliver DataOps by defining the core pillars that must underpin a successful approach. Combining this approach with an agile and governed platform like Snowflake’s Data Cloud allows organizations to indeed balance these seemingly competing goals while still delivering value at scale.
Given in Montreal on 14-Dec-2021
Data Governance, Compliance and Security in Hadoop with ClouderaCaserta
In our recent Big Data Warehousing Meetup, we discussed Data Governance, Compliance and Security in Hadoop.
As the Big Data paradigm becomes more commonplace, we must apply enterprise-grade governance capabilities for critical data that is highly regulated and adhere to stringent compliance requirements. Caserta and Cloudera shared techniques and tools that enables data governance, compliance and security on Big Data.
For more information, visit www.casertaconcepts.com
Modern Data Management for Federal ModernizationDenodo
Watch full webinar here: https://bit.ly/2QaVfE7
Faster, more agile data management is at the heart of government modernization. However, Traditional data delivery systems are limited in realizing a modernized and future-proof data architecture.
This webinar will address how data virtualization can modernize existing systems and enable new data strategies. Join this session to learn how government agencies can use data virtualization to:
- Enable governed, inter-agency data sharing
- Simplify data acquisition, search and tagging
- Streamline data delivery for transition to cloud, data science initiatives, and more
How to Optimize Sales Analytics Using 10x the Data at 1/10th the CostAtScale
Being able to analyze sales at the most granular level with up-to-date data, provides a competitive advantage for unlocking additional revenue -- especially for e-commerce and retail companies heading into the holiday season.
It Takes a Village: Organizational Alignment to Deliver Big Data Value in Hea...DataWorks Summit
The business and technology teams within a health insurer must align the company’s central data platform with its data strategy. That requires substantial organizational alignment. Hear the firsthand perspective from Health Care Service Corporation (HCSC), the largest customer-owned health insurance company in the United States. The speaker will cover how they integrated membership information, regulatory compliance, and the general ledger, to improve overall healthcare management. At HCSC, the strong alignment between executive leadership, business portfolio direction, architectural strategy, technology delivery, and program management have helped create leading-edge capabilities which help the company respond nimbly to a quickly evolving healthcare industry.
In this document, we will present a very brief introduction to BigData (what is BigData?), Hadoop (how does Hadoop fits the picture?) and Cloudera Hadoop (what is the difference between Cloudera Hadoop and regular Hadoop?).
Please note that this document is for Hadoop beginners looking for a place to start.
Big Data International Keynote Speaker Mark van Rijmenam shared his vision on Hadoop Data Lakes during a Zaloni Webinar. What are the Hadoop Data Lake trends for 2016, what are the data lake challenges and how can organizations benefit from data lakes.
Cloudera Federal Forum 2014: The Building Blocks of the Enterprise Data HubCloudera, Inc.
Chief Technologist, Office of the CTO at Cloudera Eli Collins, shares the story of the enterprise data hub and how it relates to the enterprise data warehouse.
Couchbase and Apache Kafka - Bridging the gap between RDBMS and NoSQLDATAVERSITY
Thousands of companies, from Uber and Netflix to Goldman Sachs and Cisco, use Apache Kafka to transform and reshape their data architectures. Kafka is frequently used as the bridge between legacy RDBMS and new NoSQL database systems, effectively transforming SQL table data into JSON documents and vice versa. Many companies also use Kafka for business-critical applications that drive real-time stream processing and analytics, intersystem messaging, high-volume data ingestion, and operational metrics collection.
Couchbase and Kafka can be used together to address high throughput, distributed data management, and transformation challenges.
In this webinar we’ll explore:
Where Kafka fits into the big data ecosystem
How companies are using Kafka for both real-time processing and as a bus for data exchange
An example of how Kafka can bridge legacy RDBMS and new NoSQL database systems
Several real-world use case architectures
Data volumes have experienced explosive growth in recent years, and that data is being generated from sources that are increasingly complex and varied. Harnessing and refining value from this data requires a new approach as data extraction, transformation, and loading (ETL) becoming increasingly more costly and difficult to scale.
Organizations are looking to leverage Hadoop as an enterprise data hub—also called a “data lake” or “data reservoir”—as a key component of their data architecture to augment their data warehouse, ETL and analytical systems in order to maximize their existing investments, reduce costs, and unlock new business value from their data.
In this webinar, you will learn:
Real-world examples that illustrate why Hadoop is the best low-cost data hub, data lake, or data landing zone (staging area) option for ETL processing
Proof points that demonstrate advantages of Hadoop and its ability to scale to manage increasing data volumes and support exploratory big data analytics
Proven best practices for a cost-effective, reliable way to implement a data management platform for your entire big data analytical ecosystem
Hidden issues to be aware of in deploying your data hub/data lake
Webinar - Delivering Enhanced Message Processing at Scale With an Always-on D...DataStax
Managing 3.8 million e-prescriptions daily for more than 1 million healthcare professionals is no small feat. And, with rapid growth in the number of digital transactions and expansion of its network, Surescripts needed to replace its legacy relational database system to address a new set of data management challenges while meeting their customers’ demanding SLAs. Join us for this on-demand webinar to hear from Keith Willard, Chief Architect at Surescripts, to learn how and why Surescripts leverages DataStax Enterprise to deliver enhanced message processing at scale.
View recording: https://youtu.be/1T6V1XAoaJQ
Explore all DataStax webinars: https://www.datastax.com/resources/webinars
Simplifying Real-Time Architectures for IoT with Apache KuduCloudera, Inc.
3 Things to Learn About:
*Building scalable real time architectures for managing data from IoT
*Processing data in real time with components such as Kudu & Spark
*Customer case studies highlighting real-time IoT use cases
Using real time big data analytics for competitive advantageAmazon Web Services
Many organisations find it challenging to successfully perform real-time data analytics using their own on premise IT infrastructure. Building a system that can adapt and scale rapidly to handle dramatic increases in transaction loads can potentially be quite a costly and time consuming exercise.
Most of the time, infrastructure is under-utilised and it’s near impossible for organisations to forecast the amount of computing power they will need in the future to serve their customers and suppliers.
To overcome these challenges, organisations can instead utilise the cloud to support their real-time data analytics activities. Scalable, agile and secure, cloud-based infrastructure enables organisations to quickly spin up infrastructure to support their data analytics projects exactly when it is needed. Importantly, they can ‘switch off’ infrastructure when it is not.
BluePi Consulting and Amazon Web Services (AWS) are giving you the opportunity to discover how organisations are using real time data analytics to gain new insights from their information to improve the customer experience and drive competitive advantage.
Choosing technologies for a big data solution in the cloudJames Serra
Has your company been building data warehouses for years using SQL Server? And are you now tasked with creating or moving your data warehouse to the cloud and modernizing it to support “Big Data”? What technologies and tools should use? That is what this presentation will help you answer. First we will cover what questions to ask concerning data (type, size, frequency), reporting, performance needs, on-prem vs cloud, staff technology skills, OSS requirements, cost, and MDM needs. Then we will show you common big data architecture solutions and help you to answer questions such as: Where do I store the data? Should I use a data lake? Do I still need a cube? What about Hadoop/NoSQL? Do I need the power of MPP? Should I build a "logical data warehouse"? What is this lambda architecture? Can I use Hadoop for my DW? Finally, we’ll show some architectures of real-world customer big data solutions. Come to this session to get started down the path to making the proper technology choices in moving to the cloud.
Verizon Centralizes Data into a Data Lake in Real Time for AnalyticsDataWorks Summit
Verizon – Global Technology Services (GTS) was challenged by a multi-tier, labor-intensive process when trying to migrate data from disparate sources into a data lake to create financial reports and business insights. Join this session to learn more about how Verizon:
• Easily accessed data from multiple sources including SAP data
• Ingested data into major targets including Hadoop
• Achieved real-time insights from data leveraging change data capture (CDC) technology
• Reduced costs and labor
Think of big data as all data, no matter what the volume, velocity, or variety. The simple truth is a traditional on-prem data warehouse will not handle big data. So what is Microsoft’s strategy for building a big data solution? And why is it best to have this solution in the cloud? That is what this presentation will cover. Be prepared to discover all the various Microsoft technologies and products from collecting data, transforming it, storing it, to visualizing it. My goal is to help you not only understand each product but understand how they all fit together, so you can be the hero who builds your companies big data solution.
Best Practices For Building and Operating A Managed Data Lake - StampedeCon 2016StampedeCon
This session will detail best practices for architecting, building, operating and managing an Analytics Data Lake platform. Key topics will include:
1) Defining next-generation Data Lake architectures. The defacto standard has been commodity DAS servers with HDFS, but there are now multiple solutions aimed at separating compute and storage, virtualizing or containerizing Hadoop applications, and utilizing Hadoop compatible or embedded HDFS filesystems. This portion will explore the options available, and the pros and cons of each.
2) Data Ingest. There are many ways to load data into a Data Lake, including standardized Apache tools (Sqoop, Flume, Kafka, Storm, Spark, NiFi), standard file and object protocols (SFTP, NFS, Rest, WebHDFS), and proprietary tools (eg, Zaloni Bedrock, DataTorrent). This section will explore these options in the context of best fit to workflows; it will also look at key gaps and challenges, particularly in the areas of data formats and integration with metadata/cataloging tools.
3) Metadata & Cataloguing. One of the biggest inhibitors of successful Data Lake deployments is Data Governance, particularly in the areas of indexing, cataloguing and metadata management. It is nearly impossible to run analytics on top of a Data Lake and get meaningful & timely results without solving these problems. This portion will explore both emerging open standards (Apache Atlas, HCatalog) and proprietary tools (Cloudera Navigator, Zaloni Bedrock/Mica, Informatica Metadata Manager), and balance the pros, cons and gaps of each.
4) Security & Access Controls. Solving these challenges are key for adoption in regulatory driven industries like Healthcare & Financial Services. There are multiple Apache projects and proprietary tools to address this, but the challenge is making security and access controls consistent across the entire application and infrastructure stack, and over the data lifecycle, and being able to audit this in the face of legal challenges. This portion will explore available options and best practices.
5) Provisioning & Workflow Management. The real promise of the Data Lake is integrating Analytics workflows and tools on converged infrastructure-with shared data-and build “As A Service” oriented architectures that are oriented towards self-service data exploration and Analytics for end users. This is an emerging and immature area, but this session will explore some potential concepts, tools and options to achieve this.
This will be a moderately technical session, with the above topics being illustrated by real world examples. Attendees should have basic familiarity with Hadoop and the associated Apache projects.
Modern apps and services are leveraging data to change the way we engage with users in a more personalized way. Skyla Loomis talks big data, analytics, NoSQL, SQL and how IBM Cloud is open for data.
Learn more by visiting our Bluemix Hybrid page: http://ibm.co/1PKN23h
Cloud-Native Data: What data questions to ask when building cloud-native appsVMware Tanzu
While a number of patterns and architectural guidelines exist for cloud-native applications, a discussion about data often leads to more questions than answers. For example, what are some of the typical data problems encountered, why are they different, and how can they be overcome?
Join Prasad Radhakrishnan from Pivotal and Dave Nielsen from Redis Labs as they discuss:
- Expectations and requirements of cloud-native data
- Common faux pas and strategies on how you can avoid them
Presenters:
Prasad Radhakrishnan, Platform Architecture for Data at Pivotal
Dave Nielsen, Head of Ecosystem Programs at Redis Labs
FINRA's Managed Data Lake: Next-Gen Analytics in the Cloud - ENT328 - re:Inve...Amazon Web Services
Financial Impact Regulatory Authority (FINRA)'s Technology Group has changed its customers' relationship with data by creating a managed data lake that enables discovery on petabytes of capital markets' data, while saving time and money over traditional analytics solutions. FINRA's managed data lake unlocks the value in its data to accelerate analytics and machine learning at scale. The data lake includes a centralized data catalog and separates storage from compute, allowing users to query from petabytes of data in seconds. Learn how FINRA uses Spot Instances and services such as Amazon S3, Amazon EMR, Amazon Redshift, and AWS Lambda to provide the right tool for the right job at each step in the data processing pipeline. All of this is done while meeting FINRA's security and compliance responsibilities as a financial regulator.
Data Integration for Both Self-Service Analytics and IT Users Senturus
See a cloud solution that enables data integration for applications such as Salesforce, NetSuite, Workday, Amazon Redshift and Microsoft Azure. View the webinar video recording and download this deck: http://www.senturus.com/resources/data-integration-tool-for-both-business-and-it-users/.
The rapid growth in self-service business analytics has created tremendous value for organizations, but in many cases has created tension between technical and business users. Technical teams have built solid data warehouses filled with trusted data from source systems such as sales, finance, and operations. Business teams are gaining tremendous insights by analyzing data warehouse information with traditional and new data discovery tools such as Cognos, Business Objects, Tableau, and Power BI.
The Informatica Cloud is a best-of-both-worlds solution that combines data integration for both business and IT users. It allows the following: 1) IT incorporates the business analyst’s data integration routines into the core, trusted data warehouse, 2) Business analysts can do data integration from both cloud-based and on-premise data sources, 3) Business analyst can use the industrial-strength data integration engine that IT teams have loved for years and 4) Integration for apps such as Salesforce, NetSuite, Workday, Amazon Redshift, Microsoft Azure, Marketo, SAP, Oracle and SQL Server.
Senturus, a business analytics consulting firm, has a resource library with hundreds of free recorded webinars, trainings, demos and unbiased product reviews. Take a look and share them with your colleagues and friends: http://www.senturus.com/resources/.
Businesses are generating more data than ever before.
Doing real time data analytics requires IT infrastructure that often needs to be scaled up quickly and running an on-premise environment in this setting has its limitations.
Organisations often require a massive amount of IT resources to analyse their data and the upfront capital cost can deter them from embarking on these projects.
What’s needed is scalable, agile and secure cloud-based infrastructure at the lowest possible cost so they can spin up servers that support their data analysis projects exactly when they are required. This infrastructure must enable them to create proof-of-concepts quickly and cheaply – to fail fast and move on.
Boost Performance with Scala – Learn From Those Who’ve Done It! Cécile Poyet
Scalding is a scala DSL for Cascading. Run on Hadoop, it’s a concise, functional, and very efficient way to build big data applications. One significant benefit of Scalding is that it allows easy porting of Scalding apps from MapReduce to newer, faster execution fabrics.
In this webinar, Cyrille Chépélov, of Transparency Rights Management, will share how his organization boosted the performance of their Scalding apps by over 50% by moving away from MapReduce to Cascading 3.0 on Apache Tez. Dhruv Kumar, Hortonworks Partner Solution Engineer, will then explain how you can interact with data on HDP using Scala and leverage Scala as a programming language to develop Big Data applications.
Boost Performance with Scala – Learn From Those Who’ve Done It! Cécile Poyet
Scalding is a scala DSL for Cascading. Run on Hadoop, it’s a concise, functional, and very efficient way to build big data applications. One significant benefit of Scalding is that it allows easy porting of Scalding apps from MapReduce to newer, faster execution fabrics.
In this webinar, Cyrille Chépélov, of Transparency Rights Management, will share how his organization boosted the performance of their Scalding apps by over 50% by moving away from MapReduce to Cascading 3.0 on Apache Tez. Dhruv Kumar, Hortonworks Partner Solution Engineer, will then explain how you can interact with data on HDP using Scala and leverage Scala as a programming language to develop Big Data applications.
Boost Performance with Scala – Learn From Those Who’ve Done It! Hortonworks
Scalding is a scala DSL for Cascading. Run on Hadoop, it’s a concise, functional, and very efficient way to build big data applications. One significant benefit of Scalding is that it allows easy porting of Scalding apps from MapReduce to newer, faster execution fabrics.
In this webinar, Cyrille Chépélov, of Transparency Rights Management, will share how his organization boosted the performance of their Scalding apps by over 50% by moving away from MapReduce to Cascading 3.0 on Apache Tez. Dhruv Kumar, Hortonworks Partner Solution Engineer, will then explain how you can interact with data on HDP using Scala and leverage Scala as a programming language to develop Big Data applications.
Take Action: The New Reality of Data-Driven BusinessInside Analysis
The Briefing Room with Dr. Robin Bloor and WebAction
Live Webcast on July 23, 2014
Watch the archive:
https://bloorgroup.webex.com/bloorgroup/lsr.php?RCID=360d371d3a49ad256942f55350aa0a8b
The waiting used to be the hardest part, but not anymore. Today’s cutting-edge enterprises can seize opportunities faster than ever, thanks to an array of technologies that enable real-time responsiveness across the spectrum of business processes. Early adopters are solving critical business challenges by enabling the rapid-fire design, development and production of very specific applications. Functionality can range from improved customer engagement to dynamic machine-to-machine interactions.
Register for this episode of The Briefing Room to learn from veteran Analyst Dr. Robin Bloor, who will tout a new era in data-driven organizations, and why a data flow architecture will soon be critical for industry leaders. He’ll be briefed by Sami Akbay of WebAction, who will showcase his company’s real-time data management platform, which combines all the component parts needed to access, process and leverage data big and small. He’ll explain how this new approach can provide game-changing power to organizations of all types and sizes.
Visit InsideAnlaysis.com for more information.
Oracle Big Data Appliance and Big Data SQL for advanced analyticsjdijcks
Overview presentation showing Oracle Big Data Appliance and Oracle Big Data SQL in combination with why this really matters. Big Data SQL brings you the unique ability to analyze data across the entire spectrum of system, NoSQL, Hadoop and Oracle Database.
DAMA & Denodo Webinar: Modernizing Data Architecture Using Data Virtualization Denodo
Watch here: https://bit.ly/2NGQD7R
In an era increasingly dominated by advancements in cloud computing, AI and advanced analytics it may come as a shock that many organizations still rely on data architectures built before the turn of the century. But that scenario is rapidly changing with the increasing adoption of real-time data virtualization - a paradigm shift in the approach that organizations take towards accessing, integrating, and provisioning data required to meet business goals.
As data analytics and data-driven intelligence takes centre stage in today’s digital economy, logical data integration across the widest variety of data sources, with proper security and governance structure in place has become mission-critical.
Attend this session to learn:
- Learn how you can meet cloud and data science challenges with data virtualization.
- Why data virtualization is increasingly finding enterprise-wide adoption
- Discover how customers are reducing costs and improving ROI with data virtualization
Using AWS to design and build your data architecture has never been easier to gain insights and uncover new opportunities to scale and grow your business. Join this workshop to learn how you can gain insights at scale with the right big data applications.
Similar to Horses for Courses: Database Roundtable (20)
Best Practices in DataOps: How to Create Agile, Automated Data PipelinesEric Kavanagh
Synthesis Webcast with Eric Kavanagh and Tamr
DataOps is an emerging set of practices, processes, and technologies for building and automating data pipelines to meet business needs quickly. As these pipelines become more complex and development teams grow in size, organizations need better collaboration and development processes to govern the flow of data and code from one step of the data lifecycle to the next – from data ingestion and transformation to analysis and reporting.
DataOps is not something that can be implemented all at once or in a short period of time. DataOps is a journey that requires a cultural shift. DataOps teams continuously search for new ways to cut waste, streamline steps, automate processes, increase output, and get it right the first time. The goal is to increase agility and cycle times, while reducing data defects, giving developers and business users greater confidence in data analytic output.
This webcast examines how organizations adopt DataOps practices in the field. It will review results of an Eckerson Group survey that sheds light on the rate and scope of DataOps adoption. It will also describe case studies of organizations that have successfully implemented DataOps practices, the challenges they have encountered and benefits they’ve received.
Tune into our webcast to learn:
- User perceptions of DataOps
- The rate of DataOps adoption by industry and other demographic variables
- DataOps adoption by technique and component (i.e., agile, test automation, orchestration, continuous development/continuous integration)
- Key challenges organizations face with DataOps
- Key benefits organizations experience with DataOps
- Best practices in doing DataOps
- Case studies and anecdotes of DataOps at companies
Expediting the Path to Discovery with Multi-Source AnalysisEric Kavanagh
The Briefing Room with Eric Kavanagh and Zoomdata
In the realm of complex analysis, rarely does one source of data provide everything the analyst needs. Data Warehouses were designed to pull data from multiple sources, to enable that kind of cross-system discovery. But that traditional model typically required stripping the data of significant context, essentially watering down the end result, and at times obfuscating the most meaningful facets.
Thanks to several advances in real-time data exploration, companies can now access raw data where it lives, and begin the analysis process often within seconds of connecting to a source. And new innovations allow for multi-source analytics, where disparate systems can be accessed simultaneously, allowing real-time discovery across multiple sources, creating a kind of analytical depth perception. Register for this special episode of The Briefing Room to hear Bloor Group CEO Eric Kavanagh, and Zoomdata speakers explain this remarkable new capability.
Metadata Mastery: A Big Step for BI ModernizationEric Kavanagh
Modernizing data management is on everyone’s mind today. Making the shift from data management practices of the BI era to modern data management is essential but it is also challenging. Whether you’re updating the back end by migrating your data warehouses to the cloud or advancing the front end with a shift from legacy BI tools to self-service analysis and visualization, it is critical to know the data that you have and to understand data lineage. Data inventory, data glossary, and data lineage are all metadata dependent. But legacy BI metadata is typically proprietary, non-integrated, and collected inconsistently by a variety of disparate tools. The metadata muddle is a serious inhibitor to modernization efforts. Metadata consolidation and centralization are the keys to overcoming this barrier. What if all this were automated?
Join us to learn:
- How a smart and innovative new technology resolves metadata disparity
- How metadata management automation accelerates modernization efforts
- How metadata management automation reduces errors and improves quality of results from data management modernization projects
- How metadata management automation and data cataloging work together to help you move rapidly to the next generation of BI and analytics
Database is the new black. Ever the backbone of information architectures, database technology continually evolves to meet growing and changing business needs. New types of data and applications make the database more important than ever, and understanding which technology best serves your use case is paramount to building durable systems. These days, the choices are many, so users should be careful when deciding which direction to go. Register for this Exploratory Webcast to hear veteran database Analyst Dr. Robin Bloor explain why the database market has exploded in recent years. He'll outline the current database landscape, and provide insights about which kinds of technologies are suitable for the growing variety of business needs today. He'll also focus on key auxiliary technologies that enable modern databases to do perform efficiently.
Better to Ask Permission? Best Practices for Privacy and SecurityEric Kavanagh
Hot Technologies with The Bloor Group and IDERA
If security was once a nice-to-have, those days have long gone. Between data breaches and privacy regulations, organizations today face immense pressure to protect their systems and their sensitive data. When giants like Yahoo! and Target can get hacked, so can any other company. What can you do about it? How can you protect your company and clients?
Register for this episode of Hot Technologies to hear Analysts Eric Kavanagh and Dr. Robin Bloor provide insights about the many ways that companies can buttress their defenses and stay ahead of the bad guys. They'll be briefed by Vicky Harp of IDERA who will demonstrate how to identify vulnerabilities, track sensitive data, successfully pass audits, and protect your SQL Server databases.
Best Laid Plans: Saving Time, Money and Trouble with Optimal ForecastingEric Kavanagh
Expectations have changed. That's true for users, executives and customers alike. There's no time for systems running slowly, or cost overruns. That's why fundamentals like capacity planning have become mission-critical. By paying attention to the details, and doing effective forecasts, companies can optimize their information architecture, keeping everyone happy. Register for this episode of Hot Technologies to learn from veteran Analysts Dr. Robin Bloor and Rick Sherman who will offer insights about how and why to do capacity planning. They'll be briefed by Bullett Manale of IDERA, who will explain how his company's SQL Diagnostic Manager can track a wide range of usages metrics which can be used for accurate forecasting.
A Winning Strategy for the Digital EconomyEric Kavanagh
The speed of innovation today creates tremendous opportunities for some, existential threats for others. Companies that win create their own success by leveraging modern data platforms. While architectures vary, the foundation is often in-memory, and the latency is real-time. Register for this Special Edition of The Briefing Room to hear veteran Analyst Dr. Robin Bloor explain how today's data platforms enable the modern enterprise in groundbreaking ways. He'll be briefed by Chris Hallenbeck of SAP who will demonstrate how forward-looking companies are leveraging real-time data platforms to achieve operational excellence, make decisions faster, and find new ways to innovate.
Discovering Big Data in the Fog: Why Catalogs MatterEric Kavanagh
The Briefing Room with Dr. Robin Bloor and Waterline Data
Good enterprise data can drive positive business outcomes. But if that data isn’t organized and accessible, information workers are left with an incomplete picture. Knowing the location, lineage and permissions of data across the enterprise can lead to more accurate and insightful searches, and ultimately, knowledge discovery.
Register for this episode of The Briefing Room to learn from veteran Analyst Dr. Robin Bloor as he discusses how the success of big data projects relies on understanding your data. He’ll be briefed by Todd Goldman and Mohan Sadashiva of Waterline Data, who will explain how their solution can facilitate discovery via automation and crowd sourcing. They’ll demonstrate how combining the value of tribal knowledge with rationalized data can enable self-service analytics, improve data governance, and reduce data redundancy.
Health Check: Maintaining Enterprise BIEric Kavanagh
Hot Technologies with The Bloor Group and IDERA
Most companies realize the value of business intelligence. Advanced analytics, data mining, dashboards – all surface useful insights. With so many moving parts in play, it’s crucial to provide visibility across the entire BI environment, thus delivering solid system and service performance.
Register for this episode of Hot Technologies to learn from Analyst Dr. Robin Bloor and Eric Kavanagh as they discuss why operational and strategic business intelligence are the cornerstones of any organization. They’ll be briefed by Stan Geiger of IDERA, who will showcase his company’s SQL BI Manager, and end-to-end solution designed to provide a single view into numerous running processes. He will explain that by optimizing system health and availability, users can eliminate downtime and improve efficiency.
Rapid Response: Debugging and Profiling to the RescueEric Kavanagh
Bad code happens. And when it does, developers often spend far too much time trying to find and fix the error. Debugging is a common solution, but in a complex environment, running multiple applications on multiple platforms, it can be easier said than done. Developers need instant visibility across all machines, ultimately leading to faster and higher quality insights. Register for this episode of Hot Technologies to learn from Analyst Dr. Robin Bloor and Data Scientist Dez Blanchfield as they discuss how errant code can inevitably disrupt systems and performance. They’ll be briefed by Bert Scalzo of IDERA, who will explain how his company’s Rapid SQL can facilitate the debugging and profiling of stored procedures and functions.
Solving the Really Big Tech Problems with IoTEric Kavanagh
The Briefing Room with Dr. Robin Bloor and HPE Security
The Internet of Things brings new technological problems: sensor communications are bi-directional, the scale of data generation points has no precedent and, in this new world, security, privacy and data protection need to go out to the edge. Likely, most of that data lands in Hadoop and Big Data platforms. With the need for rapid analytics never greater, companies try to seize opportunities in tighter time windows. Yet, cyber-threats are at an all-time high, targeting the most valuable of assets—the data.
Register for this episode of The Briefing Room to hear Analyst Dr. Robin Bloor explain the implications of today's divergent data forces. He’ll be briefed by Reiner Kappenberger of HPE, who will discuss how a recent innovation -- NiFi -- is revolutionizing the big data ecosystem. He’ll explain how this technology dramatically simplifies data flow design, enabling a new era of business-driven analysis, while also protecting sensitive data.
Beyond the Platform: Enabling Fluid AnalysisEric Kavanagh
When the analysts aren’t happy, no one is happy. That’s because these days, practically every aspect of the business is driven by insights. And because information architectures are increasingly complex, any number of issues can cause a slowdown in queries, or even basic reporting. How can your organization ensure that all systems are go?
Register for this episode of The Briefing Room to learn from veteran Analyst Dr. Robin Bloor as he explains the common roadblocks to successful BI and analytics. He'll be briefed by Stan Geiger of IDERA, who previously demonstrated how his company’s SQL BI Manager can optimize platform health and performance. In this episode, he will dive deeper into how IDERA’s solution resolves resource constraints, user activity and capacity issues, making tiresome troubleshooting a thing of the past.
Protect Your Database: High Availability for High Demand DataEric Kavanagh
Hot Technologies with Dr. Robin Bloor, Dez Blanchfield and IDERA
Your company’s data is mission-critical. While protecting it from outside attack or catastrophe has become a standard business requirement, it’s not enough these days to rely solely on simple backup and recovery techniques. Today’s enterprise requires high availability and uninterrupted operational performance, meaning the DBA toolbox must provide more than traditional solutions.
Register for this episode of Hot Technologies to hear from Analyst Dr. Robin Bloor and Data Scientist Dez Blanchfield as they discuss the necessary components of a modern solution architecture. They’ll be briefed by IDERA’s Oracle ACE Bert Scalzo, who will explain some innovative options for ensuring high availability in a demanding database environment.
A Better Understanding: Solving Business Challenges with DataEric Kavanagh
Good decisions make great companies. That's why the data-driven mantra keeps gaining momentum. Increasingly, smart business people are taking a data-first approach for both strategic planning and tactical decision-making. They spend ample time exploring their data to better understand their options. In doing so, they capitalize on real opportunities, while avoiding low-value projects.
Register for this episode of The Briefing Room to hear veteran Analyst Dr. Robin Bloor explain why a data-first mindset can help companies optimize their resources and thus make better decisions. He'll be briefed by Rishi Patel and Erin Haselkorn of
The Briefing Room with Dr. Robin Bloor and Experian
Experian, who will showcase Experian Pandora, which enables the kind of discovery that businesses need to better understand their data. They'll explain how Pandora can help professionals build a business case for their ideas and plans.
The Key to Effective Analytics: Fast-Returning QueriesEric Kavanagh
The best business analysts understand the value of having a "conversation" with their data. The idea is that they can pose queries, examine results, then quickly modify their questions to home in on a desired answer. This kind of iterative process creates a fluid environment that is highly conducive for identifying meaningful patterns in data. Register for this episode of Hot Technologies to hear Bloor Group Chief Analyst Dr. Robin Bloor and Data Scientist Dez Blanchfield as they outline why fluid analytics should be the norm and which hurdles still stand in the way. They'll be briefed by Bullett Manale of IDERA who will demonstrate his company's diagnostic platform for analytics. He'll provide context, and also deliver a demo that shows real-world solutions that enable iterative analytics.
A Tight Ship: How Containers and SDS Optimize the EnterpriseEric Kavanagh
The Briefing Room with Dez Blanchfield and Red Hat
Think of containers as the drones of modern computing. They're small, agile, and can carry a significant payload. In many ways, they represent the fruition of the last two major paradigm shifts in enterprise software: SOA and virtualization. However, for companies to fully leverage this innovative approach, a persistent storage platform is needed that is as flexible and scalable as containers themselves.
Register for this episode of The Briefing Room to hear Bloor Group Data Scientist Dez Blanchfield, who will explain the significance of container technology, and the relevance of software-defined storage (SDS) in a constantly evolving IT world. He'll be briefed by Steve Watt and Sayan Saha of Red Hat, who will demonstrate how open-source technology can help organizations take advantage of this brave new world of enterprise computing. They will explain how containers are the next step in the evolution of the operating system, and why SDS is now the optimal solution.
Application Acceleration: Faster Performance for End Users Eric Kavanagh
Hot Technologies with Dr. Robin Bloor, Dez Blanchfield and IDERA
Application performance issues impact end users the hardest, and too often, IT doesn’t know about it until after the fact. With many applications served by a variety of disparate technologies, troubleshooting bottlenecks can be onerous and time consuming, ultimately causing frustration and missed SLAs. How can IT quickly discover what process affected SQL execution time and keep end users focused on the bottom line?
Register for this episode of Hot Technologies to learn from Analyst Dr. Robin Bloor and Data Scientist Dez Blanchfield as they discuss the complexities of the data pipeline. They’ll be briefed by Bill Ellis of IDERA, who will explain the importance of identifying and resolving the root cause of performance problems. He’ll show how IDERA’s Precise Application Performance Platform can isolate transactions and usage patterns, thus giving IT the necessary tools to provide a consistent end user experience.
Time's Up! Getting Value from Big Data NowEric Kavanagh
The Briefing Room with Dr. Robin Bloor and CASK
We all know the promise of big data, but who gets the value? There are plenty of success stories already, and most of them involve one key ingredient: facilitated access to important data sets. Most research studies suggest that the Pareto principle applies: 80 percent goes to data integration, and only 20 to analysis. Inverting that balance is the Holy Grail.
Register for this episode of The Briefing Room to hear veteran Analyst Dr. Robin Bloor explain why the time has finally come for turning the tables on the status quo in analytics. He'll be briefed by CASK CEO Jonathan Gray, who will showcase his company's big data integration platform, CDAP, which was specifically designed to expedite time-to-value for big data.
The New Normal: Dealing with the Reality of an Unsecure WorldEric Kavanagh
Hot Technologies with Dr. Robin Bloor, Dez Blanchfield and IDERA
The data is staggering: breaches of epic proportion continue to rock the business world. Massive amounts of personal information have been hacked, then sold to all manner of bad actors. Another wave of attacks is on the way, in which those stolen IDs will be used to compromise any corporate system that can be found. What can your organization do?
Register for this episode of Hot Technologies to hear veteran Analyst Dr. Robin Bloor and Data Scientist Dez Blanchfield explain why security and compliance have entered a whole new era, and why innovative approaches are necessary to mitigate risk. They'll be briefed by Ignacio Rodriguez of IDERA, who will demonstrate how the company's SQL Secure can help organizations stay one step ahead of the bad guys, while also facilitating compliance audits.
UiPath Test Automation using UiPath Test Suite series, part 4DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 4. In this session, we will cover Test Manager overview along with SAP heatmap.
The UiPath Test Manager overview with SAP heatmap webinar offers a concise yet comprehensive exploration of the role of a Test Manager within SAP environments, coupled with the utilization of heatmaps for effective testing strategies.
Participants will gain insights into the responsibilities, challenges, and best practices associated with test management in SAP projects. Additionally, the webinar delves into the significance of heatmaps as a visual aid for identifying testing priorities, areas of risk, and resource allocation within SAP landscapes. Through this session, attendees can expect to enhance their understanding of test management principles while learning practical approaches to optimize testing processes in SAP environments using heatmap visualization techniques
What will you get from this session?
1. Insights into SAP testing best practices
2. Heatmap utilization for testing
3. Optimization of testing processes
4. Demo
Topics covered:
Execution from the test manager
Orchestrator execution result
Defect reporting
SAP heatmap example with demo
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
The Art of the Pitch: WordPress Relationships and SalesLaura Byrne
Clients don’t know what they don’t know. What web solutions are right for them? How does WordPress come into the picture? How do you make sure you understand scope and timeline? What do you do if sometime changes?
All these questions and more will be explored as we talk about matching clients’ needs with what your agency offers without pulling teeth or pulling your hair out. Practical tips, and strategies for successful relationship building that leads to closing the deal.
UiPath Test Automation using UiPath Test Suite series, part 3DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 3. In this session, we will cover desktop automation along with UI automation.
Topics covered:
UI automation Introduction,
UI automation Sample
Desktop automation flow
Pradeep Chinnala, Senior Consultant Automation Developer @WonderBotz and UiPath MVP
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Accelerate your Kubernetes clusters with Varnish CachingThijs Feryn
A presentation about the usage and availability of Varnish on Kubernetes. This talk explores the capabilities of Varnish caching and shows how to use the Varnish Helm chart to deploy it to Kubernetes.
This presentation was delivered at K8SUG Singapore. See https://feryn.eu/presentations/accelerate-your-kubernetes-clusters-with-varnish-caching-k8sug-singapore-28-2024 for more details.
DevOps and Testing slides at DASA ConnectKari Kakkonen
My and Rik Marselis slides at 30.5.2024 DASA Connect conference. We discuss about what is testing, then what is agile testing and finally what is Testing in DevOps. Finally we had lovely workshop with the participants trying to find out different ways to think about quality and testing in different parts of the DevOps infinity loop.
Dev Dives: Train smarter, not harder – active learning and UiPath LLMs for do...UiPathCommunity
💥 Speed, accuracy, and scaling – discover the superpowers of GenAI in action with UiPath Document Understanding and Communications Mining™:
See how to accelerate model training and optimize model performance with active learning
Learn about the latest enhancements to out-of-the-box document processing – with little to no training required
Get an exclusive demo of the new family of UiPath LLMs – GenAI models specialized for processing different types of documents and messages
This is a hands-on session specifically designed for automation developers and AI enthusiasts seeking to enhance their knowledge in leveraging the latest intelligent document processing capabilities offered by UiPath.
Speakers:
👨🏫 Andras Palfi, Senior Product Manager, UiPath
👩🏫 Lenka Dulovicova, Product Program Manager, UiPath
Generating a custom Ruby SDK for your web service or Rails API using Smithyg2nightmarescribd
Have you ever wanted a Ruby client API to communicate with your web service? Smithy is a protocol-agnostic language for defining services and SDKs. Smithy Ruby is an implementation of Smithy that generates a Ruby SDK using a Smithy model. In this talk, we will explore Smithy and Smithy Ruby to learn how to generate custom feature-rich SDKs that can communicate with any web service, such as a Rails JSON API.
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
Essentials of Automations: Optimizing FME Workflows with ParametersSafe Software
Are you looking to streamline your workflows and boost your projects’ efficiency? Do you find yourself searching for ways to add flexibility and control over your FME workflows? If so, you’re in the right place.
Join us for an insightful dive into the world of FME parameters, a critical element in optimizing workflow efficiency. This webinar marks the beginning of our three-part “Essentials of Automation” series. This first webinar is designed to equip you with the knowledge and skills to utilize parameters effectively: enhancing the flexibility, maintainability, and user control of your FME projects.
Here’s what you’ll gain:
- Essentials of FME Parameters: Understand the pivotal role of parameters, including Reader/Writer, Transformer, User, and FME Flow categories. Discover how they are the key to unlocking automation and optimization within your workflows.
- Practical Applications in FME Form: Delve into key user parameter types including choice, connections, and file URLs. Allow users to control how a workflow runs, making your workflows more reusable. Learn to import values and deliver the best user experience for your workflows while enhancing accuracy.
- Optimization Strategies in FME Flow: Explore the creation and strategic deployment of parameters in FME Flow, including the use of deployment and geometry parameters, to maximize workflow efficiency.
- Pro Tips for Success: Gain insights on parameterizing connections and leveraging new features like Conditional Visibility for clarity and simplicity.
We’ll wrap up with a glimpse into future webinars, followed by a Q&A session to address your specific questions surrounding this topic.
Don’t miss this opportunity to elevate your FME expertise and drive your projects to new heights of efficiency.
Slack (or Teams) Automation for Bonterra Impact Management (fka Social Soluti...Jeffrey Haguewood
Sidekick Solutions uses Bonterra Impact Management (fka Social Solutions Apricot) and automation solutions to integrate data for business workflows.
We believe integration and automation are essential to user experience and the promise of efficient work through technology. Automation is the critical ingredient to realizing that full vision. We develop integration products and services for Bonterra Case Management software to support the deployment of automations for a variety of use cases.
This video focuses on the notifications, alerts, and approval requests using Slack for Bonterra Impact Management. The solutions covered in this webinar can also be deployed for Microsoft Teams.
Interested in deploying notification automations for Bonterra Impact Management? Contact us at sales@sidekicksolutionsllc.com to discuss next steps.
Elevating Tactical DDD Patterns Through Object CalisthenicsDorra BARTAGUIZ
After immersing yourself in the blue book and its red counterpart, attending DDD-focused conferences, and applying tactical patterns, you're left with a crucial question: How do I ensure my design is effective? Tactical patterns within Domain-Driven Design (DDD) serve as guiding principles for creating clear and manageable domain models. However, achieving success with these patterns requires additional guidance. Interestingly, we've observed that a set of constraints initially designed for training purposes remarkably aligns with effective pattern implementation, offering a more ‘mechanical’ approach. Let's explore together how Object Calisthenics can elevate the design of your tactical DDD patterns, offering concrete help for those venturing into DDD for the first time!
2. An Early Thought
Data Lakes and Databases are not
very different things…
Irrespective of what the data lake
enthusiasts claim
3. Everything in flux
u Hardware (network,
storage, servers)
u Data Sources
u Data Staging
u Data Volumes
u Data Flow
u Data Governance
u Query Languages
u Data Usage
u Data Structures
u Schema definition
u Ingest speeds
u Data Workloads
u Applications
4. The Data Lake Picture
Data
Cleansing
Data
Security
Ingest
Metadata
Mgt
Real-Time
Apps
Transform &
Aggregate
Search &
Query
BI, Visual'n
& Analytics
Other
Apps
Data Lake
Mgt
Data
Governance
DATA LAKE
To
Databases
Data Marts
Other Apps
Archive
Life Cycle
Mgt Extracts
Servers, Desktops, Mobile, Network Devices, Embedded
Chips, RFID, IoT, The Cloud, Oses, VMs, Log Files, Sys
Mgt Apps, ESBs, Web Services, SaaS, Business Apps,
Office Apps, BI Apps, Workflow, Data Streams, Social...
u Data Lakes (Yes!):
u Ingest points for data for the
sake of governance
u Analytics sandboxes
u Good places for cool and cold
data – and hence archive
u Data Lakes (No!):
u OLTP databases
u Fast query engines
u High user concurrency
u Bid Data analytics apps
u Unusually structured data
(NoSQL, graph, etc.)
You don’t have one data lake you have
many
Data lakes do not manage data well.
5. Streaming
There’s a spectrum of streaming
capability and thus a spectrum of
streaming platforms:
Spark, in-memory DBMS, SQLstream
6. Database Workload Parameters
q Read-intensive vs. write-
intensive
q Mutable vs. immutable data
q Immediate vs. eventual
consistency
q Short vs. long data latency
q Predictable vs.
unpredictable data access
patterns
q Simple vs. complex data
types
7. Horses for Courses
q Relational row store databases for
conventional OLTP
q Relational databases for ACID
requirements
q Parallel databases (row or column)
for unpredictable or variable query
workloads
q Specialized databases for complex
data query workloads (graph, etc.)
q NoSQL (KVS, DHT) for high scale
OLTP
q NoSQL (KVS, DHT) for low latency
read-mostly data access
q NoSQL / Hadoop /Spark for scale-
out batch analytic workloads
q Cloud Databases can be any of the
above
8. Database Tools
q Have you noticed how databases
are not self-running.
q DBA’s are in short supply and the
need for them is increasing
q Database diversity doesn’t help
in this area.
q DBA Tools:
q SQL analysis
q Performance analysis
q Security management
q Capacity planning
q Database deployment
q We meet the same problem with
data lakes – except that there
are very few tools
27. Picking a DB
3
Structure
• Does the data fit into a nice
clean data model
• Will the schema lack clarity
or be dynamic?
Analytics
• What question(s) do
you want to ask of the
data?
• Short running queries
• Long, deep analytics
including predictive
Size
• Is the data “Big Data”
or will it ever be big
data?
Also:
• Cost per Terabyte
• Staffing considerations
• Familiarity with
technologies
• Company Financials
• Company Ancillary
Portfolio
• Community & Openness
28. Security Analytics
– Are there any attacks happening
right now?
Needing different kinds of analysis is common
Weather Application
– Tell me the current
temperature and pressure
Short, fast queries
Deeper analytics with
bigger data sets
Machine learning and
predictive
– What was the high/low for my
area?
– What was the high/low for my
region?
– What was the average
temperature?
– Highest and lowest of all time?
– Can we predict conditions
tomorrow?
– What IP and where are most of my
events coming from?
– Has traffic spiked compared to
historical?
– Has any event happened liked this
over the last three years
– What new events should we be
tracking to predict security events?
29. HPE Vertica Enterprise
– Columnar storage and advanced compression
– Maximum performance and scalability
HPE Vertica
All built on the same trusted and proven HPE Vertica Core SQL Engine
5
Core HPE Vertica SQL Engine
• Advanced Analytics
• Open ANSI SQL Standards ++
• R, Python, Java, Spark. Scala
• In-database machine learning
HPE Vertica for SQL on Hadoop
– Native support for ORC and Parquet
– Support for industry-leading distributions
– No helper node or single point of failure
HPE Vertica In the Cloud
– Get up and running quickly in the cloud
– Flexible, enterprise-class cloud deployment options
30. The Appeal of Vertica
Requirement Proof
Extreme Optimization
• Columnar design for high performance analytics
• Aggressive compression
• Scalable to petabyte scale
Total Cost of Ownership
• Simply and predictable pricing
• No penalty for additional hardware or connected users
Ready for your Enterprise
• SQL compliant to 100% of the TPC-DS benchmark queries
• Secure and ACID compliant
• No single point of failure
Open and Compatible
• Open platform – Standards compliant SQL, Python, Java
• Working with open source community on Spark, Hadoop, Kafka, etc.
6
31. Vertica Enterprise Unique Value to expand the data warehouse
7
Hadoop Data Lake Vertica Big Data Warehouse
CREATE TABLE customer_visits (
customer_id bigint,
visit_num int)
PARTITIONED BY (page_view_dt date)
STORED AS ORC;
Customer information in Hadoop Customer information in Data Warehouse
SELECT customers.customer_id FROM orders RIGHT OUTER JOIN customers
ON orders.customer_id = customers.customer_id
GROUP BY customers.customer_id HAVING COUNT(orders.customer_id) = 0;
Vertica Engine
Querying data that sits
BOTH in the data
warehouse and Hadoop
is our unique value.
Most solutions require that
you move the data.
ROS
§ Leveraging Web Logs to gain customer insight
§ Sensor and IOT data for pre-emptive service
§ Marketing Programs Tracking
§ Tracking impact of application updates
§ Many more uses
32. Machine Learning in Vertica 8.0.1
Algorithm Example
Linear Regression Demand Forecasting
Model the demand for a service or good (response) based on its features (predictors) for
example; demand for different models of laptops based on monitor size, weight, price,
operating system, etc.
Logistic Regression Engineering
Predicting the likelihood that a particular mechanical part of a system will malfunction or
require maintenance (response) based on operating conditions and diagnostic
measurements (predictors)
K-means Fraud Detection
Identify individual observations that don’t align to a distinct group (cluster) and identify
types of clusters that are more likely to be at risk of
Naïve Bayes Categorization
Using fuzzy logic, identify items that in one group or another. Used in email spam
detection, language detection, sentiment analysis and document sorting
Support the whole workflow of predictive analytics
33. Perhaps the ultimate architecture is all-inclusive
Apache Spark, Hadoop and Kafka
Data Warehouse (Vertica)
Optimal Use Case
– Deep Analysis
– Massive scale
– Many concurrent users
Kafka
Data Lake (Hadoop)
Optimal Use Case
– Data lake
– Warm, cold storage
– Data discovery
– ETL
Operational Analytics (Spark)
Optimal Use Case
– Small, fast running queries
– ETL and complex event processing
– Operational analytics
Features:
– Vertica performs optimized data load from
Spark
– Spark runs queries on Vertica data
Features:
– Analyze-in-place without data movement
via native ORC and Parquet readers
– Any Hadoop
– Run ON the Hadoop cluster or ON Vertica
cluster
Features:
– Share data between
applications that support
Kafka
– Data streaming into Vertica
34. Vertica makes data matter
Purpose built for Big Data from the first line of code
Gain insight into your data 50x-1,000x
faster than legacy products
Fast Analytics
Infinitely scale your solution by addingan
unlimited number of low cost nodes
Massive scalability
Built-in support for Hadoop, R, and a
range of ETL and BI tools
Open architecture
Store 10x-30x more data per server than
row databases with patentedcolumnar
compression
Optimized data storage
HPE Vertica Community
Edition
Download and install community
edition.Manage and analyze up to 1
TB of data across three nodes for an
unlimited time.
Try it on my.vertica.com