Being able to analyze sales at the most granular level with up-to-date data, provides a competitive advantage for unlocking additional revenue -- especially for e-commerce and retail companies heading into the holiday season.
When it comes to creating an enterprise AI strategy: if your company isn’t good at analytics, it’s not ready for AI. Succeeding in AI requires being good at data engineering AND analytics. Unfortunately, management teams often assume they can leapfrog best practices for basic data analytics by directly adopting advanced technologies such as ML/AI – setting themselves up for failure from the get-go. This presentation explains how to get basic data engineering and the right technology in place to create and maintain data pipelines so that you can solve problems with AI successfully.
2020 Big Data & Analytics Maturity Survey ResultsAtScale
Together with Cloudera and ODPI.org, AtScale surveyed over 150 data & analytics leaders. This presentation reveals the results of the survey. To download the report, go to: https://tinyurl.com/qmwofof
How to Build Business Forecasts With Microsoft Excel Using 10x the Data at 20...AtScale
Watch this presentation from the experts from SafeGraph and AtScale to learn how to turn Microsoft Excel into a crystal ball for your business forecasting - painlessly. You’ll learn how to:
combine your data with public or purchased data to enrich insights; build sophisticated time-relative analyses like period-to-date calculations; use Excel pivot tables against billions of data points for data exploration; and, build a model that will automatically refresh at the cell level.
And then you’ll be able to:
Understand the product mix and product level by store location
Model and forecast revenue and expenses
Use semi-additive measure for tracking inventory levels
Calculate per member per month KPIs
Cloud Storage Spring Cleaning: A Treasure HuntSteven Moy
This is a talk by Zach and me on how to analyze your S3 storage access pattern to save storage cost by lifecycle objects at the right time to the right cost tier.
From Traditional Data Warehouse To Real Time Data WarehouseOsama Hussein
Summarising the 'From Traditional Data Warehouse To Real Time Data Warehouse' paper.
1. S. Bouaziz, A. Nabli and F. Gargouri, "From Traditional Data Warehouse To Real Time Data Warehouse", 2017.
OLAP on the Cloud with Azure Databricks and Azure SynapseAtScale
This presentation was part of the 2020 Global Summer Azure Data Fest. It explains how Cloud OLAP helps you to analyze large amounts of data on Azure Databricks, Azure Synapse and other data platforms without moving it. And, shows how to leverage AtScale’s Cloud OLAP perform multidimensional analysis – and derive business insights – on data sets from multiple providers – with no data prep or data engineering required.
This white paper will present the opportunities laid down by
data lake and advanced analytics, as well as, the challenges
in integrating, mining and analyzing the data collected from
these sources. It goes over the important characteristics of
the data lake architecture and Data and Analytics as a
Service (DAaaS) model. It also delves into the features of a
successful data lake and its optimal designing. It goes over
data, applications, and analytics that are strung together to
speed-up the insight brewing process for industry’s
improvements with the help of a powerful architecture for
mining and analyzing unstructured data – data lake.
When it comes to creating an enterprise AI strategy: if your company isn’t good at analytics, it’s not ready for AI. Succeeding in AI requires being good at data engineering AND analytics. Unfortunately, management teams often assume they can leapfrog best practices for basic data analytics by directly adopting advanced technologies such as ML/AI – setting themselves up for failure from the get-go. This presentation explains how to get basic data engineering and the right technology in place to create and maintain data pipelines so that you can solve problems with AI successfully.
2020 Big Data & Analytics Maturity Survey ResultsAtScale
Together with Cloudera and ODPI.org, AtScale surveyed over 150 data & analytics leaders. This presentation reveals the results of the survey. To download the report, go to: https://tinyurl.com/qmwofof
How to Build Business Forecasts With Microsoft Excel Using 10x the Data at 20...AtScale
Watch this presentation from the experts from SafeGraph and AtScale to learn how to turn Microsoft Excel into a crystal ball for your business forecasting - painlessly. You’ll learn how to:
combine your data with public or purchased data to enrich insights; build sophisticated time-relative analyses like period-to-date calculations; use Excel pivot tables against billions of data points for data exploration; and, build a model that will automatically refresh at the cell level.
And then you’ll be able to:
Understand the product mix and product level by store location
Model and forecast revenue and expenses
Use semi-additive measure for tracking inventory levels
Calculate per member per month KPIs
Cloud Storage Spring Cleaning: A Treasure HuntSteven Moy
This is a talk by Zach and me on how to analyze your S3 storage access pattern to save storage cost by lifecycle objects at the right time to the right cost tier.
From Traditional Data Warehouse To Real Time Data WarehouseOsama Hussein
Summarising the 'From Traditional Data Warehouse To Real Time Data Warehouse' paper.
1. S. Bouaziz, A. Nabli and F. Gargouri, "From Traditional Data Warehouse To Real Time Data Warehouse", 2017.
OLAP on the Cloud with Azure Databricks and Azure SynapseAtScale
This presentation was part of the 2020 Global Summer Azure Data Fest. It explains how Cloud OLAP helps you to analyze large amounts of data on Azure Databricks, Azure Synapse and other data platforms without moving it. And, shows how to leverage AtScale’s Cloud OLAP perform multidimensional analysis – and derive business insights – on data sets from multiple providers – with no data prep or data engineering required.
This white paper will present the opportunities laid down by
data lake and advanced analytics, as well as, the challenges
in integrating, mining and analyzing the data collected from
these sources. It goes over the important characteristics of
the data lake architecture and Data and Analytics as a
Service (DAaaS) model. It also delves into the features of a
successful data lake and its optimal designing. It goes over
data, applications, and analytics that are strung together to
speed-up the insight brewing process for industry’s
improvements with the help of a powerful architecture for
mining and analyzing unstructured data – data lake.
A few months back I spoke with some graduate students about "what is data warehousing". In this talk I covered the past, present, and probably future of what data warehousing is and how it can add value to a company.
A Data Lake is a storage repository that can store large amount of structured, semi-structured, and unstructured data. It is a place to store every type of data in its native format with no fixed limits on account size or file. It offers high data quantity to increase analytic performance and native integration.
Data Lake is like a large container which is very similar to real lake and rivers. Just like in a lake you have multiple tributaries coming in, a data lake has structured data, unstructured data, machine to machine, logs flowing through in real-time.
Big Data Day LA 2015 - Data Lake - Re Birth of Enterprise Data Thinking by Ra...Data Con LA
Why and How has the Big Data based Enterprise Data Lake solution based on No-SQL and SQL technologies has become significantly effective in solving enterprise data challenges than its predecessor EDW which had tried and failed to solve the same problem entirely based on SQL database only.
O'Reilly ebook: Operationalizing the Data LakeVasu S
Best practices for building a cloud data lake operation—from people and tools to processes
https://www.qubole.com/resources/ebooks/ebook-operationalizing-the-data-lake
In this presentation at DAMA New York, Joe started by asking a key question: why are we doing this? Why analyze and share all these massive amounts of data? Basically, it comes down to the belief that in any organization, in any situation, if we can get the data and make it correct and timely, insights from it will become instantly actionable for companies to function more nimbly and successfully. Enabling the use of data can be a world-changing, world-improving activity and this session presents the steps necessary to get you there. Joe explained the concept of the "data lake" and also emphasizes the role of a strong data governance strategy that incorporates seven components needed for a successful program.
For more information on this presentation or Caserta Concepts, visit our website at http://casertaconcepts.com/.
How to select a modern data warehouse and get the most out of it?Slim Baltagi
In the first part of this talk, we will give a setup and definition of modern cloud data warehouses as well as outline problems with legacy and on-premise data warehouses.
We will speak to selecting, technically justifying, and practically using modern data warehouses, including criteria for how to pick a cloud data warehouse and where to start, how to use it in an optimum way and use it cost effectively.
In the second part of this talk, we discuss the challenges and where people are not getting their investment. In this business-focused track, we cover how to get business engagement, identifying the business cases/use cases, and how to leverage data as a service and consumption models.
Data Lakehouse, Data Mesh, and Data Fabric (r2)James Serra
So many buzzwords of late: Data Lakehouse, Data Mesh, and Data Fabric. What do all these terms mean and how do they compare to a modern data warehouse? In this session I’ll cover all of them in detail and compare the pros and cons of each. They all may sound great in theory, but I'll dig into the concerns you need to be aware of before taking the plunge. I’ll also include use cases so you can see what approach will work best for your big data needs. And I'll discuss Microsoft version of the data mesh.
Whether you are interested in healthcare data analytics or looking to get started with big data and marketing, these fundamental principles from data experts will contribute to your success. http://www.qubole.com/new-series-big-data-tips/
Building the Modern Data Hub: Beyond the Traditional Enterprise Data WarehouseFormant
Datavail and SlamData present on how to use NoSQL technologies (MongoDB and SlamData) to build a Data Hub -- the fast and easy way to real-time business insight.
Creating a Next-Generation Big Data ArchitecturePerficient, Inc.
If you’ve spent time investigating Big Data, you quickly realize that the issues surrounding Big Data are often complex to analyze and solve. The sheer volume, velocity and variety changes the way we think about data – including how enterprises approach data architecture.
Significant reduction in costs for processing, managing, and storing data, combined with the need for business agility and analytics, requires CIOs and enterprise architects to rethink their enterprise data architecture and develop a next-generation approach to solve the complexities of Big Data.
Creating the data architecture while integrating Big Data into the heart of the enterprise data architecture is a challenge. This webinar covered:
-Why Big Data capabilities must be strategically integrated into an enterprise’s data architecture
-How a next-generation architecture can be conceptualized
-The key components to a robust next generation architecture
-How to incrementally transition to a next generation data architecture
Say goodbye to data silos! Analytics in a Day will simplify and accelerate your journey towards the modern data warehouse. Join CCG and Microsoft for a two-day virtual workshop, hosted by James McAuliffe.
Webinar: Data Modeling and Shortcuts to Success in Scaling Time Series Applic...DATAVERSITY
Join Basho technologies and Databricks, creators of Apache Spark, as we share lessons learned by both organizations in building scalable applications for IoT and time series use cases. We'll be discussing some of the data modeling considerations unique to time series data and some of the key factors developers and architects need to take into consideration as data moves through the pipeline. You'll learn:
Challenges in building apps to leverage data being generated by IoT devices
What you need to think about before you start modeling your IoT data
Shortcuts to success in building IoT apps
The webinar will also give a live demonstration of how to store and retrieve IoT data as well as a demonstration of integrated data store with analytics engine using a live Notebook as a guide.
Data Mesh in Practice: How Europe’s Leading Online Platform for Fashion Goes ...Databricks
The Data Lake paradigm is often considered the scalable successor of the more curated Data Warehouse approach when it comes to democratization of data. However, many who went out to build a centralized Data Lake came out with a data swamp of unclear responsibilities, a lack of data ownership, and sub-par data availability.
2015 02 12 talend hortonworks webinar challenges to hadoop adoptionHortonworks
Hadoop is no longer optional. Companies of all sizes are in various phases of their own Big Data journey. Whether you are just starting to explore the platform or have multiple clusters up and running, everyone is presented with a similar challenge - developing their internal skillset. Hadoop specialists are hard to find. Hand coding is too prone to error when it comes to storing, integrating or analyzing your data. However, it doesn’t need to be this difficult.
In this recorded webinar, Talend and Hortonworks help you learn how to unify all your data in Hadoop, with no specialized Big Data skills.
Find the recording here. www.talend.com/resources/webinars/challenges-to-hadoop-adoption-if-you-can-dream-it-you-can-build-it
This webinar covers: How Hadoop opens a new world of analytic applications, How to bridge the skills gap with our Big Data solutions, Experience a real-world, simple technical demo
A few months back I spoke with some graduate students about "what is data warehousing". In this talk I covered the past, present, and probably future of what data warehousing is and how it can add value to a company.
A Data Lake is a storage repository that can store large amount of structured, semi-structured, and unstructured data. It is a place to store every type of data in its native format with no fixed limits on account size or file. It offers high data quantity to increase analytic performance and native integration.
Data Lake is like a large container which is very similar to real lake and rivers. Just like in a lake you have multiple tributaries coming in, a data lake has structured data, unstructured data, machine to machine, logs flowing through in real-time.
Big Data Day LA 2015 - Data Lake - Re Birth of Enterprise Data Thinking by Ra...Data Con LA
Why and How has the Big Data based Enterprise Data Lake solution based on No-SQL and SQL technologies has become significantly effective in solving enterprise data challenges than its predecessor EDW which had tried and failed to solve the same problem entirely based on SQL database only.
O'Reilly ebook: Operationalizing the Data LakeVasu S
Best practices for building a cloud data lake operation—from people and tools to processes
https://www.qubole.com/resources/ebooks/ebook-operationalizing-the-data-lake
In this presentation at DAMA New York, Joe started by asking a key question: why are we doing this? Why analyze and share all these massive amounts of data? Basically, it comes down to the belief that in any organization, in any situation, if we can get the data and make it correct and timely, insights from it will become instantly actionable for companies to function more nimbly and successfully. Enabling the use of data can be a world-changing, world-improving activity and this session presents the steps necessary to get you there. Joe explained the concept of the "data lake" and also emphasizes the role of a strong data governance strategy that incorporates seven components needed for a successful program.
For more information on this presentation or Caserta Concepts, visit our website at http://casertaconcepts.com/.
How to select a modern data warehouse and get the most out of it?Slim Baltagi
In the first part of this talk, we will give a setup and definition of modern cloud data warehouses as well as outline problems with legacy and on-premise data warehouses.
We will speak to selecting, technically justifying, and practically using modern data warehouses, including criteria for how to pick a cloud data warehouse and where to start, how to use it in an optimum way and use it cost effectively.
In the second part of this talk, we discuss the challenges and where people are not getting their investment. In this business-focused track, we cover how to get business engagement, identifying the business cases/use cases, and how to leverage data as a service and consumption models.
Data Lakehouse, Data Mesh, and Data Fabric (r2)James Serra
So many buzzwords of late: Data Lakehouse, Data Mesh, and Data Fabric. What do all these terms mean and how do they compare to a modern data warehouse? In this session I’ll cover all of them in detail and compare the pros and cons of each. They all may sound great in theory, but I'll dig into the concerns you need to be aware of before taking the plunge. I’ll also include use cases so you can see what approach will work best for your big data needs. And I'll discuss Microsoft version of the data mesh.
Whether you are interested in healthcare data analytics or looking to get started with big data and marketing, these fundamental principles from data experts will contribute to your success. http://www.qubole.com/new-series-big-data-tips/
Building the Modern Data Hub: Beyond the Traditional Enterprise Data WarehouseFormant
Datavail and SlamData present on how to use NoSQL technologies (MongoDB and SlamData) to build a Data Hub -- the fast and easy way to real-time business insight.
Creating a Next-Generation Big Data ArchitecturePerficient, Inc.
If you’ve spent time investigating Big Data, you quickly realize that the issues surrounding Big Data are often complex to analyze and solve. The sheer volume, velocity and variety changes the way we think about data – including how enterprises approach data architecture.
Significant reduction in costs for processing, managing, and storing data, combined with the need for business agility and analytics, requires CIOs and enterprise architects to rethink their enterprise data architecture and develop a next-generation approach to solve the complexities of Big Data.
Creating the data architecture while integrating Big Data into the heart of the enterprise data architecture is a challenge. This webinar covered:
-Why Big Data capabilities must be strategically integrated into an enterprise’s data architecture
-How a next-generation architecture can be conceptualized
-The key components to a robust next generation architecture
-How to incrementally transition to a next generation data architecture
Say goodbye to data silos! Analytics in a Day will simplify and accelerate your journey towards the modern data warehouse. Join CCG and Microsoft for a two-day virtual workshop, hosted by James McAuliffe.
Webinar: Data Modeling and Shortcuts to Success in Scaling Time Series Applic...DATAVERSITY
Join Basho technologies and Databricks, creators of Apache Spark, as we share lessons learned by both organizations in building scalable applications for IoT and time series use cases. We'll be discussing some of the data modeling considerations unique to time series data and some of the key factors developers and architects need to take into consideration as data moves through the pipeline. You'll learn:
Challenges in building apps to leverage data being generated by IoT devices
What you need to think about before you start modeling your IoT data
Shortcuts to success in building IoT apps
The webinar will also give a live demonstration of how to store and retrieve IoT data as well as a demonstration of integrated data store with analytics engine using a live Notebook as a guide.
Data Mesh in Practice: How Europe’s Leading Online Platform for Fashion Goes ...Databricks
The Data Lake paradigm is often considered the scalable successor of the more curated Data Warehouse approach when it comes to democratization of data. However, many who went out to build a centralized Data Lake came out with a data swamp of unclear responsibilities, a lack of data ownership, and sub-par data availability.
2015 02 12 talend hortonworks webinar challenges to hadoop adoptionHortonworks
Hadoop is no longer optional. Companies of all sizes are in various phases of their own Big Data journey. Whether you are just starting to explore the platform or have multiple clusters up and running, everyone is presented with a similar challenge - developing their internal skillset. Hadoop specialists are hard to find. Hand coding is too prone to error when it comes to storing, integrating or analyzing your data. However, it doesn’t need to be this difficult.
In this recorded webinar, Talend and Hortonworks help you learn how to unify all your data in Hadoop, with no specialized Big Data skills.
Find the recording here. www.talend.com/resources/webinars/challenges-to-hadoop-adoption-if-you-can-dream-it-you-can-build-it
This webinar covers: How Hadoop opens a new world of analytic applications, How to bridge the skills gap with our Big Data solutions, Experience a real-world, simple technical demo
Accelerate Self-Service Analytics with Data Virtualization and VisualizationDenodo
Watch full webinar here: https://bit.ly/3fpitC3
Enterprise organizations are shifting to self-service analytics as business users need real-time access to holistic and consistent views of data regardless of its location, source or type for arriving at critical decisions.
Data Virtualization and Data Visualization work together through a universal semantic layer. Learn how they enable self-service data discovery and improve performance of your reports and dashboards.
In this session, you will learn:
- Challenges faced by business users
- How data virtualization enables self-service analytics
- Use case and lessons from customer success
- Overview of the highlight features in Tableau
Big Data Tools: A Deep Dive into Essential ToolsFredReynolds2
Today, practically every firm uses big data to gain a competitive advantage in the market. With this in mind, freely available big data tools for analysis and processing are a cost-effective and beneficial choice for enterprises. Hadoop is the sector’s leading open-source initiative and big data tidal roller. Moreover, this is not the final chapter! Numerous other businesses pursue Hadoop’s free and open-source path.
Against the backdrop of Big Data, the Chief Data Officer, by any name, is emerging as the central player in the business of data, including cybersecurity. The MITCDOIQ Symposium explored the developing landscape, from local organizational issues to global challenges, through case studies from industry, academic, government and healthcare leaders.
Joe Caserta, president at Caserta Concepts, presented "Big Data's Impact on the Enterprise" at the MITCDOIQ Symposium.
Presentation Abstract: Organizations are challenged with managing an unprecedented volume of structured and unstructured data coming into the enterprise from a variety of verified and unverified sources. With that is the urgency to rapidly maximize value while also maintaining high data quality.
Today we start with some history and the components of data governance and information quality necessary for successful solutions. I then bring it all to life with 2 client success stories, one in healthcare and the other in banking and financial services. These case histories illustrate how accurate, complete, consistent and reliable data results in a competitive advantage and enhanced end-user and customer satisfaction.
To learn more, visit www.casertaconcepts.com
Smarter Analytics: Supporting the Enterprise with AutomationInside Analysis
The Briefing Room with Barry Devlin and WhereScape
Live Webcast on June 10, 2014
Watch the archive:
https://bloorgroup.webex.com/bloorgroup/lsr.php?RCID=5230c31ab287778c73b56002bc2c51a
The data warehouse is intended to support analysis by making the right data available to the right people in a timely fashion. But conditions change all the time, and when data doesn’t keep up with the business, analysts quickly turn to workarounds. This leads to ungoverned and largely un-managed side projects, which trade short-term wins for long-term trouble. One way to keep everyone happy is by creating an integrated environment that pulls data from all sources, and is capable of automating both the model development and delivery of analyst-ready data.
Register for this episode of The Briefing Room to hear data warehousing pioneer and Analyst Barry Devlin as he explains the critical components of a successful data warehouse environment, and how traditional approaches must be augmented to keep up with the times. He’ll be briefed by WhereScape CEO Michael Whitehead, who will showcase his company’s data warehousing automation solutions. He’ll discuss how a fast, well-managed and automated infrastructure is the key to empowering faster, smarter, repeatable decision making.
Visit InsideAnlaysis.com for more information.
Eliminating the Challenges of Big Data Management Inside HadoopHortonworks
Your Big Data strategy is only as good as the quality of your data. Today, deriving business value from data depends on how well your company can capture, cleanse, integrate and manage data. During this webinar, we discuss how to eliminate the challenges to Big Data management inside Hadoop.
Eliminating the Challenges of Big Data Management Inside HadoopHortonworks
Your Big Data strategy is only as good as the quality of your data. Today, deriving business value from data depends on how well your company can capture, cleanse, integrate and manage data. During this webinar, we discussed how to eliminate the challenges to Big Data management inside Hadoop.
Go over these slides to learn:
· How to use the scalability and flexibility of Hadoop to drive faster access to usable information across the enterprise.
· Why a pure-YARN implementation for data integration, quality and management delivers competitive advantage.
· How to use the flexibility of RedPoint and Hortonworks to create an enterprise data lake where data is captured, cleansed, linked and structured in a consistent way.
These slides - based on the webinar - shed light on how business stakeholders make the most of information from their big data environments and the requirements those stakeholders have to turn big data into business impact.
Using recent big data end-user research from leading IT analyst firm Enterprise Management (EMA), data from Vertica’s recent benchmarks on SQL on Hadoop, and firsthand customer experiences, viewers will learn:
- Use cases where end users around the world are using big data in their organizations
- How maturity with big data strategies impact why and how business stakeholders use information from their big data environments
- How Vertica empowers the use of information from big data environments
Big Data and BI Tools - BI Reporting for Bay Area Startups User GroupScott Mitchell
This presentation was presented at the July 8th 2014 user group meeting for BI Reporting for Bay Area Start Ups
Content - Creation Infocepts/DWApplications
Presented by: Scott Mitchell - DWApplications
The Right Data Warehouse: Automation Now, Business Value ThereafterInside Analysis
The Briefing Room with Dr. Robin Bloor and WhereScape
Live Webcast on April 1, 2014
Watch the archive: https://bloorgroup.webex.com/bloorgroup/lsr.php?RCID=7b23b14b532bd7be60a70f6bd5209f03
In the Big Data shuffle, everyone is looking at Hadoop as “the answer” to collect interesting data from a new set of sources. While Hadoop has given organizations the power to gather more information assets than ever before, the question still looms: which data, regardless of source, structure, volume and all the rest, are significant for affecting business value – and how do we harness it? One effective approach is to bolster the data warehouse environment with a solution capable of integrating all the data sources, including Hadoop, and automating delivery of key information into the rights hands.
Register for this episode of The Briefing Room to hear veteran Analyst Robin Bloor as he explains how a rapidly changing information landscape impacts data management. He will be briefed by Mark Budzinski of WhereScape, who will tout his company’s data warehouse automation solutions. Budzinski will discuss how automation can be the cornerstone for closing the gap between those responsible for data management and the people driving business decisions.
Visit InsideAnlaysis.com for more information.
Learn how when an organizations combine HP and Vertica Analytics Platform and Hortonworks, they can quickly explore and analyze broad variety of data types to transform to actionable information that allows them to better understand how their customers and site visitors interact with their business, offline and online.
Similar to How to Optimize Sales Analytics Using 10x the Data at 1/10th the Cost (20)
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdf91mobiles
91mobiles recently conducted a Smart TV Buyer Insights Survey in which we asked over 3,000 respondents about the TV they own, aspects they look at on a new TV, and their TV buying preferences.
JMeter webinar - integration with InfluxDB and GrafanaRTTS
Watch this recorded webinar about real-time monitoring of application performance. See how to integrate Apache JMeter, the open-source leader in performance testing, with InfluxDB, the open-source time-series database, and Grafana, the open-source analytics and visualization application.
In this webinar, we will review the benefits of leveraging InfluxDB and Grafana when executing load tests and demonstrate how these tools are used to visualize performance metrics.
Length: 30 minutes
Session Overview
-------------------------------------------
During this webinar, we will cover the following topics while demonstrating the integrations of JMeter, InfluxDB and Grafana:
- What out-of-the-box solutions are available for real-time monitoring JMeter tests?
- What are the benefits of integrating InfluxDB and Grafana into the load testing stack?
- Which features are provided by Grafana?
- Demonstration of InfluxDB and Grafana using a practice web application
To view the webinar recording, go to:
https://www.rttsweb.com/jmeter-integration-webinar
Accelerate your Kubernetes clusters with Varnish CachingThijs Feryn
A presentation about the usage and availability of Varnish on Kubernetes. This talk explores the capabilities of Varnish caching and shows how to use the Varnish Helm chart to deploy it to Kubernetes.
This presentation was delivered at K8SUG Singapore. See https://feryn.eu/presentations/accelerate-your-kubernetes-clusters-with-varnish-caching-k8sug-singapore-28-2024 for more details.
Kubernetes & AI - Beauty and the Beast !?! @KCD Istanbul 2024Tobias Schneck
As AI technology is pushing into IT I was wondering myself, as an “infrastructure container kubernetes guy”, how get this fancy AI technology get managed from an infrastructure operational view? Is it possible to apply our lovely cloud native principals as well? What benefit’s both technologies could bring to each other?
Let me take this questions and provide you a short journey through existing deployment models and use cases for AI software. On practical examples, we discuss what cloud/on-premise strategy we may need for applying it to our own infrastructure to get it to work from an enterprise perspective. I want to give an overview about infrastructure requirements and technologies, what could be beneficial or limiting your AI use cases in an enterprise environment. An interactive Demo will give you some insides, what approaches I got already working for real.
Connector Corner: Automate dynamic content and events by pushing a buttonDianaGray10
Here is something new! In our next Connector Corner webinar, we will demonstrate how you can use a single workflow to:
Create a campaign using Mailchimp with merge tags/fields
Send an interactive Slack channel message (using buttons)
Have the message received by managers and peers along with a test email for review
But there’s more:
In a second workflow supporting the same use case, you’ll see:
Your campaign sent to target colleagues for approval
If the “Approve” button is clicked, a Jira/Zendesk ticket is created for the marketing design team
But—if the “Reject” button is pushed, colleagues will be alerted via Slack message
Join us to learn more about this new, human-in-the-loop capability, brought to you by Integration Service connectors.
And...
Speakers:
Akshay Agnihotri, Product Manager
Charlie Greenberg, Host
Securing your Kubernetes cluster_ a step-by-step guide to success !KatiaHIMEUR1
Today, after several years of existence, an extremely active community and an ultra-dynamic ecosystem, Kubernetes has established itself as the de facto standard in container orchestration. Thanks to a wide range of managed services, it has never been so easy to set up a ready-to-use Kubernetes cluster.
However, this ease of use means that the subject of security in Kubernetes is often left for later, or even neglected. This exposes companies to significant risks.
In this talk, I'll show you step-by-step how to secure your Kubernetes cluster for greater peace of mind and reliability.
The Art of the Pitch: WordPress Relationships and SalesLaura Byrne
Clients don’t know what they don’t know. What web solutions are right for them? How does WordPress come into the picture? How do you make sure you understand scope and timeline? What do you do if sometime changes?
All these questions and more will be explored as we talk about matching clients’ needs with what your agency offers without pulling teeth or pulling your hair out. Practical tips, and strategies for successful relationship building that leads to closing the deal.
GraphRAG is All You need? LLM & Knowledge GraphGuy Korland
Guy Korland, CEO and Co-founder of FalkorDB, will review two articles on the integration of language models with knowledge graphs.
1. Unifying Large Language Models and Knowledge Graphs: A Roadmap.
https://arxiv.org/abs/2306.08302
2. Microsoft Research's GraphRAG paper and a review paper on various uses of knowledge graphs:
https://www.microsoft.com/en-us/research/blog/graphrag-unlocking-llm-discovery-on-narrative-private-data/
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
Essentials of Automations: Optimizing FME Workflows with ParametersSafe Software
Are you looking to streamline your workflows and boost your projects’ efficiency? Do you find yourself searching for ways to add flexibility and control over your FME workflows? If so, you’re in the right place.
Join us for an insightful dive into the world of FME parameters, a critical element in optimizing workflow efficiency. This webinar marks the beginning of our three-part “Essentials of Automation” series. This first webinar is designed to equip you with the knowledge and skills to utilize parameters effectively: enhancing the flexibility, maintainability, and user control of your FME projects.
Here’s what you’ll gain:
- Essentials of FME Parameters: Understand the pivotal role of parameters, including Reader/Writer, Transformer, User, and FME Flow categories. Discover how they are the key to unlocking automation and optimization within your workflows.
- Practical Applications in FME Form: Delve into key user parameter types including choice, connections, and file URLs. Allow users to control how a workflow runs, making your workflows more reusable. Learn to import values and deliver the best user experience for your workflows while enhancing accuracy.
- Optimization Strategies in FME Flow: Explore the creation and strategic deployment of parameters in FME Flow, including the use of deployment and geometry parameters, to maximize workflow efficiency.
- Pro Tips for Success: Gain insights on parameterizing connections and leveraging new features like Conditional Visibility for clarity and simplicity.
We’ll wrap up with a glimpse into future webinars, followed by a Q&A session to address your specific questions surrounding this topic.
Don’t miss this opportunity to elevate your FME expertise and drive your projects to new heights of efficiency.
Essentials of Automations: Optimizing FME Workflows with Parameters
How to Optimize Sales Analytics Using 10x the Data at 1/10th the Cost
1. HOW TO OPTIMIZE SALES
ANALYTICS USING 10X THE DATA
AT 1/10TH THE COST
2. 2
Today, we’ll learn how to
● Perform sales analysis on billions of rows of data
● Add new dimensions and hierarchies for drill-down analysis in seconds
● Build “what-if” analyses using 20x faster using OLAP
● Analyze near real-time data for more accurate KPIs
● Reported consistent KPIs across BI tools
3. 3
Today’s Speakers
AD of Product, Wayfair
@wayfairtech
Matt is the Associate Director of
Product Management on the Data
Infrastructure team at Wayfair where
he has worked for over 5 years.
Prior to Wayfair, he worked at
Verifone and Curb in product
management roles.
He’s a graduate of Dickinson and
holds a Data Science specialization
from Johns Hopkins.
Matt Hartwig
Chief Strategy Officer, AtScale
@dmariani
Dave is one of the co-founders of
AtScale and is currently the Chief
Strategy Officer.
Prior to AtScale, Dave was VP of
Engineering at Klout & at Yahoo!
where he built the world's largest
multi-dimensional cube for BI on
Hadoop.
Dave is a Big Data visionary & serial
entrepreneur.
Dave Mariani
4. What is Wayfair?
4
Wayfair is a Clear Leader in Home Goods
~$600B+
total addressable market
rapidly moving from
brick and mortar to
online
Utilizing in-house
software development
capabilities to build and
leverage proprietary
technology
Highly recognized brand in
North America and Europe
with increasing
engagement from repeat
customers
Partnering with fragmented
and largely unbranded
supplier base of over
12,000 suppliers
Investing in specialized
logistics network,
international markets, and
existing teams to continue
outsized share-taking
Co-founders are largest
shareholders, with focus on
sustainable long-term
growth, operational
discipline, and customer-
first orientation
6. 6
Optimizing Sales Analytics
What do I do?
Data Infrastructure
Team
applications/users
Data Platform
Access Enrich
Store
Product Merch Ad Tech Storefront Operations BI/ DS
core data platform
We provide
application datastores,
data movement, and
analytics & data science
tools to enable developers and
analysts across Wayfair to store, secure,
enrich, and present data.
7. 7
Optimizing Sales Analytics
What is Velocity:
Velocity is how we talk about speed and scale in all things at Wayfair - design,
development, decision making, etc. It’s not enough to grow today; it’s about building
our growth in a sustainable way that enables continued momentum.
What about for Data:
The speed with which Wayfair can go from data collection to driving business
decisions, outcomes and insights.
8. 8
Optimizing Sales Analytics
Storefront Analysis Decision
A customer clicks on a
product page but
doesn’t proceed to
purchase
An analyst identifies
that we’re seeing lower
conversion rate after a
recent deploy
We roll back that recent
deploy and see
conversion rate recover
to previous baseline
Now do this better and faster on repeat with ever
increasing system complexity, size, and organizational
sprawl. That is Data Velocity.
9. 9
Optimizing Sales Analytics
Big data..
Typical Problems
1
Data Everywhere
Existing data warehouse and data lake systems
store hundreds of thousands of data sets, many
of which were copies of one another and not
intended for others to use. Hard to find what
you need.
Long lead times
Scaling on-premise infrastructure had long lead
times, challenges with physical hardware,
power/network constraints.
2
3
Fragmented Tool Space
Mix of legacy BI tools, relational databases, and
open-source big data tooling.
Fragmented IAM
Patchwork access control, no central identity
provider. Employees often stuck in ticket hell.
Rapid Data Volume Growth
Over 100% YoY growth in both data volume
produced and data accessed.
4
5
10. 10
A lot goes into solving this at size and scale
Data Curation / Transformation:
That data is further enriched,
transformed, and curated downstream.
Often to power decision support and
business intelligence systems but also
other software apps.
Application Data Exchange: Data
needs to flow from production
applications into many downstream
processes across software, analytics,
and data science.
Self Service Tooling: Once data is
curated and enriched, it need to be
accessible through self-service BI Tools
that enable uniform and equal access to
data at Wayfair.
Data Literacy: Every employee at Wayfair needs to be
empowered to make data informed decisions through training and
support. Employees need opportunities to develop their data
instincts.
Scalable Infrastructure: At the base layer is infrastructure
that can power the exchange, enrichment, and access of our
data at increased and accelerating scale.
The Pillars of Data Velocity at Wayfair
11. 11
A lot goes into solving this at size and scale
Data Curation / Transformation:
That data is further enriched,
transformed, and curated downstream.
Often to power decision support and
business intelligence systems but also
other software apps.
Application Data Exchange: Data
needs to flow from production
applications into many downstream
processes across software, analytics,
and data science.
Self Service Tooling: Once data is
curated and enriched, it need to be
accessible through self-service BI Tools
that enable uniform and equal access to
data at Wayfair.
Data Literacy: Every employee at Wayfair needs to be
empowered to make data informed decisions through training and
support. Employees need opportunities to develop their data
instincts.
Scalable Infrastructure: At the base layer is infrastructure
that can power the exchange, enrichment, and access of our
data at increased and accelerating scale.
The Pillars of Data Velocity at Wayfair
13. 13
How To Optimize Sales Analytics
Using 10X the Data at
1/10th the Cost
Dave Mariani, Founder and Chief Strategy Officer, AtScale
14. The Cloud Analytics Stack
14
COMPONENT
CONSUMPTION
VISUALIZATION, ANALYSIS, REPORTING
SEMANTIC LAYER
QUERY ACCESS, FILTERING, MASKING, AUDITING
PREPARED DATA
DATA PROCESSING, MODELING
RAW DATA
DATA STORAGE, ENCRYPTION
DATA TRANSFORMATION
ETL,MERGING, AGGREGATION
LAYER (FUNCTION)
BI Tools AI/ML Tools Applications
Multi-dimensional Engine
Data Governance Engine
Virtualization Engine
Data Warehouse File Access Engine
ETL Engine
File System (Data Lake)
Data
Catalog
15. Today’s Use Case
15
Using Excel, create a model that will forecast inventory
quantities for the 2020-Q4 using SafeGraph’s foot
traffic data
17. Challenge #1: Data Integration is Slow & Cumbersome
17
DEMOED SOLUTION
Leverage data virtualization to access data quickly & easily
ALTERNATIVES
1. Build a data pipeline using tools like Hive, Databricks, etc.
2. Use ETL/ELT tools like Informatica, Talend, Matillion, etc.
19. Challenge #2: Complex Calculations are Hard to Share
19
DEMOED SOLUTION
Leverage OLAP & MDX to compute calculations server-side
ALTERNATIVES
1. Use Excel spreadsheets to compute cell-based calculations
2. Use advanced SQL functions to calculate metrics
21. Challenge #3: Getting Up to Date Data is Slow & Manual
21
DEMOED SOLUTION
Leverage Time Relative functions & direct connections to data
ALTERNATIVES
1. Update data manually by repeating data preparation
2. Build logic & data prep into a custom application
22. Summary
22
▵ Leverage virtualization to deliver faster time to insight
▵ Leverage OLAP to share “single source of truth” calculations
▵ Leverage “live” (direct) data connections to reduce data latency
▵ Build upon a cloud-based, scalable data platform
AtScale is built to leverage the efficiencies and performance of the cloud for the data consumer whether you’re on premise or in the cloud (or both).
We connect people to data. We do that without moving data and without complexity—leveraging existing investments in big data platforms, applications and tools.
We also do that consistently, securely and with one set of semantics—and without interrupting existing data usage so that data workers no longer have to understand how or where it is stored.
Performance
Optimizing performance is difficult and that’s where we focus our energies. AtScale’s data warehouse virtualization can reduce queries performance from 5 weeks to 5 seconds—automatically optimizing each time a user queries the database.
Security
Because we haven’t copied the data and applied new code or embedded rules, we’ve reduced the amount of complexity and maintain consistent data lineage throughout the data lifecycle. AtScale not only leverages existing data security and governance but applies an additional layer so that data can be ported to new data tools, applications and platforms.
Agility
What’s more powerful is we create simple interface to querying data and building models for data science and analytics data workers with deep integrations with BI and AI/ML tools. For the first time, users (and IT) have visibilities into how data is being queried and used throughout the organization (no more data silos).
Today we'll show you how to increase your data velocity to report on sales. This will include reporting on billions of rows of data in a popular BI tool that can be used across the business and performs at conversational speeds