Recently the concept of a ‘data mesh’ was introduced by Zhamak Deghani to solve architectural and organizational challenges with getting value from data at scale more logically and efficiently, built around four principles:
* Domain-oriented decentralized data ownership
* Data as a product
* Self-serve data infrastructure as a platform
* Federated computational governance
This presentation will initially deep-dive into the ‘data mesh’ and how it fundamentally differs from the typical data lake architectures used today. Subsequently, it describes OLX Europe’s current data platform state aimed partially towards a more decentralized data architecture, covering its analytical data platform, data infrastructure, data discovery, and data privacy.
Finally, it will see to what extent the main principles around the ‘data mesh’ can be applied to a future vision for our data platform and what advantages and challenges implementing such a vision can bring for OLX and other companies.
For more information on data mesh principles, check out the original article by Zhamak: https://martinfowler.com/articles/data-mesh-principles.html.
This is Part 4 of the GoldenGate series on Data Mesh - a series of webinars helping customers understand how to move off of old-fashioned monolithic data integration architecture and get ready for more agile, cost-effective, event-driven solutions. The Data Mesh is a kind of Data Fabric that emphasizes business-led data products running on event-driven streaming architectures, serverless, and microservices based platforms. These emerging solutions are essential for enterprises that run data-driven services on multi-cloud, multi-vendor ecosystems.
Join this session to get a fresh look at Data Mesh; we'll start with core architecture principles (vendor agnostic) and transition into detailed examples of how Oracle's GoldenGate platform is providing capabilities today. We will discuss essential technical characteristics of a Data Mesh solution, and the benefits that business owners can expect by moving IT in this direction. For more background on Data Mesh, Part 1, 2, and 3 are on the GoldenGate YouTube channel: https://www.youtube.com/playlist?list=PLbqmhpwYrlZJ-583p3KQGDAd6038i1ywe
Webinar Speaker: Jeff Pollock, VP Product (https://www.linkedin.com/in/jtpollock/)
Mr. Pollock is an expert technology leader for data platforms, big data, data integration and governance. Jeff has been CTO at California startups and a senior exec at Fortune 100 tech vendors. He is currently Oracle VP of Products and Cloud Services for Data Replication, Streaming Data and Database Migrations. While at IBM, he was head of all Information Integration, Replication and Governance products, and previously Jeff was an independent architect for US Defense Department, VP of Technology at Cerebra and CTO of Modulant – he has been engineering artificial intelligence based data platforms since 2001. As a business consultant, Mr. Pollock was a Head Architect at Ernst & Young’s Center for Technology Enablement. Jeff is also the author of “Semantic Web for Dummies” and "Adaptive Information,” a frequent keynote at industry conferences, author for books and industry journals, formerly a contributing member of W3C and OASIS, and an engineering instructor with UC Berkeley’s Extension for object-oriented systems, software development process and enterprise architecture.
Presentation on Data Mesh: The paradigm shift is a new type of eco-system architecture, which is a shift left towards a modern distributed architecture in which it allows domain-specific data and views “data-as-a-product,” enabling each domain to handle its own data pipelines.
Enabling a Data Mesh Architecture with Data VirtualizationDenodo
Watch full webinar here: https://bit.ly/3rwWhyv
The Data Mesh architectural design was first proposed in 2019 by Zhamak Dehghani, principal technology consultant at Thoughtworks, a technology company that is closely associated with the development of distributed agile methodology. A data mesh is a distributed, de-centralized data infrastructure in which multiple autonomous domains manage and expose their own data, called “data products,” to the rest of the organization.
Organizations leverage data mesh architecture when they experience shortcomings in highly centralized architectures, such as the lack domain-specific expertise in data teams, the inflexibility of centralized data repositories in meeting the specific needs of different departments within large organizations, and the slow nature of centralized data infrastructures in provisioning data and responding to changes.
In this session, Pablo Alvarez, Global Director of Product Management at Denodo, explains how data virtualization is your best bet for implementing an effective data mesh architecture.
You will learn:
- How data mesh architecture not only enables better performance and agility, but also self-service data access
- The requirements for “data products” in the data mesh world, and how data virtualization supports them
- How data virtualization enables domains in a data mesh to be truly autonomous
- Why a data lake is not automatically a data mesh
- How to implement a simple, functional data mesh architecture using data virtualization
this is part 3 of the series on Data Mesh ... looking at the intersection of microservices architecture concepts, data integration / replication technologies and log-based stream integration techniques. This webinar was mostly a demonstration, but several slides used to setup the demo are included here as a PDF for viewers.
Architect’s Open-Source Guide for a Data Mesh ArchitectureDatabricks
Data Mesh is an innovative concept addressing many data challenges from an architectural, cultural, and organizational perspective. But is the world ready to implement Data Mesh?
In this session, we will review the importance of core Data Mesh principles, what they can offer, and when it is a good idea to try a Data Mesh architecture. We will discuss common challenges with implementation of Data Mesh systems and focus on the role of open-source projects for it. Projects like Apache Spark can play a key part in standardized infrastructure platform implementation of Data Mesh. We will examine the landscape of useful data engineering open-source projects to utilize in several areas of a Data Mesh system in practice, along with an architectural example. We will touch on what work (culture, tools, mindset) needs to be done to ensure Data Mesh is more accessible for engineers in the industry.
The audience will leave with a good understanding of the benefits of Data Mesh architecture, common challenges, and the role of Apache Spark and other open-source projects for its implementation in real systems.
This session is targeted for architects, decision-makers, data-engineers, and system designers.
Data Lakehouse, Data Mesh, and Data Fabric (r1)James Serra
So many buzzwords of late: Data Lakehouse, Data Mesh, and Data Fabric. What do all these terms mean and how do they compare to a data warehouse? In this session I’ll cover all of them in detail and compare the pros and cons of each. I’ll include use cases so you can see what approach will work best for your big data needs.
This is Part 4 of the GoldenGate series on Data Mesh - a series of webinars helping customers understand how to move off of old-fashioned monolithic data integration architecture and get ready for more agile, cost-effective, event-driven solutions. The Data Mesh is a kind of Data Fabric that emphasizes business-led data products running on event-driven streaming architectures, serverless, and microservices based platforms. These emerging solutions are essential for enterprises that run data-driven services on multi-cloud, multi-vendor ecosystems.
Join this session to get a fresh look at Data Mesh; we'll start with core architecture principles (vendor agnostic) and transition into detailed examples of how Oracle's GoldenGate platform is providing capabilities today. We will discuss essential technical characteristics of a Data Mesh solution, and the benefits that business owners can expect by moving IT in this direction. For more background on Data Mesh, Part 1, 2, and 3 are on the GoldenGate YouTube channel: https://www.youtube.com/playlist?list=PLbqmhpwYrlZJ-583p3KQGDAd6038i1ywe
Webinar Speaker: Jeff Pollock, VP Product (https://www.linkedin.com/in/jtpollock/)
Mr. Pollock is an expert technology leader for data platforms, big data, data integration and governance. Jeff has been CTO at California startups and a senior exec at Fortune 100 tech vendors. He is currently Oracle VP of Products and Cloud Services for Data Replication, Streaming Data and Database Migrations. While at IBM, he was head of all Information Integration, Replication and Governance products, and previously Jeff was an independent architect for US Defense Department, VP of Technology at Cerebra and CTO of Modulant – he has been engineering artificial intelligence based data platforms since 2001. As a business consultant, Mr. Pollock was a Head Architect at Ernst & Young’s Center for Technology Enablement. Jeff is also the author of “Semantic Web for Dummies” and "Adaptive Information,” a frequent keynote at industry conferences, author for books and industry journals, formerly a contributing member of W3C and OASIS, and an engineering instructor with UC Berkeley’s Extension for object-oriented systems, software development process and enterprise architecture.
Presentation on Data Mesh: The paradigm shift is a new type of eco-system architecture, which is a shift left towards a modern distributed architecture in which it allows domain-specific data and views “data-as-a-product,” enabling each domain to handle its own data pipelines.
Enabling a Data Mesh Architecture with Data VirtualizationDenodo
Watch full webinar here: https://bit.ly/3rwWhyv
The Data Mesh architectural design was first proposed in 2019 by Zhamak Dehghani, principal technology consultant at Thoughtworks, a technology company that is closely associated with the development of distributed agile methodology. A data mesh is a distributed, de-centralized data infrastructure in which multiple autonomous domains manage and expose their own data, called “data products,” to the rest of the organization.
Organizations leverage data mesh architecture when they experience shortcomings in highly centralized architectures, such as the lack domain-specific expertise in data teams, the inflexibility of centralized data repositories in meeting the specific needs of different departments within large organizations, and the slow nature of centralized data infrastructures in provisioning data and responding to changes.
In this session, Pablo Alvarez, Global Director of Product Management at Denodo, explains how data virtualization is your best bet for implementing an effective data mesh architecture.
You will learn:
- How data mesh architecture not only enables better performance and agility, but also self-service data access
- The requirements for “data products” in the data mesh world, and how data virtualization supports them
- How data virtualization enables domains in a data mesh to be truly autonomous
- Why a data lake is not automatically a data mesh
- How to implement a simple, functional data mesh architecture using data virtualization
this is part 3 of the series on Data Mesh ... looking at the intersection of microservices architecture concepts, data integration / replication technologies and log-based stream integration techniques. This webinar was mostly a demonstration, but several slides used to setup the demo are included here as a PDF for viewers.
Architect’s Open-Source Guide for a Data Mesh ArchitectureDatabricks
Data Mesh is an innovative concept addressing many data challenges from an architectural, cultural, and organizational perspective. But is the world ready to implement Data Mesh?
In this session, we will review the importance of core Data Mesh principles, what they can offer, and when it is a good idea to try a Data Mesh architecture. We will discuss common challenges with implementation of Data Mesh systems and focus on the role of open-source projects for it. Projects like Apache Spark can play a key part in standardized infrastructure platform implementation of Data Mesh. We will examine the landscape of useful data engineering open-source projects to utilize in several areas of a Data Mesh system in practice, along with an architectural example. We will touch on what work (culture, tools, mindset) needs to be done to ensure Data Mesh is more accessible for engineers in the industry.
The audience will leave with a good understanding of the benefits of Data Mesh architecture, common challenges, and the role of Apache Spark and other open-source projects for its implementation in real systems.
This session is targeted for architects, decision-makers, data-engineers, and system designers.
Data Lakehouse, Data Mesh, and Data Fabric (r1)James Serra
So many buzzwords of late: Data Lakehouse, Data Mesh, and Data Fabric. What do all these terms mean and how do they compare to a data warehouse? In this session I’ll cover all of them in detail and compare the pros and cons of each. I’ll include use cases so you can see what approach will work best for your big data needs.
Uma introdução à malha de dados e as motivações por trás dela: os modos de falhas de paradigmas anteriores de gerenciamento de big data. A proposta de Zhamak Dehghani é comparar e contrastar a malha de dados com as abordagens existentes de gerenciamento de big data, apresentando os componentes técnicos que sustentam a arquitetura de software.
Wonder what this data mesh stuff is all about? What are the principles of data mesh? Can you or should you consider data mesh as the approach for your analytics platform? And most important - how can Snowflake help?
Given in Montreal on 14-Dec-2021
Modernizing to a Cloud Data ArchitectureDatabricks
Organizations with on-premises Hadoop infrastructure are bogged down by system complexity, unscalable infrastructure, and the increasing burden on DevOps to manage legacy architectures. Costs and resource utilization continue to go up while innovation has flatlined. In this session, you will learn why, now more than ever, enterprises are looking for cloud alternatives to Hadoop and are migrating off of the architecture in large numbers. You will also learn how elastic compute models’ benefits help one customer scale their analytics and AI workloads and best practices from their experience on a successful migration of their data and workloads to the cloud.
Watch full webinar here: https://bit.ly/3mdj9i7
You will often hear that "data is the new gold"? In this context, data management is one of the areas that has received more attention from the software community in recent years. From Artificial Intelligence and Machine Learning to new ways to store and process data, the landscape for data management is in constant evolution. From the privileged perspective of an enterprise middleware platform, we at Denodo have the advantage of seeing many of these changes happen.
In this webinar, we will discuss the technology trends that will drive the enterprise data strategies in the years to come. Don't miss it if you want to keep yourself informed about how to convert your data to strategic assets in order to complete the data-driven transformation in your company.
Watch this on-demand webinar as we cover:
- The most interesting trends in data management
- How to build a data fabric architecture?
- How to manage your data integration strategy in the new hybrid world
- Our predictions on how those trends will change the data management world
- How can companies monetize the data through data-as-a-service infrastructure?
- What is the role of voice computing in future data analytic
Five Things to Consider About Data Mesh and Data GovernanceDATAVERSITY
Data mesh was among the most discussed and controversial enterprise data management topics of 2021. One of the reasons people struggle with data mesh concepts is we still have a lot of open questions that we are not thinking about:
Are you thinking beyond analytics? Are you thinking about all possible stakeholders? Are you thinking about how to be agile? Are you thinking about standardization and policies? Are you thinking about organizational structures and roles?
Join data.world VP of Product Tim Gasper and Principal Scientist Juan Sequeda for an honest, no-bs discussion about data mesh and its role in data governance.
Intuit's Data Mesh - Data Mesh Leaning Community meetup 5.13.2021Tristan Baker
Past, present and future of data mesh at Intuit. This deck describes a vision and strategy for improving data worker productivity through a Data Mesh approach to organizing data and holding data producers accountable. Delivered at the inaugural Data Mesh Leaning meetup on 5/13/2021.
Master the Multi-Clustered Data Warehouse - SnowflakeMatillion
Snowflake is one of the most powerful, efficient data warehouses on the market today—and we joined forces with the Snowflake team to show you how it works!
In this webinar:
- Learn how to optimize Snowflake
- Hear insider tips and tricks on how to improve performance
- Get expert insights from Craig Collier, Technical Architect from Snowflake, and Kalyan Arangam, Solution Architect from Matillion
- Find out how leading brands like Converse, Duo Security, and Pets at Home use Snowflake and Matillion ETL to make data-driven decisions
- Discover how Matillion ETL and Snowflake work together to modernize your data world
- Learn how to utilize the impressive scalability of Snowflake and Matillion
A Work of Zhamak Dehghani
Principal consultant
ThoughtWorks
https://martinfowler.com/articles/data-monolith-to-mesh.html
https://fast.wistia.net/embed/iframe/vys2juvzc3?videoFoam
How to Move Beyond a Monolithic Data Lake to a Distributed Data Mesh
Many enterprises are investing in their next generation data lake, with the hope of democratizing data at scale to provide business insights and ultimately make automated intelligent decisions. Data platforms based on the data lake architecture have common failure modes that lead to unfulfilled promises at scale. To address these failure modes we need to shift from the centralized paradigm of a lake, or its predecessor data warehouse. We need to shift to a paradigm that draws from modern distributed architecture: considering domains as the first class concern, applying platform thinking to create self-serve data infrastructure, and treating data as a product.
The world of data architecture began with applications. Next came data warehouses. Then text was organized into a data warehouse.
Then one day the world discovered a whole new kind of data that was being generated by organizations. The world found that machines generated data that could be transformed into valuable insights. This was the origin of what is today called the data lakehouse. The evolution of data architecture continues today.
Come listen to industry experts describe this transformation of ordinary data into a data architecture that is invaluable to business. Simply put, organizations that take data architecture seriously are going to be at the forefront of business tomorrow.
This is an educational event.
Several of the authors of the book Building the Data Lakehouse will be presenting at this symposium.
Data Architecture, Solution Architecture, Platform Architecture — What’s the ...DATAVERSITY
A solid data architecture is critical to the success of any data initiative. But what is meant by “data architecture”? Throughout the industry, there are many different “flavors” of data architecture, each with its own unique value and use cases for describing key aspects of the data landscape. Join this webinar to demystify the various architecture styles and understand how they can add value to your organization.
[DSC Europe 22] Overview of the Databricks Platform - Petar ZecevicDataScienceConferenc1
Databricks' founders caused a seismic shift in data analysis community when they created Apache Spark which has become a cornerstone of Big Data processing pipelines and tools in large and small companies all around the world. Now they've built a revolutionary, comprehensive and easy-to-use platform around Apache Spark and their other inventions, such as MLFlow and Koalas frameworks and most importantly the Data Lakehouse: a concept of fusing data warehouse and data lake architectures into a single versatile and fast platform. Technical foundation for Databricks Data Lakehouse is Delta Lake. More than 7000 organizations today rely on Databricks to enable massive-scale data engineering, collaborative data science, full-lifecycle machine learning and business analytics. Come to the talk and see the demo to find out why.
Data Mesh is a new socio-technical approach to data architecture, first described by Zhamak Dehghani and popularised through a guest blog post on Martin Fowler's site.
Since then, community interest has grown, due to Data Mesh's ability to explain and address the frustrations that many organisations are experiencing as they try to get value from their data. The 2022 publication of Zhamak's book on Data Mesh further provoked conversation, as have the growing number of experience reports from companies that have put Data Mesh into practice.
So what's all the fuss about?
On one hand, Data Mesh is a new approach in the field of big data. On the other hand, Data Mesh is application of the lessons we have learned from domain-driven design and microservices to a data context.
In this talk, Chris and Pablo will explain how Data Mesh relates to current thinking in software architecture and the historical development of data architecture philosophies. They will outline what benefits Data Mesh brings, what trade-offs it comes with and when organisations should and should not consider adopting it.
[DSC Europe 22] Lakehouse architecture with Delta Lake and Databricks - Draga...DataScienceConferenc1
Dragan Berić will take a deep dive into Lakehouse architecture, a game-changing concept bridging the best elements of data lake and data warehouse. The presentation will focus on the Delta Lake format as the foundation of the Lakehouse philosophy, and Databricks as the primary platform for its implementation.
Every business today wants to leverage data to drive strategic initiatives with machine learning, data science and analytics — but runs into challenges from siloed teams, proprietary technologies and unreliable data.
That’s why enterprises are turning to the lakehouse because it offers a single platform to unify all your data, analytics and AI workloads.
Join our How to Build a Lakehouse technical training, where we’ll explore how to use Apache SparkTM, Delta Lake, and other open source technologies to build a better lakehouse. This virtual session will include concepts, architectures and demos.
Here’s what you’ll learn in this 2-hour session:
How Delta Lake combines the best of data warehouses and data lakes for improved data reliability, performance and security
How to use Apache Spark and Delta Lake to perform ETL processing, manage late-arriving data, and repair corrupted data directly on your lakehouse
Myth Busters VII: I’m building a data mesh, so I don’t need data virtualizationDenodo
Watch full webinar here: https://bit.ly/3DBA4EP
A data mesh architecture offers a lot of promise to change the way we manage data – and for the better. But there’s a lot of confusion about a data mesh. People will tell you that you can build a data mesh on top of a data lake or on top of a data warehouse, and that you don’t need data virtualization to build a data mesh.
Many vendors are jumping on to the data mesh bandwagon and are claiming that they inherently support a data mesh architecture. But do they? How much of this is hype versus reality? Is it true that you don’t need data virtualization to build a scalable, enterprise-grade data mesh?
This is the myth we will attempt to bust in this next Myth Busters webinar.
Watch this session on-demand to learn about the concepts and components of a data mesh, and hear how the logical approach to data management and integration – powered by data virtualization - is critical for a data mesh.
Unlock Your Data for ML & AI using Data VirtualizationDenodo
How Denodo Complement’s Logical Data Lake in Cloud
● Denodo does not substitute data warehouses, data lakes,
ETLs...
● Denodo enables the use of all together plus other data
sources
○ In a logical data warehouse
○ In a logical data lake
○ They are very similar, the only difference is in the main
objective
● There are also use cases where Denodo can be used as data
source in a ETL flow
Uma introdução à malha de dados e as motivações por trás dela: os modos de falhas de paradigmas anteriores de gerenciamento de big data. A proposta de Zhamak Dehghani é comparar e contrastar a malha de dados com as abordagens existentes de gerenciamento de big data, apresentando os componentes técnicos que sustentam a arquitetura de software.
Wonder what this data mesh stuff is all about? What are the principles of data mesh? Can you or should you consider data mesh as the approach for your analytics platform? And most important - how can Snowflake help?
Given in Montreal on 14-Dec-2021
Modernizing to a Cloud Data ArchitectureDatabricks
Organizations with on-premises Hadoop infrastructure are bogged down by system complexity, unscalable infrastructure, and the increasing burden on DevOps to manage legacy architectures. Costs and resource utilization continue to go up while innovation has flatlined. In this session, you will learn why, now more than ever, enterprises are looking for cloud alternatives to Hadoop and are migrating off of the architecture in large numbers. You will also learn how elastic compute models’ benefits help one customer scale their analytics and AI workloads and best practices from their experience on a successful migration of their data and workloads to the cloud.
Watch full webinar here: https://bit.ly/3mdj9i7
You will often hear that "data is the new gold"? In this context, data management is one of the areas that has received more attention from the software community in recent years. From Artificial Intelligence and Machine Learning to new ways to store and process data, the landscape for data management is in constant evolution. From the privileged perspective of an enterprise middleware platform, we at Denodo have the advantage of seeing many of these changes happen.
In this webinar, we will discuss the technology trends that will drive the enterprise data strategies in the years to come. Don't miss it if you want to keep yourself informed about how to convert your data to strategic assets in order to complete the data-driven transformation in your company.
Watch this on-demand webinar as we cover:
- The most interesting trends in data management
- How to build a data fabric architecture?
- How to manage your data integration strategy in the new hybrid world
- Our predictions on how those trends will change the data management world
- How can companies monetize the data through data-as-a-service infrastructure?
- What is the role of voice computing in future data analytic
Five Things to Consider About Data Mesh and Data GovernanceDATAVERSITY
Data mesh was among the most discussed and controversial enterprise data management topics of 2021. One of the reasons people struggle with data mesh concepts is we still have a lot of open questions that we are not thinking about:
Are you thinking beyond analytics? Are you thinking about all possible stakeholders? Are you thinking about how to be agile? Are you thinking about standardization and policies? Are you thinking about organizational structures and roles?
Join data.world VP of Product Tim Gasper and Principal Scientist Juan Sequeda for an honest, no-bs discussion about data mesh and its role in data governance.
Intuit's Data Mesh - Data Mesh Leaning Community meetup 5.13.2021Tristan Baker
Past, present and future of data mesh at Intuit. This deck describes a vision and strategy for improving data worker productivity through a Data Mesh approach to organizing data and holding data producers accountable. Delivered at the inaugural Data Mesh Leaning meetup on 5/13/2021.
Master the Multi-Clustered Data Warehouse - SnowflakeMatillion
Snowflake is one of the most powerful, efficient data warehouses on the market today—and we joined forces with the Snowflake team to show you how it works!
In this webinar:
- Learn how to optimize Snowflake
- Hear insider tips and tricks on how to improve performance
- Get expert insights from Craig Collier, Technical Architect from Snowflake, and Kalyan Arangam, Solution Architect from Matillion
- Find out how leading brands like Converse, Duo Security, and Pets at Home use Snowflake and Matillion ETL to make data-driven decisions
- Discover how Matillion ETL and Snowflake work together to modernize your data world
- Learn how to utilize the impressive scalability of Snowflake and Matillion
A Work of Zhamak Dehghani
Principal consultant
ThoughtWorks
https://martinfowler.com/articles/data-monolith-to-mesh.html
https://fast.wistia.net/embed/iframe/vys2juvzc3?videoFoam
How to Move Beyond a Monolithic Data Lake to a Distributed Data Mesh
Many enterprises are investing in their next generation data lake, with the hope of democratizing data at scale to provide business insights and ultimately make automated intelligent decisions. Data platforms based on the data lake architecture have common failure modes that lead to unfulfilled promises at scale. To address these failure modes we need to shift from the centralized paradigm of a lake, or its predecessor data warehouse. We need to shift to a paradigm that draws from modern distributed architecture: considering domains as the first class concern, applying platform thinking to create self-serve data infrastructure, and treating data as a product.
The world of data architecture began with applications. Next came data warehouses. Then text was organized into a data warehouse.
Then one day the world discovered a whole new kind of data that was being generated by organizations. The world found that machines generated data that could be transformed into valuable insights. This was the origin of what is today called the data lakehouse. The evolution of data architecture continues today.
Come listen to industry experts describe this transformation of ordinary data into a data architecture that is invaluable to business. Simply put, organizations that take data architecture seriously are going to be at the forefront of business tomorrow.
This is an educational event.
Several of the authors of the book Building the Data Lakehouse will be presenting at this symposium.
Data Architecture, Solution Architecture, Platform Architecture — What’s the ...DATAVERSITY
A solid data architecture is critical to the success of any data initiative. But what is meant by “data architecture”? Throughout the industry, there are many different “flavors” of data architecture, each with its own unique value and use cases for describing key aspects of the data landscape. Join this webinar to demystify the various architecture styles and understand how they can add value to your organization.
[DSC Europe 22] Overview of the Databricks Platform - Petar ZecevicDataScienceConferenc1
Databricks' founders caused a seismic shift in data analysis community when they created Apache Spark which has become a cornerstone of Big Data processing pipelines and tools in large and small companies all around the world. Now they've built a revolutionary, comprehensive and easy-to-use platform around Apache Spark and their other inventions, such as MLFlow and Koalas frameworks and most importantly the Data Lakehouse: a concept of fusing data warehouse and data lake architectures into a single versatile and fast platform. Technical foundation for Databricks Data Lakehouse is Delta Lake. More than 7000 organizations today rely on Databricks to enable massive-scale data engineering, collaborative data science, full-lifecycle machine learning and business analytics. Come to the talk and see the demo to find out why.
Data Mesh is a new socio-technical approach to data architecture, first described by Zhamak Dehghani and popularised through a guest blog post on Martin Fowler's site.
Since then, community interest has grown, due to Data Mesh's ability to explain and address the frustrations that many organisations are experiencing as they try to get value from their data. The 2022 publication of Zhamak's book on Data Mesh further provoked conversation, as have the growing number of experience reports from companies that have put Data Mesh into practice.
So what's all the fuss about?
On one hand, Data Mesh is a new approach in the field of big data. On the other hand, Data Mesh is application of the lessons we have learned from domain-driven design and microservices to a data context.
In this talk, Chris and Pablo will explain how Data Mesh relates to current thinking in software architecture and the historical development of data architecture philosophies. They will outline what benefits Data Mesh brings, what trade-offs it comes with and when organisations should and should not consider adopting it.
[DSC Europe 22] Lakehouse architecture with Delta Lake and Databricks - Draga...DataScienceConferenc1
Dragan Berić will take a deep dive into Lakehouse architecture, a game-changing concept bridging the best elements of data lake and data warehouse. The presentation will focus on the Delta Lake format as the foundation of the Lakehouse philosophy, and Databricks as the primary platform for its implementation.
Every business today wants to leverage data to drive strategic initiatives with machine learning, data science and analytics — but runs into challenges from siloed teams, proprietary technologies and unreliable data.
That’s why enterprises are turning to the lakehouse because it offers a single platform to unify all your data, analytics and AI workloads.
Join our How to Build a Lakehouse technical training, where we’ll explore how to use Apache SparkTM, Delta Lake, and other open source technologies to build a better lakehouse. This virtual session will include concepts, architectures and demos.
Here’s what you’ll learn in this 2-hour session:
How Delta Lake combines the best of data warehouses and data lakes for improved data reliability, performance and security
How to use Apache Spark and Delta Lake to perform ETL processing, manage late-arriving data, and repair corrupted data directly on your lakehouse
Myth Busters VII: I’m building a data mesh, so I don’t need data virtualizationDenodo
Watch full webinar here: https://bit.ly/3DBA4EP
A data mesh architecture offers a lot of promise to change the way we manage data – and for the better. But there’s a lot of confusion about a data mesh. People will tell you that you can build a data mesh on top of a data lake or on top of a data warehouse, and that you don’t need data virtualization to build a data mesh.
Many vendors are jumping on to the data mesh bandwagon and are claiming that they inherently support a data mesh architecture. But do they? How much of this is hype versus reality? Is it true that you don’t need data virtualization to build a scalable, enterprise-grade data mesh?
This is the myth we will attempt to bust in this next Myth Busters webinar.
Watch this session on-demand to learn about the concepts and components of a data mesh, and hear how the logical approach to data management and integration – powered by data virtualization - is critical for a data mesh.
Unlock Your Data for ML & AI using Data VirtualizationDenodo
How Denodo Complement’s Logical Data Lake in Cloud
● Denodo does not substitute data warehouses, data lakes,
ETLs...
● Denodo enables the use of all together plus other data
sources
○ In a logical data warehouse
○ In a logical data lake
○ They are very similar, the only difference is in the main
objective
● There are also use cases where Denodo can be used as data
source in a ETL flow
Data Mesh in Practice: How Europe’s Leading Online Platform for Fashion Goes ...Databricks
The Data Lake paradigm is often considered the scalable successor of the more curated Data Warehouse approach when it comes to democratization of data. However, many who went out to build a centralized Data Lake came out with a data swamp of unclear responsibilities, a lack of data ownership, and sub-par data availability.
Watch here: https://bit.ly/3i2iJbu
You will often hear that "data is the new gold". In this context, data management is one of the areas that has received more attention by the software community in recent years. From Artificial Intelligence and Machine Learning to new ways to store and process data, the landscape for data management is in constant evolution. From the privileged perspective of an enterprise middleware platform, we at Denodo have the advantage of seeing many of these changes happen.
Join us for an exciting session that will cover:
- The most interesting trends in data management.
- Our predictions on how those trends will change the data management world.
- How these trends are shaping the future of data virtualization and our own software.
Building a Logical Data Fabric using Data Virtualization (ASEAN)Denodo
Watch full webinar here: https://bit.ly/3FF1ubd
In the recent Building the Unified Data Warehouse and Data Lake report by leading industry analysts TDWI, we have discovered 64% of organizations stated the objective for a unified Data Warehouse and Data Lakes is to get more business value and 84% of organizations polled felt that a unified approach to Data Warehouses and Data Lakes was either extremely or moderately important.
In this session, you will learn how your organization can apply a logical data fabric and the associated technologies of machine learning, artificial intelligence, and data virtualization can reduce time to value. Hence, increasing the overall business value of your data assets.
KEY TAKEAWAYS:
- How a Logical Data Fabric is the right approach to assist organizations to unify their data.
- The advanced features of a Logical Data Fabric that assist with the democratization of data, providing an agile and governed approach to business analytics and data science.
- How a Logical Data Fabric with Data Virtualization enhances your legacy data integration landscape to simplify data access and encourage self-service.
Watch full webinar here: https://bit.ly/2SaBj5l
You will often hear that "data is the new gold". In this context, data management is one of the areas that has received more attention by the software community in recent years. From Artificial Intelligence and Machine Learning to new ways to store and process data, the landscape for data management is in constant evolution. From the privileged perspective of an enterprise middleware platform, we at Denodo have the advantage of seeing many of these changes happen.
In this webinar we will discuss the technology trends that will drive the enterprise data strategies in the years to come. Don't miss it if you want to keep yourself informed about how to convert your data to strategic assets in order to complete the data-driven transformation in your company.
Join us for an exciting session that will cover:
- The most interesting trends in data management
- How to build a logical data fabric architecture?
- How to manage your data integration strategy in the new hybrid world?
- Our predictions on how those trends will change the data management world
- How can companies monetize the data through data-as-a-service infrastructure?
- What is the role of the voice computing in the future of data analytics?
A Logical Architecture is Always a Flexible Architecture (ASEAN)Denodo
Watch full webinar here: https://bit.ly/3joZa0a
The current data landscape is fragmented, not just in location but also in terms of processing paradigms: data lakes, IoT architectures, NoSQL, and graph data stores, SaaS applications, etc. are found coexisting with relational databases to fuel the needs of modern analytics, ML, and AI. The physical consolidation of enterprise data into a central repository, although possible, is both expensive and time-consuming. A logical data warehouse is a modern data architecture that allows organizations to leverage all of their data irrespective of where the data is stored, what format it is stored in, and what technologies or protocols are used to store and access the data.
Watch this session to understand:
- What is a logical data warehouse and how to architect one
- The benefits of logical data warehouse – speed with agility
- Customer use case depicting logical architecture implementation
Using Data Platforms That Are Fit-For-PurposeDATAVERSITY
We must grow the data capabilities of our organization to fully deal with the many and varied forms of data. This cannot be accomplished without an intense focus on the many and growing technical bases that can be used to store, view, and manage data. There are many, now more than ever, that have merit in organizations today.
This session sorts out the valuable data stores, how they work, what workloads they are good for, and how to build the data foundation for a modern competitive enterprise.
When and How Data Lakes Fit into a Modern Data ArchitectureDATAVERSITY
Whether to take data ingestion cycles off the ETL tool and the data warehouse or to facilitate competitive Data Science and building algorithms in the organization, the data lake – a place for unmodeled and vast data – will be provisioned widely in 2020.
Though it doesn’t have to be complicated, the data lake has a few key design points that are critical, and it does need to follow some principles for success. Avoid building the data swamp, but not the data lake! The tool ecosystem is building up around the data lake and soon many will have a robust lake and data warehouse. We will discuss policy to keep them straight, send data to its best platform, and keep users’ confidence up in their data platforms.
Data lakes will be built in cloud object storage. We’ll discuss the options there as well.
Get this data point for your data lake journey.
DAMA & Denodo Webinar: Modernizing Data Architecture Using Data Virtualization Denodo
Watch here: https://bit.ly/2NGQD7R
In an era increasingly dominated by advancements in cloud computing, AI and advanced analytics it may come as a shock that many organizations still rely on data architectures built before the turn of the century. But that scenario is rapidly changing with the increasing adoption of real-time data virtualization - a paradigm shift in the approach that organizations take towards accessing, integrating, and provisioning data required to meet business goals.
As data analytics and data-driven intelligence takes centre stage in today’s digital economy, logical data integration across the widest variety of data sources, with proper security and governance structure in place has become mission-critical.
Attend this session to learn:
- Learn how you can meet cloud and data science challenges with data virtualization.
- Why data virtualization is increasingly finding enterprise-wide adoption
- Discover how customers are reducing costs and improving ROI with data virtualization
Data Mesh in Practice - How Europe's Leading Online Platform for Fashion Goes...Dr. Arif Wider
A talk presented by Max Schultze from Zalando and Arif Wider from ThoughtWorks at NDC Oslo 2020.
Abstract:
The Data Lake paradigm is often considered the scalable successor of the more curated Data Warehouse approach when it comes to democratization of data. However, many who went out to build a centralized Data Lake came out with a data swamp of unclear responsibilities, a lack of data ownership, and sub-par data availability.
At Zalando - europe’s biggest online fashion retailer - we realised that accessibility and availability at scale can only be guaranteed when moving more responsibilities to those who pick up the data and have the respective domain knowledge - the data owners - while keeping only data governance and metadata information central. Such a decentralized and domain focused approach has recently been coined a Data Mesh.
The Data Mesh paradigm promotes the concept of Data Products which go beyond sharing of files and towards guarantees of quality and acknowledgement of data ownership.
This talk will take you on a journey of how we went from a centralized Data Lake to embrace a distributed Data Mesh architecture and will outline the ongoing efforts to make creation of data products as simple as applying a template.
Watch full webinar here: https://bit.ly/3puUCIc
What is Data Virtualization and why do I care? In this webinar we intend to help you understand not only what Data Virtualization is but why it's a critical component of any organization's data fabric and how it fits. How data virtualization liberates and empowers your business users via data discovery, data wrangling to generation of reusable reporting objects and data services. Digital transformation demands that we empower all consumers of data within the organization, it also demands agility too. Data Virtualization gives you meaningful access to information that can be shared by a myriad of consumers.
Watch on-demand this session to learn:
- What is Data Virtualization?
- Why do I need Data Virtualization in my organization?
- How do I implement Data Virtualization in my enterprise? Where does it fit?
Your Data is Waiting. What are the Top 5 Trends for Data in 2022? (ASEAN)Denodo
Watch full webinar here: https://bit.ly/3saONRK
COVID-19 has pushed every industry and organization to embrace digital transformation at scale, upending the way many businesses will operate for the foreseeable future. Organizations no longer tolerate monolithic and centralized data architecture; they are embracing flexibility, modularity, and distributed data architecture to help drive innovation and modernize processes.
The pandemic has compelled organizations to accelerate their digital transformation initiatives and look for smarter and more agile ways to manage and leverage their corporate data assets. Data governance has become challenging in the ever-increasing complexity and distributed nature of the data ecosystem. Interoperability, collaboration and trust in data are imperative for a business to succeed. Data needs to be easily accessible and fit for purpose.
In this session, Denodo experts will discuss 5 key trends that are expected to be top of mind for CIOs and CDOs;
- Distributed Data Environments
- Decision Intelligence
- Modern Data Architecture
- Composable Data & Analytics
- Hyper-personalized Experiences
Businesses make critical decisions using key data assets, but stakeholders often find it difficult to navigate the complex data landscape to ensure they have the right data and understand it correctly. Companies are dealing with a number of different technologies, multiple data formats, and high data volumes, along with the requirements for data security and governance.
Analyst Webinar: Discover how a logical data fabric helps organizations avoid...Denodo
Watch full webinar here: https://bit.ly/3zVUXWp
In this webinar, we’ll be tackling the question of where our data is and how we can avoid it falling into a black hole.
We’ll examine how data blackholes and silos come to be and the challenges these pose to organisations. We will also look at the impact of data silos as organisations adopt more complex multi-cloud setups. Finally, we will discuss the opportunities a logical data fabric poses to assist organisations to avoid data silos and manage data in a centrally governed and controlled environment.
Join us and Barc’s Jacqueline Bloemen on this webinar to get the answer and further insights on how to better avoid falling into a #datablackhole. Hope to see you connected!
Bridging the Last Mile: Getting Data to the People Who Need ItDenodo
Watch full webinar here: https://bit.ly/3cUA0Qi
Many organizations are embarking on strategically important journeys to embrace data and analytics. The goal can be to improve internal efficiencies, improve the customer experience, drive new business models and revenue streams, or – in the public sector – provide better services. All of these goals require empowering employees to act on data and analytics and to make data-driven decisions. However, getting data – the right data at the right time – to these employees is a huge challenge and traditional technologies and data architectures are simply not up to this task. This webinar will look at how organizations are using Data Virtualization to quickly and efficiently get data to the people that need it.
Attend this session to learn:
- The challenges organizations face when trying to get data to the business users in a timely manner
- How Data Virtualization can accelerate time-to-value for an organization’s data assets
- Examples of leading companies that used data virtualization to get the right data to the users at the right time
It is a fascinating, explosive time for enterprise analytics.
It is from the position of analytics leadership that the enterprise mission will be executed and company leadership will emerge. The data professional is absolutely sitting on the performance of the company in this information economy and has an obligation to demonstrate the possibilities and originate the architecture, data, and projects that will deliver analytics. After all, no matter what business you’re in, you’re in the business of analytics.
The coming years will be full of big changes in enterprise analytics and data architecture. William will kick off the fifth year of the Advanced Analytics series with a discussion of the trends winning organizations should build into their plans, expectations, vision, and awareness now.
Similar to To mesh or mess up your data organisation - Jochem van Grondelle (Prosus/OLX Group) (20)
UiPath Test Automation using UiPath Test Suite series, part 3DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 3. In this session, we will cover desktop automation along with UI automation.
Topics covered:
UI automation Introduction,
UI automation Sample
Desktop automation flow
Pradeep Chinnala, Senior Consultant Automation Developer @WonderBotz and UiPath MVP
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Software Delivery At the Speed of AI: Inflectra Invests In AI-Powered QualityInflectra
In this insightful webinar, Inflectra explores how artificial intelligence (AI) is transforming software development and testing. Discover how AI-powered tools are revolutionizing every stage of the software development lifecycle (SDLC), from design and prototyping to testing, deployment, and monitoring.
Learn about:
• The Future of Testing: How AI is shifting testing towards verification, analysis, and higher-level skills, while reducing repetitive tasks.
• Test Automation: How AI-powered test case generation, optimization, and self-healing tests are making testing more efficient and effective.
• Visual Testing: Explore the emerging capabilities of AI in visual testing and how it's set to revolutionize UI verification.
• Inflectra's AI Solutions: See demonstrations of Inflectra's cutting-edge AI tools like the ChatGPT plugin and Azure Open AI platform, designed to streamline your testing process.
Whether you're a developer, tester, or QA professional, this webinar will give you valuable insights into how AI is shaping the future of software delivery.
Builder.ai Founder Sachin Dev Duggal's Strategic Approach to Create an Innova...Ramesh Iyer
In today's fast-changing business world, Companies that adapt and embrace new ideas often need help to keep up with the competition. However, fostering a culture of innovation takes much work. It takes vision, leadership and willingness to take risks in the right proportion. Sachin Dev Duggal, co-founder of Builder.ai, has perfected the art of this balance, creating a company culture where creativity and growth are nurtured at each stage.
Accelerate your Kubernetes clusters with Varnish CachingThijs Feryn
A presentation about the usage and availability of Varnish on Kubernetes. This talk explores the capabilities of Varnish caching and shows how to use the Varnish Helm chart to deploy it to Kubernetes.
This presentation was delivered at K8SUG Singapore. See https://feryn.eu/presentations/accelerate-your-kubernetes-clusters-with-varnish-caching-k8sug-singapore-28-2024 for more details.
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...DanBrown980551
Do you want to learn how to model and simulate an electrical network from scratch in under an hour?
Then welcome to this PowSyBl workshop, hosted by Rte, the French Transmission System Operator (TSO)!
During the webinar, you will discover the PowSyBl ecosystem as well as handle and study an electrical network through an interactive Python notebook.
PowSyBl is an open source project hosted by LF Energy, which offers a comprehensive set of features for electrical grid modelling and simulation. Among other advanced features, PowSyBl provides:
- A fully editable and extendable library for grid component modelling;
- Visualization tools to display your network;
- Grid simulation tools, such as power flows, security analyses (with or without remedial actions) and sensitivity analyses;
The framework is mostly written in Java, with a Python binding so that Python developers can access PowSyBl functionalities as well.
What you will learn during the webinar:
- For beginners: discover PowSyBl's functionalities through a quick general presentation and the notebook, without needing any expert coding skills;
- For advanced developers: master the skills to efficiently apply PowSyBl functionalities to your real-world scenarios.
Dev Dives: Train smarter, not harder – active learning and UiPath LLMs for do...UiPathCommunity
💥 Speed, accuracy, and scaling – discover the superpowers of GenAI in action with UiPath Document Understanding and Communications Mining™:
See how to accelerate model training and optimize model performance with active learning
Learn about the latest enhancements to out-of-the-box document processing – with little to no training required
Get an exclusive demo of the new family of UiPath LLMs – GenAI models specialized for processing different types of documents and messages
This is a hands-on session specifically designed for automation developers and AI enthusiasts seeking to enhance their knowledge in leveraging the latest intelligent document processing capabilities offered by UiPath.
Speakers:
👨🏫 Andras Palfi, Senior Product Manager, UiPath
👩🏫 Lenka Dulovicova, Product Program Manager, UiPath
Generating a custom Ruby SDK for your web service or Rails API using Smithyg2nightmarescribd
Have you ever wanted a Ruby client API to communicate with your web service? Smithy is a protocol-agnostic language for defining services and SDKs. Smithy Ruby is an implementation of Smithy that generates a Ruby SDK using a Smithy model. In this talk, we will explore Smithy and Smithy Ruby to learn how to generate custom feature-rich SDKs that can communicate with any web service, such as a Rails JSON API.
Key Trends Shaping the Future of Infrastructure.pdfCheryl Hung
Keynote at DIGIT West Expo, Glasgow on 29 May 2024.
Cheryl Hung, ochery.com
Sr Director, Infrastructure Ecosystem, Arm.
The key trends across hardware, cloud and open-source; exploring how these areas are likely to mature and develop over the short and long-term, and then considering how organisations can position themselves to adapt and thrive.
JMeter webinar - integration with InfluxDB and GrafanaRTTS
Watch this recorded webinar about real-time monitoring of application performance. See how to integrate Apache JMeter, the open-source leader in performance testing, with InfluxDB, the open-source time-series database, and Grafana, the open-source analytics and visualization application.
In this webinar, we will review the benefits of leveraging InfluxDB and Grafana when executing load tests and demonstrate how these tools are used to visualize performance metrics.
Length: 30 minutes
Session Overview
-------------------------------------------
During this webinar, we will cover the following topics while demonstrating the integrations of JMeter, InfluxDB and Grafana:
- What out-of-the-box solutions are available for real-time monitoring JMeter tests?
- What are the benefits of integrating InfluxDB and Grafana into the load testing stack?
- Which features are provided by Grafana?
- Demonstration of InfluxDB and Grafana using a practice web application
To view the webinar recording, go to:
https://www.rttsweb.com/jmeter-integration-webinar
Connector Corner: Automate dynamic content and events by pushing a buttonDianaGray10
Here is something new! In our next Connector Corner webinar, we will demonstrate how you can use a single workflow to:
Create a campaign using Mailchimp with merge tags/fields
Send an interactive Slack channel message (using buttons)
Have the message received by managers and peers along with a test email for review
But there’s more:
In a second workflow supporting the same use case, you’ll see:
Your campaign sent to target colleagues for approval
If the “Approve” button is clicked, a Jira/Zendesk ticket is created for the marketing design team
But—if the “Reject” button is pushed, colleagues will be alerted via Slack message
Join us to learn more about this new, human-in-the-loop capability, brought to you by Integration Service connectors.
And...
Speakers:
Akshay Agnihotri, Product Manager
Charlie Greenberg, Host
3. 3
3
OLX GROUP IS PART OF PROSUS
A collection of leading
companies and exciting
businesses!
4. 4
4
PROSUS IS A $120B MARKET CAP. COMPANY
A global internet and entertainment
group and one of the largest
technology investors in the world.
US$120bn
Market
capitalisation
US$44.5bn
Revenues over
the last
3 years
US$3.4bn
Trading profits
over the last
3 years
US$47M
Average
invested in M&A
per annum
US$18.3bn
FY18-19
Revenues
US$3.4bn
FY18-19 Trading
profit
13 of top 20
Fastest-growing
economies*
Present in
*IMF World Economic Outlook, based on 2019E GDP growth estimates for the countries with over 50 million population
5. 5
5
OLX GROUP TODAY:
THE WORLD'S #1 CLASSIFIEDS BUSINESS
HORIZONTALS REAL ESTATE
VERTICALS
OTHER
VERTICALS
CAR
VERTICALS
global
Turkey
Russia
UAE
Africa and
Philippines
Russia
Portugal
Poland
Romania,
Egypt
Furniture,
Europe
Heavy
machinery,
global
Services,
Poland
Poland
South Africa
Romania
Portugal
CONVENIENT
TRANSACTIONS
LATAM,
Asia,
Poland
UAE
Latin
America
South
Africa
Jobs,
India
Jobs,
Poland
+
6. 6
6
WHO WE ARE - OLX GROUP
We are a global product and tech group.
★ +20 brands
★ 15 time zones
★ +10,000 people
★ One mindset
We are a team of 10,000+ ambitious,
curious people building market-leading
trading platforms that empower 300
million people every month to upgrade
their lives.
7. 7
Agenda
4. Next steps and challenges
3. Our data journey
2. What is data mesh?
1. Challenges in data organizations today
8. 8
Concepts in this presentation are based on the data mesh
architecture abstracted and promoted by Zhamak Dehgani
11. 11
There has been a revolution in how operational applications
are being run
▪ For the last 20 years there is a continuous trend to move away from the
monolith to distributed domain driven architectures
12. 12
However, data engineers often stay behind by ingesting all that
data in one central data lake - the biggest monolith of them all
▪ The original data warehousing approach was getting data from all different
complex domains and putting them in one big fat database.
▪ Due to issues with scale in volume and complexity, architectures evolved into a
data lake architecture: Don't worry about that whole modeling we talked about,
just get the data out of the operational systems, bring them to this big, fat data
lake in its original form.
https://martinfowler.com/articles/data-monolith-to-mesh.html
13. 13
The data team responsible for storing big data is mostly
disconnected from consumers trying to make sense of that data.
Source: Zhamak Dehghani - Data Mesh Principles and Logical Architecture
14. 14
Data teams are trying to break down their architecture by
functional areas
Source: Zhamak Dehghani - Data Mesh Principles and Logical Architecture
15. 15
However, these data engineers are still siloed in between the
world of operational systems and the world of consumers
Source: Zhamak Dehghani - Data Mesh Principles and Logical Architecture
16. 16
Have pity on your data engineers
▪ They are often dependent from teams
who have no incentive in providing
meaningful, truthful and correct data
▪ They have little understanding of the
source domains that generate the data
and lack the domain expertise in their
teams
▪ They need to provide data for a diverse
set of needs without access to all
consuming domain's experts.
17. 17
In summary: We need to revolutionize our data strategy.
How can we apply domain-driven architecture to data?
▪ The cycle of innovation requires constant
adaptation to the data.
▪ This centralized system simply doesn't
scale.
▪ It has divided the work based on the
technical operation, implemented by one
or more silos of data engineers.
19. 19
Data mesh sets a foundation for getting value from analytical
data at scale – using 4 principles.
Domain-oriented decentralized data
ownership and architecture
Data as a product
Self-service data infrastructure
as a platform
Federated governance
20. 20
Principle 1: Domain oriented decentralized data ownership
▪ Although DDD has influenced modern
architectural thinking, the notion of business
domains have been disregarded in data.
▪ All raw data is in the lake, but there is often no
clear separation of business domains.
▪ Rather than limiting to ingesting raw data from
domains into a centrally owned data lake,
domains need to own and serve their
domain datasets in an easily consumable
way.
21. 21
Domain-driven design moves from a ‘big ball of mud’…
“A BIG BALL OF MUD is haphazardly
structured, sprawling, sloppy, duct-tape and
bailing wire, spaghetti code jungle. We’ve all
seen them. These systems show
unmistakable signs of unregulated growth,
and repeated, expedient repair. Information is
shared promiscuously among distant
elements of the system, often to the point
where nearly all the important information
becomes global or duplicated. The overall
structure of the system may never have been
well defined. If it was, it may have eroded
beyond recognition.”
22. 22
…to contextual models
Great article about DDD: https://medium.com/raa-labs/part-1-domain-driven-design-like-a-pro-f9e78d081f10
▪ Focused
▪ Small
▪ Decoupled
▪ Easy to change
▪ Enables autonomy
▪ Ubiquitous language
23. 23
Some domains are more source oriented while other domains
are more consumer oriented.
Domains aligned with the source Domains aligned with consumption
Chat messages
Browsing interactions
Deliveries
Item recommendations
Customer support tickets
User segmentation
Fraud detection
24. 24
Principle 2: Data as a product
▪ Domain data teams must consider their data assets as their products
and the rest of the organization as their customers.
– Discoverability
– Addressable
– Trustworthy and truthful
– Self-describing semantics and syntax
– Inter-operable and governed by
global standards
Design
“Build what
matters”
Marketing
“Tell people
about it”
Engineering
“Ship it!”
25. 25
Establish the responsibility of domain data product owner –
which could simply be an additional hat to any type of engineer
▪ Makes decisions around the vision and the roadmap for the data products
▪ Concerns themselves with satisfaction of their consumers
▪ Continuously measures and improves the quality of the data
▪ Responsible for the lifecycle of the domain datasets
▪ Defines success criteria and business-aligned metrics
26. 26
Principle 3: Self-service data infra as a platform
▪ High-level abstraction of infrastructure enables teams to autonomously own
their data products
▪ Must include tooling that supports a developer’s workflow of creating,
maintaining and running data products with less specialized knowledge
that existing technologies assume
▪ Domain agnostic
▪ Hides underlying complexity and designed in a self-service manner
▪ But: Treat domain data ownership as primary concern, and tooling and
pipelines secondary
27. 27
The self-service data infra platform can include many generic
elements aimed at making domain data producers more efficient
– Data product versioning
– Data product schema
– Unified data access control and logging
– Data pipeline implementation and orchestration
– Data product discovery, catalog registration and
publishing
– Data governance and standardization
– Data product lineage
– Data product monitoring/alerting/log
– Data product quality metrics (collection and
sharing)
28. 28
Principle 4: (Computational) federated governance
▪ Independent data products need to
interoperate through global standardization
▪ Naming conventions, identifiers, nulls
▪ It is an art to find a balance between what shall
be standardized globally, and what shall be left
to the domains to decide.
– For example, the semantics of ‘chat replies’
could be left to the chat team
– However, a ‘buyer’, as a population of
‘users’, is a global concern.
29. 29
In summary: The ”great divide”.
When it still looked simple.
Source: Zhamak Dehghani - Data Mesh Principles and Logical Architecture
30. 30
In summary: The ”greater divide”.
When it was still manageable.
Source: Zhamak Dehghani - Data Mesh Principles and Logical Architecture
31. 31
In summary: The ”best divide”.
When we thought we could still handle it.
Source: Zhamak Dehghani - Data Mesh Principles and Logical Architecture
32. 32
In summary: Data mesh
The paradigm shift
Source: Data mesh 101 – Everything you need to know to get started
34. 34
A mature data infrastructure is in place and managed globally for
OLX Group to serve, ingest and consume data
Catalog
Self Service tool for data
management, consumption and
discovery
Data Engineers
Ninja
Set of libraries that unify the
tracking integration.
Hydra
A platform-agnostic raw data
HTTP collector.
Lazarus
Service that synchronises the data
from databases.
Data Lake
Data storage compliance with
data protection laws.
Reservoir
A dedicated and reduced data
storage for a specific purpose
Cerberus
A service to process real-time
events
Schema Management
Easily query via Athena,
Spectrum and Presto.
Laquesis
Self Service tool for performing
Experiments, Feature Flags and
Surveys.
Odyn
Operational Data Hub consisting
of a scheduler, operators and
storage
KaaS
Packaged solution for Apache
Kylin usage
Real time process
Machine learning
Analysts
Analytics
Product Managers
Databases
User devices
Microservices
COLLECTION GOVERNANCE CONSUMPTION
DAPI
Generic and scalable DATA API
Real time
5 minutes
5 minutes - 1 hour
SERVICES
Data Scientist
35. 35
Scheduler
ETL-as-a-code platform with
advanced dependency
management system.
Odyn decides which tasks
should run, when and where to
achieve maximum efficiency.
It supports templating, step
expansion and notifications.
Operators
Odyn allows to easily transform
and query the data using SQL.
Out of the box it supports
Athena, Presto, Hive, Redshift,
AWS Batch, Kylin and Spark.
It can be easily extended with
custom operators written in
Python.
Storage
Odyn is a cheap and reliable
storage as the part of Data
Reservoir.
It contains a build in solution to
be compliant with all the data
protection laws.
It is fully integrated with other
data services.
Operational Data Hub
One place for data access so that
many point-to-point connections
between callers and data suppliers
do not need to be made.
Odyn allows blazing fast data
processing as well as
collaboration and sharing of
datasets between the users.
Odyn is the operational data hub consisting of a scheduler,
extendable set of operators and storage
180 users
200 active DAGs
11 K daily tasks
36. 36
OLX Europe’s data team leverages the global data infra and
provides additional data products adapted to regional needs
EU Data
Build and maintain a best-in-class data analytics platform
that enables easy, timely, fast data discovery and consumption
Enable top-notch data mastery in our company by providing the right training
and coaching to our team and our users
Assure top quality of prepared data for the right purpose at the right time
following product and business strategy
Solve new problems in innovative and pragmatic ways grabbing
opportunities for quick value delivery
More data played back to our
external customers
Data-driven product
development lifecycle
Business decisions
fueled by data
37. 37
For example, Sherlock as part of the analytical data platform
enables our users to discover, understand, and explore datasets
38. 38
Our internal data academy helps anyone in OLX get more
familiar with data concepts, technologies and our platforms
39. 39
Finally, Yamato is a large AWS Redshift data base empowering
many data teams to process data blazingly fast in an easy way
250 users
150 K queries daily
16xDC2.8XL - 40 TB
40. 40
Some of our consumer-oriented data teams are already
checking all the boxes for data mesh
Customer support Sales CRM integration
Recommendations Customer classification
41. 41
Challenges in data teams
▪ Data teams are often a bottleneck and cannot keep up
with product development fast enough
▪ Data teams are the go-to point for expertise about domain
data, however they are not the specialists
▪ Lack of governance across data teams
▪ Duplication of work and/or reinventing the wheel
▪ Lack of software engineering principles in data processing
42. 42
Challenges in operational teams in OLX
▪ Technical design for new features does not include requirements for analytics,
experimentation and machine learning projects –
therefore data needs not always covered from start
▪ Data ownership mostly limited to what is required to run a feature
▪ Operational teams are not always aware of the value of data
43. 43
So, although the doors are now open across the data
organization, there is still a divide between data and tech.
Source: Zhamak Dehghani - Data Mesh Principles and Logical Architecture
44. 44
So, although the doors are now open across the data
organization, there is still a divide between data and tech.
Source: Zhamak Dehghani - Data Mesh Principles and Logical Architecture
45. 45
However, many things are going well - and we are getting ready
for the paradigm shift!
▪ Self-service data infrastructure is mostly in place already
▪ Data publication possible for anyone
▪ GDPR, security and access governance is in place.
▪ There is a growing acceptance for domain data ownership in operational teams
▪ We already have 100s of product managers so no lack of product thinking
47. 47
Let’s remind ourselves of the 4 principles of data mesh – and
first assess what is in place.
Domain-oriented decentralized data
ownership and architecture
Data as a product
Self-service data infra
as a platform
Federated governance
48. 48
Start small: Pilot partial embedding of domain data engineers in
a few selected operational teams on project basis
▪ Data engineers will ensure that operational teams integrate data
requirements into the design of new features
▪ Data engineers will set the foundation for domain data sets after which the
ownership remains in these operational teams
▪ Data engineers will team up with product managers
▪ Data engineers will learn from software engineers
and adopt software engineering best practices
▪ Data engineers will facilitate trainings
▪ Ensure engineering leaders are on board!
49. 49
Make both sides aware that this is a win-win situation!
▪ Data engineers often lack software engineering
standard practices when it comes to building data
assets.
▪ Software engineers who are building operational
systems often have no experience utilizing data
engineering tool sets, or even understanding the
concept of ‘datasets’.
▪ Removing the skill set silos will lead to
creation of a larger and deeper pool of data
engineering skill sets available to the
organization!
50. 50
Meanwhile, start mapping out the major domains, identify
ownership and develop a data maturity framework by domain
Source: Zhamak Dehghani - Data Mesh Principles and Logical Architecture
51. 51
Define how to measure success and set a baseline
Data quality
Domain-data Data consumer
needs covered
Documented and
useful datasets
Data
discoverability
Usage Satisfaction
Speed/reliability Skill levels Ease of use
Risk &
Governance
Cost &
Compliance
Ubiquitous
language
Data as a product
Self-service data infra
Federated governance
52. 52
It sounds too good to be true. What is the fine print?
▪ Data mesh is primarily about mindset and
organization; technology is second
▪ Success depends on converging operational and
data roles – organization needs to be ready
▪ Organization needs to be big enough to benefit
▪ Data mesh is a vision that needs to be tailored to
your organization – no plug and play solution
▪ The data lake can still exist in this architecture, but
they become just another node in the mesh, rather
than being the center place.
53. 53
You are not alone! Other companies are setting steps towards a
data mesh architecture – and a learning community is live
▪ https://launchpass.com/data-mesh-learning
54. 54
Jochem van Grondelle
Data Engineering Manager
linkedin.com/in/jochemvangrondelle
jochem.vangrondelle@olx.com
Thank you! Feel free to reach out to discuss further