This document proposes a reference architecture for providing high availability across multiple data centers using MariaDB and related open source tools. It summarizes:
- The need for a geo-redundant highly available database architecture at Nokia to support multiple product units.
- An evaluation of alternatives including Galera clusters and master-master replication between data centers.
- A proposed architecture using MaxScale for local master-slave replication within each data center and cross-data center replication between masters for redundancy.
- Testing and development of MaxScale plugins and scripts to support automatic failover and recovery after failures within or between data centers.
- Plans for containerized deployment of the database clusters and MaxScale using Kubernetes with additional
How to migrate from Oracle Database with easeMariaDB plc
MariaDB introduced Oracle Database compatibility last May with support for Oracle Database data types, sequences, stored procedures (PL/SQL) and more, making it easier than ever to migrate to MariaDB. In this session, MariaDB's Alexander Bienemann and Wagner Bianchi share best practices and lessons learned from our experiences helping customers migrate from Oracle Database. They explain how MariaDB approaches migrations, what’s needed to complete a successful migration and the tools used to determine the level of effort required.
How to migrate from Oracle Database with easeMariaDB plc
MariaDB introduced Oracle Database compatibility last May with support for Oracle Database data types, sequences, stored procedures (PL/SQL) and more, making it easier than ever to migrate to MariaDB. In this session, MariaDB's Alexander Bienemann and Wagner Bianchi share best practices and lessons learned from our experiences helping customers migrate from Oracle Database. They explain how MariaDB approaches migrations, what’s needed to complete a successful migration and the tools used to determine the level of effort required.
What to expect from MariaDB Platform X5, part 2MariaDB plc
MariaDB Platform X5 will include MariaDB MaxScale 2.5 (with its brand-new web UI for configuration and monitoring) and MariaDB ColumnStore 1.5 (with cluster management reimplemented in MariaDB MaxScale for improved ease of use and deployment). In addition to the new web UI, MariaDB MaxScale 2.5 will be introducing support for distributed caches such as Redis, streaming to Apache Kafka, and a completely rewritten binlog router.
In this session, we’ll provide a short overview of MariaDB MaxScale and ColumnStore followed by a walkthrough of new features and a short discussion of plans for the next releases.
Configuring workload-based storage and topologiesMariaDB plc
MariaDB has multiple workload-optimized storage engines, including InnoDB for mixed workloads, MyRocks for write-intensive workloads, Spider for scalable workloads and ColumnStore for analytical workloads. In this session, Kenny Geiselhart discusses how to choose the right storage engine for individual tables, and how replication and asymmetric topologies can be used to further optimize MariaDB and the hardware it runs on for specific workloads.
How THINQ runs both transactions and analytics at scaleMariaDB plc
THINQ provides a cloud-based Communications-Platform-as-a-Service (CPaaS) that routes tens of millions of phone calls per day for customers in enterprise and telecommunications industries. In this session Sasha Vaniachine, Senior Database Administrator at THINQ, explains how he combined MariaDB Server and MariaDB ColumnStore to support both high-performance transaction processing and scalable analytics. In addition, he shares some of THINQ's best practices and lessons learned from supporting an ever-increasing database workload that currently exceeds 10,000 transactions per second.
Inside CynosDB: MariaDB optimized for the cloud at TencentMariaDB plc
Qinglin Zhang, Database Kernel Engineer at Tencent, introduces CynosDB, Tencent's self-developed database for the cloud. CynosDB is based on MariaDB Server, but separates computing and storage. Zhang goes on to provide a detailed explanation of the architecture with a focus on how Tencent implemented the computing and storage layers, and created Tencent’s MariaDB-based “Aurora”.
Migrating from InnoDB and HBase to MyRocks at FacebookMariaDB plc
Facebook created a new storage engine called MyRocks to optimize space and write performance, and recently migrated both UDB (a database for social activities, and our biggest in production) and Facebook Messenger to MyRocks. In this session, Yoshinori Matsunobu of Facebook talks about the challenges, benefits and lessons learned by migrating these applications from InnoDB to MyRocks.
Writing powerful stored procedures in PL/SQLMariaDB plc
Oracle Database compatibility in MariaDB Server lets developers choose between ANSI SQL and PL/SQL when writing stored procedures. In this session, Senior Solutions Engineer Alton Dinsmore focuses on how to write powerful stored procedures and functions with PL/SQL, whether you are migrating from Oracle Database or not.
Auto Europe's ongoing journey with MariaDB and open sourceMariaDB plc
Tom Girsch, Lead System Architect at Auto Europe, covers the use case that initially brought Auto Europe to MariaDB, as well as additional planned and ongoing projects. He goes on to discuss Auto Europe’s implementation of MariaDB using clustering, traditional replication and MaxScale. Next, he covers some of the problems and pitfalls encountered along the way, as well as some suggestions to further improve the product.
Faster, better, stronger: The new InnoDBMariaDB plc
For MariaDB Enterprise Server 10.5, the default transactional storage engine, InnoDB, has been significantly rewritten to improve the performance of writes and backups. Next, we removed a number of parameters to reduce unnecessary complexity, not only in terms of configuration but of the code itself. And finally, we improved crash recovery thanks to better consistency checks and we reduced memory consumption and file I/O thanks to an all new log record format.
In this session, we’ll walk through all of the improvements to InnoDB, and dive deep into the implementation to explain how these improvements help everything from configuration and performance to reliability and recovery.
What to expect from MariaDB Platform X5, part 1MariaDB plc
MariaDB Platform X5 will be based on MariaDB Enterprise Server 10.5. This release includes Xpand, a fully distributed storage engine for scaling out, as well as many new features and improvements for DBAs and developers alike, including enhancements to temporal tables, additional JSON functions, a new performance schema, non-blocking schema changes with clustering and a Hashicorp Vault plugin for key management.
In this session, we’ll walk through all of the new features and enhancements available in MariaDB Enterprise Server 10.5. In addition, we will highlight those being backported to maintenance releases of MariaDB Enterprise Server 10.2, 10.3 and 10.4.
In this session, Engineer Allen Herrera describes how SpendHQ made the move to a columnar database with MariaDB. He shares every aspect of the process from setting up their first cluster and testing it within their application to automating cluster deployment, analyzing performance and refining their data import process (i.e., ETL). He finishes by discussing future plans for MariaDB at SpendHQ.
How QBerg scaled to store data longer, query it fasterMariaDB plc
The continuous increase in terms of services and countries to which QBerg delivers its services requires an ever-increasing load of resources. During the last year QBerg has reached a critical point, storing so much transactional data that standard relational databases were unable to meet the SLAs, or support the features, required by customers. As an example, they had to cap web analytics to running on a maximum of four months of history. The introduction of MariaDB ColumnStore, flanked by existing MariaDB Server databases, not only will allow them to store multiple years’ worth of historical data for analytics – it decreased overall processing time by one order of magnitude right off the bat. The move to a unified platform was incremental, using MariaDB MaxScale as both a router and a replicator. QBerg is now able to replicate full InnoDB schemas to MariaDB ColumnStore and incrementally update big tables without impacting the performance of ongoing transactions.
Ramp-Tutorial for MYSQL Cluster - Scaling with Continuous AvailabilityPythian
Rene Cannao's Ramp-Tutorial for MYSQL Cluster - Scaling with Continuous Availability. Rene, a Senior Operational DBA at PalominoDB.com, will guide attendees through a hands-on experience in the installation, configuration management and tuning of MySQL Cluster.
Agenda:
- MySQL Cluster Concepts and Architecture: we will review the principle of a fault-tolerant shared nothing architecture, and how this is implemented into NDB;
- MySQL Cluster processes : attendees will understand the various roles and interactions between Data Nodes, API Nodes and Management Nodes;
- Installation : we will install a minimal HA solution with MySQL Cluster on 3 virtual machines;
- Configuration of a basic system : upon describing the most important configuration parameters, Data/API/Management nodes will be configured and the Cluster launched;
- Loading data: the "world" schema will be imported into NDB using "in memory" and "disk based" storages; the attendees will experience how data changes are visible across API Nodes;
- Understand the NDB Storage Engine : internal implementation details will be explained, like synchronous replication, transaction coordinator, heartbeat, communication, failure detection and handling, checkpoint, etc;
- Query and schema design : attendees will understand the execution plan of queries with NDB, how SQL and Data Nodes communicate, how indexes and partitions are implemented, condition pushdown, join pushdown, query cache;
- Management and Administration: the attendees will test High Availability of NDB when a node become unavailable will learn how to read log file, how to stop/start any component of the Cluster to perform a rolling restart with no downtime, and how to handle a degraded setup;
- Backup and Recovery: attendees will be driven through the procedure of using NDB-native online backup and restore, and how this differs from mysqldump;
- Monitor and improve performance: attendee will learn how to boost performance tweaking variables according to hardware configuration and application workload
Running Dataproc At Scale in production - Searce Talk at GDG DelhiSearce Inc
From the collective experience of helping multiple customers run hundreds of dataproc clusters in production, scale them, troubleshoot issues, our engineers Manan and Rohit talk about running Dataproc At Scale in Production.
What to expect from MariaDB Platform X5, part 2MariaDB plc
MariaDB Platform X5 will include MariaDB MaxScale 2.5 (with its brand-new web UI for configuration and monitoring) and MariaDB ColumnStore 1.5 (with cluster management reimplemented in MariaDB MaxScale for improved ease of use and deployment). In addition to the new web UI, MariaDB MaxScale 2.5 will be introducing support for distributed caches such as Redis, streaming to Apache Kafka, and a completely rewritten binlog router.
In this session, we’ll provide a short overview of MariaDB MaxScale and ColumnStore followed by a walkthrough of new features and a short discussion of plans for the next releases.
Configuring workload-based storage and topologiesMariaDB plc
MariaDB has multiple workload-optimized storage engines, including InnoDB for mixed workloads, MyRocks for write-intensive workloads, Spider for scalable workloads and ColumnStore for analytical workloads. In this session, Kenny Geiselhart discusses how to choose the right storage engine for individual tables, and how replication and asymmetric topologies can be used to further optimize MariaDB and the hardware it runs on for specific workloads.
How THINQ runs both transactions and analytics at scaleMariaDB plc
THINQ provides a cloud-based Communications-Platform-as-a-Service (CPaaS) that routes tens of millions of phone calls per day for customers in enterprise and telecommunications industries. In this session Sasha Vaniachine, Senior Database Administrator at THINQ, explains how he combined MariaDB Server and MariaDB ColumnStore to support both high-performance transaction processing and scalable analytics. In addition, he shares some of THINQ's best practices and lessons learned from supporting an ever-increasing database workload that currently exceeds 10,000 transactions per second.
Inside CynosDB: MariaDB optimized for the cloud at TencentMariaDB plc
Qinglin Zhang, Database Kernel Engineer at Tencent, introduces CynosDB, Tencent's self-developed database for the cloud. CynosDB is based on MariaDB Server, but separates computing and storage. Zhang goes on to provide a detailed explanation of the architecture with a focus on how Tencent implemented the computing and storage layers, and created Tencent’s MariaDB-based “Aurora”.
Migrating from InnoDB and HBase to MyRocks at FacebookMariaDB plc
Facebook created a new storage engine called MyRocks to optimize space and write performance, and recently migrated both UDB (a database for social activities, and our biggest in production) and Facebook Messenger to MyRocks. In this session, Yoshinori Matsunobu of Facebook talks about the challenges, benefits and lessons learned by migrating these applications from InnoDB to MyRocks.
Writing powerful stored procedures in PL/SQLMariaDB plc
Oracle Database compatibility in MariaDB Server lets developers choose between ANSI SQL and PL/SQL when writing stored procedures. In this session, Senior Solutions Engineer Alton Dinsmore focuses on how to write powerful stored procedures and functions with PL/SQL, whether you are migrating from Oracle Database or not.
Auto Europe's ongoing journey with MariaDB and open sourceMariaDB plc
Tom Girsch, Lead System Architect at Auto Europe, covers the use case that initially brought Auto Europe to MariaDB, as well as additional planned and ongoing projects. He goes on to discuss Auto Europe’s implementation of MariaDB using clustering, traditional replication and MaxScale. Next, he covers some of the problems and pitfalls encountered along the way, as well as some suggestions to further improve the product.
Faster, better, stronger: The new InnoDBMariaDB plc
For MariaDB Enterprise Server 10.5, the default transactional storage engine, InnoDB, has been significantly rewritten to improve the performance of writes and backups. Next, we removed a number of parameters to reduce unnecessary complexity, not only in terms of configuration but of the code itself. And finally, we improved crash recovery thanks to better consistency checks and we reduced memory consumption and file I/O thanks to an all new log record format.
In this session, we’ll walk through all of the improvements to InnoDB, and dive deep into the implementation to explain how these improvements help everything from configuration and performance to reliability and recovery.
What to expect from MariaDB Platform X5, part 1MariaDB plc
MariaDB Platform X5 will be based on MariaDB Enterprise Server 10.5. This release includes Xpand, a fully distributed storage engine for scaling out, as well as many new features and improvements for DBAs and developers alike, including enhancements to temporal tables, additional JSON functions, a new performance schema, non-blocking schema changes with clustering and a Hashicorp Vault plugin for key management.
In this session, we’ll walk through all of the new features and enhancements available in MariaDB Enterprise Server 10.5. In addition, we will highlight those being backported to maintenance releases of MariaDB Enterprise Server 10.2, 10.3 and 10.4.
In this session, Engineer Allen Herrera describes how SpendHQ made the move to a columnar database with MariaDB. He shares every aspect of the process from setting up their first cluster and testing it within their application to automating cluster deployment, analyzing performance and refining their data import process (i.e., ETL). He finishes by discussing future plans for MariaDB at SpendHQ.
How QBerg scaled to store data longer, query it fasterMariaDB plc
The continuous increase in terms of services and countries to which QBerg delivers its services requires an ever-increasing load of resources. During the last year QBerg has reached a critical point, storing so much transactional data that standard relational databases were unable to meet the SLAs, or support the features, required by customers. As an example, they had to cap web analytics to running on a maximum of four months of history. The introduction of MariaDB ColumnStore, flanked by existing MariaDB Server databases, not only will allow them to store multiple years’ worth of historical data for analytics – it decreased overall processing time by one order of magnitude right off the bat. The move to a unified platform was incremental, using MariaDB MaxScale as both a router and a replicator. QBerg is now able to replicate full InnoDB schemas to MariaDB ColumnStore and incrementally update big tables without impacting the performance of ongoing transactions.
Ramp-Tutorial for MYSQL Cluster - Scaling with Continuous AvailabilityPythian
Rene Cannao's Ramp-Tutorial for MYSQL Cluster - Scaling with Continuous Availability. Rene, a Senior Operational DBA at PalominoDB.com, will guide attendees through a hands-on experience in the installation, configuration management and tuning of MySQL Cluster.
Agenda:
- MySQL Cluster Concepts and Architecture: we will review the principle of a fault-tolerant shared nothing architecture, and how this is implemented into NDB;
- MySQL Cluster processes : attendees will understand the various roles and interactions between Data Nodes, API Nodes and Management Nodes;
- Installation : we will install a minimal HA solution with MySQL Cluster on 3 virtual machines;
- Configuration of a basic system : upon describing the most important configuration parameters, Data/API/Management nodes will be configured and the Cluster launched;
- Loading data: the "world" schema will be imported into NDB using "in memory" and "disk based" storages; the attendees will experience how data changes are visible across API Nodes;
- Understand the NDB Storage Engine : internal implementation details will be explained, like synchronous replication, transaction coordinator, heartbeat, communication, failure detection and handling, checkpoint, etc;
- Query and schema design : attendees will understand the execution plan of queries with NDB, how SQL and Data Nodes communicate, how indexes and partitions are implemented, condition pushdown, join pushdown, query cache;
- Management and Administration: the attendees will test High Availability of NDB when a node become unavailable will learn how to read log file, how to stop/start any component of the Cluster to perform a rolling restart with no downtime, and how to handle a degraded setup;
- Backup and Recovery: attendees will be driven through the procedure of using NDB-native online backup and restore, and how this differs from mysqldump;
- Monitor and improve performance: attendee will learn how to boost performance tweaking variables according to hardware configuration and application workload
Running Dataproc At Scale in production - Searce Talk at GDG DelhiSearce Inc
From the collective experience of helping multiple customers run hundreds of dataproc clusters in production, scale them, troubleshoot issues, our engineers Manan and Rohit talk about running Dataproc At Scale in Production.
Pivotal Cloud Foundry 2.6: A First LookVMware Tanzu
Join Dan Baskette and Jared Ruckle for a view into Pivotal Cloud Foundry (PCF) 2.6 capabilities with demos and expert Q&A. We’ll review the latest features for Pivotal’s flagship app platform, including:
CUSTOM SIDECAR PROCESSES (BETA)
In Pivotal Application ServiceⓇ 2.6 (PAS), developers can run custom sidecar processes in the same container as their application. This simplifies development for all kinds of “wire” use cases, including proxy forwarding, client-side load balancing, timeouts, and retries.
MULTI-CLOUD CONTINUOUS DELIVERY WITH SPINNAKER
PCF now integrates nicely with the most popular CD tool, Spinnaker. Spinnaker 1.14 now supports several advanced CD scenarios with PCF. As a result, large development teams can more easily deploy to production to improve outcomes. Use Spinnaker with PAS as well as Enterprise PKSⓇ. (This integration is backed by community support.)
NEW PERMISSIONS MODEL IN CONCOURSE FOR PCF (coming soon) Concourse for PCF 5.2 will include a powerful new permissions model to better segment access to build pipelines. The new release will add compatibility with CredHub for secrets management as well.
MULTI-DATACENTER REPLICATION CAPABILITIES FOR MySQL (coming soon) MySQL for PCF 2.7 will add multi-DC replication capabilities as a beta feature. This will offer more stability and scalability for your database apps.
Plus much more!
Webinar Slides: MySQL HA/DR/Geo-Scale - High Noon #2: Galera ClusterContinuent
Galera Cluster vs. Continuent Tungsten Clusters
Building a Geo-Scale, Multi-Region and Highly Available MySQL Cloud Back-End
This second installment of our High Noon series of on-demand webinars is focused on Galera Cluster (including MariaDB Cluster & Percona XtraDB Cluster). It looks at some of the key characteristics of Galera Cluster and how it fares as a MySQL HA / DR / Geo-Scale solution, especially when compared to Continuent Tungsten Clustering.
Watch this webinar to learn how to do better MySQL HA / DR / Geo-Scale.
AGENDA
- Goals for the High Noon Webinar Series
- High Noon Series: Tungsten Clustering vs Others
- Galera Cluster (aka MariaDB Cluster & Percona XtraDB Cluster)
- Key Characteristics
- Certification-based Replication
- Galera Multi-Site Requirements
- Limitations Using Galera Cluster
- How to do better MySQL HA / DR / Geo-Scale?
- Galera Cluster vs Tungsten Clustering
- About Continuent & Its Solutions
PRESENTER
Matthew Lang - Customer Success Director – Americas, Continuent - has over 25 years of experience in database administration, database programming, and system architecture, including the creation of a database replication product that is still in use today. He has designed highly available, scaleable systems that have allowed startups to quickly become enterprise organizations, utilizing a variety of technologies including open source projects, virtualization and cloud.
Multi-Tenancy Kafka cluster for LINE services with 250 billion daily messagesLINE Corporation
Yuto Kawamura
LINE / Z Part Team
At LINE we've been operating Apache Kafka to provide the company-wide shared data pipeline for services using it for storing and distributing data.
Kafka is underlying many of our services in some way, not only the messaging service but also AD, Blockchain, Pay, Timeline, Cryptocurrency trading and more.
Many services feeding many data into our cluster, leading over 250 billion daily messages and 3.5GB incoming bytes in 1 second which is one of the world largest scale.
At the same time, it is required to be stable and performant all the time because many important services uses it as a backend.
In this talk I will introduce the overview of Kafka usage at LINE and how we're operating it.
I'm also going to talk about some engineerings we did for maximizing its performance, solving troubles led particularly by hosting huge data from many services, leveraging advanced techniques like kernel-level dynamic tracing.
We will show what aspects of hawkBit need to be completed or implemented to use it in a production environment.
We will look at hawkBit's extension points and how they can be used and show some practices for deploying and managing a hawkBit-based product.
Training Slides: Basics 102: Introduction to Tungsten ClusteringContinuent
This 30 minutes training session provides an introduction to how Tungsten Clustering for MySQL / MariaDB / Percona Server works, its basic principles, understanding Tungsten Clustering topologies, failover, rolling maintenance and related tools.
AGENDA
- Review the key benefits offered by Tungsten Clustering
- Examine the Tungsten Clustering architecture
- Tungsten Cluster Topologies for MySQL High Availability and Disaster Recovery
- Composite vs Multi-Site/Multi-Master
- Review automatic and manual failover
- Explore the concepts of a rolling maintenance procedure
- Study key resources to monitor and manage the cluster
Como creamos QuestDB Cloud, un SaaS basado en Kubernetes alrededor de QuestDB...javier ramirez
QuestDB es una base de datos open source de alto rendimiento. Mucha gente nos comentaba que les gustaría usarla como servicio, sin tener que gestionar las máquinas. Así que nos pusimos manos a la obra para desarrollar una solución que nos permitiese lanzar instancias de QuestDB con provisionado, monitorización, seguridad o actualizaciones totalmente gestionadas.
Unos cuantos clusters de Kubernetes más tarde, conseguimos lanzar nuestra oferta de QuestDB Cloud. Esta charla es la historia de cómo llegamos ahí. Hablaré de herramientas como Calico, Karpenter, CoreDNS, Telegraf, Prometheus, Loki o Grafana, pero también de retos como autenticación, facturación, multi-nube, o de a qué tienes que decir que no para poder sobrevivir en la nube.
CERN, the European Organization for Nuclear Research, is running for several years a large OpenStack Cloud that helps thousands of scientists to analyze the data from the LHC.
In 2012, early in the design phase of the CERN Cloud we decided to use Nova Cells to enable the infrastructure to scale to thousands of nodes. Now with more than 280K cores spread across 70 cells that are hosted in two data centres we were faced with the challenge to migrate to Nova Cells V2 required in the Pike release.
In this presentation, we will describe how Nova Cells allowed CERN to scale to thousands of nodes, its advantages and how we mitigate the implementation issues of Nova Cells V1. Next, we will cover how we upgraded Nova from Newton with Cells V1 to Pike with Cells V2. We will explain the steps that we followed and the issues that we faced during the upgrade. Finally, we will report our experience with Cells V2 at scale, its caveats and how we are mitigate them.
What can I expect to learn?
This presentation describes how CERN migrated from Cells V1 to Cells V2 when upgraded from Newton to Pike release.
You will learn the procedures followed by CERN in order to migrate Cells V1 to Cells V2 in a large production environment.
The issues found during the upgrade and how we mitigate them will be discussed.
Also, we will present how Cells V2 behaves in a large scale deployment with serveral thounsands nodes in 70 cells.
MariaDB Auto-Clustering, Vertical and Horizontal Scaling within Jelastic PaaSJelastic Multi-Cloud PaaS
Availability and performance have a direct business impact for most of the companies nowadays. No one wants to lose money because of occasional downtime or data loss. Thus, to minimize the risk and ensure an extra level of redundancy, clustering and automatic scaling should be used. In this video Ruslan Synytsky presented how Jelastic PaaS implemented auto-clustering of MariaDB by providing the customers with different replication options out-of-box with no need in manual configurations. It is also detailed how to automate vertical and horizontal scaling of databases running in the cloud.
Video recording of the session https://www.youtube.com/watch?v=6MND3feb5zM
Presenation provides the practical aspects of migrating a database setup based on traditional asynchronous replication to multi-master Galera Cluster. We will discuss the benefits Galera provides and how traditional replication settings, architecture and practices can be converted to Galera Cluster. We will show the steps that are needed to perform the migration with limited or no downtime.
Enabling Presto to handle massive scale at lightning speedShubham Tagra
Presto User Group Singapore Meetup - March 2019.
These slides talk through the current state of Presto and features that help Presto work better in cloud and a glimpse into the roadmap
Kubermatic How to Migrate 100 Clusters from On-Prem to Google Cloud Without D...Tobias Schneck
Have you ever thought about migrating your Kubernetes clusters to Google Cloud to get your services closer to your customers? Yes? We too! Join us on an interactive journey to discover the main challenges of live migration at scale of etcd's, traffic routing and application workloads from your on-premise platform to GCP. The talk will discuss the current state of the technical concept, known problems and insides of the already proven migration steps for stateless workload.
As part of the journey, we'll see the differences between migrating one or one hundred clusters with productive workloads; What parts can be automated? What steps may need to be manual? Let's see how an automated solution could look like in the future and what steps are missing.
How to Migrate 100 Clusters from On-Prem to Google Cloud Without Downtimeloodse
Have you ever thought about migrating your Kubernetes clusters to Google Cloud to get your services closer to your customers? Yes? Us too! Join us on an interactive journey to discover the main challenges of live migration at scale of etcd’s, traffic routing and application workloads from your on-premise platform to GCP. The talk will discuss the current state of the technical concept, known problems and insides of the already proven migration steps for stateless workloads.
As part of the journey, we'll see
- The differences between migrating one or one hundred clusters with productive workloads
- What parts can be automated?
- What steps may need to be done manually?
SkySQL is the first and only database-as-a-service (DBaaS) to perform workload analysis with advanced deep learning models, identifying and classifying discrete workload patterns so DBAs can better understand database workloads, identify anomalies and predict changes.
In this session, we’ll explain the concepts behind workload analysis and show how it can be used in the real world (and with sample real-world data) to improve database performance and efficiency by identifying key metrics and changes to cyclical patterns.
SkySQL uses best-of-breed software, and when it comes to metrics and monitoring that means Prometheus and Grafana. SkySQL Monitor is built on both, and provides customers with interactive dashboards for both real-time and historic metrics monitoring. In addition, it meets the same high availability and security requirements as other SkySQL components, ensuring metrics are always available and always secure.
In this session, we’ll explain how SkySQL Monitor works, walk through its dashboards and show how to monitor key metrics for performance and replication.
Introducing the R2DBC async Java connectorMariaDB plc
Not too long ago, a reactive variant of the JDBC driver was released, known as Reactive Relational Database Connectivity (R2DBC for short). While R2DBC started as an experiment to enable integration of SQL databases into systems that use reactive programming models, it now specifies a full-fledged service-provider interface that can be used to retrieve data from a target data source.
In this session, we’ll take a look at the new MariaDB R2DBC connector and examine the advantages of fully reactive, non-blocking development with MariaDB. And, of course, we’ll dive in and get a first-hand look at what it’s like to use the new connector with some live coding!
The capabilities and features of MariaDB Platform continue to expand, resulting in larger and more sophisticated production deployments – and the need for better tools. To provide DBAs with comprehensive, consolidating tooling, we created MariaDB Enterprise Tools: an easy-to-use, modular command-line interface for interacting with any part of MariaDB Platform.
In this session, we will provide a preview of the MariaDB Enterprise Client, walk through current and planned modules and discuss future plans for MariaDB Enterprise Tools – including SkySQL modules and the ability to create custom modules.
SkySQL implements a groundbreaking, state-of-the-art architecture based on Kubernetes and ServiceNow, and with a strong emphasis on cloud security – using compartmentalization and indirect access to secure and protect customer databases.
In this session, we’ll walk through the architecture of SkySQL and discuss how MariaDB leverages an advanced Kubernetes operator and powerful ServiceNow configuration/workflow management to deploy and manage databases on cloud infrastructure.
Introducing the ultimate MariaDB cloud, SkySQLMariaDB plc
SkySQL is the first and only database-as-a-service (DBaaS) engineered for MariaDB by MariaDB, to use a state-of-the-art multi-cloud architecture built on Kubernetes and ServiceNow, and to deploy databases and data warehouses for transactional, analytical and hybrid transactional/analytical workloads.
In this session, we’ll lay out the vision for SkySQL, provide an overview of its capabilities, take a tour of its architecture, and discuss the long-term roadmap. We’ll wrap things up with a live demo of SkySQL, including a preview of its deep learning–based workload analysis and visualization interface.
Beyond the basics: advanced SQL with MariaDBMariaDB plc
We've been writing SQL queries with WHERE, GROUP BY, ORDER BY, HAVING for decades, but we’re not using DOS 3.2 or Windows 1.0 anymore. Why limit yourself to SQL:86? In the past couple of releases, MariaDB has added support for features in the SQL:99 (common table expressions), SQL:2003 (window functions), SQL:2011 (system-versioned tables), and SQL:2016 (JSON) specifications – allowing you to build more complex data models (e.g., semi-structured or hierarchical) and write simpler, faster queries. In this session, Sergei Golubchik brings everyone up to speed on the latest SQL syntax supported in MariaDB.
Adjusting primitives for graph : SHORT REPORT / NOTESSubhajit Sahu
Graph algorithms, like PageRank Compressed Sparse Row (CSR) is an adjacency-list based graph representation that is
Multiply with different modes (map)
1. Performance of sequential execution based vs OpenMP based vector multiply.
2. Comparing various launch configs for CUDA based vector multiply.
Sum with different storage types (reduce)
1. Performance of vector element sum using float vs bfloat16 as the storage type.
Sum with different modes (reduce)
1. Performance of sequential execution based vs OpenMP based vector element sum.
2. Performance of memcpy vs in-place based CUDA based vector element sum.
3. Comparing various launch configs for CUDA based vector element sum (memcpy).
4. Comparing various launch configs for CUDA based vector element sum (in-place).
Sum with in-place strategies of CUDA mode (reduce)
1. Comparing various launch configs for CUDA based vector element sum (in-place).
As Europe's leading economic powerhouse and the fourth-largest hashtag#economy globally, Germany stands at the forefront of innovation and industrial might. Renowned for its precision engineering and high-tech sectors, Germany's economic structure is heavily supported by a robust service industry, accounting for approximately 68% of its GDP. This economic clout and strategic geopolitical stance position Germany as a focal point in the global cyber threat landscape.
In the face of escalating global tensions, particularly those emanating from geopolitical disputes with nations like hashtag#Russia and hashtag#China, hashtag#Germany has witnessed a significant uptick in targeted cyber operations. Our analysis indicates a marked increase in hashtag#cyberattack sophistication aimed at critical infrastructure and key industrial sectors. These attacks range from ransomware campaigns to hashtag#AdvancedPersistentThreats (hashtag#APTs), threatening national security and business integrity.
🔑 Key findings include:
🔍 Increased frequency and complexity of cyber threats.
🔍 Escalation of state-sponsored and criminally motivated cyber operations.
🔍 Active dark web exchanges of malicious tools and tactics.
Our comprehensive report delves into these challenges, using a blend of open-source and proprietary data collection techniques. By monitoring activity on critical networks and analyzing attack patterns, our team provides a detailed overview of the threats facing German entities.
This report aims to equip stakeholders across public and private sectors with the knowledge to enhance their defensive strategies, reduce exposure to cyber risks, and reinforce Germany's resilience against cyber threats.
3. Nokia Common Software Foundation (CSF)
● Central development organization for commonly used components
● Work with PUs within company to determine needs/plans
● Determine which open-source products are used/planned by multiple PUs
● Productize highly used open-source products in central organization to
eliminate silo-based development model to
○ Reduce overall corporate development costs
○ Improve time to market for products
4. CSF OS/DB Department
● Centralized OS (Linux) and DB development and distribution
● Each DB component will define common reference architecture(s)
● CSF DB component develops the following around the open-source DB for
each reference architecture:
○ Deployment and Life Cycle Management (LCM) of database in Cloud/openstack environment
○ Docker container for deployment and LCM via kubernetes/Helm → PaaS
○ Tools/scripts to support deployment and management of DB as above, including ansible
● Centralized support for OS and DB within corporation
5. CSF CMDB (MariaDB Component)
● MariaDB was selected as one of the CSF DB components
● CMDB provides DB server with following reference architectures:
○ Standalone
○ Galera Cluster w/ or w/o Arbitrator
○ Replication: Master/Master and Master/Slave
● Galera or Master-Master w/JDBC mostly used for basic HA needs
● Deployment and LCM developed and packaged for:
○ Cloud/openstack ansible automated deployment for all reference architectures
○ Automated kubernetes container deployment via Helm charts
6. HA clustered DB
Geo-Redundant HA Requirements
● Multi-datacenter deployment (dual data center
initially)
● Must have HA clustering at each datacenter
● 99.999% availability
● 5 second WAN replication lag tolerance
● 6 hours of Transaction Buffering for datacenter
disconnect
● Automatic replication recovery after datacenter
disconnect
● Procedure to recover standby data center from
active datacenter after 6 hours of replication failure
APPLICATION APPLICATION
DC-A DC-B
HA clustered DB
7. Geo-Redundant Application Assumptions
● Application will write to only one datacenter at a time
● Application is responsible for failing over to alternate datacenter on failure
● Inter-DC replication lag and binlog buffering are a function of application
transaction rate and WAN speed and reliability
● Some data loss acceptable on failure
8. Architecture Alternatives
● Galera Cluster at each DC with segment_id definition to limit inter-DC traffic
over a single link
○ Achieve HA at each DC with minimal data loss (internal)
○ Would ideally require 3 datacenters with 3 nodes each for quorum requirements
○ Synchronous replication between DCs could cause severe performance impacts for more
write-intensive applications
● Master-Master at each DC
○ Can provide intra-DC HA via JDBC
○ Auto_increment_offset/increment would have to be managed for all nodes at all DCs
○ Still have inter-DC replication path adjustments needed on local DC failover
● Still need node to run service to monitor alternate DC health and alarm
○ MaxScale makes sense here to provide proxy and manage local cluster
○ Nokia added services will monitor alternate DC health via MaxScale
9. Geo-Redundant Reference Architecture
read write
MASTERSLAVES
readwrite
MASTER SLAVES
APPLICATION APPLICATION
DC-A DC-B
Master/Slave at each
datacenter for HA and
auto-failover
Masters in each
datacenter cross-
replicate to each other
10. Datacenter HA (automatic failover)
● First started working with replication-manager to support auto-failover
○ Replication-manager runs along-side MaxScale and performs auto-failover
○ MaxScale expected to auto-detect topology changes and update status - not always
○ Replication-manager configuration very complex
● Discovered maxscale-2.2 with auto-failover built in
○ Single point topology manager (no inconsistencies)
○ No additional 3rd party open-source plugins required (fully supported)
○ Worked with MaxScale development team as beta-tester
11. MaxScale-2.2 Initial Testing
● Very simple configuration
ignore_external_masters = true
auto_failover = true
auto_rejoin = true
● Maxscale provides full datacenter HA
○ Automatic promotion of a Slave as new Master when Master fails
○ Rejoin of previous failed master as Slave when recovers
○ Maxscale manages inter-DC cluster, we would add scripts to fix replication in both directions
when Master failover occurs in either DC
○ Huge simplification over using 3rd party replication managers / deterministic behavior
○ Maxscale now has crash-safe recovery
12. MaxScale-2.2 Final Testing
● New configuration
ignore_external_masters = false
auto_failover = true
auto_rejoin = true
● Additional behavior
○ Changed to set ignore_external_masters = true
○ New Master automatically replicates from same external Master as failed Master
○ Notification script plugin on topology changes allows us to automatically fix alternate
datacenter Master to replication from new Master on “new_master” notify event
○ Failover very quick (500ms monitor interval) - 2 to 3 seconds?
13. Master/Slave Database HA with MaxScale
• Automatic Failover: election and promotion of slave by MaxScale
• Continues routing read queries to slave during failover
• Manual failover and switchover interface
readwrite
MASTER SLAVES
MAXSCALE
1
1 Master fails
2 MaxScale elects new master
3 MaxScale promote candidate slave as new master
4 MaxScale instructs rest of the slave of new master
5 MaxScale sets new master to replication from old
Master external server (if existed)
6 MaxScale calls new_master event notify - Nokia
script to fix external master to replication to new
local master
2,34
14. Containerized Delivery
● Deployment of Master/Slave container cluster
○ Automatic configuration of first container as Master, other two Slave
○ All containers will automatically recover as Slave role
○ Advertise IPs via etcd service advertisement
● Deployment of Maxscale container
○ Gets all config from kubernetes manifest file and server IPs from etcd advertisement
○ Automatically configures maxscale and starts monitor of DB M/S cluster
● Container failure and re-deployed with different IP
○ Developed service to monitor etcd IP advertisements and detect host IP change at MaxScale
○ Run: maxadmin alter server <hostname> address=<ip>
15. Nokia developed capabilities
● Notify script plugin on “new_master” event to set remote DC master to new
promoted master
● Additional service will be developed to perform the following additional
functions:
○ Monitor MaxScale in alternate datacenter to verify datacenter health and verify local master
replicating to correct external master
○ Monitor replication to generate SNMP traps when replication breaks
○ Monitor replication lag and generate SNMP traps when lag exceeds threshold
○ Possible implementation of replication CHANGE MASTER retry if replication fails due to GTID
not found
○ Since log_slave_updates must be set, configurable “slave flush interval” to flush binlogs in
slaves
16. Future work
● Still need to see how replication will hold during failure under load
conditions
● MaxScale enhancement to full automate inter-DC master failover (both
directions)
● Support Galera Cluster as HA solution in each DC
● Support MaxScale HA cluster at each DC (via keepalived)
● Support segregated database access on both databases at same time?