PGConf.ASIA 2019 Bali - How did PostgreSQL Write Load Balancing of Queries Us...Equnix Business Solutions
Atsushi Mitani from SRA Nishi-Nihon Inc. presented on how to perform write load balancing in PostgreSQL using transactions. He explained that write load distribution is important for systems with high write volumes. PostgreSQL can distribute write load using table partitioning with foreign data wrappers (FDW), which allows partitioning across database instances. Mitani created patches to automate the partitioning setup and load data in parallel to child tables to speed up benchmarking. Benchmark results showed that while increasing child databases improves performance without transactions, increasing parent databases is better with transactions to avoid lock queues. The optimal configuration depends on data size, queries, and hardware.
PGConf APAC 2018 - PostgreSQL performance comparison in various cloudsPGConf APAC
Speaker: Oskari Saarenmaa
Aiven PostgreSQL is available in five different public cloud providers' infrastructure in more than 60 regions around the world, including 18 in APAC. This has given us a unique opportunity to benchmark and compare performance of similar configurations in different environments.
We'll share our benchmark methods and results, comparing various PostgreSQL configurations and workloads across different clouds.
PGConf APAC 2018 - Monitoring PostgreSQL at ScalePGConf APAC
Speaker: Lukas Fittl
Your PostgreSQL database is one of the most important pieces of your architecture - yet the level of introspection available in Postgres is often hard to work with. Its easy to get very detailed information, but what should you really watch out for, send reports on and alert on?
In this talk we'll discuss how query performance statistics can be made accessible to application developers, critical entries one should monitor in the PostgreSQL log files, how to collect EXPLAIN plans at scale, how to watch over autovacuum and VACUUM operations, and how to flag issues based on schema statistics.
We'll also talk a bit about monitoring multi-server setups, first going into high availability and read standbys, logical replication, and then reviewing how monitoring looks like for sharded databases like Citus.
The talk will primarily describe free/open-source tools and statistics views readily available from within Postgres.
PGConf.ASIA 2019 Bali - Setup a High-Availability and Load Balancing PostgreS...Equnix Business Solutions
PGConf.ASIA 2019 Bali - 10 September 2019
Speaker: Bo Peng
Room: SQL
Title: Setup a High-Availability and Load Balancing PostgreSQL Cluster - New Features of Pgpool-II 4.1
PGConf APAC 2018 - A PostgreSQL DBAs Toolbelt for 2018PGConf APAC
There's no need to re-invent the wheel! Dozens of people have already tried...and succeeded. This talk is a categorized and illustrated overview on most popular and/or useful PostgreSQL specific scripts, utilities and whole toolsets that DBAs should be aware of for solving daily tasks. Inlcuding - performance monitoring, logs management/analyzis, identifying/fixing most common adminstration problems around areas of general performance metrics, tuning, locking, indexing, bloat, leaving out high-availability topics. Covered are venerable oldies from wiki.postgresql.org as well as my newer favourites from Github.
PGConf.ASIA 2019 Bali - How did PostgreSQL Write Load Balancing of Queries Us...Equnix Business Solutions
Atsushi Mitani from SRA Nishi-Nihon Inc. presented on how to perform write load balancing in PostgreSQL using transactions. He explained that write load distribution is important for systems with high write volumes. PostgreSQL can distribute write load using table partitioning with foreign data wrappers (FDW), which allows partitioning across database instances. Mitani created patches to automate the partitioning setup and load data in parallel to child tables to speed up benchmarking. Benchmark results showed that while increasing child databases improves performance without transactions, increasing parent databases is better with transactions to avoid lock queues. The optimal configuration depends on data size, queries, and hardware.
PGConf APAC 2018 - PostgreSQL performance comparison in various cloudsPGConf APAC
Speaker: Oskari Saarenmaa
Aiven PostgreSQL is available in five different public cloud providers' infrastructure in more than 60 regions around the world, including 18 in APAC. This has given us a unique opportunity to benchmark and compare performance of similar configurations in different environments.
We'll share our benchmark methods and results, comparing various PostgreSQL configurations and workloads across different clouds.
PGConf APAC 2018 - Monitoring PostgreSQL at ScalePGConf APAC
Speaker: Lukas Fittl
Your PostgreSQL database is one of the most important pieces of your architecture - yet the level of introspection available in Postgres is often hard to work with. Its easy to get very detailed information, but what should you really watch out for, send reports on and alert on?
In this talk we'll discuss how query performance statistics can be made accessible to application developers, critical entries one should monitor in the PostgreSQL log files, how to collect EXPLAIN plans at scale, how to watch over autovacuum and VACUUM operations, and how to flag issues based on schema statistics.
We'll also talk a bit about monitoring multi-server setups, first going into high availability and read standbys, logical replication, and then reviewing how monitoring looks like for sharded databases like Citus.
The talk will primarily describe free/open-source tools and statistics views readily available from within Postgres.
PGConf.ASIA 2019 Bali - Setup a High-Availability and Load Balancing PostgreS...Equnix Business Solutions
PGConf.ASIA 2019 Bali - 10 September 2019
Speaker: Bo Peng
Room: SQL
Title: Setup a High-Availability and Load Balancing PostgreSQL Cluster - New Features of Pgpool-II 4.1
PGConf APAC 2018 - A PostgreSQL DBAs Toolbelt for 2018PGConf APAC
There's no need to re-invent the wheel! Dozens of people have already tried...and succeeded. This talk is a categorized and illustrated overview on most popular and/or useful PostgreSQL specific scripts, utilities and whole toolsets that DBAs should be aware of for solving daily tasks. Inlcuding - performance monitoring, logs management/analyzis, identifying/fixing most common adminstration problems around areas of general performance metrics, tuning, locking, indexing, bloat, leaving out high-availability topics. Covered are venerable oldies from wiki.postgresql.org as well as my newer favourites from Github.
PGConf.ASIA 2019 - PGSpider High Performance Cluster Engine - Shigeo HiroseEqunix Business Solutions
PGSpider is a high-performance SQL cluster engine developed by Toshiba Corporation. It allows distributed querying of heterogeneous data sources using standard SQL. PGSpider improves retrieval performance through parallel queries across nodes and supports multi-tenant querying to retrieve records from the same table across nodes. It utilizes techniques like pushdown of conditional expressions and aggregation functions to nodes to reduce network traffic.
PGConf APAC 2018 - Managing replication clusters with repmgr, Barman and PgBo...PGConf APAC
Speaker: Ian Barwick
PostgreSQL and reliability go hand-in-hand - but your data is only truly safe with a solid and trusted backup system in place, and no matter how good your application is, it's useless if it can't talk to your database.
In this talk we'll demonstrate how to set up a reliable replication
cluster using open source tools closely associated with the PostgreSQL project. The talk will cover following areas:
- how to set up and manage a replication cluster with `repmgr`
- how to set up and manage reliable backups with `Barman`
- how to manage failover and application connections with `repmgr` and `PgBouncer`
Ian Barwick has worked for 2ndQuadrant since 2014, and as well as making various contributions to PostgreSQL itself, is lead `repmgr` developer. He lives in Tokyo, Japan.
PGConf APAC 2018 - PostgreSQL HA with Pgpool-II and whats been happening in P...PGConf APAC
Speaker: Muhammad Usama
Pgpool-II has been around to complement PostgreSQL over a decade and provides many features like connection pooling, failover, query caching, load balancing, and HA. High Availability (HA) is very critical to most enterprise application, the clients needs the ability to automatically reconnect with a secondary node when the master nodes goes down.
This is where Pgpool-II watchdog feature comes in, the core feature of Pgpool-II provides HA by eliminating the SPOF is the Watchdog. This watchdog feature has been around for a while but it went through major overhauling and enhancements in recent releases. This talk aims to explain the watchdog feature, the recent enhancements went into the watchdog and describe how it can be used to provide PostgreSQL HA and automatic failover.
Their is rising trend of enterprise deployment shifting to cloud based environment, Pgpool II can be used in the cloud without any issues. In this talk we will give some ideas how Pgpool-II is used to provide PostgreSQL HA in cloud environment.
Finally we will summarise the major features that have been added in the recent major release of Pgpool II and whats in the pipeline for the next major release.
The document discusses PostgreSQL version 11 and future development. It provides a history of PostgreSQL and its predecessors, describing the development process and community. It summarizes key features committed to version 11, including improvements to partitioning, parallelization, performance and logical replication. It also outlines features proposed for future versions, with a focus on continued enhancements to partitioning and query planning.
This document provides an introduction to HeteroDB, Inc. and its chief architect, KaiGai Kohei. It discusses PG-Strom, an open source PostgreSQL extension developed by HeteroDB for high performance data processing using heterogeneous architectures like GPUs. PG-Strom uses techniques like SSD-to-GPU direct data transfer and a columnar data store to accelerate analytics and reporting workloads on terabyte-scale log data using GPUs and NVMe SSDs. Benchmark results show PG-Strom can process terabyte workloads at throughput nearing the hardware limit of the storage and network infrastructure.
About a year ago I was caught up in line-of-fire when a production system started behaving abruptly
- A batch process which would finish in 15minutes started taking 1.5 hours
- We started facing OLTP read queries on standby being cancelled
- We faced a sudden slowness on the Primary server and we were forced to do a forceful switch to standby.
We were able to figure out that some peculiarities of the application code and batch process were responsible for this. But we could not fix the application code (as it is packaged application).
In this talk I would like to share more details of how we debugged, what was the problem we were facing and how we applied a work around for it. We also learnt that a query returning in 10minutes may not be as dangerous as a query returning in 10sec but executed 100s of times in an hour.
I will share in detail-
- How to map the process/top stats from OS with pg_stat_activity
- How to get and read explain plan
- How to judge if a query is costly
- What tools helped us
- A peculiar autovacuum/vacuum Vs Replication conflict we ran into
- Various parameters to tune autvacuum and auto-analyze process
- What we have done to work-around the problem
- What we have put in place for better monitoring and information gathering
At a blistering pace and for a variety of reasons, companies are migrating their on-premise database infrastructures to cloud-based solutions - to save costs on hardware, tame the impact of disaster recovery, or even to improve security. Zalando is not an exception: more than two years ago we migrated our first production services to AWS.
In addition to the fully managed database services like RDS and Aurora, Amazon offers a wide spectra of EC2 Instances with different types of performance and price. Without a lot of experience in running cloud databases it’s not easy to make the right choice, and as a result you will either have pure database performance or will overpay for over-provisioned resources.
In this talk I will compare different ways of running PostgreSQL on AWS, explain why we decided to run most of our databases on EC2 Instances instead of RDS, how we chose EC2 Instance types and EBS Volumes, which AWS CloudWatch metrics MUST be monitored (and why), and what problems we hit plus how to avoid them.
Taming GC Pauses for Humongous Java Heaps in Spark Graph Computing-(Eric Kacz...Spark Summit
The document discusses tuning the Garbage First (G1) garbage collector in Java 8 to reduce garbage collection pauses for large heaps used in Spark graph computing workloads. It was found that the default G1 settings resulted in lengthy full garbage collections over 100 seconds. After analyzing the garbage collection logs, the main issue was identified as the concurrent marking phase not completing before a full collection was needed. Increasing the number of concurrent marking threads from 8 to 20 addressed this by speeding up the concurrent phase. With this tuning, no full collections occurred and total stop-the-world pause time was reduced to under a minute, a significant improvement over the original implementation.
(Nina Hanzlikova, Zalando) Kafka Summit SF 2018
My team at Zalando fell in love with KStreams and their programming model straight out of the gate. However, as a small team of developers, building out and supporting our infrastructure while still trying to deliver solutions for our business has not always resulted in a smooth journey.
Can a small team of a couple of developers run their own Kafka infrastructure confidently and still spend most of their time developing code?
In this talk, we will dive into some of the problems we experienced while running Kafka brokers and Kafka Streams applications, as well as the consultations we had with other teams around this matter. We will outline some of the pragmatic decisions we made regarding backups, monitoring and operations to minimize our time spent administering our Kafka brokers and various stream applications.
Understanding Memory Management In Spark For Fun And ProfitSpark Summit
1) The document discusses memory management in Spark applications and summarizes different approaches tried by developers to address out of memory errors in Spark executors.
2) It analyzes the root causes of memory issues like executor overheads and data sizes, and evaluates fixes like increasing memory overhead, reducing cores, frequent garbage collection.
3) The document dives into Spark and JVM level configuration options for memory like storage pool sizes, caching formats, and garbage collection settings to improve reliability, efficiency and performance of Spark jobs.
Speaker: Alexander Kukushkin
Kubernetes is a solid leader among different cloud orchestration engines and its adoption rate is growing on a daily basis. Naturally people want to run both their applications and databases on the same infrastructure.
There are a lot of ways to deploy and run PostgreSQL on Kubernetes, but most of them are not cloud-native. Around one year ago Zalando started to run HA setup of PostgreSQL on Kubernetes managed by Patroni. Those experiments were quite successful and produced a Helm chart for Patroni. That chart was useful, albeit a single problem: Patroni depended on Etcd, ZooKeeper or Consul.
Few people look forward to deploy two applications instead of one and support them later on. In this talk I would like to introduce Kubernetes-native Patroni. I will explain how Patroni uses Kubernetes API to run a leader election and store the cluster state. I’m going to live-demo a deployment of HA PostgreSQL cluster on Minikube and share our own experience of running more than 130 clusters on Kubernetes.
Patroni is a Python open-source project developed by Zalando in cooperation with other contributors on GitHub: https://github.com/zalando/patroni
Reduce Redundant Producers from Partitioned Producer - Pulsar Summit NA 2021StreamNative
The document discusses reducing redundant Apache Pulsar producers from partitioned topics. It presents a solution that limits the number of internal producers per partitioned producer and lazily loads producers. Benchmark results show the proposed approach reduces client-side resource usage like heap and number of TCP connections compared to the existing approach, while having little impact on broker-side resources. The conclusion is that implementing these producer changes can improve efficiency.
hbaseconasia2017: HBase Practice At XiaoMiHBaseCon
Zheng Hu
We'll share some HBase experience at XiaoMi:
1. How did we tuning G1GC for HBase Clusters.
2. Development and performance of Async HBase Client.
hbaseconasia2017 hbasecon hbase xiaomi https://www.eventbrite.com/e/hbasecon-asia-2017-tickets-34935546159#
Take Kafka-on-Pulsar to Production at Internet Scale: Improvements Made for P...StreamNative
This document discusses improvements made to Kafka-on-Pulsar (KoP) for Pulsar 2.8.0 to help take it to production at internet scale. Key updates include continuous offset support, Kafka entry formatting, heap memory optimizations, exposed metrics, and OAuth 2.0 authentication. KoP is now generally available and sees use at Tencent supporting 60 trillion daily messages through optimizations like speeding up reboots, producer throttling, and implementing continuous offsets.
SparklingWater provides the ability to pass data between Spark and H2O, allowing machine learning to be done using both Spark and H2O algorithms and at scale. It supports popular languages like R, Python, Java and Scala and common ML algorithms like generalized linear models, K-means clustering, and random forests. Models can be exported without needing H2O, simplifying the typical ML workflow by performing all stages within the same Spark/H2O environment.
HBaseCon2017 Improving HBase availability in a multi tenant environmentHBaseCon
The document discusses improvements made by Hubspot's Big Data Team to increase the availability of HBase in a multi-tenant environment. It outlines reducing the cost of region server failures by improving mean time to recovery, addressing issues that slowed recovery, and optimizing the load balancer. It also details eliminating workload-driven failures through service limits and improving hardware monitoring to reduce impacts of failures. The changes resulted in 8-10x faster balancing, reduced recovery times from 90 to 30 seconds, and consistently achieving 99.99% availability across clusters.
While the physical replication in PostgreSQL is quite robust, however, it doesn’t fit well in the picture when:
- You need partial replication only
- You want to replicate between different major versions of PostgreSQL
- You need to replicate multiple databases to the same target
- Transformation of the data is needed
- You want to replicate in order to upgrade without downtime
The answer to these use cases is logical replication
This talk will discuss and cover these use cases followed by a logical replication demo.
Webinar: PostgreSQL continuous backup and PITR with BarmanGabriele Bartolini
How can you achieve an RTO of 5 minutes for the backups of your PostgreSQL databases? And what about RPO=0 for zero data loss backups?
This webinar gave an answer to those questions, by providing an overview of Disaster Recovery of PostgreSQL databases with Barman, covering its major features.
Barman, Backup and Recovery Manager for PostgreSQL, is an open source tool that was conceived by 2ndQuadrant about 10 years ago and released open source in 2012 under GNU GPL 3.
It is now one of the most popular backup and recovery tools in the PostgreSQL ecosystem.
Video available at https://resources.2ndquadrant.com/en/webinar-postgresql-continuous-backup-and-pitr-with-barman
PGConf.ASIA 2019 - PGSpider High Performance Cluster Engine - Shigeo HiroseEqunix Business Solutions
PGSpider is a high-performance SQL cluster engine developed by Toshiba Corporation. It allows distributed querying of heterogeneous data sources using standard SQL. PGSpider improves retrieval performance through parallel queries across nodes and supports multi-tenant querying to retrieve records from the same table across nodes. It utilizes techniques like pushdown of conditional expressions and aggregation functions to nodes to reduce network traffic.
PGConf APAC 2018 - Managing replication clusters with repmgr, Barman and PgBo...PGConf APAC
Speaker: Ian Barwick
PostgreSQL and reliability go hand-in-hand - but your data is only truly safe with a solid and trusted backup system in place, and no matter how good your application is, it's useless if it can't talk to your database.
In this talk we'll demonstrate how to set up a reliable replication
cluster using open source tools closely associated with the PostgreSQL project. The talk will cover following areas:
- how to set up and manage a replication cluster with `repmgr`
- how to set up and manage reliable backups with `Barman`
- how to manage failover and application connections with `repmgr` and `PgBouncer`
Ian Barwick has worked for 2ndQuadrant since 2014, and as well as making various contributions to PostgreSQL itself, is lead `repmgr` developer. He lives in Tokyo, Japan.
PGConf APAC 2018 - PostgreSQL HA with Pgpool-II and whats been happening in P...PGConf APAC
Speaker: Muhammad Usama
Pgpool-II has been around to complement PostgreSQL over a decade and provides many features like connection pooling, failover, query caching, load balancing, and HA. High Availability (HA) is very critical to most enterprise application, the clients needs the ability to automatically reconnect with a secondary node when the master nodes goes down.
This is where Pgpool-II watchdog feature comes in, the core feature of Pgpool-II provides HA by eliminating the SPOF is the Watchdog. This watchdog feature has been around for a while but it went through major overhauling and enhancements in recent releases. This talk aims to explain the watchdog feature, the recent enhancements went into the watchdog and describe how it can be used to provide PostgreSQL HA and automatic failover.
Their is rising trend of enterprise deployment shifting to cloud based environment, Pgpool II can be used in the cloud without any issues. In this talk we will give some ideas how Pgpool-II is used to provide PostgreSQL HA in cloud environment.
Finally we will summarise the major features that have been added in the recent major release of Pgpool II and whats in the pipeline for the next major release.
The document discusses PostgreSQL version 11 and future development. It provides a history of PostgreSQL and its predecessors, describing the development process and community. It summarizes key features committed to version 11, including improvements to partitioning, parallelization, performance and logical replication. It also outlines features proposed for future versions, with a focus on continued enhancements to partitioning and query planning.
This document provides an introduction to HeteroDB, Inc. and its chief architect, KaiGai Kohei. It discusses PG-Strom, an open source PostgreSQL extension developed by HeteroDB for high performance data processing using heterogeneous architectures like GPUs. PG-Strom uses techniques like SSD-to-GPU direct data transfer and a columnar data store to accelerate analytics and reporting workloads on terabyte-scale log data using GPUs and NVMe SSDs. Benchmark results show PG-Strom can process terabyte workloads at throughput nearing the hardware limit of the storage and network infrastructure.
About a year ago I was caught up in line-of-fire when a production system started behaving abruptly
- A batch process which would finish in 15minutes started taking 1.5 hours
- We started facing OLTP read queries on standby being cancelled
- We faced a sudden slowness on the Primary server and we were forced to do a forceful switch to standby.
We were able to figure out that some peculiarities of the application code and batch process were responsible for this. But we could not fix the application code (as it is packaged application).
In this talk I would like to share more details of how we debugged, what was the problem we were facing and how we applied a work around for it. We also learnt that a query returning in 10minutes may not be as dangerous as a query returning in 10sec but executed 100s of times in an hour.
I will share in detail-
- How to map the process/top stats from OS with pg_stat_activity
- How to get and read explain plan
- How to judge if a query is costly
- What tools helped us
- A peculiar autovacuum/vacuum Vs Replication conflict we ran into
- Various parameters to tune autvacuum and auto-analyze process
- What we have done to work-around the problem
- What we have put in place for better monitoring and information gathering
At a blistering pace and for a variety of reasons, companies are migrating their on-premise database infrastructures to cloud-based solutions - to save costs on hardware, tame the impact of disaster recovery, or even to improve security. Zalando is not an exception: more than two years ago we migrated our first production services to AWS.
In addition to the fully managed database services like RDS and Aurora, Amazon offers a wide spectra of EC2 Instances with different types of performance and price. Without a lot of experience in running cloud databases it’s not easy to make the right choice, and as a result you will either have pure database performance or will overpay for over-provisioned resources.
In this talk I will compare different ways of running PostgreSQL on AWS, explain why we decided to run most of our databases on EC2 Instances instead of RDS, how we chose EC2 Instance types and EBS Volumes, which AWS CloudWatch metrics MUST be monitored (and why), and what problems we hit plus how to avoid them.
Taming GC Pauses for Humongous Java Heaps in Spark Graph Computing-(Eric Kacz...Spark Summit
The document discusses tuning the Garbage First (G1) garbage collector in Java 8 to reduce garbage collection pauses for large heaps used in Spark graph computing workloads. It was found that the default G1 settings resulted in lengthy full garbage collections over 100 seconds. After analyzing the garbage collection logs, the main issue was identified as the concurrent marking phase not completing before a full collection was needed. Increasing the number of concurrent marking threads from 8 to 20 addressed this by speeding up the concurrent phase. With this tuning, no full collections occurred and total stop-the-world pause time was reduced to under a minute, a significant improvement over the original implementation.
(Nina Hanzlikova, Zalando) Kafka Summit SF 2018
My team at Zalando fell in love with KStreams and their programming model straight out of the gate. However, as a small team of developers, building out and supporting our infrastructure while still trying to deliver solutions for our business has not always resulted in a smooth journey.
Can a small team of a couple of developers run their own Kafka infrastructure confidently and still spend most of their time developing code?
In this talk, we will dive into some of the problems we experienced while running Kafka brokers and Kafka Streams applications, as well as the consultations we had with other teams around this matter. We will outline some of the pragmatic decisions we made regarding backups, monitoring and operations to minimize our time spent administering our Kafka brokers and various stream applications.
Understanding Memory Management In Spark For Fun And ProfitSpark Summit
1) The document discusses memory management in Spark applications and summarizes different approaches tried by developers to address out of memory errors in Spark executors.
2) It analyzes the root causes of memory issues like executor overheads and data sizes, and evaluates fixes like increasing memory overhead, reducing cores, frequent garbage collection.
3) The document dives into Spark and JVM level configuration options for memory like storage pool sizes, caching formats, and garbage collection settings to improve reliability, efficiency and performance of Spark jobs.
Speaker: Alexander Kukushkin
Kubernetes is a solid leader among different cloud orchestration engines and its adoption rate is growing on a daily basis. Naturally people want to run both their applications and databases on the same infrastructure.
There are a lot of ways to deploy and run PostgreSQL on Kubernetes, but most of them are not cloud-native. Around one year ago Zalando started to run HA setup of PostgreSQL on Kubernetes managed by Patroni. Those experiments were quite successful and produced a Helm chart for Patroni. That chart was useful, albeit a single problem: Patroni depended on Etcd, ZooKeeper or Consul.
Few people look forward to deploy two applications instead of one and support them later on. In this talk I would like to introduce Kubernetes-native Patroni. I will explain how Patroni uses Kubernetes API to run a leader election and store the cluster state. I’m going to live-demo a deployment of HA PostgreSQL cluster on Minikube and share our own experience of running more than 130 clusters on Kubernetes.
Patroni is a Python open-source project developed by Zalando in cooperation with other contributors on GitHub: https://github.com/zalando/patroni
Reduce Redundant Producers from Partitioned Producer - Pulsar Summit NA 2021StreamNative
The document discusses reducing redundant Apache Pulsar producers from partitioned topics. It presents a solution that limits the number of internal producers per partitioned producer and lazily loads producers. Benchmark results show the proposed approach reduces client-side resource usage like heap and number of TCP connections compared to the existing approach, while having little impact on broker-side resources. The conclusion is that implementing these producer changes can improve efficiency.
hbaseconasia2017: HBase Practice At XiaoMiHBaseCon
Zheng Hu
We'll share some HBase experience at XiaoMi:
1. How did we tuning G1GC for HBase Clusters.
2. Development and performance of Async HBase Client.
hbaseconasia2017 hbasecon hbase xiaomi https://www.eventbrite.com/e/hbasecon-asia-2017-tickets-34935546159#
Take Kafka-on-Pulsar to Production at Internet Scale: Improvements Made for P...StreamNative
This document discusses improvements made to Kafka-on-Pulsar (KoP) for Pulsar 2.8.0 to help take it to production at internet scale. Key updates include continuous offset support, Kafka entry formatting, heap memory optimizations, exposed metrics, and OAuth 2.0 authentication. KoP is now generally available and sees use at Tencent supporting 60 trillion daily messages through optimizations like speeding up reboots, producer throttling, and implementing continuous offsets.
SparklingWater provides the ability to pass data between Spark and H2O, allowing machine learning to be done using both Spark and H2O algorithms and at scale. It supports popular languages like R, Python, Java and Scala and common ML algorithms like generalized linear models, K-means clustering, and random forests. Models can be exported without needing H2O, simplifying the typical ML workflow by performing all stages within the same Spark/H2O environment.
HBaseCon2017 Improving HBase availability in a multi tenant environmentHBaseCon
The document discusses improvements made by Hubspot's Big Data Team to increase the availability of HBase in a multi-tenant environment. It outlines reducing the cost of region server failures by improving mean time to recovery, addressing issues that slowed recovery, and optimizing the load balancer. It also details eliminating workload-driven failures through service limits and improving hardware monitoring to reduce impacts of failures. The changes resulted in 8-10x faster balancing, reduced recovery times from 90 to 30 seconds, and consistently achieving 99.99% availability across clusters.
While the physical replication in PostgreSQL is quite robust, however, it doesn’t fit well in the picture when:
- You need partial replication only
- You want to replicate between different major versions of PostgreSQL
- You need to replicate multiple databases to the same target
- Transformation of the data is needed
- You want to replicate in order to upgrade without downtime
The answer to these use cases is logical replication
This talk will discuss and cover these use cases followed by a logical replication demo.
Webinar: PostgreSQL continuous backup and PITR with BarmanGabriele Bartolini
How can you achieve an RTO of 5 minutes for the backups of your PostgreSQL databases? And what about RPO=0 for zero data loss backups?
This webinar gave an answer to those questions, by providing an overview of Disaster Recovery of PostgreSQL databases with Barman, covering its major features.
Barman, Backup and Recovery Manager for PostgreSQL, is an open source tool that was conceived by 2ndQuadrant about 10 years ago and released open source in 2012 under GNU GPL 3.
It is now one of the most popular backup and recovery tools in the PostgreSQL ecosystem.
Video available at https://resources.2ndquadrant.com/en/webinar-postgresql-continuous-backup-and-pitr-with-barman
Maarten Mulders - Mastering Microservices with Kong - Codemotion Amsterdam 2019Codemotion
When architecting microservice solutions, you'll often find yourself struggling with cross-cutting concerns. Think security, rate limiting, access control, monitoring, location-aware routing… Things can quickly become a nightmare. The API Gateway pattern can help you solve such problems in an elegant and uniform way. Using Kong, an open source product, you can get started today. In this session we'll look at the why and how of this approach. Disclaimer: This presentation may include live coding. No humans or animals will be hurt during the process.
Mastering Microservices with Kong (CodeMotion 2019)Maarten Mulders
This document discusses API management and API gateways. It introduces API management as the process of creating and publishing APIs, enforcing usage policies, and collecting analytics. It then discusses how API gateways can provide functionality like routing, authentication, rate limiting and analytics for microservices. The document demonstrates configuring Kong, an open source API gateway, to route requests to beer and brewery microservices and add authentication. It also discusses how plugins can extend Kong's functionality and how Kong was a good fit for its flexibility and community support.
Create a Custom Plugin in Burp Suite using the ExtensionNSConclave
This document discusses creating a custom plugin in Burp Suite using the extension framework. It provides advantages of using the extension, requirements, an overview of implementing request and response functions on the server, server helper functions, and a demo of creating a custom plugin that decrypts and encrypts requests and responses for a bank web application. The presentation agenda includes an introduction, block diagram, requirements, running the server, request and response functions, server helper functions, and a demo.
Disaster Recovery for Multi-Region Apache Kafka Ecosystems at Uberconfluent
Speaker: Yupeng Fu, Staff Engineer, Uber
High availability and reliability are important requirements to Uber services, and the services shall tolerate datacenter failures in a region and fail over to another region. In this talk, we will present the active-active Apache Kafka® at Uber and how it facilitates disaster discovery across regions for Uber services. In particular, we will highlight the key components including topic replication, topic aggregation, offsets sync and then walk through several use cases of their disaster recovery strategy using active-active Kafka. Lastly, we will present several interesting challenges and the future work planned.
Yupeng Fu is a staff engineer in Uber Data Org leading the streaming data platform. Previously, he worked at Alluxio and Palantir, building distributed data analysis and storage platforms. Yupeng holds a B.S. and an M.S. from Tsinghua University and did his Ph.D. research on databases at UCSD.
Ohio Linux Fest 2013: Provisioning VMs Quickly with Vagrant and CFEngineNick Anderson
During this hands-on tutorial you will learn how to quickly provision local test/development/demo environments using Vagrant and Virtualbox. will cover provisioning and configuring machines quickly using Vagrant and CFEngine. You will learn how Vagrant and Virtualbox can be used to bring up local development/test/demo environments. You will also learn how CFEngine can be leveraged to automate configuration of the environment after it has been initialized. You will take away a multi-vm test environment managed by CFEngine.
This tutorial targets technical people who need repeatable test environments and are comfortable using the Linux command-line. These environments can speed developer on-boarding, play a role in continuous integration, or just provide quick sandboxes for experimentation. No previous knowledge of Vagrant or CFEngine is required.
Attendees should bring a laptop with at least 10G of available disk space (SSD strongly recommended, but not required), and have current versions of Vagrant and Virtualbox installed.
Deploying MariaDB for HA on Google Cloud PlatformMariaDB plc
This document discusses deploying MariaDB for high availability on Google Cloud Platform. It presents two solutions: 1) a MariaDB Galera cluster within a single GCP region, and 2) asynchronous multi-master replication across regions using Orchestrator for automated failover. It describes using ProxySQL or MaxScale for read/write splitting and Consul for coordinating configuration updates between Orchestrator and proxies during failovers.
Energy domain based IIoT workshop using Serverless framework, Elasticsearch, Kibana and AWS cloud resources to configure, stream, manage, search, analyze telemetry data
By Simon Hong.
Slides at https://docs.google.com/presentation/d/1rtFJVbyJ3qIUq1V8Q1Hp_ZIokWt_0GgVH8djEdDnD9w/edit#slide=id.p.
(c) BlinkOn 10
Toronto, Ontario (Canada)
April 09 - 10, 2019
https://docs.google.com/document/u/1/d/e/2PACX-1vTgBrqyQ4KCchsymvssri1pN1BkOg3sEqHThqhvFDl9-zl-hLx1S5c8sc5gaZ_VzKEVaYj94H3m1vso/pub#h.igsyfaa103a0
This document discusses using TurboGears web application frameworks on both Python 2 and Python 3. It describes setting up separate Python 2 and Python 3 environments to develop TurboGears apps, installing TurboGears on both, and creating a basic app that renders templates. It also covers TurboGears features like object dispatch routing, template engines, database access using SQLAlchemy/Ming, and authentication.
Ryan Jarvinen Open Shift Talk @ Postgres Open 2013PostgresOpen
The document discusses writing portable PostgreSQL applications for the open cloud. It covers open cloud platforms like OpenStack that avoid lock-in, and platforms like OpenShift that provide portable PostgreSQL applications using cartridges and scaling hooks. Examples are given of simple applications like a TODO app using Flask, SQLAlchemy and PostgreSQL that can be easily deployed to OpenShift.
Check Yourself Before You Wreck Yourself: Auditing and Improving the Performa...Nicholas Jansma
Boomerang is an open-source Real User Monitoring (RUM) JavaScript library used by thousands of websites to measure their visitor's experiences. The developers behind Boomerang take pride in building a reliable and performant third-party library that everyone can use without being concerned about its measurements affecting their site. We recently performed and shared an audit of Boomerang's performance, to help communicate its "cost of doing business", and in doing so we found several areas of code that we wanted to improve. We'll discuss how we performed the audit, some of the improvements we've made, how we're testing and validating our changes, and the real-time telemetry we capture for our library to ensure we're having as little of an impact as possible on the sites we're included on.
Boomerang is an open-source Real User Monitoring (RUM) JavaScript library used by thousands of websites to measure their visitor's experiences.
Boomerang runs on billions of page loads a day, either via the open-source library or as part of Akamai's mPulse RUM service. The developers behind Boomerang take pride in building a reliable and performant third-party library that everyone can use without being concerned about its measurements affecting their site.
Recently, we performed and shared an audit of Boomerang's performance, to help communicate the "cost of doing business" of including Boomerang on a page while it takes its measurements. In doing the audit, we found several areas of code that we wanted to improve and have been making continuous improvements ever since. We've taken ideas and contributions from the OSS community, and have built a Performance Lab that helps "lock in" our improvements by continuously measuring the metrics that are important to us.
We'll discuss how we performed the audit, some of the improvements we've made, how we're testing and validating our changes, and the real-time telemetry we capture on our library to ensure we're having as little of an impact as possible on the sites we're included on.
Mailchimp to the Edge - Establishing Akamai Best Practices at MailchimpBob Strecansky
The document discusses best practices for using Akamai technologies at Mailchimp, covering three topics: Kubernetes and Akamai, continuous integration and continuous deployment with Akamai, and using BigQuery as an ingestion point for Akamai Datastream. It includes diagrams, code snippets, and glossaries to explain concepts like Kubernetes, Google Cloud Platform services, and how they interact with Akamai services.
The document outlines 15 ways to optimize Spring Boot applications for the cloud. It recommends using services provided by cloud platforms for monitoring, Spring Cloud Sleuth for request tracing, Spring Boot Actuator for metrics and health checks, and circuit breakers to prevent failures from cascading. It also suggests keeping dependencies up-to-date, using Eclipse OpenJ9 to reduce memory usage, enabling zero-downtime configuration changes, and introducing chaos testing into production environments.
Following simple patterns of good application design can allow you to scale your application for your customers easily. We'll dive into the 12 factor application design and demo how this applies to containers and deployments on Amazon ECS and Fargate. We'll take a look at tooling that can be used to simplify your work flow and help you adopt the principles of the 12 factor application.
Fluent 2018: Tracking Performance of the Web with HTTP ArchivePaul Calvano
Have you ever thought about how your site’s performance compares to the web as a whole? Or maybe you’re curious how popular a particular web feature is. How much is too much JavaScript? The HTTP Archive has been keeping track of how the web is built since 2010. It enables you to find answers to questions about the state of the web past and present.
Paul Calvano explores how the HTTP Archive works, how people are using this dataset, and some ways that Akamai has leveraged data within the HTTP Archive to help its customers.
OSCamp 2019 | #3 Ansible: Directing the Director by Martin SchurzNETWAYS
I will show how we designed our new central monitoring system based on Icinga2. Currently we are in the process of migrating our projects to this new system. While doing this, our teams learn a lot of new stuff and discover solutions to their specific problems. After describing, how we share our acquired knowledge to all teams, I will give a closer insight in one particular approach where we used Ansible to provision all our Icinga configuration via the Director interface.
An overview of the latest news and interesting developments across PHP from January 2019: new language features, RFCs, community news, upcoming conferences, and the latest versions of popular frameworks and platforms.
Movable Type 5.2 Overview at MTDDC 2012Yuji Takayama
The document provides an overview of new features in Movable Type 5.2 from the perspective of a software engineer. Key new features include a rich text editor, ability for system administrators to restrict publishing paths, SMTP authentication, an upgraded tools interface, and support for running Movable Type on cloud platforms. It also discusses upgrades to the mail sending backend to support SSL/TLS, submision ports, and debugging. Additional sections cover implementing a PSGI interface for Movable Type to support pluggable web servers and reloading of configurations.
Similar to PGConf.ASIA 2019 Bali - Your Business Continuity Matrix and PostgreSQL's Disaster Recovery - Muhammad Haroon (20)
Equnix Business Solutions (Equnix) is an IT Solution provider in Indonesia, providing comprehensive solution services especially on the infrastructure side for corporate business needs based on research and Open Source. Equnix has 3 (three) main services known as the Trilogy of Services: Support (Maintenance/Managed), World class level of Software Development, and Expert Consulting and Assessment for High Performance Transactions System. Equnix is customer oriented, not product or principal. Equal opportunity based on merit is our credo in managing HR development.
Kebocoran Data_ Tindakan Hacker atau Kriminal_ Bagaimana kita mengantisipasi...Equnix Business Solutions
Equnix Business Solutions (Equnix) is an IT Solution provider in Indonesia, providing comprehensive solution services especially on the infrastructure side for corporate business needs based on research and Open Source. Equnix has 3 (three) main services known as the Trilogy of Services: Support (Maintenance/Managed), World class level of Software Development, and Expert Consulting and Assessment for High Performance Transactions System. Equnix is customer oriented, not product or principal. Equal opportunity based on merit is our credo in managing HR development.
Equnix Business Solutions (Equnix) is an IT Solution provider in Indonesia, providing comprehensive solution services especially on the infrastructure side for corporate business needs based on research and Open Source. Equnix has 3 (three) main services known as the Trilogy of Services: Support (Maintenance/Managed), World class level of Software Development, and Expert Consulting and Assessment for High Performance Transactions System. Equnix is customer oriented, not product or principal. Equal opportunity based on merit is our credo in managing HR development.
Equnix Business Solutions (Equnix) is an IT Solution provider in Indonesia, providing comprehensive solution services especially on the infrastructure side for corporate business needs based on research and Open Source. Equnix has 3 (three) main services known as the Trilogy of Services: Support (Maintenance/Managed), World class level of Software Development, and Expert Consulting and Assessment for High Performance Transactions System. Equnix is customer oriented, not product or principal. Equal opportunity based on merit is our credo in managing HR development.
Equnix Business Solutions (Equnix) is an IT Solution provider in Indonesia, providing comprehensive solution services especially on the infrastructure side for corporate business needs based on research and Open Source. Equnix has 3 (three) main services known as the Trilogy of Services: Support (Maintenance/Managed), World class level of Software Development, and Expert Consulting and Assessment for High Performance Transactions System. Equnix is customer oriented, not product or principal. Equal opportunity based on merit is our credo in managing HR development.
[EWTT2022] Strategi Implementasi Database dalam Microservice Architecture.pdfEqunix Business Solutions
Equnix Business Solutions (Equnix) is an IT Solution provider in Indonesia, providing comprehensive solution services especially on the infrastructure side for corporate business needs based on research and Open Source. Equnix has 3 (three) main services known as the Trilogy of Services: Support (Maintenance/Managed), World class level of Software Development, and Expert Consulting and Assessment for High Performance Transactions System. Equnix is customer oriented, not product or principal. Equal opportunity based on merit is our credo in managing HR development.
Equnix Business Solutions (Equnix) is an IT Solution provider in Indonesia, providing comprehensive solution services especially on the infrastructure side for corporate business needs based on research and Open Source. Equnix has 3 (three) main services known as the Trilogy of Services: Support (Maintenance/Managed), World class level of Software Development, and Expert Consulting and Assessment for High Performance Transactions System. Equnix is customer oriented, not product or principal. Equal opportunity based on merit is our credo in managing HR development.
Equnix Business Solutions (Equnix) is an IT Solution provider in Indonesia, providing comprehensive solution services especially on the infrastructure side for corporate business needs based on research and Open Source. Equnix has 3 (three) main services known as the Trilogy of Services: Support (Maintenance/Managed), World class level of Software Development, and Expert Consulting and Assessment for High Performance Transactions System. Equnix is customer oriented, not product or principal. Equal opportunity based on merit is our credo in managing HR development.
Equnix Business Solutions (Equnix) is an IT Solution provider in Indonesia, providing comprehensive solution services especially on the infrastructure side for corporate business needs based on research and Open Source. Equnix has 3 (three) main services known as the Trilogy of Services: Support (Maintenance/Managed), World class level of Software Development, and Expert Consulting and Assessment for High Performance Transactions System. Equnix is customer oriented, not product or principal. Equal opportunity based on merit is our credo in managing HR development.
Equnix Business Solutions (Equnix) is an IT Solution provider in Indonesia, providing comprehensive solution services especially on the infrastructure side for corporate business needs based on research and Open Source. Equnix has 3 (three) main services known as the Trilogy of Services: Support (Maintenance/Managed), World class level of Software Development, and Expert Consulting and Assessment for High Performance Transactions System. Equnix is customer oriented, not product or principal. Equal opportunity based on merit is our credo in managing HR development.
Equnix Business Solutions (Equnix) is an IT Solution provider in Indonesia, providing comprehensive solution services especially on the infrastructure side for corporate business needs based on research and Open Source. Equnix has 3 (three) main services known as the Trilogy of Services: Support (Maintenance/Managed), World class level of Software Development, and Expert Consulting and Assessment for High Performance Transactions System. Equnix is customer oriented, not product or principal. Equal opportunity based on merit is our credo in managing HR development.
Equnix Appliance- Jawaban terbaik untuk kebutuhan komputasi yang mumpuni.pdfEqunix Business Solutions
Equnix Business Solutions (Equnix) is an IT Solution provider in Indonesia, providing comprehensive solution services especially on the infrastructure side for corporate business needs based on research and Open Source. Equnix has 3 (three) main services known as the Trilogy of Services: Support (Maintenance/Managed), World class level of Software Development, and Expert Consulting and Assessment for High Performance Transactions System. Equnix is customer oriented, not product or principal. Equal opportunity based on merit is our credo in managing HR development.
Equnix Business Solutions (Equnix) is an IT Solution provider in Indonesia, providing comprehensive solution services especially on the infrastructure side for corporate business needs based on research and Open Source. Equnix has 3 (three) main services known as the Trilogy of Services: Support (Maintenance/Managed), World class level of Software Development, and Expert Consulting and Assessment for High Performance Transactions System. Equnix is customer oriented, not product or principal. Equal opportunity based on merit is our credo in managing HR development.
Equnix Business Solutions (Equnix) is an IT Solution provider in Indonesia, providing comprehensive solution services especially on the infrastructure side for corporate business needs based on research and Open Source. Equnix has 3 (three) main services known as the Trilogy of Services: Support (Maintenance/Managed), World class level of Software Development, and Expert Consulting and Assessment for High Performance Transactions System. Equnix is customer oriented, not product or principal. Equal opportunity based on merit is our credo in managing HR development.
Enchancing adoption of Open Source Libraries. A case study on Albumentations.AIVladimir Iglovikov, Ph.D.
Presented by Vladimir Iglovikov:
- https://www.linkedin.com/in/iglovikov/
- https://x.com/viglovikov
- https://www.instagram.com/ternaus/
This presentation delves into the journey of Albumentations.ai, a highly successful open-source library for data augmentation.
Created out of a necessity for superior performance in Kaggle competitions, Albumentations has grown to become a widely used tool among data scientists and machine learning practitioners.
This case study covers various aspects, including:
People: The contributors and community that have supported Albumentations.
Metrics: The success indicators such as downloads, daily active users, GitHub stars, and financial contributions.
Challenges: The hurdles in monetizing open-source projects and measuring user engagement.
Development Practices: Best practices for creating, maintaining, and scaling open-source libraries, including code hygiene, CI/CD, and fast iteration.
Community Building: Strategies for making adoption easy, iterating quickly, and fostering a vibrant, engaged community.
Marketing: Both online and offline marketing tactics, focusing on real, impactful interactions and collaborations.
Mental Health: Maintaining balance and not feeling pressured by user demands.
Key insights include the importance of automation, making the adoption process seamless, and leveraging offline interactions for marketing. The presentation also emphasizes the need for continuous small improvements and building a friendly, inclusive community that contributes to the project's growth.
Vladimir Iglovikov brings his extensive experience as a Kaggle Grandmaster, ex-Staff ML Engineer at Lyft, sharing valuable lessons and practical advice for anyone looking to enhance the adoption of their open-source projects.
Explore more about Albumentations and join the community at:
GitHub: https://github.com/albumentations-team/albumentations
Website: https://albumentations.ai/
LinkedIn: https://www.linkedin.com/company/100504475
Twitter: https://x.com/albumentations
Dr. Sean Tan, Head of Data Science, Changi Airport Group
Discover how Changi Airport Group (CAG) leverages graph technologies and generative AI to revolutionize their search capabilities. This session delves into the unique search needs of CAG’s diverse passengers and customers, showcasing how graph data structures enhance the accuracy and relevance of AI-generated search results, mitigating the risk of “hallucinations” and improving the overall customer journey.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/06/building-and-scaling-ai-applications-with-the-nx-ai-manager-a-presentation-from-network-optix/
Robin van Emden, Senior Director of Data Science at Network Optix, presents the “Building and Scaling AI Applications with the Nx AI Manager,” tutorial at the May 2024 Embedded Vision Summit.
In this presentation, van Emden covers the basics of scaling edge AI solutions using the Nx tool kit. He emphasizes the process of developing AI models and deploying them globally. He also showcases the conversion of AI models and the creation of effective edge AI pipelines, with a focus on pre-processing, model conversion, selecting the appropriate inference engine for the target hardware and post-processing.
van Emden shows how Nx can simplify the developer’s life and facilitate a rapid transition from concept to production-ready applications.He provides valuable insights into developing scalable and efficient edge AI solutions, with a strong focus on practical implementation.
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...SOFTTECHHUB
The choice of an operating system plays a pivotal role in shaping our computing experience. For decades, Microsoft's Windows has dominated the market, offering a familiar and widely adopted platform for personal and professional use. However, as technological advancements continue to push the boundaries of innovation, alternative operating systems have emerged, challenging the status quo and offering users a fresh perspective on computing.
One such alternative that has garnered significant attention and acclaim is Nitrux Linux 3.5.0, a sleek, powerful, and user-friendly Linux distribution that promises to redefine the way we interact with our devices. With its focus on performance, security, and customization, Nitrux Linux presents a compelling case for those seeking to break free from the constraints of proprietary software and embrace the freedom and flexibility of open-source computing.
A tale of scale & speed: How the US Navy is enabling software delivery from l...sonjaschweigert1
Rapid and secure feature delivery is a goal across every application team and every branch of the DoD. The Navy’s DevSecOps platform, Party Barge, has achieved:
- Reduction in onboarding time from 5 weeks to 1 day
- Improved developer experience and productivity through actionable findings and reduction of false positives
- Maintenance of superior security standards and inherent policy enforcement with Authorization to Operate (ATO)
Development teams can ship efficiently and ensure applications are cyber ready for Navy Authorizing Officials (AOs). In this webinar, Sigma Defense and Anchore will give attendees a look behind the scenes and demo secure pipeline automation and security artifacts that speed up application ATO and time to production.
We will cover:
- How to remove silos in DevSecOps
- How to build efficient development pipeline roles and component templates
- How to deliver security artifacts that matter for ATO’s (SBOMs, vulnerability reports, and policy evidence)
- How to streamline operations with automated policy checks on container images
Cosa hanno in comune un mattoncino Lego e la backdoor XZ?Speck&Tech
ABSTRACT: A prima vista, un mattoncino Lego e la backdoor XZ potrebbero avere in comune il fatto di essere entrambi blocchi di costruzione, o dipendenze di progetti creativi e software. La realtà è che un mattoncino Lego e il caso della backdoor XZ hanno molto di più di tutto ciò in comune.
Partecipate alla presentazione per immergervi in una storia di interoperabilità, standard e formati aperti, per poi discutere del ruolo importante che i contributori hanno in una comunità open source sostenibile.
BIO: Sostenitrice del software libero e dei formati standard e aperti. È stata un membro attivo dei progetti Fedora e openSUSE e ha co-fondato l'Associazione LibreItalia dove è stata coinvolta in diversi eventi, migrazioni e formazione relativi a LibreOffice. In precedenza ha lavorato a migrazioni e corsi di formazione su LibreOffice per diverse amministrazioni pubbliche e privati. Da gennaio 2020 lavora in SUSE come Software Release Engineer per Uyuni e SUSE Manager e quando non segue la sua passione per i computer e per Geeko coltiva la sua curiosità per l'astronomia (da cui deriva il suo nickname deneb_alpha).
UiPath Test Automation using UiPath Test Suite series, part 6DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 6. In this session, we will cover Test Automation with generative AI and Open AI.
UiPath Test Automation with generative AI and Open AI webinar offers an in-depth exploration of leveraging cutting-edge technologies for test automation within the UiPath platform. Attendees will delve into the integration of generative AI, a test automation solution, with Open AI advanced natural language processing capabilities.
Throughout the session, participants will discover how this synergy empowers testers to automate repetitive tasks, enhance testing accuracy, and expedite the software testing life cycle. Topics covered include the seamless integration process, practical use cases, and the benefits of harnessing AI-driven automation for UiPath testing initiatives. By attending this webinar, testers, and automation professionals can gain valuable insights into harnessing the power of AI to optimize their test automation workflows within the UiPath ecosystem, ultimately driving efficiency and quality in software development processes.
What will you get from this session?
1. Insights into integrating generative AI.
2. Understanding how this integration enhances test automation within the UiPath platform
3. Practical demonstrations
4. Exploration of real-world use cases illustrating the benefits of AI-driven test automation for UiPath
Topics covered:
What is generative AI
Test Automation with generative AI and Open AI.
UiPath integration with generative AI
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Threats to mobile devices are more prevalent and increasing in scope and complexity. Users of mobile devices desire to take full advantage of the features
available on those devices, but many of the features provide convenience and capability but sacrifice security. This best practices guide outlines steps the users can take to better protect personal devices and information.
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024Albert Hoitingh
In this session I delve into the encryption technology used in Microsoft 365 and Microsoft Purview. Including the concepts of Customer Key and Double Key Encryption.
GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...Neo4j
Leonard Jayamohan, Partner & Generative AI Lead, Deloitte
This keynote will reveal how Deloitte leverages Neo4j’s graph power for groundbreaking digital twin solutions, achieving a staggering 100x performance boost. Discover the essential role knowledge graphs play in successful generative AI implementations. Plus, get an exclusive look at an innovative Neo4j + Generative AI solution Deloitte is developing in-house.
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdfPaige Cruz
Monitoring and observability aren’t traditionally found in software curriculums and many of us cobble this knowledge together from whatever vendor or ecosystem we were first introduced to and whatever is a part of your current company’s observability stack.
While the dev and ops silo continues to crumble….many organizations still relegate monitoring & observability as the purview of ops, infra and SRE teams. This is a mistake - achieving a highly observable system requires collaboration up and down the stack.
I, a former op, would like to extend an invitation to all application developers to join the observability party will share these foundational concepts to build on:
Removing Uninteresting Bytes in Software FuzzingAftab Hussain
Imagine a world where software fuzzing, the process of mutating bytes in test seeds to uncover hidden and erroneous program behaviors, becomes faster and more effective. A lot depends on the initial seeds, which can significantly dictate the trajectory of a fuzzing campaign, particularly in terms of how long it takes to uncover interesting behaviour in your code. We introduce DIAR, a technique designed to speedup fuzzing campaigns by pinpointing and eliminating those uninteresting bytes in the seeds. Picture this: instead of wasting valuable resources on meaningless mutations in large, bloated seeds, DIAR removes the unnecessary bytes, streamlining the entire process.
In this work, we equipped AFL, a popular fuzzer, with DIAR and examined two critical Linux libraries -- Libxml's xmllint, a tool for parsing xml documents, and Binutil's readelf, an essential debugging and security analysis command-line tool used to display detailed information about ELF (Executable and Linkable Format). Our preliminary results show that AFL+DIAR does not only discover new paths more quickly but also achieves higher coverage overall. This work thus showcases how starting with lean and optimized seeds can lead to faster, more comprehensive fuzzing campaigns -- and DIAR helps you find such seeds.
- These are slides of the talk given at IEEE International Conference on Software Testing Verification and Validation Workshop, ICSTW 2022.
Introducing Milvus Lite: Easy-to-Install, Easy-to-Use vector database for you...Zilliz
Join us to introduce Milvus Lite, a vector database that can run on notebooks and laptops, share the same API with Milvus, and integrate with every popular GenAI framework. This webinar is perfect for developers seeking easy-to-use, well-integrated vector databases for their GenAI apps.
Pushing the limits of ePRTC: 100ns holdover for 100 daysAdtran
At WSTS 2024, Alon Stern explored the topic of parametric holdover and explained how recent research findings can be implemented in real-world PNT networks to achieve 100 nanoseconds of accuracy for up to 100 days.
5. https://www.2ndQuadrant.com
PGConf ASIA
Bali, Sept, 2019
$psql~#: d haroon
● Working in PostgreSQL space @ 2ndQuadrant
● Part of PostgreSQL family for nearly a decade and a half
● Development, support, consulting, professional services and administration
● Past stints with PostgreSQL family include
○ EnterpriseDB
○ OpenSCG (now Amazon Web Services)
● Led Engineering Ops efforts @ IBEX Group
● Principal/Architect/Product Owner @ TRG
Email: haroon@2ndQuadrant.com
Skype: contact.haroon
7. https://www.2ndQuadrant.com
PGConf ASIA
Bali, Sept, 2019
Objectives
- Recovery Point Objective (RPO)
- How much data can I afford to lose?
- Recovery Time Objective (RTO)
- How long will it take me to recover?
8. https://www.2ndQuadrant.com
PGConf ASIA
Bali, Sept, 2019
Service RELIABILITY
● Cost of downtime
○ How many €/$/£/AUD/AED/…?
● Risk management
● SLI (Service Level Indicator), SLO (Service Level Objective) and SLA
(Service Level Agreement)
15. https://www.2ndQuadrant.com
PGConf ASIA
Bali, Sept, 2019
Recap
● How do you feel now?
● Still: RPO = ∞ and RTO = n/a. Why?
● A backup is valid only if you have tested it
● Unfortunately, this is very common
18. https://www.2ndQuadrant.com
PGConf ASIA
Bali, Sept, 2019
Are we any better now ?
● RPO = backup frequency
● RTO = maximum time of recovery
○ Provision another server
○ Configure another server (automated, right?)
○ Time to restore the last backup (measure it)
● Still missing something ?
20. https://www.2ndQuadrant.com
PGConf ASIA
Bali, Sept, 2019
Recap
● Can this architecture work for you?
● We need reliable monitoring
a. From now on, we assume we have it in place!
● We need to reduce both RPO and RTO
22. https://www.2ndQuadrant.com
PGConf ASIA
Bali, Sept, 2019
PostgreSQL’s PITR
● Part of core (fully open source)
● Rebuild a cluster at a point in time
● From crash recovery to sync streamrep (physical/logical)
● RPO = 0 (zero data loss)
● Hot base backup, continuous WAL archiving, Recovery
23. https://www.2ndQuadrant.com
PGConf ASIA
Bali, Sept, 2019
Basic concepts
● Continuous copy of WAL data (continuous archiving)
● Physical base backups
● Recovery:
○ copy base backup to another location
○ recovery mode (replay of WALs until target)
24. https://www.2ndQuadrant.com
PGConf ASIA
Bali, Sept, 2019
Tools
- Custom written scripts
- pg_backrest
- Pg_probackup
- WAL-G
- Barman
So yeah, any tool is fine as long as you know what you are trying to achieve ...
25. https://www.2ndQuadrant.com
PGConf ASIA
Bali, Sept, 2019
BARMAN
● Latest released version: Barman 2.9
● Open Source (GNU GPL 3)
● Written in Python
● Developed and maintained by 2ndQuadrant
● Available at www.pgbarman.org and
www.2ndquadrant.com/en/resources/barman/
29. https://www.2ndQuadrant.com
PGConf ASIA
Bali, Sept, 2019
Copy method
● PostgreSQL streaming
○ Windows/Docker
● Rsync/SSH
○ Incremental backup and recovery (via hard links)
○ Parallel backup and recovery
○ Network compression and bandwidth limitation
30. https://www.2ndQuadrant.com
PGConf ASIA
Bali, Sept, 2019
WAL shipping method
● “archiving”, through “archive_command”:
○ RPO ~ 16MB of WAL data, or
○ “archive_timeout”
● “streaming”, through streaming replication:
○ “pg_receivewal” or “pg_receivexlog”
○ continuous stream, RPO ~ 0
○ PostgreSQL 9.2+ required
31. https://www.2ndQuadrant.com
PGConf ASIA
Bali, Sept, 2019
Example from postgresql.conf
archive_mode = on
wal_level = logical
max_wal_senders = 10
max_replication_slots = 10
archive_command = 'rsync -a %p
barman@HOST:/var/lib/barman/ID/incoming'
32. https://www.2ndQuadrant.com
PGConf ASIA
Bali, Sept, 2019
Example from barman.conf
[angus]
description = “Angus Young database"
ssh_command = ssh postgres@angus
conninfo = user=barman-acdc dbname=postgres host=angus
retention_policy = RECOVERY WINDOW OF 6 MONTHS
copy_method = rsync
reuse_backup = link
parallel_jobs = 4
archiver = true
streaming_archiver = true
slot_name = barman_streaming_acdc
33. https://www.2ndQuadrant.com
PGConf ASIA
Bali, Sept, 2019
RECAP
● How do you feel now?
● Still: RPO = ∞ and RTO = n/a. Why?
● A backup is valid only if you have tested it
● Barman reduces backup risks, does not exclude them
○ Systematic tests (especially custom scripts)
○ Business risk is very high
37. https://www.2ndQuadrant.com
PGConf ASIA
Bali, Sept, 2019
Example of Recovery script
● Write a bash script that:
○ connects to a remote server via SSH
○ stops the PostgreSQL server
○ issues a “barman recover” with target “immediate”
○ starts PostgreSQL
● Set it as post-backup script
38. https://www.2ndQuadrant.com
PGConf ASIA
Bali, Sept, 2019
Some food for thought
● Outcomes:
○ Systematically test your backup
○ Measure your recovery time
○ Identical server? This is a backup server ready to start
● You can use a different data centre
● Be creative, PostgreSQL gives you infinite freedom!
39. https://www.2ndQuadrant.com
PGConf ASIA
Bali, Sept, 2019
RECAP
● RPO ~ 0 (your backups work, every time)
● RTO = Time of reaction + Recovery time
● Example: RPO ~0 and RTO < 1 day
○ Acceptable or not acceptable?
● Entry level architecture for business continuity
● Priority now: improve RTO
41. https://www.2ndQuadrant.com
PGConf ASIA
Bali, Sept, 2019
PostgreSQL’s Replication
● Part of core (fully open source)
● One master, multiple standby servers
● Evolution of PITR
○ Standby server is in continuous recovery mode
○ Hot standby (read-only)
○ Both streaming (9.0+) and file based pulling of WAL
● Cascading from a standby
42. https://www.2ndQuadrant.com
PGConf ASIA
Bali, Sept, 2019
Synchronous replication
● Fine control (from global down to transaction level)
● 2-safe replication
○ COMMIT of a write transactions waits until written on both the
master and a standby (or more from 9.6)
■ More than a synchronous client is required
○ Read consistency of a cluster
● RPO = 0 (zero data loss)
45. https://www.2ndQuadrant.com
PGConf ASIA
Bali, Sept, 2019
Excerpt from postgresql’s configuration
postgresql.conf:
hot_standby = on
recovery.conf:
standby_mode = ‘on'
# Streaming
primary_conninfo = 'host=angus user=replica application_name=ha
sslmode=require’
# Fallback via Barman
restore_command = 'barman-wal-restore -U barman acdc angus %f %p'
46. https://www.2ndQuadrant.com
PGConf ASIA
Bali, Sept, 2019
Switchover (planned)
● Applications are paused (start of downtime)
● Shut down the master
● Allow the standby to catch up with the master
● Promote the standby
● Switch virtual IPs
● Resume applications (end of downtime)
● Reconfigure the former master as standby
49. https://www.2ndQuadrant.com
PGConf ASIA
Bali, Sept, 2019
RECAP
● RPO ~ 0 (your backups work, every time)
● RTO = Time of reaction + Time of promotion
● Criticality: manual intervention
○ Reliable monitoring
○ Trained people (practice & docs!)
50. https://www.2ndQuadrant.com
PGConf ASIA
Bali, Sept, 2019
Manual failover vs automated failover
● Risk management
○ Split brain nightmare
○ Automated is built on manual (test!)
○ Your choice
● Very good solution for business continuity
● Uptime > 99.99% in a year
53. https://www.2ndQuadrant.com
PGConf ASIA
Bali, Sept, 2019
Synchronous replication
● Primary: Barman
○ Zero data loss backup
● Primary: Standby
○ Zero data loss cluster (reduce RTO)
● Just one configuration line in PostgreSQL
○ synchronous_standby_names = '1 (ha, barman_receive_wal)'
57. https://www.2ndQuadrant.com
PGConf ASIA
Bali, Sept, 2019
Push the boundaries
● Repeatable architectures in multiple data centres
● PgBouncer
● Virtual IPs
● S3 relay via Barman hook scripts
● Multiple standby servers and cascading replication
● Docker containers
● Logical replication backups
58. https://www.2ndQuadrant.com
PGConf ASIA
Bali, Sept, 2019
Conclusions
● Babysteps and KISS
● New? Explore and learn
● Practice is the only way to mastery (drills)
● Plan regular healthy downtimes
○ Use switchovers to perform PostgreSQL updates
○ Smart downtimes increase long-term uptime