This document provides an overview of logical replication in PostgreSQL 10. It discusses the history of PostgreSQL replication, the key concepts of logical replication including publications, subscriptions, and replication slots. It presents several use cases for logical replication such as cross-platform, cross-version, and write operations at the subscriber. The document also covers configuration settings, a quick setup demonstration, monitoring, resolving conflicts, and limitations of logical replication.
PostgreSQL 10 had added built-in logical replication which tackles some of the limitations of physical replication and opens up the possibility of promising new areas of replication. In this webinar, we will introduce the concept of logical replication and demonstrate how one can configure a logical replication in no time.
Highlights include:
- Basic architecture, including the publisher and subscriber model
- Configuration, administration, and monitoring
- Limitations and future plans
PGDay.Amsterdam 2018 - Stefan Fercot - Save your data with pgBackRestPGDay.Amsterdam
Ever heard of Point-in-time recovery? pgBackRest is an awsome tool to handle backups, restores and even helps you build streaming replication ! This talk will introduce the tool, its basic features and how to use it.
Built in physical and logical replication in postgresql-Firat GulecFIRAT GULEC
What is Replication?
Why do we need Replication?
How many replication layers do we have?
Understanding milestones of built-in Database Physical Replication.
What is the purpose of replication? and How to rescue system in case of failover?
What is Streaming Replication and what is its advantages? Async vs Sync, Hot standby etc.
How to configurate Master and Standby Servers? And What is the most important parameters? Example of topoloji.
What is Cascading Replication and how to configurate it? Live Demo on Terminal.
What is Logical Replication coming with PostgreSQL 10? And What is its advantages?
Logical Replication vs Physical Replication
Limitations of Logical Replication
Quorum Commit for Sync Replication etc.
What is coming up with PostgreSQL 11 about replication?
10 Questions quiz and giving some gifts to participants according to their success.
This document discusses PostgreSQL replication. It provides an overview of replication, including its history and features. Replication allows data to be copied from a primary database to one or more standby databases. This allows for high availability, load balancing, and read scaling. The document describes asynchronous and synchronous replication modes.
Security Best Practices for your Postgres DeploymentPGConf APAC
These slides were used by Sameer Kumar of Ashnik for presenting his topic at pgDay Asia 2016. He took audience through some of the security best practices for deploying and hardening PostgreSQL
PostgreSQL Replication High Availability MethodsMydbops
This slides illustrates the need for replication in PostgreSQL, why do you need a replication DB topology, terminologies, replication nodes and many more.
PostgreSQL 10 had added built-in logical replication which tackles some of the limitations of physical replication and opens up the possibility of promising new areas of replication. In this webinar, we will introduce the concept of logical replication and demonstrate how one can configure a logical replication in no time.
Highlights include:
- Basic architecture, including the publisher and subscriber model
- Configuration, administration, and monitoring
- Limitations and future plans
PGDay.Amsterdam 2018 - Stefan Fercot - Save your data with pgBackRestPGDay.Amsterdam
Ever heard of Point-in-time recovery? pgBackRest is an awsome tool to handle backups, restores and even helps you build streaming replication ! This talk will introduce the tool, its basic features and how to use it.
Built in physical and logical replication in postgresql-Firat GulecFIRAT GULEC
What is Replication?
Why do we need Replication?
How many replication layers do we have?
Understanding milestones of built-in Database Physical Replication.
What is the purpose of replication? and How to rescue system in case of failover?
What is Streaming Replication and what is its advantages? Async vs Sync, Hot standby etc.
How to configurate Master and Standby Servers? And What is the most important parameters? Example of topoloji.
What is Cascading Replication and how to configurate it? Live Demo on Terminal.
What is Logical Replication coming with PostgreSQL 10? And What is its advantages?
Logical Replication vs Physical Replication
Limitations of Logical Replication
Quorum Commit for Sync Replication etc.
What is coming up with PostgreSQL 11 about replication?
10 Questions quiz and giving some gifts to participants according to their success.
This document discusses PostgreSQL replication. It provides an overview of replication, including its history and features. Replication allows data to be copied from a primary database to one or more standby databases. This allows for high availability, load balancing, and read scaling. The document describes asynchronous and synchronous replication modes.
Security Best Practices for your Postgres DeploymentPGConf APAC
These slides were used by Sameer Kumar of Ashnik for presenting his topic at pgDay Asia 2016. He took audience through some of the security best practices for deploying and hardening PostgreSQL
PostgreSQL Replication High Availability MethodsMydbops
This slides illustrates the need for replication in PostgreSQL, why do you need a replication DB topology, terminologies, replication nodes and many more.
This document summarizes a presentation about Presto, an open source distributed SQL query engine. It discusses Presto's distributed and plug-in architecture, query planning process, and cluster configuration options. For architecture, it explains that Presto uses coordinators, workers, and connectors to distribute queries across data sources. For query planning, it shows how SQL queries are converted into logical and physical query plans with stages, tasks, and splits. For configuration, it reviews single-server, multi-worker, and multi-coordinator cluster topologies. It also provides an overview of Presto's recent updates.
Koalas is an open source project that provides pandas APIs on top of Apache Spark. Pandas is the standard tool for data science and it is typically the first step to explore and manipulate a data set, but pandas does not scale well to big data. Koalas fills the gap by providing pandas equivalent APIs that work on Apache Spark.
There are also many libraries trying to scale pandas APIs, such as Vaex, Modin, and so on. Dask is one of them and very popular among pandas users, and also works on its own cluster similar to Koalas which is on top of Spark cluster. In this talk, we will introduce Koalas and its current status, and the comparison between Koalas and Dask, including benchmarking.
This document discusses streaming replication in PostgreSQL. It covers how streaming replication works, including the write-ahead log and replication processes. It also discusses setting up replication between a primary and standby server, including configuring the servers and verifying replication is working properly. Monitoring replication is discussed along with views and functions for checking replication status. Maintenance tasks like adding or removing standbys and pausing replication are also mentioned.
Postgresql 12 streaming replication holVijay Kumar N
This is a step by step hands on lab for PostgreSQL 12 , setup of replication, replication slot, failover (promoting) to standby as new master cluster and also covering the scenario where old master has to be reinstated using the utility "pg_rewind"
Open Policy Agent Deep Dive Seattle 2018Torin Sandall
Topics:
* Background on Open Policy Agent project: users, use cases, and stats.
* How OPA works (decoupling policy decision-making from enforcement)
* Hands-on example: Users can view their own account details and support staff can view accounts they have are assigned to via a ticketing system.
* SQL data filtering use case: writing policy in OPA and enforcing policy in SQL.
* WebAssembly compiler.
Webinar: PostgreSQL continuous backup and PITR with BarmanGabriele Bartolini
How can you achieve an RTO of 5 minutes for the backups of your PostgreSQL databases? And what about RPO=0 for zero data loss backups?
This webinar gave an answer to those questions, by providing an overview of Disaster Recovery of PostgreSQL databases with Barman, covering its major features.
Barman, Backup and Recovery Manager for PostgreSQL, is an open source tool that was conceived by 2ndQuadrant about 10 years ago and released open source in 2012 under GNU GPL 3.
It is now one of the most popular backup and recovery tools in the PostgreSQL ecosystem.
Video available at https://resources.2ndquadrant.com/en/webinar-postgresql-continuous-backup-and-pitr-with-barman
PostgreSQL major version upgrade using built in Logical ReplicationAtsushi Torikoshi
This document discusses upgrading a PostgreSQL database from a major version to another using logical replication. It begins with an introduction to logical replication and its advantages over other upgrade methods. It then covers the architecture of logical replication, including how the walsender reads WAL files and applies changes to the subscriber. Finally, it addresses some limitations of logical replication for upgrades, such as objects that are not replicated, and provides strategies for completing the upgrade while minimizing downtime.
Integrating Apache Kafka and Elastic Using the Connect Frameworkconfluent
As a streaming platform, Apache Kafka provides low-latency, high-throughput, fault-tolerant publish and subscribe pipelines and excels at processing streams of real-time events. Kafka provides reliable, millisecond delivery for connecting downstream systems with real-time data.
In this talk, we will show how easy it is to leverage Kafka and the Elasticsearch connector to keep your indices populated with the latest data from the rest of your enterprise, as it changes.
Smarter Together - Bringing Relational Algebra, Powered by Apache Calcite, in...Julian Hyde
What if Looker saw the queries you just executed and could predict your next query? Could it make those queries faster, by smarter caching, or aggregate navigation? Could it read your past SQL queries and help you write your LookML model? Those are some of the reasons to add relational algebra into Looker’s query engine, and why Looker hired Julian Hyde, author of Apache Calcite, to lead the effort. In this talk about the internals of Looker’s query engine, Julian Hyde will describe how the engine works, how Looker queries are described in Calcite’s relational algebra, and some features that it makes possible.
A talk by Julian Hyde at JOIN 2019 in San Francisco.
A look at what HA is and what PostgreSQL has to offer for building an open source HA solution. Covers various aspects in terms of Recovery Point Objective and Recovery Time Objective. Includes backup and restore, PITR (point in time recovery) and streaming replication concepts.
Flink Forward Berlin 2017: Piotr Nowojski - "Hit me, baby, just one time" - B...Flink Forward
Getting data in and out of Flink is by far the most important aspect, and an everyday typical requirement of building Flink applications. Doing so in an end-to-end exactly-once manner, however, can be tricky. Being able to reliably consume data from the outside world without any duplicate processing and guaranteeing consistent distributed state, and at the same time provide computed results back to the outside world also without introducing duplicates, is crucial for the consistency and correctness of applications built upon stream processors. In this talk, we will talk about how end-to-end exactly-once guarantees can be achieved with Apache Flink. We will talk about Flink’s checkpointing mechanism, and how exactly to leverage it when consuming and producing data from your Flink streaming pipelines. In particular, we will be having a detailed review on how our supported connectors do so, with the aim to provide reference implementations for your own custom consumers and sinks.
Flink Forward San Francisco 2022.
This talk will take you on the long journey of Apache Flink into the cloud-native era. It started all the way from where Hadoop and YARN were the standard way of deploying and operating data applications.
We're going to deep dive into the cloud-native set of principles and how they map to the Apache Flink internals and recent improvements. We'll cover fast checkpointing, fault tolerance, resource elasticity, minimal infrastructure dependencies, industry-standard tooling, ease of deployment and declarative APIs.
After this talk you'll get a broader understanding of the operational requirements for a modern streaming application and where the current limits are.
by
David Moravek
Lightweight Transactions at Lightning SpeedScyllaDB
This talk will outline the Scylla implementation of Lightweight Transactions (LWT) that brings us to parity with Apache Cassandra. We will cover how to use it, what is working, and what is left to be done. We will also cover what other improvements are in store to improve Scylla's transactional capabilities and why it matters.
The document discusses installing and configuring MySQL on Linux. It provides steps to install MySQL using RPM files, set passwords for security, test the installation, and configure applications to connect to the database. It also covers basic and advanced MySQL commands like CREATE TABLE, SELECT, JOIN, and more.
Demystifying postgres logical replication percona live scEmanuel Calvo
This document provides an overview of logical replication in PostgreSQL, including:
- The different types of replication in PostgreSQL and how logical replication works
- How logical replication compares to MySQL replication and the elements involved
- What logical replication can be used for and some limitations
- Key concepts like publications, subscriptions, replication slots, and conflict handling
- Monitoring and configuration options for logical replication
This document summarizes a presentation about Presto, an open source distributed SQL query engine. It discusses Presto's distributed and plug-in architecture, query planning process, and cluster configuration options. For architecture, it explains that Presto uses coordinators, workers, and connectors to distribute queries across data sources. For query planning, it shows how SQL queries are converted into logical and physical query plans with stages, tasks, and splits. For configuration, it reviews single-server, multi-worker, and multi-coordinator cluster topologies. It also provides an overview of Presto's recent updates.
Koalas is an open source project that provides pandas APIs on top of Apache Spark. Pandas is the standard tool for data science and it is typically the first step to explore and manipulate a data set, but pandas does not scale well to big data. Koalas fills the gap by providing pandas equivalent APIs that work on Apache Spark.
There are also many libraries trying to scale pandas APIs, such as Vaex, Modin, and so on. Dask is one of them and very popular among pandas users, and also works on its own cluster similar to Koalas which is on top of Spark cluster. In this talk, we will introduce Koalas and its current status, and the comparison between Koalas and Dask, including benchmarking.
This document discusses streaming replication in PostgreSQL. It covers how streaming replication works, including the write-ahead log and replication processes. It also discusses setting up replication between a primary and standby server, including configuring the servers and verifying replication is working properly. Monitoring replication is discussed along with views and functions for checking replication status. Maintenance tasks like adding or removing standbys and pausing replication are also mentioned.
Postgresql 12 streaming replication holVijay Kumar N
This is a step by step hands on lab for PostgreSQL 12 , setup of replication, replication slot, failover (promoting) to standby as new master cluster and also covering the scenario where old master has to be reinstated using the utility "pg_rewind"
Open Policy Agent Deep Dive Seattle 2018Torin Sandall
Topics:
* Background on Open Policy Agent project: users, use cases, and stats.
* How OPA works (decoupling policy decision-making from enforcement)
* Hands-on example: Users can view their own account details and support staff can view accounts they have are assigned to via a ticketing system.
* SQL data filtering use case: writing policy in OPA and enforcing policy in SQL.
* WebAssembly compiler.
Webinar: PostgreSQL continuous backup and PITR with BarmanGabriele Bartolini
How can you achieve an RTO of 5 minutes for the backups of your PostgreSQL databases? And what about RPO=0 for zero data loss backups?
This webinar gave an answer to those questions, by providing an overview of Disaster Recovery of PostgreSQL databases with Barman, covering its major features.
Barman, Backup and Recovery Manager for PostgreSQL, is an open source tool that was conceived by 2ndQuadrant about 10 years ago and released open source in 2012 under GNU GPL 3.
It is now one of the most popular backup and recovery tools in the PostgreSQL ecosystem.
Video available at https://resources.2ndquadrant.com/en/webinar-postgresql-continuous-backup-and-pitr-with-barman
PostgreSQL major version upgrade using built in Logical ReplicationAtsushi Torikoshi
This document discusses upgrading a PostgreSQL database from a major version to another using logical replication. It begins with an introduction to logical replication and its advantages over other upgrade methods. It then covers the architecture of logical replication, including how the walsender reads WAL files and applies changes to the subscriber. Finally, it addresses some limitations of logical replication for upgrades, such as objects that are not replicated, and provides strategies for completing the upgrade while minimizing downtime.
Integrating Apache Kafka and Elastic Using the Connect Frameworkconfluent
As a streaming platform, Apache Kafka provides low-latency, high-throughput, fault-tolerant publish and subscribe pipelines and excels at processing streams of real-time events. Kafka provides reliable, millisecond delivery for connecting downstream systems with real-time data.
In this talk, we will show how easy it is to leverage Kafka and the Elasticsearch connector to keep your indices populated with the latest data from the rest of your enterprise, as it changes.
Smarter Together - Bringing Relational Algebra, Powered by Apache Calcite, in...Julian Hyde
What if Looker saw the queries you just executed and could predict your next query? Could it make those queries faster, by smarter caching, or aggregate navigation? Could it read your past SQL queries and help you write your LookML model? Those are some of the reasons to add relational algebra into Looker’s query engine, and why Looker hired Julian Hyde, author of Apache Calcite, to lead the effort. In this talk about the internals of Looker’s query engine, Julian Hyde will describe how the engine works, how Looker queries are described in Calcite’s relational algebra, and some features that it makes possible.
A talk by Julian Hyde at JOIN 2019 in San Francisco.
A look at what HA is and what PostgreSQL has to offer for building an open source HA solution. Covers various aspects in terms of Recovery Point Objective and Recovery Time Objective. Includes backup and restore, PITR (point in time recovery) and streaming replication concepts.
Flink Forward Berlin 2017: Piotr Nowojski - "Hit me, baby, just one time" - B...Flink Forward
Getting data in and out of Flink is by far the most important aspect, and an everyday typical requirement of building Flink applications. Doing so in an end-to-end exactly-once manner, however, can be tricky. Being able to reliably consume data from the outside world without any duplicate processing and guaranteeing consistent distributed state, and at the same time provide computed results back to the outside world also without introducing duplicates, is crucial for the consistency and correctness of applications built upon stream processors. In this talk, we will talk about how end-to-end exactly-once guarantees can be achieved with Apache Flink. We will talk about Flink’s checkpointing mechanism, and how exactly to leverage it when consuming and producing data from your Flink streaming pipelines. In particular, we will be having a detailed review on how our supported connectors do so, with the aim to provide reference implementations for your own custom consumers and sinks.
Flink Forward San Francisco 2022.
This talk will take you on the long journey of Apache Flink into the cloud-native era. It started all the way from where Hadoop and YARN were the standard way of deploying and operating data applications.
We're going to deep dive into the cloud-native set of principles and how they map to the Apache Flink internals and recent improvements. We'll cover fast checkpointing, fault tolerance, resource elasticity, minimal infrastructure dependencies, industry-standard tooling, ease of deployment and declarative APIs.
After this talk you'll get a broader understanding of the operational requirements for a modern streaming application and where the current limits are.
by
David Moravek
Lightweight Transactions at Lightning SpeedScyllaDB
This talk will outline the Scylla implementation of Lightweight Transactions (LWT) that brings us to parity with Apache Cassandra. We will cover how to use it, what is working, and what is left to be done. We will also cover what other improvements are in store to improve Scylla's transactional capabilities and why it matters.
The document discusses installing and configuring MySQL on Linux. It provides steps to install MySQL using RPM files, set passwords for security, test the installation, and configure applications to connect to the database. It also covers basic and advanced MySQL commands like CREATE TABLE, SELECT, JOIN, and more.
Demystifying postgres logical replication percona live scEmanuel Calvo
This document provides an overview of logical replication in PostgreSQL, including:
- The different types of replication in PostgreSQL and how logical replication works
- How logical replication compares to MySQL replication and the elements involved
- What logical replication can be used for and some limitations
- Key concepts like publications, subscriptions, replication slots, and conflict handling
- Monitoring and configuration options for logical replication
DBA Brasil 1.0 - DBA Commands and Concepts That Every Developer Should KnowAlex Zaballa
This document summarizes a presentation on DBA commands and concepts that every developer should know. The presentation covers topics such as parallel processing, explain plans, flashback queries, pending statistics, virtual columns, and online table redefinition. It demonstrates several commands and concepts to help developers better understand database administration tasks.
DBA Brasil 1.0 - DBA Commands and Concepts That Every Developer Should KnowAlex Zaballa
This document summarizes a presentation on DBA commands and concepts that every developer should know. The presentation covers topics such as:
- Using SQLcl and DBMS_XPLAN to analyze query performance
- Leveraging parallelism to improve query performance
- Using Flashback Query and Flashback Table to view and recover past data
- Performing table recovery with RMAN
- Managing database schemas and statistics
- Implementing security with Virtual Private Database
DBA Brasil 1.0 - DBA Commands and Concepts That Every Developer Should KnowAlex Zaballa
This document summarizes a presentation on DBA commands and concepts that every developer should know. The presentation covers topics such as:
- Using SQLcl and tools like dbms_xplan to analyze queries and explain plans
- Leveraging parallelism to improve query performance
- Using flashback queries to view past data states
- Recovering dropped or corrupted tables using flashback and recycle bin features
- Migrating and restoring statistics to support testing and troubleshooting
The presentation provides examples and demonstrations of commands for tasks like analyzing queries, improving performance, recovering data, and managing database metadata.
The document discusses using plProxy and pgBouncer to split a PostgreSQL database horizontally and vertically to improve scalability. It describes how plProxy allows functions to make remote calls to other databases and how pgBouncer can be used for connection pooling. The RUN ON clause of plProxy is also summarized, which allows queries to execute on all partitions or on a specific partition.
Presentation from kafka meetup 13-SEP-2013. including some notes to clarify some slides. enjoy
Avi Levi
123avi@gmail.com
https://www.linkedin.com/in/leviavi/
This document provides an introduction and overview of Apache Kafka. It discusses Kafka's core concepts including producers, consumers, topics, partitions and brokers. It also covers how to install and run Kafka, producer and consumer configuration settings, and how data is distributed in a Kafka cluster. Examples of creating topics, producing and consuming messages are also included.
This document provides an overview of built-in physical and logical replication in PostgreSQL. It discusses:
- The different layers of replication including database physical, database logical, and operating system layers.
- The milestones of physical replication including file-based replication from PostgreSQL 8, streaming replication introduced in 9.0, and synchronous replication and quorum commit features added later.
- How to set up streaming replication including configuring the master and standby servers, synchronizing data, and creating a recovery configuration file.
- Features of streaming replication like asynchronous vs synchronous modes and hot standby vs warm standby.
- Logical replication introduced in PostgreSQL 10 and how it differs from physical replication by replic
PostgreSQL is a free and open-source relational database management system that provides high performance and reliability. It supports replication through various methods including log-based asynchronous master-slave replication, which the presenter recommends as a first option. The upcoming PostgreSQL 9.4 release includes improvements to replication such as logical decoding and replication slots. Future releases may add features like logical replication consumers and SQL MERGE statements. The presenter took questions at the end and provided additional resources on PostgreSQL replication.
PL/proxy is a PostgreSQL extension that allows writing functions to query other PostgreSQL databases, enabling cross-database queries. It can be used to implement horizontal partitioning by running functions on any, all, or an exact node. However, PL/proxy should still be considered alpha software with limited documentation and stability for some use cases.
MySQL replication allows data from a master database server to be copied to one or more slave database servers. It provides advantages like improving performance through load balancing, increasing data security with backups on slaves, and enabling analytics on slaves without impacting the master. Basic replication involves setting up a master server and slave server with unique IDs, configuring the master to log binary changes, and pointing the slave to the master so it can copy the binary log entries.
Webinar: MariaDB Provides the Solution to Ease Multi-Source ReplicationWagner Bianchi
MariaDB provides the solution to ease Multi-Source Replication aimed to show up the main characteristics of the
feature that was lunched together with MariaDB 10.0.1.
This document provides instructions on installing and configuring MySQL on Linux. It discusses downloading and installing the MySQL RPM package, setting the root password for security, starting the MySQL server and client, and running basic queries to test the installation. It also covers additional MySQL commands and configurations including user privileges, database design, backups, and restoring data.
This document provides a summary of PostgreSQL 9.1 features presented at the Postgres Open conference in September 2011. It discusses synchronous and asynchronous replication, per-column collations, writable common table expressions, serialized snapshot isolation, unlogged tables, extensions, and other features. Sessions at the conference focused on unlogged tables, accelerating local search, PostgreSQL in data management, and serializable transactions.
Oracle Database 12.1.0.2 introduced several new features including approximate count distinct, full database caching, pluggable database (PDB) improvements like cloning and state management, JSON support, data redaction, SQL query row limits and offsets, invisible columns, SQL text expansion, calling PL/SQL from SQL, session level sequences, and extended data types support.
This document describes how to configure MySQL database replication between a master and slave server. The key steps are:
1. Configure the master server by editing its configuration file to enable binary logging and set the server ID. Create a replication user and grant privileges.
2. Export the databases from the master using mysqldump.
3. Configure the slave server by editing its configuration file to point to the master server. Import the database dump. Start replication on the slave.
4. Verify replication is working by inserting data on the master and checking it is replicated to the slave.
DBA Commands and Concepts That Every Developer Should KnowAlex Zaballa
DBA Commands and Concepts That Every Developer Should Know was presented by Alex Zaballa, an Oracle DBA with experience in Brazil and Angola. The presentation covered Oracle Flashback Query, Flashback Table, RMAN table recovery, pending statistics, explain plan, DBMS_APPLICATION_INFO, row-by-row vs bulk processing, Virtual Private Database, extended data types, SQL text expansion, identity columns, UTL_CALL_STACK, READ privileges vs SELECT privileges, and online table redefinition. The presentation included demonstrations of many of these concepts.
This document provides information about installing and using the Firebird RDBMS, including:
- The two main types of Firebird servers and how to start/stop the Superserver.
- Default username and password for administration, and how to add/modify user accounts.
- Using the isql tool to connect to databases and execute SQL statements.
- Basic troubleshooting for common errors.
- Security measures like logging login attempts and restricting access after failed logins.
- Using the GBAK tool to backup and restore entire Firebird databases.
Similar to Basics of Logical Replication,Streaming replication vs Logical Replication ,Use Cases in PostgreSQL10 (20)
Goodbye Windows 11: Make Way for Nitrux Linux 3.5.0!SOFTTECHHUB
As the digital landscape continually evolves, operating systems play a critical role in shaping user experiences and productivity. The launch of Nitrux Linux 3.5.0 marks a significant milestone, offering a robust alternative to traditional systems such as Windows 11. This article delves into the essence of Nitrux Linux 3.5.0, exploring its unique features, advantages, and how it stands as a compelling choice for both casual users and tech enthusiasts.
HCL Notes und Domino Lizenzkostenreduzierung in der Welt von DLAUpanagenda
Webinar Recording: https://www.panagenda.com/webinars/hcl-notes-und-domino-lizenzkostenreduzierung-in-der-welt-von-dlau/
DLAU und die Lizenzen nach dem CCB- und CCX-Modell sind für viele in der HCL-Community seit letztem Jahr ein heißes Thema. Als Notes- oder Domino-Kunde haben Sie vielleicht mit unerwartet hohen Benutzerzahlen und Lizenzgebühren zu kämpfen. Sie fragen sich vielleicht, wie diese neue Art der Lizenzierung funktioniert und welchen Nutzen sie Ihnen bringt. Vor allem wollen Sie sicherlich Ihr Budget einhalten und Kosten sparen, wo immer möglich. Das verstehen wir und wir möchten Ihnen dabei helfen!
Wir erklären Ihnen, wie Sie häufige Konfigurationsprobleme lösen können, die dazu führen können, dass mehr Benutzer gezählt werden als nötig, und wie Sie überflüssige oder ungenutzte Konten identifizieren und entfernen können, um Geld zu sparen. Es gibt auch einige Ansätze, die zu unnötigen Ausgaben führen können, z. B. wenn ein Personendokument anstelle eines Mail-Ins für geteilte Mailboxen verwendet wird. Wir zeigen Ihnen solche Fälle und deren Lösungen. Und natürlich erklären wir Ihnen das neue Lizenzmodell.
Nehmen Sie an diesem Webinar teil, bei dem HCL-Ambassador Marc Thomas und Gastredner Franz Walder Ihnen diese neue Welt näherbringen. Es vermittelt Ihnen die Tools und das Know-how, um den Überblick zu bewahren. Sie werden in der Lage sein, Ihre Kosten durch eine optimierte Domino-Konfiguration zu reduzieren und auch in Zukunft gering zu halten.
Diese Themen werden behandelt
- Reduzierung der Lizenzkosten durch Auffinden und Beheben von Fehlkonfigurationen und überflüssigen Konten
- Wie funktionieren CCB- und CCX-Lizenzen wirklich?
- Verstehen des DLAU-Tools und wie man es am besten nutzt
- Tipps für häufige Problembereiche, wie z. B. Team-Postfächer, Funktions-/Testbenutzer usw.
- Praxisbeispiele und Best Practices zum sofortigen Umsetzen
GraphRAG for Life Science to increase LLM accuracyTomaz Bratanic
GraphRAG for life science domain, where you retriever information from biomedical knowledge graphs using LLMs to increase the accuracy and performance of generated answers
Driving Business Innovation: Latest Generative AI Advancements & Success StorySafe Software
Are you ready to revolutionize how you handle data? Join us for a webinar where we’ll bring you up to speed with the latest advancements in Generative AI technology and discover how leveraging FME with tools from giants like Google Gemini, Amazon, and Microsoft OpenAI can supercharge your workflow efficiency.
During the hour, we’ll take you through:
Guest Speaker Segment with Hannah Barrington: Dive into the world of dynamic real estate marketing with Hannah, the Marketing Manager at Workspace Group. Hear firsthand how their team generates engaging descriptions for thousands of office units by integrating diverse data sources—from PDF floorplans to web pages—using FME transformers, like OpenAIVisionConnector and AnthropicVisionConnector. This use case will show you how GenAI can streamline content creation for marketing across the board.
Ollama Use Case: Learn how Scenario Specialist Dmitri Bagh has utilized Ollama within FME to input data, create custom models, and enhance security protocols. This segment will include demos to illustrate the full capabilities of FME in AI-driven processes.
Custom AI Models: Discover how to leverage FME to build personalized AI models using your data. Whether it’s populating a model with local data for added security or integrating public AI tools, find out how FME facilitates a versatile and secure approach to AI.
We’ll wrap up with a live Q&A session where you can engage with our experts on your specific use cases, and learn more about optimizing your data workflows with AI.
This webinar is ideal for professionals seeking to harness the power of AI within their data management systems while ensuring high levels of customization and security. Whether you're a novice or an expert, gain actionable insights and strategies to elevate your data processes. Join us to see how FME and AI can revolutionize how you work with data!
Removing Uninteresting Bytes in Software FuzzingAftab Hussain
Imagine a world where software fuzzing, the process of mutating bytes in test seeds to uncover hidden and erroneous program behaviors, becomes faster and more effective. A lot depends on the initial seeds, which can significantly dictate the trajectory of a fuzzing campaign, particularly in terms of how long it takes to uncover interesting behaviour in your code. We introduce DIAR, a technique designed to speedup fuzzing campaigns by pinpointing and eliminating those uninteresting bytes in the seeds. Picture this: instead of wasting valuable resources on meaningless mutations in large, bloated seeds, DIAR removes the unnecessary bytes, streamlining the entire process.
In this work, we equipped AFL, a popular fuzzer, with DIAR and examined two critical Linux libraries -- Libxml's xmllint, a tool for parsing xml documents, and Binutil's readelf, an essential debugging and security analysis command-line tool used to display detailed information about ELF (Executable and Linkable Format). Our preliminary results show that AFL+DIAR does not only discover new paths more quickly but also achieves higher coverage overall. This work thus showcases how starting with lean and optimized seeds can lead to faster, more comprehensive fuzzing campaigns -- and DIAR helps you find such seeds.
- These are slides of the talk given at IEEE International Conference on Software Testing Verification and Validation Workshop, ICSTW 2022.
UiPath Test Automation using UiPath Test Suite series, part 5DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 5. In this session, we will cover CI/CD with devops.
Topics covered:
CI/CD with in UiPath
End-to-end overview of CI/CD pipeline with Azure devops
Speaker:
Lyndsey Byblow, Test Suite Sales Engineer @ UiPath, Inc.
In his public lecture, Christian Timmerer provides insights into the fascinating history of video streaming, starting from its humble beginnings before YouTube to the groundbreaking technologies that now dominate platforms like Netflix and ORF ON. Timmerer also presents provocative contributions of his own that have significantly influenced the industry. He concludes by looking at future challenges and invites the audience to join in a discussion.
How to Get CNIC Information System with Paksim Ga.pptxdanishmna97
Pakdata Cf is a groundbreaking system designed to streamline and facilitate access to CNIC information. This innovative platform leverages advanced technology to provide users with efficient and secure access to their CNIC details.
UiPath Test Automation using UiPath Test Suite series, part 6DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 6. In this session, we will cover Test Automation with generative AI and Open AI.
UiPath Test Automation with generative AI and Open AI webinar offers an in-depth exploration of leveraging cutting-edge technologies for test automation within the UiPath platform. Attendees will delve into the integration of generative AI, a test automation solution, with Open AI advanced natural language processing capabilities.
Throughout the session, participants will discover how this synergy empowers testers to automate repetitive tasks, enhance testing accuracy, and expedite the software testing life cycle. Topics covered include the seamless integration process, practical use cases, and the benefits of harnessing AI-driven automation for UiPath testing initiatives. By attending this webinar, testers, and automation professionals can gain valuable insights into harnessing the power of AI to optimize their test automation workflows within the UiPath ecosystem, ultimately driving efficiency and quality in software development processes.
What will you get from this session?
1. Insights into integrating generative AI.
2. Understanding how this integration enhances test automation within the UiPath platform
3. Practical demonstrations
4. Exploration of real-world use cases illustrating the benefits of AI-driven test automation for UiPath
Topics covered:
What is generative AI
Test Automation with generative AI and Open AI.
UiPath integration with generative AI
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Cosa hanno in comune un mattoncino Lego e la backdoor XZ?Speck&Tech
ABSTRACT: A prima vista, un mattoncino Lego e la backdoor XZ potrebbero avere in comune il fatto di essere entrambi blocchi di costruzione, o dipendenze di progetti creativi e software. La realtà è che un mattoncino Lego e il caso della backdoor XZ hanno molto di più di tutto ciò in comune.
Partecipate alla presentazione per immergervi in una storia di interoperabilità, standard e formati aperti, per poi discutere del ruolo importante che i contributori hanno in una comunità open source sostenibile.
BIO: Sostenitrice del software libero e dei formati standard e aperti. È stata un membro attivo dei progetti Fedora e openSUSE e ha co-fondato l'Associazione LibreItalia dove è stata coinvolta in diversi eventi, migrazioni e formazione relativi a LibreOffice. In precedenza ha lavorato a migrazioni e corsi di formazione su LibreOffice per diverse amministrazioni pubbliche e privati. Da gennaio 2020 lavora in SUSE come Software Release Engineer per Uyuni e SUSE Manager e quando non segue la sua passione per i computer e per Geeko coltiva la sua curiosità per l'astronomia (da cui deriva il suo nickname deneb_alpha).
In the rapidly evolving landscape of technologies, XML continues to play a vital role in structuring, storing, and transporting data across diverse systems. The recent advancements in artificial intelligence (AI) present new methodologies for enhancing XML development workflows, introducing efficiency, automation, and intelligent capabilities. This presentation will outline the scope and perspective of utilizing AI in XML development. The potential benefits and the possible pitfalls will be highlighted, providing a balanced view of the subject.
We will explore the capabilities of AI in understanding XML markup languages and autonomously creating structured XML content. Additionally, we will examine the capacity of AI to enrich plain text with appropriate XML markup. Practical examples and methodological guidelines will be provided to elucidate how AI can be effectively prompted to interpret and generate accurate XML markup.
Further emphasis will be placed on the role of AI in developing XSLT, or schemas such as XSD and Schematron. We will address the techniques and strategies adopted to create prompts for generating code, explaining code, or refactoring the code, and the results achieved.
The discussion will extend to how AI can be used to transform XML content. In particular, the focus will be on the use of AI XPath extension functions in XSLT, Schematron, Schematron Quick Fixes, or for XML content refactoring.
The presentation aims to deliver a comprehensive overview of AI usage in XML development, providing attendees with the necessary knowledge to make informed decisions. Whether you’re at the early stages of adopting AI or considering integrating it in advanced XML development, this presentation will cover all levels of expertise.
By highlighting the potential advantages and challenges of integrating AI with XML development tools and languages, the presentation seeks to inspire thoughtful conversation around the future of XML development. We’ll not only delve into the technical aspects of AI-powered XML development but also discuss practical implications and possible future directions.
Dr. Sean Tan, Head of Data Science, Changi Airport Group
Discover how Changi Airport Group (CAG) leverages graph technologies and generative AI to revolutionize their search capabilities. This session delves into the unique search needs of CAG’s diverse passengers and customers, showcasing how graph data structures enhance the accuracy and relevance of AI-generated search results, mitigating the risk of “hallucinations” and improving the overall customer journey.
Unlock the Future of Search with MongoDB Atlas_ Vector Search Unleashed.pdfMalak Abu Hammad
Discover how MongoDB Atlas and vector search technology can revolutionize your application's search capabilities. This comprehensive presentation covers:
* What is Vector Search?
* Importance and benefits of vector search
* Practical use cases across various industries
* Step-by-step implementation guide
* Live demos with code snippets
* Enhancing LLM capabilities with vector search
* Best practices and optimization strategies
Perfect for developers, AI enthusiasts, and tech leaders. Learn how to leverage MongoDB Atlas to deliver highly relevant, context-aware search results, transforming your data retrieval process. Stay ahead in tech innovation and maximize the potential of your applications.
#MongoDB #VectorSearch #AI #SemanticSearch #TechInnovation #DataScience #LLM #MachineLearning #SearchTechnology
Programming Foundation Models with DSPy - Meetup SlidesZilliz
Prompting language models is hard, while programming language models is easy. In this talk, I will discuss the state-of-the-art framework DSPy for programming foundation models with its powerful optimizers and runtime constraint system.
Let's Integrate MuleSoft RPA, COMPOSER, APM with AWS IDP along with Slackshyamraj55
Discover the seamless integration of RPA (Robotic Process Automation), COMPOSER, and APM with AWS IDP enhanced with Slack notifications. Explore how these technologies converge to streamline workflows, optimize performance, and ensure secure access, all while leveraging the power of AWS IDP and real-time communication via Slack notifications.
5. Overview of PostgreSQL replication 9.0 onwards
cont..
9.4
Replication
Slots, Logical
Decoding (LD)
9.5 pg_rewind
9.6
10
Native support
for Logical
replication.
More support
to LD
6. Logical Replication: Overview
Replicating/capturing logical changes based upon their replica identity
using replication slots.
Propagate any changes (Insert, Update, Delete, all, combination) to
replica.
Uses a publish and subscribe model.
Publisher-Sender
Subscriber- Receiver
Subscribers pulls data from publications.
Generally ‘asynchronous’ in nature.
7. Why Logical Replication over Streaming Replication?
Features Streaming Replication Logical Replication
Replication Type Binary replication(byte-by-byte) Selective row level changes (logical
change)
Hardware/OS Same (Linux-Linux) Can be different(Linux-Windows)
Replication between major
versions
Not supported Supported.
Replication Level Instance level Object level(table)
Consolidation No consolidation Consolidating is possible
Replica/ Subscriber- Open
Mode
Standby is in read-only mode Subscriber can be used for write
operations
8. Use Case1: Cross platform replication
Cross Platform : Linux Windows
Db1 Db2
Db2
Logical Replication
PostgreSQL 10 server1 PostgreSQL 10 server2
tb1 tb1
11. Use Case2: Cross version:
Cross Version: Replication between different major versions of PostgreSQL
Db1 Db2 Db3
Logical Replication
PostgreSQL 11 server2PostgreSQL10 server1
tb1 tb1
12. Use Case3: Write operation @Subscriber:
Modifying objects and adding new tables, insert records on subscriber.
Tbl1 Tbl1
Logical ReplicationID Name Name ID Phno
ID Title Name
Tbl2
PostgreSQL 10 server2PostgreSQL 10 server1
PUB SUB
13. Use Case4: Consolidating multiple databases
PostgreSQL Server1 PostgreSQL
Server2(reporting)
PostgreSQL Server3
Can be helpful for reporting or analytical purposes
Can replicate even few tables from one DB to another on different servers.
db1
db3
db2
db4
db5
Reporting db6
tb1 tb1 tb2 tb2
14. Terminology used in Logical Replication:
Publisher/Publication
Subscriber/Subscription
Replication Slot
Replica Identity
16. Publisher/Publication
Publisher node where a publication is defined.
Publication set of changes defined for tables.
Publication can be defined on any physical replication master.
DB name can be different from Subscriber.
Create publication using the ‘CREATE’ command and can be later ALTERED and
DROPPED.
A published table should exists before creating publication and must have a “replica
identity“ (can be primary key or index).
Operations possible on tables- INSERT, UPDATE, and DELETE, combinations, or
ALL(default)
UPDATE or DELETE operations on publisher without replica identity will give error.
Same replica identity has to be set at subscription side as well.
Currently replicates only tables.
17. Commands related to Publication:
Create:
• CREATE PUBLICATION testpub FOR TABLE tb1, tb2;
• CREATE PUBLICATION testpub FOR TABLE tbl3 WITH (publish =
'insert, update');
• CREATE PUBLICATION testpub FOR ALL TABLES;
Alter :
• ALTER PUBLICATION testpub ADD TABLE tb4, tbl5;
• ALTER PUBLICATION testpub SET TABLE tbl4, tbl5;
• ALTER PUBLICATION testpub DROP TABLE tbl4, tbl5;
Drop:
• DROP PUBLICATION testpub;
• DROP PUBLICATION testpub, testpub1;
19. Subscription:
Subscription Downstream side of logical replication and this node is called
subscriber.
It can be used as a publisher for other databases.
Multiple subscriptions on one subscriber is possible.
Subscription should be created using superuser.
Same table names at publisher and subscriber.
Tables should be created first before creating a subscription.
Same columns in source and target tables with same data type but can have
different order.
20. Commands related to Subscription:
Create:
• CREATE SUBSCRIPTION testsub1 CONNECTION 'host=<remote>
'dbname=<db> user=<user>' PUBLICATION testpub1,testpub2;
• CREATE SUBSCRIPTION mysub1 CONNECTION '...' PUBLICATION
...WITH (enabled = false,create_slot = false,slot_name =
'myslot', copy_data = false, ...);
Alter:
• ALTER SUBSCRIPTION testsub ENABLE/DISABLE;
• ALTER SUBSCRIPTION testsub CONNECTION 'host=newhost ...';
• ALTER SUBSCRIPTION testsub SET(slot_name = 'newslot‘/’NONE’);
• ALTER SUBSCRIPTION testsub REFRESH PUBLICATION;
Drop:
• DROP SUBSCRIPTION testsub;
21. Security
For creating publication : CREATE privilege in the database.
To add tables to a publication, the user must have Ownership rights on the table.
To create a subscription, the user must be a superuser.
Role used for the replication connection must have the REPLICATION attribute (or
be a superuser)
Entry for replication role in pg_hba.conf
24. Settings required for demo:
1. Setting postgresql.conf @Publisher
wal_level = logical
Note: Default values of other required parameters will suffice for this demo.
2. Use replication role(CREATE ROLE replication WITH REPLICATION PASSWORD ‘’ LOGIN;
3. Setting pg_hba.conf on both nodes.
host all postgres 0.0.0.0/0 SCRAM-SHA-256
host all replication 0.0.0.0/0 SCRAM-SHA-256
* Note :The values here depend on your actual network configuration
25. Overview of Demo:
Server1=pg10srv1 Server2=pg10srv2
SubscriberPublisher
PUB SUB
test emptest
Publication: testpub Subscription: testsub
User
Subscriber
39. * Default operation performed on tables is ‘All’ i.e. Insert, update and delete.
List of publications
Name | Owner | All tables | Inserts | Updates | Deletes
---------+----------+------------+---------+---------+--------
-
testpub | postgres | f | t | t | t
(1 row)
68. What if we insert a record with id=4 on Publisher?
User
Publisher
Db1PUB
Db3
SUB
testtest
Publication: testpub Subscription: testsub
Id=4
Inserted later
Id=4
inserted
first
Server1=pg10srv1 Server2=pg10srv2
conflict
Subscriber
69. Check conflict status in subscriber log:
Error in log file:
On Subscriber On Publisher
sub=# SELECT * FROM TEST;
name | id
-------------------+----
TOM | 1
DICK | 2
AFTER REPLICATION | 3
CONFLICT | 4
(4 rows)
pub=# INSERT INTO TEST
VALUES(4,'RESOLVE');
sub=# SELECT * FROM
TEST;
name | id
-------------------+----
TOM | 1
DICK | 2
AFTER REPLICATION | 3
CONFLICT | 4
(4 rows)
70. Tab1
Tab1
Tab1
Delete a column on subscriber:
ID Name
ID Name
ID Name
PUBLISHER SUBSCRIBER
Tab1(different order)
Tab1(add column)
Tab1(delete column)
Name ID
Name ID Phno
ID Phno
conflict
When
insert
71. Resolving Conflict:
Change data on subscriber i.e. delete conflicting key.
Skip the conflicting transaction from replication by calling the
pg_replication_origin_advance() function.
The current position of origins can be seen in the
pg_replication_origin_status system view.
73. Conclusion:
Good feature but still room to include more features.
No need to replicate whole instance.
Subscriber can be used for Write operations but with care.
Good for cross platform replication, different major version replication, write
operations on subscriber, like use cases.
Not the replacement of SR.