Presentation by Rajkumar Sen, Chief Technical Architect, Founder - BlitzzIO, recorded at Distributed SQL Summit on Sept 20, 2019.
https://vimeo.com/362348541
distributedsql.org/
Distributed Databases Deconstructed: CockroachDB, TiDB and YugaByte DBYugabyteDB
Slides for Amey Banarse's, Principal Data Architect at Yugabyte, "Distributed Databases Deconstructed: CockroachDB, TiDB and YugaByte DB" webinar recorded on Oct 30, 2019 at 11 AM Pacific.
Playback here: https://vimeo.com/369929255
Docker containers are great for running stateless microservices, but what about stateful applications such as databases and persistent queues? Kubernetes provides the StatefulSets controller for such applications that have to manage data in some form of persistent storage. While StatefulSets is a great start, a lot more goes into ensuring high performance, data durability and high availability for stateful apps in Kubernetes. Following are 5 best practices that developers and operations engineers should be aware of.
1. Ensure high performance with local persistent volumes and pod anti-affinity rules.
2. Achieve data resilience with auto-failover and multi-zone pod scheduling.
3. Integrate StatefulSet services with other application services through NodePorts & LoadBalancer services.
4. Run Day 2 operations such as monitoring, elastic scaling, capacity re-sizing, backups with caution.
5. Automate operations through Kubernetes Operators that extend the StatefulSets controller.
We will demonstrate how to run a complete E-Commerce application powered by YugaByte DB, when all services are deployed in Kubernetes.
This document discusses using StatefulSets in Kubernetes to deploy the distributed database YugaByte DB. It provides examples of deploying YugaByte DB masters and tservers as StatefulSets, ensuring high performance through local storage and data resilience through pod anti-affinity. It also demonstrates a real-world e-commerce application deployed on this stack and discusses extending StatefulSets with operators.
Scale Transactional Apps Across Multiple Regions with Low LatencyYugabyte
User-facing transactional apps in verticals such as Retail, Finance and SaaS are increasingly moving from a single-region, monolithic architecture to a multi-region, cloud-native architecture. Enhancing customer satisfaction with low latency access, protecting data through geo-redundancy and satisfying compliance requirements such as GDPR are some of the major drivers for this move. Unfortunately, the DB tier powering the above apps has remained as a high-latency, hard-to-scale master-slave RDBMS for a long time. Multi-master deployments as well as the use of a separate NoSQL DB for multi-region data distribution are simply band-aids to this problem and do not deliver the desired business outcomes.
This talk shows how to use YugaByte DB to scale transactional apps across multiple regions with low latency.
How YugaByte DB Implements Distributed PostgreSQLYugabyte
Building applications on PostgreSQL that require automatic data sharding and replication, fault tolerance, distributed transactions and geographic data distribution has been hard. In this 3 hour workshop, we will look at how to do this using a real-world example running on top of YugaByte DB, a distributed database that is fully wire-compatible with PostgreSQL and NoSQL APIs (Apache Cassandra and Redis). We will look at the design and architecture of YugaByte DB and how it reuses the PostgreSQL codebase to achieve full API compatibility. YugaByte DB support for PostgreSQL includes most data types, queries, stored procedures, etc. We will also take a look at how to build applications that are planet scale (requiring geographic distribution of data) and how to run them in cloud-native environments (for example, Kubernetes, hybrid or multi-cloud deployments).
SQL is a popular database language for modern applications, given its flexibility in modelling workloads and how widely it is understood by developers. However, most modern applications running in the clouds require fault tolerance, the ability to scale out and geographic data distribution of data. These are hard to achieve with traditional SQL databases, which is paving the way for distributed SQL databases.
Google Spanner is arguably the world's first truly distributed SQL database. Given its fully decentralized architecture, it delivers higher performance and availability for geo-distributed SQL workloads than other specialized transactional databases such as Amazon Aurora. Now, there are a number of open source derivatives of Google Spanner such as YugaByte DB, CockroachDB and TiDB. This talk will focus on the common architectural paradigms that these databases are built on (using YugaByte DB as an example). Learn about the concepts these databases leverage, how to evaluate if these will meet your needs and the questions to ask to differentiate among these databases.
YugaByte DB on Kubernetes - An IntroductionYugabyte
This document summarizes YugaByte DB, a distributed SQL and NoSQL database. It discusses how YugaByte DB provides ACID transactions, strong consistency, and high performance at a planet scale. It also describes how to deploy YugaByte DB and an example e-commerce application called Yugastore on Kubernetes. The document outlines the database architecture and components, and provides steps to deploy the system and run a sample workload.
Distributed Databases Deconstructed: CockroachDB, TiDB and YugaByte DBYugabyteDB
Slides for Amey Banarse's, Principal Data Architect at Yugabyte, "Distributed Databases Deconstructed: CockroachDB, TiDB and YugaByte DB" webinar recorded on Oct 30, 2019 at 11 AM Pacific.
Playback here: https://vimeo.com/369929255
Docker containers are great for running stateless microservices, but what about stateful applications such as databases and persistent queues? Kubernetes provides the StatefulSets controller for such applications that have to manage data in some form of persistent storage. While StatefulSets is a great start, a lot more goes into ensuring high performance, data durability and high availability for stateful apps in Kubernetes. Following are 5 best practices that developers and operations engineers should be aware of.
1. Ensure high performance with local persistent volumes and pod anti-affinity rules.
2. Achieve data resilience with auto-failover and multi-zone pod scheduling.
3. Integrate StatefulSet services with other application services through NodePorts & LoadBalancer services.
4. Run Day 2 operations such as monitoring, elastic scaling, capacity re-sizing, backups with caution.
5. Automate operations through Kubernetes Operators that extend the StatefulSets controller.
We will demonstrate how to run a complete E-Commerce application powered by YugaByte DB, when all services are deployed in Kubernetes.
This document discusses using StatefulSets in Kubernetes to deploy the distributed database YugaByte DB. It provides examples of deploying YugaByte DB masters and tservers as StatefulSets, ensuring high performance through local storage and data resilience through pod anti-affinity. It also demonstrates a real-world e-commerce application deployed on this stack and discusses extending StatefulSets with operators.
Scale Transactional Apps Across Multiple Regions with Low LatencyYugabyte
User-facing transactional apps in verticals such as Retail, Finance and SaaS are increasingly moving from a single-region, monolithic architecture to a multi-region, cloud-native architecture. Enhancing customer satisfaction with low latency access, protecting data through geo-redundancy and satisfying compliance requirements such as GDPR are some of the major drivers for this move. Unfortunately, the DB tier powering the above apps has remained as a high-latency, hard-to-scale master-slave RDBMS for a long time. Multi-master deployments as well as the use of a separate NoSQL DB for multi-region data distribution are simply band-aids to this problem and do not deliver the desired business outcomes.
This talk shows how to use YugaByte DB to scale transactional apps across multiple regions with low latency.
How YugaByte DB Implements Distributed PostgreSQLYugabyte
Building applications on PostgreSQL that require automatic data sharding and replication, fault tolerance, distributed transactions and geographic data distribution has been hard. In this 3 hour workshop, we will look at how to do this using a real-world example running on top of YugaByte DB, a distributed database that is fully wire-compatible with PostgreSQL and NoSQL APIs (Apache Cassandra and Redis). We will look at the design and architecture of YugaByte DB and how it reuses the PostgreSQL codebase to achieve full API compatibility. YugaByte DB support for PostgreSQL includes most data types, queries, stored procedures, etc. We will also take a look at how to build applications that are planet scale (requiring geographic distribution of data) and how to run them in cloud-native environments (for example, Kubernetes, hybrid or multi-cloud deployments).
SQL is a popular database language for modern applications, given its flexibility in modelling workloads and how widely it is understood by developers. However, most modern applications running in the clouds require fault tolerance, the ability to scale out and geographic data distribution of data. These are hard to achieve with traditional SQL databases, which is paving the way for distributed SQL databases.
Google Spanner is arguably the world's first truly distributed SQL database. Given its fully decentralized architecture, it delivers higher performance and availability for geo-distributed SQL workloads than other specialized transactional databases such as Amazon Aurora. Now, there are a number of open source derivatives of Google Spanner such as YugaByte DB, CockroachDB and TiDB. This talk will focus on the common architectural paradigms that these databases are built on (using YugaByte DB as an example). Learn about the concepts these databases leverage, how to evaluate if these will meet your needs and the questions to ask to differentiate among these databases.
YugaByte DB on Kubernetes - An IntroductionYugabyte
This document summarizes YugaByte DB, a distributed SQL and NoSQL database. It discusses how YugaByte DB provides ACID transactions, strong consistency, and high performance at a planet scale. It also describes how to deploy YugaByte DB and an example e-commerce application called Yugastore on Kubernetes. The document outlines the database architecture and components, and provides steps to deploy the system and run a sample workload.
YugaByte DB Internals - Storage Engine and Transactions Yugabyte
This document introduces YugaByte DB, a high-performance, distributed, transactional database. It is built to scale horizontally on commodity servers across data centers for mission-critical applications. YugaByte DB uses a transactional document store based on RocksDB, Raft-based replication for resilience, and automatic sharding and rebalancing. It supports ACID transactions across documents, provides APIs compatible with Cassandra and Redis, and is open source. The architecture is designed for high performance, strong consistency, and cloud-native deployment.
In this webinar, we will discuss different open-source models and different ways open source communities are organized. Understanding these key concepts is essential when selecting a strategic open-source platform. We will explore how the PostgreSQL community ensures that it stays independent, remains vibrant, drives innovation, and provides a reliable long-term platform for strategic IT projects.
Always upgrade! There are hundreds of fixes between each PostgreSQL release, and an important number of them are security fixes! Logical replication allows making major upgrades with minimal downtime and feasible cons.
True Postgres High Availability Architecture with Industry Standard Open-Sou...EDB
With use of modern deployment tools like Ansible, a high availability stack is built with only open-source products. Automatic fail-over, Backup- recovery, Monitoring, SSL, Virtual IP and an environment to manage the stack is part of the setup. Complete setup with industry-standard tools like Barman, Repmgr, Prometheus/Grafana, HAProxy and Keepalived. This talk will deliver a live demo of building the environment, and testing a demo application while doing a switch over, and prove that no transactions are lost.
One of the main reasons we hear for not being able to migrate from Oracle to PostgreSQL is the fact that “We absolutely, definitively must have RAC”. Despite its cost and complexity, RAC is widely used in response to a variety of (real or imagined) performance, high availability and/or scalability requirements. PostgreSQL does not have an equivalent to RAC. But does anyone really need it?
This talk investigates whether or not RAC is actually required in most situations and looks at the alternatives that are available to us in PostgreSQL.
PostgreSQL continuous backup and PITR with BarmanEDB
How can I achieve an RPO of 5 minutes for the backups of my PostgreSQL databases? And what about RPO=0 for zero data loss backups? This talk will give you answers to those questions, by guiding you through an overview of Disaster Recovery of PostgreSQL databases with Barman, covering its key concepts and providing useful patterns and tips.
Apache AGE and the synergy effect in the combination of Postgres and NoSQLEDB
This document provides an overview and agenda for a presentation on Apache AGE and the synergy between PostgreSQL and NoSQL databases. The agenda includes introducing the speaker, discussing graph databases and Apache AGE as an open source project, an introduction to Apache AGE including its internal logic and roadmap, advantages of using Apache AGE, and how to contribute to the project. Key points are that Apache AGE is a multi-model graph database built on PostgreSQL that allows users to integrate relational and graph data models, leveraging the features of PostgreSQL and providing a graph query interface using openCypher. The goal of the project is to become an Apache Software Foundation top-level project.
This document outlines Pivotal's Greenplum roadmap. It discusses plans to enhance Greenplum's open source strategy by continuing to align with PostgreSQL functionality. It also describes strategies for multi-cloud support on AWS, Azure, Google Cloud and others as well as integration with Kubernetes. Near term focus areas include improving the query planner, adding resource groups and containerized Python/R support. The roadmap outlines enhancements through 2018 and a major release in 2019 incorporating additional PostgreSQL features while retaining MPP performance and scale. Long term initiatives include disaster recovery, foreign data and integrating new data types like spatial and time series data.
PostgreSQL is versatile and used for a wide range of applications and use cases in the enterprise. It is more than just database technology, it is an accelerator for innovation. Much innovation today is happening in new application development, application modernization, and re-platforming to the cloud across the information architecture landscape. In this webinar, you will learn how EDB supercharges PostgreSQL to re-platform to cloud and containers more efficiently and develop new applications that are more scalable and secure.
The talk will be about the project to find a replacement for all IBM products in the company with the example for the databases. What was the goal of the project, the learning, a short overview about the options
we migrated about 500 db2 databases to EnterpriseDB. The database size was from a small size up to 4 TB and we implemented a completely new fully automated deployment of VM and database. Databases are now 11 month in production. The talk will have an overview of the project, the learnings, a few parameters and technical parameters that were found for stability and performance.
Apache Flink is a popular stream computing framework for real-time stream computing. Many stream compute algorithms require trailing data in order to compute the intended result. One example is computing the number of user logins in the last 7 days. This creates a dilemma where the results of the stream program are incomplete until the runtime of the program exceeds 7 days. The alternative is to bootstrap the program using historic data to seed the state before shifting to use real-time data.
This talk will discuss alternatives to bootstrap programs in Flink. Some alternatives rely on technologies exogenous to the stream program, such as enhancements to the pub/sub layer, that are more generally applicable to other stream compute engines. Other alternatives include enhancements to Flink source implementations. Lyft is exploring another alternative using orchestration of multiple Flink programs. The talk will cover why Lyft pursued this alternative and future directions to further enhance bootstrapping support in Flink.
Speaker
Gregory Fee, Principal Engineer, Lyft
Large Table Partitioning with PostgreSQL and DjangoEDB
With great DB Table comes great responsibility". Our email messages table was growing too much and we needed to do something about it. We will talk about how we integrated PostgreSQL Declarative partitioning with our Django based Customer Portal to solve the problem.
Join Marc Linster andDave Page as they shows you how to manage large scale Postgres deployments. This webinar will highlight how EnterpriseDB Postgres Enterprise Manager can be used for monitoring, alerting and administration of your Postgres estate - no matter where it is deployed.
Highlights include:
Tools and techniques for Postgres management
Key database performance metrics and monitoring
Operating Postgres on premises, in containers, and in the cloud
They will conclude with a live demo of Postgres Enterprise Manager.
SpringCamp 2016 - Apache Geode 와 Spring Data GemfireJay Lee
The document discusses Apache Geode and Spring integration. It provides an overview of Apache Geode, an open source distributed in-memory database. It then covers Spring Data Gemfire, which allows using Geode with Spring's programming model. It also discusses using Spring Session to manage user sessions in a stateless manner by storing them in Geode. The presentation includes demos of integrating Geode with Spring applications.
A webinarban megtudhatják milyen kihívásokkal kell szembenézni Oracle adatbázis PostgreSQL-re migrálása során. Bemutatjuk az utóbbi két év nagy komplexitású Oracle kompatibilitási vizsgálatainak tapasztalatait, beleértve az idei évben az EDB migrációs portálján keresztül vizsgált több mint 2 200 000 Oracle DDL konstrukciót.
Az előadás alatt kitérünk az alábbiakra:
- Tárhely (storage) definiciók
- Csomagok
- Tárolt eljárások
- PL/SQL kód
- Gyártói adatbázis API-k
- Komplex adatbázis migrációk
Előadásunkat olyan migrációs eszközök bemutatásával zárjuk, amelyek jelentősen leegyszerűsítik az Oracle-PostgeSQL migrációt és csökkentik annak kockázatait.
EDB 13 - New Enhancements for Security and Usability - APJEDB
Database security is always of paramount importance to all organizations. In this webinar, we will explore the security, usability, and portability updates of the latest version of the EDB database server and tools.
Join us in this webinar to learn:
- The new security features such as SCRAM and the encryption of database passwords and traffic between Failover Manager agents
- Usability updates that automate partitioning, verify backup integrity, and streamline the management of failover and backups
- Portability improvements that simplify running PostgreSQL across on-premise and cloud environments
Why you should use native packages to install PostgreSQL on LinuxEDB
When it comes to installing PostgreSQL on a Linux distribution, using a native package is considered to be a better choice compared to installing from the source. In this webinar, we will discuss some benefits of using native packages to install PostgreSQL.
Optimizing Performance: Installers are shipped with “the lowest common denominator” and native packages use the OS libraries resulting in performance and maintenance benefits.
Security: Native packages depend on the OS, thus updating OS packages are sufficient for security issues.
Usability: Native packages integrate with the OS packaging system (dnf, yum, apt, etc), increasing usability.
Present & Future of Greenplum Database A massively parallel Postgres Database...VMware Tanzu
Greenplum Database is Pivotal's massively parallel Postgres database. Version 5 has proven features for mission critical use cases. Version 6 adds improvements like row-level locking, foreign data wrappers, and online expansion to make Greenplum a superset of Postgres. It also provides up to 50x faster OLTP performance. Version 7 will focus on capabilities beyond the cluster like streaming replication and using Greenplum as a source for data integration tools.
Introducing Apache Geode and Spring Data GemFireJohn Blum
This document introduces Apache Geode, an open source distributed in-memory data management platform. It discusses what Geode is, how it is implemented, and some key features like high availability, scalability and low latency. It also introduces Spring Data GemFire, which simplifies using Geode with Spring applications through features like repositories and caching. Finally, it outlines the project roadmap and opportunities to get involved in the Geode community.
EnterpriseDB Postgres Plus Advanced Server provides Oracle compatibility with enterprise performance features built upon the legendary open source PostgreSQL platform, all certified on IBM’s latest Linux on Power servers.
The highlights of this presentation include:
* An overview of the database landscape – past, present and future
* Postgres NoSQL capabilities for document and key-value store work loads
* How you can lower your Total-Cost-of-Ownership (TCO) with Postgres in conjunction with your current database
* What resources are available to assess the right decision
* How the IBM Power Systems™ platform is fueling performance, reliability, security, TCO and virtualization for new applications, markets and geographies.
* Suggested audience: This presentation is intended for strategic IT and Business Decision-Makers involved in IT infrastructure and application development.
Healthcare Claim Reimbursement using Apache SparkDatabricks
The document discusses rewriting a claims reimbursement system using Spark. It describes how Spark provides better performance, scalability and cost savings compared to the previous Oracle-based system. Key points include using Spark for ETL to load data into a Delta Lake data lake, implementing the business logic in a reusable Java library, and seeing significant increases in processing volumes and speeds compared to the prior system. Challenges and tips for adoption are also provided.
YugaByte DB Internals - Storage Engine and Transactions Yugabyte
This document introduces YugaByte DB, a high-performance, distributed, transactional database. It is built to scale horizontally on commodity servers across data centers for mission-critical applications. YugaByte DB uses a transactional document store based on RocksDB, Raft-based replication for resilience, and automatic sharding and rebalancing. It supports ACID transactions across documents, provides APIs compatible with Cassandra and Redis, and is open source. The architecture is designed for high performance, strong consistency, and cloud-native deployment.
In this webinar, we will discuss different open-source models and different ways open source communities are organized. Understanding these key concepts is essential when selecting a strategic open-source platform. We will explore how the PostgreSQL community ensures that it stays independent, remains vibrant, drives innovation, and provides a reliable long-term platform for strategic IT projects.
Always upgrade! There are hundreds of fixes between each PostgreSQL release, and an important number of them are security fixes! Logical replication allows making major upgrades with minimal downtime and feasible cons.
True Postgres High Availability Architecture with Industry Standard Open-Sou...EDB
With use of modern deployment tools like Ansible, a high availability stack is built with only open-source products. Automatic fail-over, Backup- recovery, Monitoring, SSL, Virtual IP and an environment to manage the stack is part of the setup. Complete setup with industry-standard tools like Barman, Repmgr, Prometheus/Grafana, HAProxy and Keepalived. This talk will deliver a live demo of building the environment, and testing a demo application while doing a switch over, and prove that no transactions are lost.
One of the main reasons we hear for not being able to migrate from Oracle to PostgreSQL is the fact that “We absolutely, definitively must have RAC”. Despite its cost and complexity, RAC is widely used in response to a variety of (real or imagined) performance, high availability and/or scalability requirements. PostgreSQL does not have an equivalent to RAC. But does anyone really need it?
This talk investigates whether or not RAC is actually required in most situations and looks at the alternatives that are available to us in PostgreSQL.
PostgreSQL continuous backup and PITR with BarmanEDB
How can I achieve an RPO of 5 minutes for the backups of my PostgreSQL databases? And what about RPO=0 for zero data loss backups? This talk will give you answers to those questions, by guiding you through an overview of Disaster Recovery of PostgreSQL databases with Barman, covering its key concepts and providing useful patterns and tips.
Apache AGE and the synergy effect in the combination of Postgres and NoSQLEDB
This document provides an overview and agenda for a presentation on Apache AGE and the synergy between PostgreSQL and NoSQL databases. The agenda includes introducing the speaker, discussing graph databases and Apache AGE as an open source project, an introduction to Apache AGE including its internal logic and roadmap, advantages of using Apache AGE, and how to contribute to the project. Key points are that Apache AGE is a multi-model graph database built on PostgreSQL that allows users to integrate relational and graph data models, leveraging the features of PostgreSQL and providing a graph query interface using openCypher. The goal of the project is to become an Apache Software Foundation top-level project.
This document outlines Pivotal's Greenplum roadmap. It discusses plans to enhance Greenplum's open source strategy by continuing to align with PostgreSQL functionality. It also describes strategies for multi-cloud support on AWS, Azure, Google Cloud and others as well as integration with Kubernetes. Near term focus areas include improving the query planner, adding resource groups and containerized Python/R support. The roadmap outlines enhancements through 2018 and a major release in 2019 incorporating additional PostgreSQL features while retaining MPP performance and scale. Long term initiatives include disaster recovery, foreign data and integrating new data types like spatial and time series data.
PostgreSQL is versatile and used for a wide range of applications and use cases in the enterprise. It is more than just database technology, it is an accelerator for innovation. Much innovation today is happening in new application development, application modernization, and re-platforming to the cloud across the information architecture landscape. In this webinar, you will learn how EDB supercharges PostgreSQL to re-platform to cloud and containers more efficiently and develop new applications that are more scalable and secure.
The talk will be about the project to find a replacement for all IBM products in the company with the example for the databases. What was the goal of the project, the learning, a short overview about the options
we migrated about 500 db2 databases to EnterpriseDB. The database size was from a small size up to 4 TB and we implemented a completely new fully automated deployment of VM and database. Databases are now 11 month in production. The talk will have an overview of the project, the learnings, a few parameters and technical parameters that were found for stability and performance.
Apache Flink is a popular stream computing framework for real-time stream computing. Many stream compute algorithms require trailing data in order to compute the intended result. One example is computing the number of user logins in the last 7 days. This creates a dilemma where the results of the stream program are incomplete until the runtime of the program exceeds 7 days. The alternative is to bootstrap the program using historic data to seed the state before shifting to use real-time data.
This talk will discuss alternatives to bootstrap programs in Flink. Some alternatives rely on technologies exogenous to the stream program, such as enhancements to the pub/sub layer, that are more generally applicable to other stream compute engines. Other alternatives include enhancements to Flink source implementations. Lyft is exploring another alternative using orchestration of multiple Flink programs. The talk will cover why Lyft pursued this alternative and future directions to further enhance bootstrapping support in Flink.
Speaker
Gregory Fee, Principal Engineer, Lyft
Large Table Partitioning with PostgreSQL and DjangoEDB
With great DB Table comes great responsibility". Our email messages table was growing too much and we needed to do something about it. We will talk about how we integrated PostgreSQL Declarative partitioning with our Django based Customer Portal to solve the problem.
Join Marc Linster andDave Page as they shows you how to manage large scale Postgres deployments. This webinar will highlight how EnterpriseDB Postgres Enterprise Manager can be used for monitoring, alerting and administration of your Postgres estate - no matter where it is deployed.
Highlights include:
Tools and techniques for Postgres management
Key database performance metrics and monitoring
Operating Postgres on premises, in containers, and in the cloud
They will conclude with a live demo of Postgres Enterprise Manager.
SpringCamp 2016 - Apache Geode 와 Spring Data GemfireJay Lee
The document discusses Apache Geode and Spring integration. It provides an overview of Apache Geode, an open source distributed in-memory database. It then covers Spring Data Gemfire, which allows using Geode with Spring's programming model. It also discusses using Spring Session to manage user sessions in a stateless manner by storing them in Geode. The presentation includes demos of integrating Geode with Spring applications.
A webinarban megtudhatják milyen kihívásokkal kell szembenézni Oracle adatbázis PostgreSQL-re migrálása során. Bemutatjuk az utóbbi két év nagy komplexitású Oracle kompatibilitási vizsgálatainak tapasztalatait, beleértve az idei évben az EDB migrációs portálján keresztül vizsgált több mint 2 200 000 Oracle DDL konstrukciót.
Az előadás alatt kitérünk az alábbiakra:
- Tárhely (storage) definiciók
- Csomagok
- Tárolt eljárások
- PL/SQL kód
- Gyártói adatbázis API-k
- Komplex adatbázis migrációk
Előadásunkat olyan migrációs eszközök bemutatásával zárjuk, amelyek jelentősen leegyszerűsítik az Oracle-PostgeSQL migrációt és csökkentik annak kockázatait.
EDB 13 - New Enhancements for Security and Usability - APJEDB
Database security is always of paramount importance to all organizations. In this webinar, we will explore the security, usability, and portability updates of the latest version of the EDB database server and tools.
Join us in this webinar to learn:
- The new security features such as SCRAM and the encryption of database passwords and traffic between Failover Manager agents
- Usability updates that automate partitioning, verify backup integrity, and streamline the management of failover and backups
- Portability improvements that simplify running PostgreSQL across on-premise and cloud environments
Why you should use native packages to install PostgreSQL on LinuxEDB
When it comes to installing PostgreSQL on a Linux distribution, using a native package is considered to be a better choice compared to installing from the source. In this webinar, we will discuss some benefits of using native packages to install PostgreSQL.
Optimizing Performance: Installers are shipped with “the lowest common denominator” and native packages use the OS libraries resulting in performance and maintenance benefits.
Security: Native packages depend on the OS, thus updating OS packages are sufficient for security issues.
Usability: Native packages integrate with the OS packaging system (dnf, yum, apt, etc), increasing usability.
Present & Future of Greenplum Database A massively parallel Postgres Database...VMware Tanzu
Greenplum Database is Pivotal's massively parallel Postgres database. Version 5 has proven features for mission critical use cases. Version 6 adds improvements like row-level locking, foreign data wrappers, and online expansion to make Greenplum a superset of Postgres. It also provides up to 50x faster OLTP performance. Version 7 will focus on capabilities beyond the cluster like streaming replication and using Greenplum as a source for data integration tools.
Introducing Apache Geode and Spring Data GemFireJohn Blum
This document introduces Apache Geode, an open source distributed in-memory data management platform. It discusses what Geode is, how it is implemented, and some key features like high availability, scalability and low latency. It also introduces Spring Data GemFire, which simplifies using Geode with Spring applications through features like repositories and caching. Finally, it outlines the project roadmap and opportunities to get involved in the Geode community.
EnterpriseDB Postgres Plus Advanced Server provides Oracle compatibility with enterprise performance features built upon the legendary open source PostgreSQL platform, all certified on IBM’s latest Linux on Power servers.
The highlights of this presentation include:
* An overview of the database landscape – past, present and future
* Postgres NoSQL capabilities for document and key-value store work loads
* How you can lower your Total-Cost-of-Ownership (TCO) with Postgres in conjunction with your current database
* What resources are available to assess the right decision
* How the IBM Power Systems™ platform is fueling performance, reliability, security, TCO and virtualization for new applications, markets and geographies.
* Suggested audience: This presentation is intended for strategic IT and Business Decision-Makers involved in IT infrastructure and application development.
Healthcare Claim Reimbursement using Apache SparkDatabricks
The document discusses rewriting a claims reimbursement system using Spark. It describes how Spark provides better performance, scalability and cost savings compared to the previous Oracle-based system. Key points include using Spark for ETL to load data into a Delta Lake data lake, implementing the business logic in a reusable Java library, and seeing significant increases in processing volumes and speeds compared to the prior system. Challenges and tips for adoption are also provided.
The Real Scoop on Migrating from Oracle DatabasesEDB
During this presentation you will be provided with actionable guidelines to:
• Identify the right applications to migrate
• Easily and safely migrate your applications
• Leverage resources before, during and after your migration
• Learn how to achieve independence from Oracle databases - without sacrificing performance.
Oracle Database 19c - poslední z rodiny 12.2 a co přináší novéhoMarketingArrowECS_CZ
The document provides an overview of Oracle Database 19c, highlighting its key features and capabilities. It notes that Oracle Database 19c is Oracle's recommended release for all database upgrades. New features in 19c include fast data ingestion support for IoT workloads, SQL statement quarantine, and enhancements to JSON and high availability functionality.
The document discusses the benefits and challenges of running big data workloads on cloud native platforms. Some key points discussed include:
- Big data workloads are migrating to the cloud to take advantage of scalability, flexibility and cost effectiveness compared to on-premises solutions.
- Enterprise cloud platforms need to provide centralized management and monitoring of multiple clusters, secure data access, and replication capabilities.
- Running big data on cloud introduces challenges around storage, networking, compute resources, and security that systems need to address, such as consistency issues with object storage, network throughput reductions, and hardware variations across cloud vendors.
- The open source community is helping users address these challenges to build cloud native data architectures
A5 oracle exadata-the game changer for online transaction processing data w...Dr. Wilfred Lin (Ph.D.)
The document discusses Oracle Exadata and how it can transform online transaction processing, data warehousing, and database consolidation. It describes Exadata as a scale-out platform that integrates servers, storage, and networking optimized for Oracle Database. Exadata delivers extreme performance through special software that brings database intelligence to storage, flash, and networking. It is suitable for all database workloads including OLTP, data warehousing, and database clouds.
Recent advances in Postgres have propelled the database forward to meet today’s data challenges. At some of the world’s largest companies, Postgres plays a major role in controlling costs and reducing dependence on traditional providers.
This presentation addresses:
* What workloads are best suited for introducing Postgres into your environment
* The success milestones for evaluating the ‘when and how’ of expanding Postgres deployments
* Key advances in recent Postgres releases that support new data types and evolving data challenges
This presentation is intended for strategic IT and Business Decision-Makers involved in data infrastructure decisions and cost-savings.
The document summarizes Oracle's SuperCluster engineered system. It provides consolidated application and database deployment with in-memory performance. Key features include Exadata intelligent storage, Oracle M6 and T5 servers, a high-speed InfiniBand network, and Oracle VM virtualization. The SuperCluster enables database as a service with automated provisioning and security for multi-tenant deployment across industries.
The document outlines an Exadata implementation strategy with the following key steps:
1. Plan the migration by analyzing the current architecture, making deployment decisions, and considering migration and high availability.
2. Optimize the database on Exadata by using features like smart scan, partitioning, compression, and parallelism.
3. Migrate the database using various physical and logical migration methods.
4. Test the optimized database configuration on Exadata and capture performance statistics.
This document summarizes a presentation on tuning Oracle GoldenGate for optimal performance in real-world environments. It discusses architectural changes in GoldenGate 12c including a microservices architecture and parallel replication. It also outlines several areas and tools for tuning performance at the host, database, and GoldenGate configuration levels including the use of AWR, STATS commands, and health check scripts.
[db tech showcase Tokyo 2018] #dbts2018 #B31 『1,2,3 and Done! 3 easy ways to ...Insight Technology, Inc.
[db tech showcase Tokyo 2018] #dbts2018 #B31
『1,2,3 and Done! 3 easy ways to migrate to the cloud!』
Data Intensity - Director of Innovation Francisco Munoz Alvarez 氏
This document summarizes a presentation about new features and changes coming in MySQL 5.7. Key points include: MySQL 5.7 will include performance improvements, more robust transaction handling and memory instrumentation. However, some backwards incompatible changes will be needed to improve the architecture. The presentation outlines several proposed changes, such as making replication more durable by default and changing the default SQL mode to STRICT. It also discusses new features for InnoDB and the optimizer.
The Oracle GoldenGate software package delivers low-impact, real-time data integration and transactional data replication across heterogeneous systems for continuous availability, zero-downtime migration, and business intelligence.
Join the Webinar to learn Golden Gate 12c New Features
• Expanded heterogeneous Support
• Multitenant Container Database (CDB) Support
• Oracle Universal Installer (OUI) Support
• Support for Public and Private Clouds
• Integrated Replicat
• Security
• Coordinated Replicat
• New 32K VARCHAR2 Support
• High Availability (HA) enhancements
• Support for Other Oracle products
• Improvements to feature Functionality
This document provides a summary of Chaitanya Prati's work experience and qualifications. He has over 10 years of experience as an Oracle DBA providing support for multi-terabyte Oracle databases. Currently he works as an onsite technical lead for Wipro Technologies providing Oracle database administration support to Citigroup. His responsibilities include managing critical compliance applications, implementing GoldenGate replication, and resolving performance issues. He is proficient in technologies like Oracle RAC, ASM, GoldenGate and tools like SQL*Plus and Toad.
Benchmarking for postgresql workloads in kubernetesDoKC
ABSTRACT OF THE TALK
6 months have passed since our last DoK webinar about benchmarking PostgreSQL workloads in a Kubernetes environment. In the meantime, many things have happened at EDB, and we’re happy to share what we’ve learned in this timeframe. We’ll use cnp-bench and cnp-sandbox to help us describe some of the challenges we might face when running PostgreSQL workloads, how to spot them, and what actions to take to make your databases healthier and more longeve.
cnp-bench is a collection of Helm charts that help run storage and database benchmarks, using popular open source tools like fio, pgbench, and HammerDB. cnp-sandbox is a Helm chart that sets up a Prometheus/Grafana stack, including basic metrics and dashboards for Cloud Native PostgreSQL, the Kubernetes operator developed by EDB. Both cnp-sandbox and cnp-bench are open source and recommended for development, testing, and pre-production environments only.
BIO
A long time open-source programmer and entrepreneur, Gabriele has a degree in Statistics from the University of Florence. After having consistently contributed to the growth of 2ndQuadrant and its members through nurturing a lean and devops culture, he is now leading the Cloud Native initiative at EDB. Gabriele lives in Prato, a small but vibrant city located in the northern part of Tuscany, Italy - famous for having hosted the first European PostgreSQL conferences. His second home is Melbourne, Australia, where he studied at Monash University and worked in the ICT sector. He loves playing the Blues with his Fender Stratocaster, but his major passions are called Elisabeth and Charlotte!
KEY TAKE-AWAYS FROM THE TALK
- A methodology for benchmarking a PostgreSQL database in Kubernetes
- Open source set of tools for benchmarking a PostgreSQL database in Kubernetes
- Reasons why benchmarking both the storage and the database is important
https://github.com/EnterpriseDB/cnp-sandbox
https://github.com/EnterpriseDB/cnp-bench
Oracle RAC provides high availability, scalability and performance for databases across clustered servers with no application changes required. It uses a shared cache architecture to overcome limitations of traditional shared-nothing and shared-disk approaches. iONE provides Oracle RAC implementation and maintenance services to deliver continuous uptime for database applications through server pool management, datacenter HA, and scaling to 100 nodes.
An overview of reference architectures for PostgresEDB
EDB Reference Architectures are designed to help new and existing users alike to quickly design a deployment architecture that suits their needs. They can be used as either the blueprint for a deployment, or as the basis for a design that enhances and extends the functionality and features offered.
Add-on architectures allow users to easily extend their core database server deployment to add additional features and functionality "building block" style.
In this webinar, we will review the following architectures:
- Single Node
- Multi Node with Asynchronous Replication
- Multi Node with Synchronous Replication
- Add-on Architectures
Speaker:
Michael Willer
Sales Engineer, EDB
State of Artificial intelligence Report 2023kuntobimo2016
Artificial intelligence (AI) is a multidisciplinary field of science and engineering whose goal is to create intelligent machines.
We believe that AI will be a force multiplier on technological progress in our increasingly digital, data-driven world. This is because everything around us today, ranging from culture to consumer products, is a product of intelligence.
The State of AI Report is now in its sixth year. Consider this report as a compilation of the most interesting things we’ve seen with a goal of triggering an informed conversation about the state of AI and its implication for the future.
We consider the following key dimensions in our report:
Research: Technology breakthroughs and their capabilities.
Industry: Areas of commercial application for AI and its business impact.
Politics: Regulation of AI, its economic implications and the evolving geopolitics of AI.
Safety: Identifying and mitigating catastrophic risks that highly-capable future AI systems could pose to us.
Predictions: What we believe will happen in the next 12 months and a 2022 performance review to keep us honest.
Global Situational Awareness of A.I. and where its headedvikram sood
You can see the future first in San Francisco.
Over the past year, the talk of the town has shifted from $10 billion compute clusters to $100 billion clusters to trillion-dollar clusters. Every six months another zero is added to the boardroom plans. Behind the scenes, there’s a fierce scramble to secure every power contract still available for the rest of the decade, every voltage transformer that can possibly be procured. American big business is gearing up to pour trillions of dollars into a long-unseen mobilization of American industrial might. By the end of the decade, American electricity production will have grown tens of percent; from the shale fields of Pennsylvania to the solar farms of Nevada, hundreds of millions of GPUs will hum.
The AGI race has begun. We are building machines that can think and reason. By 2025/26, these machines will outpace college graduates. By the end of the decade, they will be smarter than you or I; we will have superintelligence, in the true sense of the word. Along the way, national security forces not seen in half a century will be un-leashed, and before long, The Project will be on. If we’re lucky, we’ll be in an all-out race with the CCP; if we’re unlucky, an all-out war.
Everyone is now talking about AI, but few have the faintest glimmer of what is about to hit them. Nvidia analysts still think 2024 might be close to the peak. Mainstream pundits are stuck on the wilful blindness of “it’s just predicting the next word”. They see only hype and business-as-usual; at most they entertain another internet-scale technological change.
Before long, the world will wake up. But right now, there are perhaps a few hundred people, most of them in San Francisco and the AI labs, that have situational awareness. Through whatever peculiar forces of fate, I have found myself amongst them. A few years ago, these people were derided as crazy—but they trusted the trendlines, which allowed them to correctly predict the AI advances of the past few years. Whether these people are also right about the next few years remains to be seen. But these are very smart people—the smartest people I have ever met—and they are the ones building this technology. Perhaps they will be an odd footnote in history, or perhaps they will go down in history like Szilard and Oppenheimer and Teller. If they are seeing the future even close to correctly, we are in for a wild ride.
Let me tell you what we see.
The Building Blocks of QuestDB, a Time Series Databasejavier ramirez
Talk Delivered at Valencia Codes Meetup 2024-06.
Traditionally, databases have treated timestamps just as another data type. However, when performing real-time analytics, timestamps should be first class citizens and we need rich time semantics to get the most out of our data. We also need to deal with ever growing datasets while keeping performant, which is as fun as it sounds.
It is no wonder time-series databases are now more popular than ever before. Join me in this session to learn about the internal architecture and building blocks of QuestDB, an open source time-series database designed for speed. We will also review a history of some of the changes we have gone over the past two years to deal with late and unordered data, non-blocking writes, read-replicas, or faster batch ingestion.
Predictably Improve Your B2B Tech Company's Performance by Leveraging DataKiwi Creative
Harness the power of AI-backed reports, benchmarking and data analysis to predict trends and detect anomalies in your marketing efforts.
Peter Caputa, CEO at Databox, reveals how you can discover the strategies and tools to increase your growth rate (and margins!).
From metrics to track to data habits to pick up, enhance your reporting for powerful insights to improve your B2B tech company's marketing.
- - -
This is the webinar recording from the June 2024 HubSpot User Group (HUG) for B2B Technology USA.
Watch the video recording at https://youtu.be/5vjwGfPN9lw
Sign up for future HUG events at https://events.hubspot.com/b2b-technology-usa/
Enhanced Enterprise Intelligence with your personal AI Data Copilot.pdfGetInData
Recently we have observed the rise of open-source Large Language Models (LLMs) that are community-driven or developed by the AI market leaders, such as Meta (Llama3), Databricks (DBRX) and Snowflake (Arctic). On the other hand, there is a growth in interest in specialized, carefully fine-tuned yet relatively small models that can efficiently assist programmers in day-to-day tasks. Finally, Retrieval-Augmented Generation (RAG) architectures have gained a lot of traction as the preferred approach for LLMs context and prompt augmentation for building conversational SQL data copilots, code copilots and chatbots.
In this presentation, we will show how we built upon these three concepts a robust Data Copilot that can help to democratize access to company data assets and boost performance of everyone working with data platforms.
Why do we need yet another (open-source ) Copilot?
How can we build one?
Architecture and evaluation
06-04-2024 - NYC Tech Week - Discussion on Vector Databases, Unstructured Data and AI
Discussion on Vector Databases, Unstructured Data and AI
https://www.meetup.com/unstructured-data-meetup-new-york/
This meetup is for people working in unstructured data. Speakers will come present about related topics such as vector databases, LLMs, and managing data at scale. The intended audience of this group includes roles like machine learning engineers, data scientists, data engineers, software engineers, and PMs.This meetup was formerly Milvus Meetup, and is sponsored by Zilliz maintainers of Milvus.
Beyond the Basics of A/B Tests: Highly Innovative Experimentation Tactics You...Aggregage
This webinar will explore cutting-edge, less familiar but powerful experimentation methodologies which address well-known limitations of standard A/B Testing. Designed for data and product leaders, this session aims to inspire the embrace of innovative approaches and provide insights into the frontiers of experimentation!
The Ipsos - AI - Monitor 2024 Report.pdfSocial Samosa
According to Ipsos AI Monitor's 2024 report, 65% Indians said that products and services using AI have profoundly changed their daily life in the past 3-5 years.
4th Modern Marketing Reckoner by MMA Global India & Group M: 60+ experts on W...Social Samosa
The Modern Marketing Reckoner (MMR) is a comprehensive resource packed with POVs from 60+ industry leaders on how AI is transforming the 4 key pillars of marketing – product, place, price and promotions.