Which Change Data Capture Strategy is Right for You?Precisely
Change Data Capture or CDC is the practice of moving the changes made in an important transactional system to other systems, so that data is kept current and consistent across the enterprise. CDC keeps reporting and analytic systems working on the latest, most accurate data.
Many different CDC strategies exist. Each strategy has advantages and disadvantages. Some put an undue burden on the source database. They can cause queries or applications to become slow or even fail. Some bog down network bandwidth, or have big delays between change and replication.
Each business process has different requirements, as well. For some business needs, a replication delay of more than a second is too long. For others, a delay of less than 24 hours is excellent.
Which CDC strategy will match your business needs? How do you choose?
View this webcast on-demand to learn:
• Advantages and disadvantages of different CDC methods
• The replication latency your project requires
• How to keep data current in Big Data technologies like Hadoop
This document discusses IBM's Integrated Analytics System (IIAS), which is a next generation hybrid data warehouse appliance. Some key points:
- IIAS provides high performance analytics capabilities along with data warehousing and management functions.
- It utilizes a common SQL engine to allow workloads and skills to be portable across public/private clouds and on-premises.
- The system is designed for flexibility with the ability to independently scale compute and storage capacity. It also supports a variety of workloads including reporting, analytics, and operational analytics.
- IBM is positioning IIAS to address top customer requirements around broader workloads, higher concurrency, in-place expansion, and availability solutions.
Michael Howard, CEO of MariaDB, presented on the company's usage and growth. MariaDB has become a top open-source database according to rankings, has received $54M in funding, and aims to differentiate itself from Oracle through compatibility with PL/SQL and temporal queries while also focusing on machine learning, distributed computing, and new technologies through its MariaDB Labs division.
Move to Hadoop, Go Faster and Save Millions - Mainframe Legacy ModernizationDataWorks Summit
In spite of recent advances in computing, many core business processes are batch-oriented running on Mainframes. Annual Mainframe costs are counted in 6+ figure Dollars per year, potentially growing with capacity needs. In order to tackle the cost challenge, many organizations have considered or attempted multi-year mainframe migration/re-hosting strategies. Traditional approaches to Mainframe elimination call for large initial investments and carry significant risks – It is hard to match Mainframe performance and reliability. Using Hadoop, Sears/MetaScale developed an innovative alternative that enables batch processing migration to Hadoop, without the risks, time and costs of other methods. This solution has been adopted in multiple businesses with excellent results and associated cost savings, as Mainframes are physically eliminated or downsized: Millions of dollars in savings based on MIP reductions have been seen – A reduction of 200 MIPS can yield $1 million in annual savings. MetaScale eliminated over 900 MIPs and an entire Mainframe system for one fortune 500 client. This presentation illustrates reference architecture and approach successfully used by MetaScale to move mainframe processing to the Hadoop platform without altering user-facing business applications.
EDBT 2013 - Near Realtime Analytics with IBM DB2 Analytics AcceleratorDaniel Martin
The document discusses IBM's DB2 Analytics Accelerator (IDAA) which uses incremental updates to synchronize data between DB2 and the IDAA appliance in near real-time. It describes the architecture of using log-based capture and propagation to minimize latency. The user interface allows controlling replication at the subsystem and table level. High availability is ensured through failover capabilities. Tuning options and evaluation of query impact are also covered.
Einführung: MariaDB heute und unsere Vision für die ZukunftMariaDB plc
MariaDB is presenting at a roadshow in Hannover on building an easy to use, extendable, and deployable database. They highlight how MariaDB reduces costs compared to Oracle and is the default database on major Linux distributions and cloud platforms. MariaDB has an extensible architecture and over 40 plugins contributed by an active community to provide features like analytics, search, and replication.
Caching for Microservices Architectures: Session IVMware Tanzu
This document discusses how caching can help address performance, scalability, and autonomy challenges for microservices architectures. It introduces Pivotal Cloud Cache (PCC) as a caching solution for microservices on Pivotal Cloud Foundry. PCC provides an in-memory cache that can scale horizontally and increase performance. It also allows for data autonomy between microservices and teams while providing high availability. PCC offers an easy and cost-effective way to cache data and adopt microservices on Pivotal Cloud Foundry.
This document discusses IBM's DB2 tools and solutions including the DB2 Performance Solution Pack, DB2 Utilities Solution Pack, IBM DB2 Analytics Accelerator (IDAA), and QMF for z/OS. It provides an overview of each solution's components and capabilities for optimizing DB2 performance, managing utilities, identifying accelerated queries, and workload analysis. The document also demonstrates how IBM tools like Query Monitor can identify eligible queries for acceleration with IDAA and quantify the potential CPU savings.
Which Change Data Capture Strategy is Right for You?Precisely
Change Data Capture or CDC is the practice of moving the changes made in an important transactional system to other systems, so that data is kept current and consistent across the enterprise. CDC keeps reporting and analytic systems working on the latest, most accurate data.
Many different CDC strategies exist. Each strategy has advantages and disadvantages. Some put an undue burden on the source database. They can cause queries or applications to become slow or even fail. Some bog down network bandwidth, or have big delays between change and replication.
Each business process has different requirements, as well. For some business needs, a replication delay of more than a second is too long. For others, a delay of less than 24 hours is excellent.
Which CDC strategy will match your business needs? How do you choose?
View this webcast on-demand to learn:
• Advantages and disadvantages of different CDC methods
• The replication latency your project requires
• How to keep data current in Big Data technologies like Hadoop
This document discusses IBM's Integrated Analytics System (IIAS), which is a next generation hybrid data warehouse appliance. Some key points:
- IIAS provides high performance analytics capabilities along with data warehousing and management functions.
- It utilizes a common SQL engine to allow workloads and skills to be portable across public/private clouds and on-premises.
- The system is designed for flexibility with the ability to independently scale compute and storage capacity. It also supports a variety of workloads including reporting, analytics, and operational analytics.
- IBM is positioning IIAS to address top customer requirements around broader workloads, higher concurrency, in-place expansion, and availability solutions.
Michael Howard, CEO of MariaDB, presented on the company's usage and growth. MariaDB has become a top open-source database according to rankings, has received $54M in funding, and aims to differentiate itself from Oracle through compatibility with PL/SQL and temporal queries while also focusing on machine learning, distributed computing, and new technologies through its MariaDB Labs division.
Move to Hadoop, Go Faster and Save Millions - Mainframe Legacy ModernizationDataWorks Summit
In spite of recent advances in computing, many core business processes are batch-oriented running on Mainframes. Annual Mainframe costs are counted in 6+ figure Dollars per year, potentially growing with capacity needs. In order to tackle the cost challenge, many organizations have considered or attempted multi-year mainframe migration/re-hosting strategies. Traditional approaches to Mainframe elimination call for large initial investments and carry significant risks – It is hard to match Mainframe performance and reliability. Using Hadoop, Sears/MetaScale developed an innovative alternative that enables batch processing migration to Hadoop, without the risks, time and costs of other methods. This solution has been adopted in multiple businesses with excellent results and associated cost savings, as Mainframes are physically eliminated or downsized: Millions of dollars in savings based on MIP reductions have been seen – A reduction of 200 MIPS can yield $1 million in annual savings. MetaScale eliminated over 900 MIPs and an entire Mainframe system for one fortune 500 client. This presentation illustrates reference architecture and approach successfully used by MetaScale to move mainframe processing to the Hadoop platform without altering user-facing business applications.
EDBT 2013 - Near Realtime Analytics with IBM DB2 Analytics AcceleratorDaniel Martin
The document discusses IBM's DB2 Analytics Accelerator (IDAA) which uses incremental updates to synchronize data between DB2 and the IDAA appliance in near real-time. It describes the architecture of using log-based capture and propagation to minimize latency. The user interface allows controlling replication at the subsystem and table level. High availability is ensured through failover capabilities. Tuning options and evaluation of query impact are also covered.
Einführung: MariaDB heute und unsere Vision für die ZukunftMariaDB plc
MariaDB is presenting at a roadshow in Hannover on building an easy to use, extendable, and deployable database. They highlight how MariaDB reduces costs compared to Oracle and is the default database on major Linux distributions and cloud platforms. MariaDB has an extensible architecture and over 40 plugins contributed by an active community to provide features like analytics, search, and replication.
Caching for Microservices Architectures: Session IVMware Tanzu
This document discusses how caching can help address performance, scalability, and autonomy challenges for microservices architectures. It introduces Pivotal Cloud Cache (PCC) as a caching solution for microservices on Pivotal Cloud Foundry. PCC provides an in-memory cache that can scale horizontally and increase performance. It also allows for data autonomy between microservices and teams while providing high availability. PCC offers an easy and cost-effective way to cache data and adopt microservices on Pivotal Cloud Foundry.
This document discusses IBM's DB2 tools and solutions including the DB2 Performance Solution Pack, DB2 Utilities Solution Pack, IBM DB2 Analytics Accelerator (IDAA), and QMF for z/OS. It provides an overview of each solution's components and capabilities for optimizing DB2 performance, managing utilities, identifying accelerated queries, and workload analysis. The document also demonstrates how IBM tools like Query Monitor can identify eligible queries for acceleration with IDAA and quantify the potential CPU savings.
This document provides an overview of MariaDB's 2017 roadshow, including what they are doing, where they are going, and who the field CTO is. It discusses trends in the database market moving away from expensive proprietary databases toward lower-cost open source options with subscriptions and community involvement. It highlights cost savings of MariaDB compared to Oracle and MariaDB's extensible architecture and community contributions. It also summarizes MariaDB products and technologies like the database server, MaxScale proxy, and ColumnStore, as well as MariaDB's customers, use cases, services, and how to get started with MariaDB.
Caching for Microservices Architectures: Session II - Caching PatternsVMware Tanzu
In the first webinar of the series we covered the importance of caching in microservice-based application architectures—in addition to improving performance it also aids in making content available from legacy systems, promotes loose coupling and team autonomy, and provides air gaps that can limit failures from cascading through a system.
To reap these benefits, though, the right caching patterns must be employed. In this webinar, we will examine various caching patterns and shed light on how they deliver the capabilities needed by our microservices. What about rapidly changing data, and concurrent updates to data? What impact do these and other factors have to various use cases and patterns?
Understanding data access patterns, covered in this webinar, will help you make the right decisions for each use case. Beyond the simplest of use cases, caching can be tricky business—join us for this webinar to see how best to use them.
Jagdish Mirani, Cornelia Davis, Michael Stolz, Pulkit Chandra, Pivotal
Development of concurrent services using In-Memory Data Gridsjlorenzocima
As part of OTN Tour 2014 believes this presentation which is intented for covers the basic explanation of a solution of IMDG, explains how it works and how it can be used within an architecture and shows some use cases. Enjoy
How to Manage Scale-Out Environments with MariaDB MaxScaleMariaDB plc
MariaDB MaxScale is a database proxy that provides high availability, scalability, and security for MariaDB and MySQL database infrastructures. It implements read/write splitting to route read queries to slave servers and write queries to the master server. The document provides instructions on installing and configuring MariaDB MaxScale, including creating a service for read/write splitting, defining servers, adding authentication, and testing the split routing functionality.
Maximizing performance via tuning and optimizationMariaDB plc
This document provides an overview of best practices for maximizing performance of MariaDB Server through tuning and optimization. It discusses general best practices like service level agreements and metrics collection. It also covers specific areas like server, storage, and network configuration, connection pooling, MariaDB configuration settings, query tuning using indexes and EXPLAIN, and monitoring tools like performance schema. The goal is to help users get the most out of their MariaDB deployment through performance analysis and tuning.
How Pixid dropped Oracle and went hybrid with MariaDBMariaDB plc
Pixid replaced Oracle Database with MySQL in 2011, then soon migrated to MariaDB to get better performance, more features and synchronous clustering for high availability. In addition to high-performance transactions, their customers needed access to fast analytics for self-service reporting and data exploration. Pixid started with a separate columnar database for analytics, but with the release of MariaDB ColumnStore, they found a more elegant solution – deploying a single database platform to handle both transactions and analytics. In this session, Antoine Gosset and Jérôme Mouret share how Pixid went from Oracle Database to handling both transactional and analytical workloads with MariaDB.
The talk will be about the project to find a replacement for all IBM products in the company with the example for the databases. What was the goal of the project, the learning, a short overview about the options
we migrated about 500 db2 databases to EnterpriseDB. The database size was from a small size up to 4 TB and we implemented a completely new fully automated deployment of VM and database. Databases are now 11 month in production. The talk will have an overview of the project, the learnings, a few parameters and technical parameters that were found for stability and performance.
MariaDB AX is engineered to simplify the process of ingesting and analyzing data, removing the need for complex, time-consuming batch processes or data models constrained to, or optimized for, a limited set of queries – modern requirements when predictive/prescriptive analytics and time-to-insight become competitive differentiators.
MariaDB AX is now based on MariaDB ColumnStore 1.1, a distributed and columnar storage engine for high-performance analytics at scale, and introduces simplified data ingestion via data adapters, semi-structured/unstructured data analysis via text and binary columns, and custom analytical functions via a user-defined aggregate/window API.
In this webinar, Dipti Joshi, MariaDB’s Director of Product Management, will provide an overview of MariaDB AX, highlight innovative customer use cases, and provide a detailed explanation of new features and the benefits they provide.
How Orwell built a geo-distributed Bank-as-a-Service with microservicesMariaDB plc
Orwell Group shares how they leveraged microservices, an event driven architecture and both master and reference data management methodologies to build a new banking system for high retail banking customers and corporate banks requiring cross border payments and cash flow management – and scaled it to handle customers with millions of clients. In particular they explain how they built a high availability, geo-distributed and consistent platform on top of MariaDB. The result was a secure and distributed platform with high cost efficiency, and the data accuracy and consistency needed to create high quality data pipelines from transactions to analytics and ensure regulatory compliance (e.g., GDPR).
IBM World of Watson 2016 - DB2 Analytics Accelerator on CloudDaniel Martin
IBM is introducing a new deployment option for the DB2 Analytics Accelerator on Cloud using dashDB as the acceleration engine. This provides customers with a hybrid cloud offering that gives the flexibility of running the Accelerator either on-premises or in the cloud. The Cloud deployment offers benefits like monthly pricing, hardware provisioning by IBM, and fast provisioning time. Initial focus areas include basic Accelerator functionality for offloading queries to the cloud, with a roadmap to continuously expand features and functionality.
Transform DBMS to Drive Apps of Engagement InnovationEDB
IT leaders face ever-increasing challenges to fund and deliver innovative business solutions for better customer and stakeholder engagement. While much of the infrastructure stack has been commoditized, the DBMS remains an expensive and growing drain on IT resources that otherwise could drive innovation.
This presentation outlines how money can be freed up in IT from expensive database spend by transforming applications and the DBMS to a subscription-based, cloud ready EnterpriseDB Postgres offering.
Further, learn how the same DBMS used for these traditional workloads with familiar tools and a large skill base can also be used for new applications of engagement, which rely on a wider variety of data including NoSQL, semi-structured along with relational data.
Visit EnterpriseDB > Resources > Webcasts to listen to the presentation recording.
This presentation is intended for strategic IT and business decision-makers involved in data infrastructure decisions and cost savings.
Temporal Tables, Transparent Archiving in DB2 for z/OS and IDAACuneyt Goksu
The document discusses several data archiving solutions for z/OS systems including temporal tables, transparent archiving, and IDAA technology. Temporal tables allow querying and updating historical data using system time periods. Transparent archiving moves old data to other storage platforms while still allowing dynamic queries. IDAA provides accelerated query performance for temporal tables by routing queries to an accelerator system. The solutions can be combined for different use cases depending on data retention and access needs.
Open innovation and collaboration between IBM and other technology companies is fueling advances in cloud computing, big data analytics, and software development. This includes contributions to open source projects like Linux as well as partnerships through organizations like the OpenPOWER Foundation. New systems based on IBM's Power architecture and optimized for Linux are helping customers improve the performance and efficiency of their analytics, database, and application workloads.
- Centering your enterprise for growth discusses recent innovations with IMS that improve performance, affordability, and simplicity. These enhancements help clients optimize and innovate with IMS to support business growth in an era of data, cloud, and mobile engagement.
- Key highlights include a 117,292 TPS and 130% increased workload throughput on z13, as well as pricing innovations like IMS Value Unit Editions to enable new analytics and mobile workloads.
- The document emphasizes how optimizing IMS can help clients accelerate insights from transaction data, rapidly enable cloud and mobile access, and expand the IMS talent population to continue delivering value.
Big Data Analytics with MariaDB ColumnStoreMariaDB plc
MariaDB ColumnStore is a massively parallel columnar storage engine for MariaDB that provides high performance analytics on large datasets. It uses a distributed columnar architecture where each column is stored separately and data is partitioned horizontally across nodes. This allows for very fast analytical queries by only accessing the relevant columns and partitions. Some key features include built-in analytics functions, high speed data ingestion, and support for running on-premises or on cloud platforms like AWS. The latest 1.1 version adds capabilities like streaming data ingestion APIs, improved high availability with GlusterFS, and performance optimizations.
Better performance and cost effectiveness empower better results in the cognitive era. For more information, visit: http://www.ibm.com/systems/power/hardware/linux-lc.html
Join Postgres experts Bruce Momjian and Marc Linster as they preview everything new in Postgres 12. You don’t want to miss this!
Highlights include:
- New compatibility features
- PostgreSQL: Table access methods
- Partitioning Improvements
GemFire Enterprise Data Fabric provides a scalable and high-performance data infrastructure that caches and distributes data across multiple formats and distributed applications. It connects to backend systems, provides high data availability, supports distributed transactions and querying. GemFire solves challenges in SOA and distributed systems by enhancing scalability, performance and reliability while potentially reducing hardware and software costs.
Intel's Data Center & Connected Systems Group and Diane Bryant shares the latest news on the latest Intel Xeon E5v2 family of processors and technologies like Intel Network Builders to enable the re-architecture of the Data Center.
The document discusses accelerating Ceph storage performance using SPDK. SPDK introduces optimizations like asynchronous APIs, userspace I/O stacks, and polling mode drivers to reduce software overhead and better utilize fast storage devices. This allows Ceph to better support high performance networks and storage like NVMe SSDs. The document provides an example where SPDK helped XSKY's BlueStore object store achieve significant performance gains over the standard Ceph implementation.
This document provides an overview of MariaDB's 2017 roadshow, including what they are doing, where they are going, and who the field CTO is. It discusses trends in the database market moving away from expensive proprietary databases toward lower-cost open source options with subscriptions and community involvement. It highlights cost savings of MariaDB compared to Oracle and MariaDB's extensible architecture and community contributions. It also summarizes MariaDB products and technologies like the database server, MaxScale proxy, and ColumnStore, as well as MariaDB's customers, use cases, services, and how to get started with MariaDB.
Caching for Microservices Architectures: Session II - Caching PatternsVMware Tanzu
In the first webinar of the series we covered the importance of caching in microservice-based application architectures—in addition to improving performance it also aids in making content available from legacy systems, promotes loose coupling and team autonomy, and provides air gaps that can limit failures from cascading through a system.
To reap these benefits, though, the right caching patterns must be employed. In this webinar, we will examine various caching patterns and shed light on how they deliver the capabilities needed by our microservices. What about rapidly changing data, and concurrent updates to data? What impact do these and other factors have to various use cases and patterns?
Understanding data access patterns, covered in this webinar, will help you make the right decisions for each use case. Beyond the simplest of use cases, caching can be tricky business—join us for this webinar to see how best to use them.
Jagdish Mirani, Cornelia Davis, Michael Stolz, Pulkit Chandra, Pivotal
Development of concurrent services using In-Memory Data Gridsjlorenzocima
As part of OTN Tour 2014 believes this presentation which is intented for covers the basic explanation of a solution of IMDG, explains how it works and how it can be used within an architecture and shows some use cases. Enjoy
How to Manage Scale-Out Environments with MariaDB MaxScaleMariaDB plc
MariaDB MaxScale is a database proxy that provides high availability, scalability, and security for MariaDB and MySQL database infrastructures. It implements read/write splitting to route read queries to slave servers and write queries to the master server. The document provides instructions on installing and configuring MariaDB MaxScale, including creating a service for read/write splitting, defining servers, adding authentication, and testing the split routing functionality.
Maximizing performance via tuning and optimizationMariaDB plc
This document provides an overview of best practices for maximizing performance of MariaDB Server through tuning and optimization. It discusses general best practices like service level agreements and metrics collection. It also covers specific areas like server, storage, and network configuration, connection pooling, MariaDB configuration settings, query tuning using indexes and EXPLAIN, and monitoring tools like performance schema. The goal is to help users get the most out of their MariaDB deployment through performance analysis and tuning.
How Pixid dropped Oracle and went hybrid with MariaDBMariaDB plc
Pixid replaced Oracle Database with MySQL in 2011, then soon migrated to MariaDB to get better performance, more features and synchronous clustering for high availability. In addition to high-performance transactions, their customers needed access to fast analytics for self-service reporting and data exploration. Pixid started with a separate columnar database for analytics, but with the release of MariaDB ColumnStore, they found a more elegant solution – deploying a single database platform to handle both transactions and analytics. In this session, Antoine Gosset and Jérôme Mouret share how Pixid went from Oracle Database to handling both transactional and analytical workloads with MariaDB.
The talk will be about the project to find a replacement for all IBM products in the company with the example for the databases. What was the goal of the project, the learning, a short overview about the options
we migrated about 500 db2 databases to EnterpriseDB. The database size was from a small size up to 4 TB and we implemented a completely new fully automated deployment of VM and database. Databases are now 11 month in production. The talk will have an overview of the project, the learnings, a few parameters and technical parameters that were found for stability and performance.
MariaDB AX is engineered to simplify the process of ingesting and analyzing data, removing the need for complex, time-consuming batch processes or data models constrained to, or optimized for, a limited set of queries – modern requirements when predictive/prescriptive analytics and time-to-insight become competitive differentiators.
MariaDB AX is now based on MariaDB ColumnStore 1.1, a distributed and columnar storage engine for high-performance analytics at scale, and introduces simplified data ingestion via data adapters, semi-structured/unstructured data analysis via text and binary columns, and custom analytical functions via a user-defined aggregate/window API.
In this webinar, Dipti Joshi, MariaDB’s Director of Product Management, will provide an overview of MariaDB AX, highlight innovative customer use cases, and provide a detailed explanation of new features and the benefits they provide.
How Orwell built a geo-distributed Bank-as-a-Service with microservicesMariaDB plc
Orwell Group shares how they leveraged microservices, an event driven architecture and both master and reference data management methodologies to build a new banking system for high retail banking customers and corporate banks requiring cross border payments and cash flow management – and scaled it to handle customers with millions of clients. In particular they explain how they built a high availability, geo-distributed and consistent platform on top of MariaDB. The result was a secure and distributed platform with high cost efficiency, and the data accuracy and consistency needed to create high quality data pipelines from transactions to analytics and ensure regulatory compliance (e.g., GDPR).
IBM World of Watson 2016 - DB2 Analytics Accelerator on CloudDaniel Martin
IBM is introducing a new deployment option for the DB2 Analytics Accelerator on Cloud using dashDB as the acceleration engine. This provides customers with a hybrid cloud offering that gives the flexibility of running the Accelerator either on-premises or in the cloud. The Cloud deployment offers benefits like monthly pricing, hardware provisioning by IBM, and fast provisioning time. Initial focus areas include basic Accelerator functionality for offloading queries to the cloud, with a roadmap to continuously expand features and functionality.
Transform DBMS to Drive Apps of Engagement InnovationEDB
IT leaders face ever-increasing challenges to fund and deliver innovative business solutions for better customer and stakeholder engagement. While much of the infrastructure stack has been commoditized, the DBMS remains an expensive and growing drain on IT resources that otherwise could drive innovation.
This presentation outlines how money can be freed up in IT from expensive database spend by transforming applications and the DBMS to a subscription-based, cloud ready EnterpriseDB Postgres offering.
Further, learn how the same DBMS used for these traditional workloads with familiar tools and a large skill base can also be used for new applications of engagement, which rely on a wider variety of data including NoSQL, semi-structured along with relational data.
Visit EnterpriseDB > Resources > Webcasts to listen to the presentation recording.
This presentation is intended for strategic IT and business decision-makers involved in data infrastructure decisions and cost savings.
Temporal Tables, Transparent Archiving in DB2 for z/OS and IDAACuneyt Goksu
The document discusses several data archiving solutions for z/OS systems including temporal tables, transparent archiving, and IDAA technology. Temporal tables allow querying and updating historical data using system time periods. Transparent archiving moves old data to other storage platforms while still allowing dynamic queries. IDAA provides accelerated query performance for temporal tables by routing queries to an accelerator system. The solutions can be combined for different use cases depending on data retention and access needs.
Open innovation and collaboration between IBM and other technology companies is fueling advances in cloud computing, big data analytics, and software development. This includes contributions to open source projects like Linux as well as partnerships through organizations like the OpenPOWER Foundation. New systems based on IBM's Power architecture and optimized for Linux are helping customers improve the performance and efficiency of their analytics, database, and application workloads.
- Centering your enterprise for growth discusses recent innovations with IMS that improve performance, affordability, and simplicity. These enhancements help clients optimize and innovate with IMS to support business growth in an era of data, cloud, and mobile engagement.
- Key highlights include a 117,292 TPS and 130% increased workload throughput on z13, as well as pricing innovations like IMS Value Unit Editions to enable new analytics and mobile workloads.
- The document emphasizes how optimizing IMS can help clients accelerate insights from transaction data, rapidly enable cloud and mobile access, and expand the IMS talent population to continue delivering value.
Big Data Analytics with MariaDB ColumnStoreMariaDB plc
MariaDB ColumnStore is a massively parallel columnar storage engine for MariaDB that provides high performance analytics on large datasets. It uses a distributed columnar architecture where each column is stored separately and data is partitioned horizontally across nodes. This allows for very fast analytical queries by only accessing the relevant columns and partitions. Some key features include built-in analytics functions, high speed data ingestion, and support for running on-premises or on cloud platforms like AWS. The latest 1.1 version adds capabilities like streaming data ingestion APIs, improved high availability with GlusterFS, and performance optimizations.
Better performance and cost effectiveness empower better results in the cognitive era. For more information, visit: http://www.ibm.com/systems/power/hardware/linux-lc.html
Join Postgres experts Bruce Momjian and Marc Linster as they preview everything new in Postgres 12. You don’t want to miss this!
Highlights include:
- New compatibility features
- PostgreSQL: Table access methods
- Partitioning Improvements
GemFire Enterprise Data Fabric provides a scalable and high-performance data infrastructure that caches and distributes data across multiple formats and distributed applications. It connects to backend systems, provides high data availability, supports distributed transactions and querying. GemFire solves challenges in SOA and distributed systems by enhancing scalability, performance and reliability while potentially reducing hardware and software costs.
Intel's Data Center & Connected Systems Group and Diane Bryant shares the latest news on the latest Intel Xeon E5v2 family of processors and technologies like Intel Network Builders to enable the re-architecture of the Data Center.
The document discusses accelerating Ceph storage performance using SPDK. SPDK introduces optimizations like asynchronous APIs, userspace I/O stacks, and polling mode drivers to reduce software overhead and better utilize fast storage devices. This allows Ceph to better support high performance networks and storage like NVMe SSDs. The document provides an example where SPDK helped XSKY's BlueStore object store achieve significant performance gains over the standard Ceph implementation.
Red Hat Storage Day Atlanta - Designing Ceph Clusters Using Intel-Based Hardw...Red_Hat_Storage
This document discusses the need for storage modernization driven by trends like mobile, social media, IoT and big data. It outlines how scale-out architectures using open source Ceph software can help meet this need more cost effectively than traditional scale-up storage. Specific optimizations for IOPS, throughput and capacity are described. Intel is presented as helping advance the industry through open source contributions and optimized platforms, software and SSD technologies. Real-world examples are given showing the wide performance range Ceph can provide.
Intel® Xeon® Scalable Processors Enabled Applications Marketing GuideIntel IT Center
The Future-Ready Data Center platform is here. Whether you navigate in the High Performance Computing, Enterprise, Cloud, or Communications spheres, you will find an Intel® Xeon® processor that is ready to power your data center now and well into the future. An innovative approach to platform design in the Intel® Xeon® Scalable processor platform unlocks the power of scalable performance for today’s data centers—from the smallest workloads to your most mission-critical applications. Powerful convergence and capabilities across compute, storage, memory, network and security deliver unprecedented scale and highly optimized performance across a broad range of workloads—from high performance computing (HPC) and network functions virtualization, to advanced analytics and artificial intelligence (AI). Many examples here show how our software partner ecosystem has optimized their applications and/or taken advantage of inherent platform enhancements to deliver dramatic performance gains, that can translate into tangible business benefits.
Accelerating Virtual Machine Access with the Storage Performance Development ...Michelle Holley
Abstract: Although new non-volatile media inherently offers very low latency, remote access
using protocols such as NVMe-oF and presenting the data to VMs via virtualized interfaces such as virtio
adds considerable software overhead. One way to reduce the overhead is to use the Storage
Performance Development Kit (SPDK), an open-source software project that provides building blocks for
scalable and efficient storage applications with breakthrough performance. Comparing the software
paths for virtualizing block storage I/O illustrates the advantages of the SPDK-based approach. Empirical
data shows that using SPDK can improve CPU efficiency by up to 10 x and reduce latency up to 50% over
existing methods. Future enhancements for SPDK will make its advantages even greater.
Speaker Bio: Anu Rao is Product line manager for storage software in Data center Group. She helps
customer ease into and adopt open source Storage software like Storage Performance Development Kit
(SPDK) and Intelligent Software Acceleration-Library (ISA-L).
The document summarizes Intel's new Solid-State Drive Data Center Family for PCIe. It provides an overview of Intel's SSD product families for different market segments. It then focuses on the new Data Center Family for PCIe, highlighting its native PCIe interface, performance benefits over SAS/SATA, endurance, reliability features, and product lineup. Finally, it lists upcoming events where Intel will promote the new data center SSD family.
1. The document introduces the Intel Xeon Scalable platform, which provides the foundation for data center innovation with a 1.65x average performance boost over previous generations.
2. It highlights key advantages of the platform including scalable performance, agility in rapid service delivery, and hardware-enhanced security with near-zero performance overhead.
3. Various workload-optimized solutions are discussed that leverage the platform's performance to accelerate insights from analytics, deploy cloud infrastructure more quickly, and transform networks.
The document discusses reimagining the datacenter through software defined infrastructure. This allows datacenters to become more dynamic, automated and efficient by treating compute, storage and networking resources as composable blocks that can be allocated on demand. This approach breaks down traditional silos and allows simpler deployment and maintenance while improving agility, automation and efficiency. The software defined approach is compared to the traditional rigid infrastructure model and examples are given of how it can improve provisioning times, utilization rates and flexibility.
Streamline End-to-End AI Pipelines with Intel, Databricks, and OmniSciIntel® Software
Preprocess, visualize, and Build AI Faster at-Scale on Intel Architecture. Develop end-to-end AI pipelines for inferencing including data ingestion, preprocessing, and model inferencing with tabular, NLP, RecSys, video and image using Intel oneAPI AI Analytics Toolkit and other optimized libraries. Build at-scale performant pipelines with Databricks and end-to-end Xeon optimizations. Learn how to visualize with the OmniSci Immerse Platform and experience a live demonstration of the Intel Distribution of Modin and OmniSci.
Virtualization is an increasingly critical part of data center computing. Selecting a server that excels at virtualization makes good business sense. Two Lenovo ThinkServer RD630 servers, paired with Dot Hill AssuredSAN Pro5720 tiered storage, ran 10 VMmark tiles for a total of 80 running VMs and achieved a score of 11.17@10 tiles, placing it in the top 8 percent of the 32-core server configurations. This makes the Lenovo ThinkServer RD630 an excellent choice for any enterprise that uses virtualization.
The document discusses Intel's new Xeon E5 v3 processors and vision for hybrid cloud and software defined infrastructure. It highlights key features of the new processors like AVX2 for improved performance, VMCS shadowing for virtualization efficiency, and power optimizations. It argues these technologies help deliver agility, efficiency and insights for software defined datacenters.
Yashi dealer meeting settembre 2016 tecnologie xeon intel italiaYashi Italia
The document discusses new infrastructure solutions for evolving needs, including private cloud, data analytics, performance improvement, and energy efficiency. It describes application-driven allocation of orchestrated compute, network, and storage resources that can be automatically provisioned and managed. The document also covers customized Intel hardware and software optimizations for network applications, including Intel QuickAssist technology, Intel Ethernet controllers, and reliability profiles.
Intel® Xeon® Processor E5-2600 v3 Product Family Application Showcase - Fin...Intel IT Center
This Intel® Xeon® Processor E5-2600 v3 Product Family Application Showcase focuses on Financial Services software companies who have seen preformance increases with Intel products.
Accelerate Your Apache Spark with Intel Optane DC Persistent MemoryDatabricks
The capacity of data grows rapidly in big data area, more and more memory are consumed either in the computation or holding the intermediate data for analytic jobs. For those memory intensive workloads, end-point users have to scale out the computation cluster or extend memory with storage like HDD or SSD to meet the requirement of computing tasks. For scaling out the cluster, the extra cost from cluster management, operation and maintenance will increase the total cost if the extra CPU resources are not fully utilized. To address the shortcoming above, Intel Optane DC persistent memory (Optane DCPM) breaks the traditional memory/storage hierarchy and scale up the computing server with higher capacity persistent memory. Also it brings higher bandwidth & lower latency than storage like SSD or HDD. And Apache Spark is widely used in the analytics like SQL and Machine Learning on the cloud environment. For cloud environment, low performance of remote data access is typical a stop gap for users especially for some I/O intensive queries. For the ML workload, it's an iterative model which I/O bandwidth is the key to the end-2-end performance. In this talk, we will introduce how to accelerate Spark SQL with OAP (https://github.com/Intel-bigdata/OAP) to accelerate SQL performance on Cloud to archive 8X performance gain and RDD cache to improve K-means performance with 2.5X performance gain leveraging Intel Optane DCPM. Also we will have a deep dive how Optane DCPM for these performance gains.
Speakers: Cheng Xu, Piotr Balcer
Intel® Xeon® Processor E5-2600 v3 Product Family Application Showcase - Telec...Intel IT Center
This Intel® Xeon® Processor E5-2600 v3 Product Family Application Showcase focuses on Telecommunications and Cloud software companies who have seen performance increases with Intel products.
Ceph on Intel: Intel Storage Components, Benchmarks, and ContributionsColleen Corrice
At Red Hat Storage Day Minneapolis on 4/12/16, Intel's Dan Ferber presented on Intel storage components, benchmarks, and contributions as they relate to Ceph.
Ceph on Intel: Intel Storage Components, Benchmarks, and ContributionsRed_Hat_Storage
At Red Hat Storage Day Minneapolis on 4/12/16, Intel's Dan Ferber presented on Intel storage components, benchmarks, and contributions as they relate to Ceph.
DAOS - Scale-Out Software-Defined Storage for HPC/Big Data/AI Convergenceinside-BigData.com
In this deck, Johann Lombardi from Intel presents: DAOS - Scale-Out Software-Defined Storage for HPC/Big Data/AI Convergence.
"Intel has been building an entirely open source software ecosystem for data-centric computing, fully optimized for Intel® architecture and non-volatile memory (NVM) technologies, including Intel Optane DC persistent memory and Intel Optane DC SSDs. Distributed Asynchronous Object Storage (DAOS) is the foundation of the Intel exascale storage stack. DAOS is an open source software-defined scale-out object store that provides high bandwidth, low latency, and high I/O operations per second (IOPS) storage containers to HPC applications. It enables next-generation data-centric workflows that combine simulation, data analytics, and AI."
Unlike traditional storage stacks that were primarily designed for rotating media, DAOS is architected from the ground up to make use of new NVM technologies, and it is extremely lightweight because it operates end-to-end in user space with full operating system bypass. DAOS offers a shift away from an I/O model designed for block-based, high-latency storage to one that inherently supports fine- grained data access and unlocks the performance of next- generation storage technologies.
Watch the video: https://youtu.be/wnGBW31yhLM
Learn more: https://www.intel.com/content/www/us/en/high-performance-computing/daos-high-performance-storage-brief.html
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
Spring Hill (NNP-I 1000): Intel's Data Center Inference Chipinside-BigData.com
- SpringHill (NNP-I1000) is Intel's new data center inference chip that provides best-in-class performance per watt for major data center inference workloads.
- It delivers 4.8 TOPs/watt of performance and can scale from 10-50 watts to boost performance.
- The chip features 12 inference compute engines, 24MB of shared cache, and Intel architecture cores to drive AI innovation while maintaining high performance and efficiency.
Similar to M|18 Intel and MariaDB: Strategic Collaboration to Enhance MariaDB Functionality, Performance and TCO (20)
MariaDB Paris Workshop 2023 - MaxScale 23.02.xMariaDB plc
The document discusses the benefits of exercise for mental health. Regular physical activity can help reduce anxiety and depression and improve mood and cognitive functioning. Exercise causes chemical changes in the brain that may help boost feelings of calmness, happiness and focus.
MariaDB Paris Workshop 2023 - NewpharmaMariaDB plc
This document summarizes Newpharma's transition from a standalone database server to an enterprise MariaDB Galera cluster configuration between 2018-2023. It discusses the business needs that drove the change, including increased traffic and access to multiple data sources. Key benefits of the Galera cluster are highlighted like synchronous replication, read/write access from any node, and automatic node joining. Challenges of migrating like converting table types and splitting large transactions are also outlined. The transition has supported Newpharma's growth to over 100 million euro in turnover.
The document discusses the benefits of exercise for mental health. Regular physical activity can help reduce anxiety and depression and improve mood and cognitive functioning. Exercise causes chemical changes in the brain that may help boost feelings of calmness and well-being.
MariaDB Paris Workshop 2023 - MariaDB EnterpriseMariaDB plc
The document discusses the benefits of exercise for mental health. Regular physical activity can help reduce anxiety and depression and improve mood and cognitive functioning. Exercise causes chemical changes in the brain that may help protect against mental illness and improve symptoms.
MariaDB Paris Workshop 2023 - Performance OptimizationMariaDB plc
MariaDB is an open-source database that is highly tunable and modular. It allows for various storage engines, plugins, and configurations to optimize performance depending on usage. Key aspects that impact performance include memory allocation, disk access, query optimization, and architecture choices like replication, sharding, or using ColumnStore for analytics. Solutions like MyRocks, Spider, MaxScale can improve performance for transactional or large scale workloads by optimizing resources, adding high availability, and distributing load.
MariaDB Paris Workshop 2023 - MaxScale MariaDB plc
The document outlines requirements and criteria for a database solution involving two buildings 30km apart with a WAN link. The chosen solution was MariaDB with Galera cluster for high availability and synchronous replication across sites, along with Maxscale for read/write splitting and failover. Maxscale instances on each site allow for zero downtime database patching and upgrades per site, while the Galera cluster provides structure-independent synchronous replication between sites.
MariaDB Tech und Business Update Hamburg 2023 - MariaDB Enterprise Server MariaDB plc
MariaDB Enterprise Server 10.6 includes the following key features:
- New JSON functions and data types like UUID and INET4.
- Improved Oracle compatibility with function parameters.
- Enhanced partitioning capabilities like converting partitions.
- Optimistic ALTER TABLE for replicas to reduce downtime.
- Online schema changes without locking tables for improved performance.
- Security enhancements including password policies and privilege changes.
MariaDB SkySQL is a cloud database service that provides autonomous scaling, observability, and cloud backup capabilities. It offers multi-cloud and hybrid operations across AWS, Google Cloud, and on-premises databases. The service includes features like the Remote Observability Service (ROS) for monitoring across environments, and a Cloud Backup Service. It aims to provide a simple yet advanced service for scaling databases from small to extreme sizes with tools for automation, self-service, and unified operations.
The document discusses high availability solutions for MariaDB databases. It begins by defining high availability and concepts like Recovery Time Objective (RTO) and Recovery Point Objective (RPO). It then presents different MariaDB and MaxScale architectures that provide high availability, including single node, primary-replica, Galera cluster, and SkySQL solutions. Key aspects covered are automatic failover, load balancing, data filtering, and service level agreements.
Die Neuheiten in MariaDB Enterprise ServerMariaDB plc
This document summarizes new features in MariaDB Enterprise Server. Key points include:
- MariaDB Enterprise Server is geared toward enterprise customers and focuses on stability, robustness, and predictability.
- It has a longer release cycle than Community Server, with new versions every 2 years and long maintenance cycles. New features from Community Server are backported.
- Recent additions include analytics functions, JSON support, bi-temporal modeling, schema changes, database compatibility features, and security enhancements.
- The upcoming 23.x release will include new JSON functions, data types like UUID and INET4, Oracle compatibility features, partitioning improvements, and Galera enhancements.
Global Data Replication with Galera for Ansell Guardian®MariaDB plc
Ansell Guardian® faced challenges with their previous database replication solution as their data and usage grew globally. They evaluated MariaDB/Galera and implemented it to replace their legacy solution. The implementation was smooth using automation scripts. MariaDB/Galera provided increased performance, faster deployment times, and more reliable data synchronization across their 3 data centers compared to their previous solution. It helped resolve a critical data divergence issue and improved the user experience. They plan to further enhance their database infrastructure using MaxScale in the future.
SkySQL is the first and only database-as-a-service (DBaaS) to perform workload analysis with advanced deep learning models, identifying and classifying discrete workload patterns so DBAs can better understand database workloads, identify anomalies and predict changes.
In this session, we’ll explain the concepts behind workload analysis and show how it can be used in the real world (and with sample real-world data) to improve database performance and efficiency by identifying key metrics and changes to cyclical patterns.
SkySQL uses best-of-breed software, and when it comes to metrics and monitoring that means Prometheus and Grafana. SkySQL Monitor is built on both, and provides customers with interactive dashboards for both real-time and historic metrics monitoring. In addition, it meets the same high availability and security requirements as other SkySQL components, ensuring metrics are always available and always secure.
In this session, we’ll explain how SkySQL Monitor works, walk through its dashboards and show how to monitor key metrics for performance and replication.
Introducing the R2DBC async Java connectorMariaDB plc
Not too long ago, a reactive variant of the JDBC driver was released, known as Reactive Relational Database Connectivity (R2DBC for short). While R2DBC started as an experiment to enable integration of SQL databases into systems that use reactive programming models, it now specifies a full-fledged service-provider interface that can be used to retrieve data from a target data source.
In this session, we’ll take a look at the new MariaDB R2DBC connector and examine the advantages of fully reactive, non-blocking development with MariaDB. And, of course, we’ll dive in and get a first-hand look at what it’s like to use the new connector with some live coding!
The capabilities and features of MariaDB Platform continue to expand, resulting in larger and more sophisticated production deployments – and the need for better tools. To provide DBAs with comprehensive, consolidating tooling, we created MariaDB Enterprise Tools: an easy-to-use, modular command-line interface for interacting with any part of MariaDB Platform.
In this session, we will provide a preview of the MariaDB Enterprise Client, walk through current and planned modules and discuss future plans for MariaDB Enterprise Tools – including SkySQL modules and the ability to create custom modules.
Faster, better, stronger: The new InnoDBMariaDB plc
For MariaDB Enterprise Server 10.5, the default transactional storage engine, InnoDB, has been significantly rewritten to improve the performance of writes and backups. Next, we removed a number of parameters to reduce unnecessary complexity, not only in terms of configuration but of the code itself. And finally, we improved crash recovery thanks to better consistency checks and we reduced memory consumption and file I/O thanks to an all new log record format.
In this session, we’ll walk through all of the improvements to InnoDB, and dive deep into the implementation to explain how these improvements help everything from configuration and performance to reliability and recovery.
SkySQL implements a groundbreaking, state-of-the-art architecture based on Kubernetes and ServiceNow, and with a strong emphasis on cloud security – using compartmentalization and indirect access to secure and protect customer databases.
In this session, we’ll walk through the architecture of SkySQL and discuss how MariaDB leverages an advanced Kubernetes operator and powerful ServiceNow configuration/workflow management to deploy and manage databases on cloud infrastructure.
Build applications with generative AI on Google CloudMárton Kodok
We will explore Vertex AI - Model Garden powered experiences, we are going to learn more about the integration of these generative AI APIs. We are going to see in action what the Gemini family of generative models are for developers to build and deploy AI-driven applications. Vertex AI includes a suite of foundation models, these are referred to as the PaLM and Gemini family of generative ai models, and they come in different versions. We are going to cover how to use via API to: - execute prompts in text and chat - cover multimodal use cases with image prompts. - finetune and distill to improve knowledge domains - run function calls with foundation models to optimize them for specific tasks. At the end of the session, developers will understand how to innovate with generative AI and develop apps using the generative ai industry trends.
Open Source Contributions to Postgres: The Basics POSETTE 2024ElizabethGarrettChri
Postgres is the most advanced open-source database in the world and it's supported by a community, not a single company. So how does this work? How does code actually get into Postgres? I recently had a patch submitted and committed and I want to share what I learned in that process. I’ll give you an overview of Postgres versions and how the underlying project codebase functions. I’ll also show you the process for submitting a patch and getting that tested and committed.
Beyond the Basics of A/B Tests: Highly Innovative Experimentation Tactics You...Aggregage
This webinar will explore cutting-edge, less familiar but powerful experimentation methodologies which address well-known limitations of standard A/B Testing. Designed for data and product leaders, this session aims to inspire the embrace of innovative approaches and provide insights into the frontiers of experimentation!
4th Modern Marketing Reckoner by MMA Global India & Group M: 60+ experts on W...Social Samosa
The Modern Marketing Reckoner (MMR) is a comprehensive resource packed with POVs from 60+ industry leaders on how AI is transforming the 4 key pillars of marketing – product, place, price and promotions.
"Financial Odyssey: Navigating Past Performance Through Diverse Analytical Lens"sameer shah
Embark on a captivating financial journey with 'Financial Odyssey,' our hackathon project. Delve deep into the past performance of two companies as we employ an array of financial statement analysis techniques. From ratio analysis to trend analysis, uncover insights crucial for informed decision-making in the dynamic world of finance."
5. Cloud &
DATA Center
Things &
Devices
MEM
ORY
FPGA
Cloud and Data Center,
AI, SD Infrastructure,
Big Data Analytics
Manufacturing, Memory,
FPGA, Software,
Security
Intel Capital leads $20 million
investment to grow MariaDB
Open source database gets funding to boost
adoption and develop product range
Intel® Xeon® Processor
Scalable Family
3D XPointTM
Intel® OptaneTM SSD
DC P4800X
Intel® Stratix® 10 FPGA
Virtuous Cycle of Growth
6. Intel’s Software Group:
MariaDB Team
6
Shanghai
Intel DCG database
performance lab
(Intel® Xeon®)
Moscow
MariaDB database
performance
Early Access lab
London
Database performance
tools & customer
engagements
Munich
MariaDB - Intel
Global Relationship
Management
7. 7
Working together to optimize
the combination of:
• MariaDB* and MyRocks*
storage engine
• Intel® Xeon® Scalable Platform
• Intel® Optane™ SSDs
Co-Marketing Collaborations
MariaDB Intel
Collaboration
Early access to
Intel pre-production H/W
GoToMarket support
and customer Trusted Advisors
Joint performance
analysis and tuning
S/W development
tools
Full tech. support
Long-term development
project collaboration
MyRocks
8. delivers 1.65x average performance boost over
prior Generation1
Intel® Xeon® Scalable
platform
http://www.intel.com/performance
Perform
ance
Agilit
y
Securi
tyPervasive through compute,
storage and network
Pervasive data security with
no performance overhead
Rapid service delivery
The Foundation of Data Center Innovation: Agile & Trusted Infrastructure
9. MariaDB OLTP Performance on Intel® Xeon®
Scalable PLATFORM
9
0
175000
350000
525000
700000
Xeon E5-2697v3 Xeon E5-2699v4 Xeon Platinum 8180
NOPM
1.26X
1.34X
1.7X
HammerDB test provides a complex OLTP workload
with write/write locking contention to simulate real-world
applications and compare performance across
database platforms
http://www.intel.com/performance
MariaDB OLTP Performance on Intel® Xeon®
Scalable PLATFORM
MariaDB 10.2 InnoDB OLTP Workload
10. 10
0.
0.3
0.6
0.9
1.2
1.5
MariaDB Commercial
$
0
0.35
0.7
1.05
1.4
MariaDB Commercial
$
0.
0.25
0.5
0.75
1.
MariaDB Commercial
$
32X 41X 42X
• Software Total Cost of Ownership (TCO) over 3 years MariaDB and commercial database
• ‘Cost per Transaction’ means cost for the same unit of work
• MariaDB becomes more compelling with each CPU generation
$0.038
$1.20
$1.23
$0.030
$0.92
$0.022
http://www.intel.com/performance
Lower Cost Lower Cost Lower Cost
MariaDB OLTP Cost on Intel® Xeon®
Scalable Platform
2014 Cost per
Transaction
E5-2697v3
2016 Cost per
Transaction
E5-2699v4
2017 Cost per
Transaction
Platinum 8180
11. A Glimpse Inside the
Intel® Xeon® Scalable platform
SSD
Intel® Optane™ SSD
DC P4800X
Complementary
Intel® FPGA
Workload optimized frameworks
Integrated Options
Fabric
Intel® Omni-Path
Architecture
Networking
Intel® Ethernet
Accelerators
Intel® QuickAssist
Intel® AVX-512
Advancing virtually every aspect:
Brand New core,
cache, on-die interconnects, memory
controller & more
performa
nce
Agilitysecurity
12. 12
Combining the attributes of memory and storage with
low latency, high endurance, outstanding QoS and high
throughput+
Intel® optane™ SSD DC P4800X series is
delivering on data center needs
Enabling a new data tier to accelerate applications for fast
caching & storage and expanded memory pools, to
increase scale per server and reduce transaction cost
Bringing synergy with Intel® Xeon® processors to enable
bigger and more affordable datasets to gain new
insights from larger memory pools
13. 13
MySQL* 5.7 Sysbench internal testing
Memory
DRAM
256GB
DRAM
256GB
DRAM
256GB
CPU
Intel® Xeon®
v4 (44 core)
Intel®
Xeon™
Intel® Xeon®
v4 (44 core)
Intel®
Xeon™
Intel® Xeon®
v4 (44 core)
Intel®
Xeon™
Storage
Intel® SSD
DC S3520
Intel® SSD
DC P4500
Intel® Optane™
SSD
Transactions
Per Second
17,235
5,834
3,535
TPS1
4x
better
up to
8x
better
up to
Latency1
Intel® Optane™
ssd
vs. “good”
solution
P99
Latency
12 ms
51 ms
100 ms
1.System configuration:Server Intel® Server System R2208WT2YS, 2x Intel® Xeon® E5 2699v4, 384 GB DDR4 DRAM, boot drive- 1x Intel® SSD DC S3710 Series (400 GB), database drives- 1x Intel® SSD DC P3700 Series (400 GB) and 1x Intel® Optane™ SSD DC P4800X Series (140 GB prototype),CentOS 7.2, MySQL Server 5.7.14,
Sysbench 0.5 configuredfor 70/30 Read/WriteOLTP transactionsplit using a 100GB database.Cost per transactiondetermined by totalMSRP for each configurationdivided by the transactionsper second. Estimatedresultswereobtainedpriortoimplementationofrecentsoftwarepatchesandfirmwareupdatesintendedtoaddressexploitsreferredtoas"Spectre"and
"Meltdown".Implementationoftheseupdatesmaymaketheseresultsinapplicabletoyourdeviceorsystem.
*Other names and brands names may be claimed as thepropertyof others
14. 14
Xeon E5-2699v4 Xeon Platinum 8180
1.78X
MariaDB*: Sysbench OLTP workload. OS: Ubuntu Server 17.04 x64. Sysbench 0.4.12, MariaDB 10.2.9 GIT snapshot 2017-09-12. Testing by Intel, September 2017.
BASELINE: 2S Intel® Xeon® E5-2699 v4, 2.2GHz, 22 cores, turbo and HT on, BIOS 63.R00, 64GB total memory, 1600 MT/s / DDR4 LRDIMM, 1TB SATA HDD, Intel® SSD DC P3700
Series 2TB . NEW: 2S Intel® Xeon® Platinum 8180 processor, 2.5 GHz, 28 cores, turbo and HT on, BIOS 1.00.0412, 192GB total memory, 12 slots / 16GB / 2666 MT/s / DDR4 LRDIMM, 1
TB SATA HDD, Intel® SSD DC P3700 Series 2TB.
http://www.intel.com/performance
“MyRocks is a storage engine that adds the RocksDB LSM flash
storage-optimized database to MariaDB. Even while under
development, it shows good performance and scalability -- close
to mature storage engines like InnoDB. Running our pre-release
MariaDB 10.2.9 with a built-in MyRocks engine under a Sysbench
RO multi-table test on the newest Intel® Xeon® Platinum 8180
processors, we saw up to 1.78x more throughput with a 1.18x
reduction of average response time vs. previous-gen Intel®
Xeon® processor E5-2699 v4 performance.”
Michael Widenius, CTO, MariaDB Corporation Ab
MYROCKS* NEXT-GENERATION STORAGE
ENGINE ON Intel® Xeon® Scalable PlatformMariaDB 10.2.9 Sysbench
OLTP Workload
17. 17
1.AWS Relational Database
Service (RDS)
2.AWS Aurora (RDS)
“Data Center Networks
Are in My Way!”
Using AWS’s RDS* vis-à-vis their RDS/Aurora* Log-Structured Storage as the example
18. 18
Matsunobu, "MyRocks: Space and write optimized OLTP database at Facebook," 2016
https://atscaleconference.com/videos/myrocks-space-and-write-optimized-oltp-database-at-facebook/
Facebook’s MyRocks* Comes Out from Behind
the Curtain
Major FEATURES in
MyRocks
Similar
Feature Sets
as InnoDB*
TRANSAC
TIONS
ONLINE
BACKUP• Atomicity
• Non-locking consistent reads
• Read committed
• Repeatable read
• Crash-safe slave and master
• Logical backup by mysqldump
• Binary backup by
myrocks_hotbackup
19. 19
Customer “Pain Points”
The Downside of Becoming a Data Company
Companies that have pivoted to AI and analytics in order to monetize their data are
experiencing accelerating data growth rates
Disaggregating Storage
This data growth challenge makes disaggregating storage from compute attractive
because the company can scale their storage capacity to match their data growth,
independent of compute.
Cloud-Like, Open-Source, and AI/Analytics-Ready
Many of these companies are big Oracle RAC*/Exadata* users. They expressed a
desire to move off this platform to something more cloud-like, open source, and readily
integrated into their AI/analytics investments.
1
2
3
20. 20
Physical Cluster
ToR Switch
Spine Switch
JBOF
Storage Servers
Compute
Servers
100GbsLinks
50GbsLinks
Resource
Mgmt
Service
Mgmt
Services
A Solution to Address
These Pain Points CLUSTER
SOFTWARE
21. 21
Revisiting Cluster Computing
in the Era of Cloud
Rack-Scale Design and Disaggregated Storage
1.Rack-Centric, Physical Cluster 2.Cluster-Wide
Service
Management
(k8s)
3.Database-as-a-
Service on Log-
Structured
Storage
website1 website2
ELB
Service (Load balancer)
Pod ingress controller Ingress
Service
(Node Port)
Service
(Node Port)
… Service
(Node Port)
Pod
(website1)
Pod
(website2)
Pod (default
backend)
Kubernetes
22. • Support for the persistence requirements of distributed databases
• Fast recovery and cloning through an automated recovery process
- Replicated log service
• Little or no interruptions to the transaction processing
- High-throughput and low-latency
- Durability, replication and strong consistency
- Optimizes both performance and TCO for transaction-centric workloads
22
Advantages of Log-
Structured Storage
23. 23
A Stack
Comparison
A Stack Comparison
Enterprise (Oracle) Cloud
(Amazon Aurora*)
Intel Solution
(MariaDB*-based
Solution)
Database Engine Proprietary
(Oracle RAC*)
Open Source
(MySQL, Postgres)
Open Source
MariaDB with the MaxScale
Layer-7 Proxy
Shared Storage Proprietary
Oracle Automatic Storage
Management (ASM) with
POSIX File System
interface
Open Source +Proprietary
(Distributed Log, S3)
Open Source
MyRocks* compiled with
Rockset’s RocksDB-Cloud*
library
Clustering Software Proprietary
(Oracle Clusterware*)
Proprietary
AWS RDS*
Open Source
Kubernetes*
Hardware Proprietary
x86 servers interconnected
via Infiniband*
(or Ethernet)
Commodity
x86 servers interconnected
via Ethernet Fabric
Commodity
x86 servers interconnected
via 100Gbs Ethernet Fabric
24. Flush sst file
to local SSD
writes
memtable cache
Persistent read
cache on SSD
Flush sst file to
cloud storage
Cloud Storage Bucket
Cloud Application
RocksDB-Cloud*
block cache
reads
Queries
Updates
Cloud Log
Storage,
Kafka* Topic
24
Provisioning a Database
Instance
WAL
25. read
write
Cloned Server
RocksDB-Cloud
read
Cloud Bucket A
Server
RocksDB-Cloud*
Clones are read-only replicas
25
fast cloning
Queries
Updates
Queries
Cloud Bucket B
write read
Cloud Log
Storage,
Kafka*
WAL writes
WAL tailer
26. • Uses MyRocks w/Rockset’s RocksDB-Cloud library
• Requires Support for Lossless-SemiSync Protocol and GTIDs
in the MaxLog Binlog Server (“Log-Tailer”)
• Need a “Lambda” function to kickoff “Switchover”
• Need Automation for carrying out Switchover
• Can use either InnoDB* or MyRocks* w/RocksDB* library
• Requires Support for Lossless-SemiSync Protocol and
GTIDs in the Slaves
• Need a “Lambda” function to kickoff “Switchover”
• Need Automation for carrying out Switchover
26
Time0 Time1
Time0 Time1
Database Instance within a
Single Cluster1.Master-Slave (Shared-Nothing) 2.Master-Only (Shared)
29. Intel® Big Data Analytics Frameworks
Accelerate innovation in big data analytics with frameworks
built on software-defined Infrastructure with open-standard
building blocks.
Intel® Frameworks & Libraries Integrated with FPGAs
Run unmodified customer applications, use runtime
orchestration with both Intel® Xeon® processor and FPGA
support, and leverage end-to-end virtualization and security.
Accelerate Relational, NoSQL and Unstructured
FPGA data access, networking and algorithm acceleration
options with a single FPGA for highly structured, semi-
structured, and unstructured data for better TCO, flexibility
and future proofing.
Analytics Landscape and Scaling
Accelerate Big Data Analytics with Existing
Interfaces and FPGAs
1
2
3
30.
31. Microsoft Scale-Out FPGA Multi-Function Accelerator
• “Diversity of cloud workloads and … rapid …
change” (weekly or monthly)
– Search, SmartNIC, Machine Learning, Encrypt,
Compress, Big Data Analytics…
• Bing Search: 2X server level perf, 29% latency
reduction
• Networking Virtualization: 10X latency improvement
• Machine Learning: Stratix 10 capable of 90 TFLOPs
8-bit floating-point
Single FPGA Algorithm, Networking & Data
Access Acceleration
32. 32
Data
Network
Streaming
Data
Integrate to Intel® Frameworks and APIs
– Run unmodified customer applications
– Orchestration run-time advantage: Xeon® or FPGA
– End-to-end security and virtualization framework
Moderate Acceleration is Common
– PCIe* look-aside acceleration (two data copies)
Significant Acceleration requires FPGA
– Multifunction and inline with a single FPGA
FPGAs Offer Unique Value for Analytics &
Streaming
Offloads algorithm, networking
and data access processing
Single Multifunction
Accelerator
33.
34. 34
Acceleration Overview
• 20X+ single table inserts/s for real time data analytics
- With modest tuning, 15M INSERT/s1
• 2X+ optimized queries for data warehousing
- Using industry standard TPC-DS benchmark
• 3X+ storage compression
- Data and tables managed by Swarm64*
Swarm64 Relational Database
Acceleration
Two Workloads: Traditional Data Warehousing, Real-Time Data Analytics
Database acceleration
with a plugin
35. 35
Swarm64 Relational Database
Acceleration
Scale-Up Data Warehousing, Real-Time Data Analytics, and Storage Compression
Database acceleration
with a plugin
Overview
• No customer application change
- Storage engine plugin
- Query Engine accelerates INSERT, SELECT, …
• Optimized indexing
• More I/O bandwidth, memory depth from compression
Significant Acceleration
√ Data access acceleration
√ Compression, filtering, replication …
√ Memory-mapped acceleration, data cache
√ “Optimized Columns” indexing
36.
37. 37
FPGACPU
User Application
and Libraries
CPU
FPGA Interface Manager (FIM)
Intel® Acceleration Engine with OPAE1
Technology
Accelerator Function
(Developer created or
provided by Intel)
UPI2/PCIe*
HSSI3
Hypervisor & OS
Optimized and simplified
hardware and software
APIs provided by IntelOPAE
FPGA
Acceleration Environment
Common Developer Interface for Intel® FPGA Data Center Products
Accelerator Function
Interfaces
38. 38
Start developing for Intel®
FPGAs with OPAE today:
http://github.com/OPAE
• Consistent API across product generations and platforms
- Abstraction for hardware-specific FPGA resource details
• Designed for minimal software overhead and latency
- Lightweight user-space library (libfpga)
• Open ecosystem for industry and developer community
- License: FPGA API (BSD), FPGA driver (GPLv2)
• FPGA driver being upstreamed into Linux* kernel
• Supports both virtual machines and bare metal platforms
• Faster development and debugging of Accelerator Functions
with the included AFU Simulation Environment (ASE)**
• Includes guides, command-line utilities and sample code
FPGA Hardware + Interface Manager
FPGA Driver
(Physical Function – PF)
FPGA API
(C) (Enumeration, Management, Access)
Applications, Frameworks, Intel® Acceleration Libraries
Bare Metal Virtual Machine
FPGA Driver
(Virtual Function - VF)
Hypervisor
FPGA Driver
(Common – AFU, Local Memory)
OS
Open Programmable Acceleration Engine
(OPAE) Technology
Simplified FPGA Programming Model for Application Developers
39. • Data analytics acceleration with no change to application required
- Relational, NoSQL and Spark*/Hadoop*
• Single workload with FPGA multifunction acceleration
• Swarm64* accelerates big-and-fast data or traditional data warehousing
• Potential to accelerate proxy access with partner rENIAC
• Intel® FPGA card with framework for interfaces, end-to-end security and virtualization
- Production in 2Q18
SUMMARY
Editor's Notes
As businesses become more and more data intensive, the cost per transaction becomes an important metric. The combination of MariaDB and Intel® technologies is extremely powerful in this age of distributed computing.
In this session, we will introduce the Intel/MariaDB collaboration team. Then we’ll discuss how the team is working to add shared, log-structured storage to support the persistence requirements of databases. Furthermore, you will learn how our cooperation supports the transformation of transaction performance and cost by optimizing MariaDB running on the combination of the Intel® Xeon® processor Scalable family and Intel® Optane™ SSDs. This platform includes the unique combination of 3D XPoint™ memory media with Intel’s advanced system memory controller, interface hardware, and software. We provide some early performance results based on an operational MariaDB/MyRocks implementation. We will also describe how Intel® FPGAs are accelerating database performance.
Our work with MariaDB* is optimizing the combination of the following technologies:
MariaDB and MyRocks* storage engine
The Intel® Xeon® Scalable Platform using Intel® Optane™ SSDs
This platform includes the unique combination of 3D XPoint™ memory media with Intel’s advanced system memory controller, interface hardware, and software.
Intel is the ONLY company that powers every segment of the smart, connected world – from the cloud, to the network, to the device – and everything in between. We alone have the assets to power the next generation of technologies for the world.
Our Strategy – A Virtuous Cycle
We call it a virtuous cycle – the cloud and the data center, the Internet of Things, memory and FPGAs -- all bound together by connectivity and enhanced by the economics of Moore’s Law.
Our strategy builds upon itself. These assets are all connected, and together, they fuel our business. With each new device that comes online and connects to the cloud – we have the tremendous opportunity to reinforce our company’s growth.
The Cloud and Data Center
The cloud is the most important trend shaping the future of the smart, connected world – and thus Intel’s future.
Intel architecture defines the infrastructure of the cloud, and we will continue to drive more and more of the footprint of the data center globally. The cloud and analytics from the data center are the greatest value drivers in technology today.
This digitization of everything will disrupt entire industries and open up new cycles of growth.
At Intel, we’ll accelerate the power and value of this data and its analytics by continuing to innovate in high-performance computing, big data and machine learning capabilities.
Internet of Things
The Internet of Things encompasses all smart devices – every PC, device, sensor, console and any other edge device – that are connected to the cloud.
The PC is foundational to our compute strategy and to our business. It’s an engine that creates critical shared IP that drives innovation across our entire product portfolio. Intel will continue to deliver an annual cadence of leadership performance and innovation in our PC and broader computing roadmap, with a focus on key growth opportunities in 2 in 1s, gaming and home gateways. We are imaging and inventing a PC that is truly immersive, assistive, with an incredibly natural interface tailored for your individual needs.
Intel is also inventing a future that is more informed, collaborative, and meaningful given the widespread interconnectedness of everyday objects. As the Internet of Things evolves, we see three distinct phases emerging.
Make everyday objects smart – this is well underway with everything from smart toothbrushes to smart car seats now available.
Connect the unconnected, with new devices connecting to the cloud and enabling new revenue, services and savings. New devices like cars and watches are being designed with connectivity and intelligence built into the device.
Deliver constant connectivity for devices that will need the intelligence to make real-time decisions based on their surroundings.
At Intel, we will focus on autonomous vehicles, industrial and retail as our primary growth drivers of the Internet of Things.
Memory and Programmable Solutions
Memory and programmable solutions such as FPGAs will deliver entirely new classes of products for the data center and the Internet of Things.
Intel® Rack Scale Design, 3D XPoint™ memory, FPGAs and silicon photonics are technologies that have been in development for several years at Intel and that we will bring to production soon.
FPGAs are now at the heart of a diverse compute model that spans the Internet of Things (IoT), communications infrastructure, to the cloud and data center, and automotive markets. Intel FPGAs are used by over 10,000 customers in applications ranging from machine learning, to Advanced Driver Assist Systems (ADAS), from the core to edge, and in factory automation and smart energy systems among many others.
Intel is driving innovations – using our deep understanding of materials science as well as computer architecture – to grow memory and storage at a much faster rate to virtually eliminate latency for data from storage.
Bigger memory and faster storage benefits many new devices by enabling more immersive experiences with natural interaction and also provides significant value to the cloud by allowing businesses to run more efficiently.
The availability of faster storage and bigger memory also unlocks more value in the cloud as we learn to automate and efficiently analyze increasing quantities of data.
Connectivity
Threading this virtuous cycle together is connectivity – the fact that providing computing power to a device and connecting it to the cloud makes it more valuable.
In the future, we will add more than 50 billion smart and connected devices, machines, autonomous vehicles, buildings and cities. These devices will be always on and connected — with a demand for lower-latency where dynamic, split-second analytics, decisions and action on the data is required.
At Intel, we recognize that 5G is more than an evolutionary step forward for our industry. As the world moves to 5G, Intel will lead because of our technological strength to deliver end-to-end 5G systems. This is why Intel is focusing on three key areas: industry partnerships, end-to-end 5G-related hardware and software development, and supporting 5G standards-setting.
Intel will continue to develop technologies such as Mobile Edge Computing, millimeter wave and NarrowBand IOT (NB-IOT). These are all important steps to bring connectivity to a variety of new IoT devices globally. We’ll also continue to forge industry partnerships to develop and transform network infrastructure technologies.
Moore’s Law
Moore’s Law will continue to progress and Intel will confidently continue to harness its value. Intel’s leadership in Moore’s Law has driven the products delivering massive computing power growth and increasingly better economics and pricing.
As we progress from 14 nanometer technology to 10 nanometer and plan for 7 nanometer and 5 nanometer and even beyond, our plans are proof that Moore’s Law is alive and well.
Amazing Experiences
Understanding this virtuous cycle, imagine a world where everything is smart and connected.
Where your wedding ring monitors your blood sugar, your coffee pot automatically orders a refill of your favorite bean, or where your work station eliminates distractions to keep you productive - soon everything will seamlessly improve and support day-to-day tasks, both simple and complex.
Where smart devices are embedded in all types of sports equipment, from jerseys, to balls, to protective gear. And these devices are generating vast amounts of data that are revolutionizing how players train, coaches teach, and scouts evaluate talent. The technologies we invent will also radically improve player safety, and transform the viewing experience. Audiences will be more engaged than ever, tracking the precise activity of their favorite athletes, from warm ups, to celebrating the victory.
Where autonomous vehicles will soon be part of an integrated transportation network where cars will communicate with each other about traffic patterns, collisions, or even let you know when walking or taking mass transit is faster. Cars are becoming intelligent in a way that will save time and save lives.
Intel transforms the technology we invent into amazing experiences.
Current major sites where we are doing MariaDB related activities:
Steve - HammerDB tool + remote testing
Mikhail - running the lab and local testing
Intel and MariaDB participate in several co-marketing collaborations, such as the Solution Brief, “Intel® Xeon® Platinum processor Accelerates MariaDB Server* and MyRocks*,” that you received in your Welcome Packet.
Key Message: Introducing Xeon Scalable Platform
Storyline: As I mentioned at the beginning, I’m excited to introduce the latest addition to our data center portfolio, the Intel Xeon Scalable Platform. This platform represents a new pinnacle in Intel’s 20+ year history of data center innovation. The Intel Xeon Scalable Processor was architected to deliver strong workload driven performance.
Performance: We are excited to be delivering a consistent outstanding performance boost of 1.6X average across a wide range of real-world applications gen over gen.
Compute, Storage, Network Optimized: It’s been a multi-year journey but finally, in this generation, we’ve implemented more features and capabilities specifically for storage and network than ever before. This enables us to win more workloads, more customers and enable new usage models or business opportunities.
Security: We’re also lowering the performance overhead for security data; making data encryption / decryption without a performance penalty a reality. This allows data center operators to deploy security pervasively with minimal impact to service delivery.
Simple and Easy to Deploy:
In the last 5 years, we’ve done a lot of work with the ecosystem to ensure that all the capabilities that went into workload optimization are industry-ready for fast adoption. Also, for the first time, we are releasing a platform that is capable of 2, 4, & 8S+, all at the same time. This will not only accelerate the availability of larger systems in the industry but now, some of the mission critical features (previously available on higher-end system only) can now be supported in 2S mainstream systems.
Efficient utilization of new enhanced processor cores and cache
Ongoing opportunity for optimization for nonvolatile memory
Software TCO 3 years
Oracle
Xeon E5-2697v3 14 Haswell Q3,14 - 28 cores $665,000 + 438900 22% support for 3 yrs = $1,103,900
Xeon E5-2699v4 22 Broadwell Q1,16 - 44 $1,045,000 + 689700 22% support for 3 years = $1,734,700
Xeon Platinum 8180 28 Skylake Q3,17 - 56 $1,330,000 + 877800 22% support for 3 years = $2,207,800
MariaDB
Xeon E5-2697v3 14 Haswell Q3,14 - 28 cores $15000
Xeon E5-2699v4 22 Broadwell Q1,16 - 44 cores $15000
Xeon Platinum 8180 28 Skylake Q3,17 - 56 cores $15000
Moores Law
MariaDB (Hammerdb oltp)
Xeon E5-2697v3 387861 NOPM - $0.038 Cost per transaction
Xeon E5-2699v4 490594 NOPM - $0.030 Cost per transaction
Xeon Platinum 8180 659185 NOPM $0.022 Cost per transaction
Performance
v3 to v4 - 1.26X
v4 to Platinum - 1.34X
v3 to Platinum - 1.7X
Oracle (Hammerdb oltp)
Xeon E5-2697v3 918060 NOPM license $1.20 cost per transaction
Xeon E5-2699v4 1404780 NOPM license $1.23 cost per transaction
Xeon Platinum 810 2393498 NOPM 0.27X $0.92 cost per transaction
Performance
v3 to v4 - 1.53X
v4 to Platinum - 1.7X
Cost difference
v3 32X
v4 41X
Platinum 42X
Key Message: A glimpse inside the Xeon Scalable Platform
Storyline: What’s inside this platform? It represents the best combination of leadership capabilities built upon Intel’s >20 year innovations across CPU and chipset and more. At the top-left, you see some of the technologies we will be offering in this generation as integrated options: Omnipath fabric, Ethernet - we’ll have Quick Assist integrated. Quick assist was born of network WL for compression and crypto acceleration. Over time, it has grown in use case where CSPs and Financial Services want to use it for their security. Another example is AVX-512. We’ve had it in other products and we are bringing it to Xeon in this generation. Hugely popular in HPC for acceleration and with FSIs for High Frequency Trading but AVX-512 benefits WLs across compute, storage, and network – we’ll show you more an real-world performance with AVX-512 optimizations in a few slides. These integrated options are for customers looking to reduce power or save space inside the system.
Our platform supports our latest and greatest SSDs including the Optane SSDs as well as our complementary products like discrete Intel FPGAs.
We have invested multiple years with the ecosystem on SW enablement to take advantage of the new core features and capabilities. Along with specialized workload frameworks like Caffe for artificial intelligence and development tools like Data Plane Development Kit for networking, we can accelerate the adoption of these new datacenter technologies and create new business opportunities.
This platform delivers on the value vectors that matter to the customer to fully embrace the new market opportunities, accelerate growth and remain competitive: performance, security, and agility…
Digging in a bit deeper… This platform provides additional choice of capability offerings and configurations. For example, for networking the Intel® Ethernet 700 series provides via an add-in card with support up to 40GbE for high speed data communications, supports enhanced Data Plane Development Kit (DPDK), which is a set of libraries and drivers to accelerate Network Functions Virtualization (NFV), and achieve higher performance for network packet processing workloads.
This level of Ethernet capability is now offered via integration at the chip level & since DPDK is a core-intensive application, the better per core performance offered with the new Intel Xeon Processor scalable family over previous generation products enables better DPDK performance
What about integrated accelerators like Intel® Advanced Vector Extensions 512 or Intel® QuickAssist?
With new Xeon® processor families further extending the breadth of vector processing capability across both floating-point and integer data domains, and also adding Fused Multiply Add (FMA) support and scalability. Intel® AVX-512 can double the FLOPS/clock vs. AVX2. Intel QAT provides hardware acceleration for compute-intensive workloads (cryptography and data compression);
To extend the platform value across the system we have new Intel® Data Center SSDs, which can be used across a continuum of data tiering needs including Storage, Caching and Memory. The new Intel® Optane™ SSDs are amazingly fast and the Intel® 3D NAND SSDs provide a terrific balance of speed and storage capacity. Both can benefit from a new platform capability we call Intel® Intel® Volume Management Device (Intel® VMD), which is designed to deliver seamless management of PCIe*-based (NVMe*) solid state drives, including enabling “hot plug” capability that minimize service interruptions during drive swaps.
For specialized workload optimizations & evolving workloads like Artificial Intelligence, enterprises can benefit by combining the new Intel Xeon Scalable processors with other products from Intel’s portfolio, ranging from FPGAs and the Intel® Xeon® Phi processor to the Intel Nervana offerings.
Snap*: The Open Telemetry Framework (Snap*): Data is everywhere around us. Telemetry, said simply, is system information: anything and everything we can gather related to the state of a process, piece of hardware, OS, virtualization layer and cloud software. This information is diverse in form and its effectiveness in driving decisions. Individual data points don’t tell much of a story. In aggregate, it can be truly insightful. Enter Snap, the open telemetry framework that provides greater exposure to system data by standardizing telemetry behind a single API.Storage Performance Development Kit (SPDK): To help storage OEMs and ISVs integrate this hardware, Intel has created a set of drivers and a complete, end-to-end reference storage architecture called the Storage Performance Development Kit (SPDK). The goal of SPDK is to highlight the outstanding efficiency and performance enabled by using Intel’s networking, processing, and storage technologies together. By running software designed from the silicon up, SPDK has demonstrated that millions of I/Os per second are easily attainable by using a few processor cores and a few NVMe drives for storage with no additional offload hardware.
The P4800X is the World’s Most Responsive Data Center SSD.
The Intel Optane SSD DC P4800X brings an industry-leading combination of low latency, outstanding QoS, and high endurance, delivering performance needed for both Memory and Storage workloads.
In storage applications, the Optane SSD will help Break Storage and caching bottlenecks, enabling faster performance and increasing scalability.
The Intel Optane SSD will also be used to replace or extend the DRAM memory pool, enabling more affordable deployments and significantly larger data sets. With these larger data sets closer to the Processor, we will see new insights and discoveries.
As you all know, at Intel, we always ensure that we have data to back up any claims. Let’s dive into the elements and data which support this bold claim.
Notes from DCG SDI Gold Deck.
Storage is probably the biggest near-term need in Data Centers. The solutions are rapidly maturing, and the value proposition is strong. Reduction in the cost per Terabyte of storage by a factor of 3 to 15 X. (Swiftstack case study)
The major changes as we move from traditional storage to SDI storage are:
BEFORE: Each major system has dedicated storage resources, including disks, networks, back-up and archives. AFTER: Storage resources are pooled across systems, with shared infrastructure, reducing cost and wasted capacity. What had been fixed-function devices become virtual appliances running on standard Xeon servers.
BEFORE: Each system kept stored data in isolated vaults, which made cross-system or cross-enterprise analysis very slow and difficult. AFTER: Unified resource pools with a single namespace managed by a master orchestrator enables greater visibility into stored data, and business insights yielded from simpler analysis.
BEFORE: Capacity was delivered through expensive scale-up systems, that often locked customers into specific vendors. AFTER: Scale-out systems allow customers to easily add capacity to their storage pools. And since they are based on standard Intel Xeon platforms, customers have more choices and lower cost.
All this, means customers’ storage systems become much more nimble and easy to configure, reduce total cost, and opens up their stored data to more analysis and insight.
Dong et al, “Optimizing Space Amplification in RocksDB,” 2017
http://cidrdb.org/cidr2017/papers/p82-dong-cidr17.pdf
MyRocks - a RocksDB storage engine with MySQL
http://myrocks.io/
Matsunobu, "MyRocks: A space- and write-optimized MySQL database," 2016
https://code.facebook.com/posts/190251048047090/myrocks-a-space-and-write-optimized-mysql-database/
Cloud-centric, shared storage architecture
Three imperatives enable the process of data center modernization.
Enable visibility, virtualization and performance management through cloud. Engaging cloud with visibility across all your workloads allows you to optimize your operations and flexibly respond to future needs while guaranteeing SLAs. Only about half of IT infrastructures are virtualized, leaving plenty of opportunities to consolidate and improve operations.
Migrate to a services-oriented business through software-defined infrastructure (SDI). Software-defined infrastructure is the data center approach that will enable fast services delivery by quickly configuring resources to meet the needs of new services. A self-service portal allows users to choose their own services, so they don’t have to wait for technologies they can leverage to innovate with. This will drive down time to market, costs and labor.
Create greater value in IT by driving business innovation with analytics. With the growing data that companies acquire, IT is in a position to leverage that data to gain insight that drives innovation.
A Reconfigurable Fabric for Accelerating Large-Scale Datacenter Services
Microsoft's Production Configurable Cloud
Accelerating Persistent Neural Networks at Datacenter Scale
Average 7.72 usec to L1 switch and average 18.71 usec latency to L2 switch
Roughly half height half length PCIe expansion card ; 29.2W maximum power consumption
Relational: 2X+ TPC-DS or TPC-H w/Swarm64
NoSQL: 4X Cassandra w/rENIAC (80/20 R/W)
Hadoop/Spark: 3X Terasort w/A3Cube (HDD)
As an example:
Relational: 2X+ TPC-DS or TPC-H w/Swarm64
• Microsoft: A Reconfigurable Fabric for Accelerating Large-Scale Datacenter Services
• Microsoft's Production Configurable Cloud
• Accelerating Persistent Neural Networks at Data Center Scale
• Average 7.72 usec to L1 switch and average 18.71 usec latency to L2 switch
• Roughly half height half length PCIe expansion card ; 29.2W maximum power consumption
Banking/Finance, Business Intelligence, Government, Healthcare, Retail
At the heart of the Acceleration Stack is the Acceleration Environment for Intel Xeon CPU with FPGAs.
This Acceleration Environment has a software component, the Intel Acceleration Engine, and an FPGA component, the FPGA Interface Manager.
Together, these provide the common developer interface with simplified and optimized APIs and interfaces from Intel.
Developer can build their own acceleration applications and Accelerator Functions directly on top of the Acceleration Environment and reuse that code on different Intel FPGA form factors in the data center.
Or they can take advantage of the Acceleration Libraries for FPGAs to fast-track their performance and time to market.
Intel® Programmable Acceleration card with Intel Arria® 10 FPGA
Sampling now, production 2Q18; offering end-to-end virtualization and security